transfusion-vae / README.md
lavinal712's picture
Update README.md
1585596 verified
---
license: mit
datasets:
- ILSVRC/imagenet-1k
language:
- en
tags:
- diffusion
---
# Transfusion - VAE
## How to use with 🧨 diffusers
```py
from diffusers.models import AutoencoderKL
vae = AutoencoderKL.from_pretrained("lavinal712/transfusion-vae")
```
## Model
This model was trained for 50 (legacy: 7) epochs on ImageNet, COCO and FFHQ (legacy: ImageNet), with training parameters following the original Transfusion paper.
$$\mathcal{L}_{\mathrm{VAE}} = \mathcal{L}_1 + \mathcal{L}_{\mathrm{LPIPS}} + 0.5\mathcal{L}_{\mathrm{GAN}} + 0.2\mathcal{L}_{\mathrm{ID}} + 0.000001\mathcal{L}_{\mathrm{KL}}$$
## Evaluation
ImageNet 2012 (256x256, val, 50000 images)
| Model | rFID | PSNR | SSIM | LPIPS |
|-----------------|-------|--------|-------|-------|
| Transfusion-VAE | 0.408 | 28.723 | 0.845 | 0.081 |
| SD-VAE | 0.692 | 26.910 | 0.772 | 0.130 |
COCO 2017 (256x256, val, 5000 images)
| Model | rFID | PSNR | SSIM | LPIPS |
|-----------------|-------|--------|-------|-------|
| Transfusion-VAE | 2.749 | 28.556 | 0.855 | 0.078 |
| SD-VAE | 4.246 | 26.622 | 0.784 | 0.127 |
## Evaluation (legacy)
ImageNet 2012 (256x256, val, 50000 images)
| Model | rFID | PSNR | SSIM | LPIPS |
|-----------------|-------|--------|-------|-------|
| Transfusion-VAE | 0.567 | 28.195 | 0.829 | 0.100 |
| SD-VAE | 0.692 | 26.910 | 0.772 | 0.130 |
Paper: [Transfusion: Predict the Next Token and Diffuse Images with One Multi-Modal Model](https://arxiv.org/abs/2408.11039)
Dataset: [ImageNet](https://image-net.org/), [COCO](https://cocodataset.org/), [FFHQ](https://github.com/NVlabs/ffhq-dataset)
Base Code: [lavinal712/AutoencoderKL](https://github.com/lavinal712/AutoencoderKL)
Training Code: [lavinal712/AutoencoderKL/tree/transfusion_vae](https://github.com/lavinal712/AutoencoderKL/tree/transfusion_vae)