metadata
pipeline_tag: image-to-image
library_name: diffusers
license: mit
EQ-VAE: Equivariance Regularized Latent Space for Improved Generative Image Modeling
Arxiv: https://arxiv.org/abs/2502.09509 Project page: https://eq-vae.github.io/ Code: https://github.com/zelaki/eqvae
EQ-VAE regularizes the latent space of pretrained autoencoders by enforcing equivariance under scaling and rotation transformations.
Model Description
This model (eq-vae-ema) is a regularized version of SD-VAE. We finetune it with EQ-VAE regularization for 44 epochs on Imagenet with EMA weights.
Model Usage
These weights are intended to be used with the EQ-VAE codebase or the CompVis Stable Diffusion codebase.
You can also use this model with the 🧨 diffusers library:
from diffusers import AutoencoderKL
eqvae = AutoencoderKL.from_pretrained("zelaki/eq-vae-ema")
Metrics
Reconstruction performance of eq-vae-ema on Imagenet Validation Set.
| Metric | Score |
|---|---|
| FID | 0.552 |
| PSNR | 26.158 |
| LPIPS | 0.133 |
| SSIM | 0.725 |