eq-vae-ema-ldm / README.md
nielsr's picture
nielsr HF Staff
Improve model card: Add metadata, external links, and usage example
a673cc8 verified
|
raw
history blame
1.37 kB
metadata
pipeline_tag: image-to-image
library_name: diffusers
license: mit

EQ-VAE: Equivariance Regularized Latent Space for Improved Generative Image Modeling

Arxiv: https://arxiv.org/abs/2502.09509 Project page: https://eq-vae.github.io/ Code: https://github.com/zelaki/eqvae

EQ-VAE regularizes the latent space of pretrained autoencoders by enforcing equivariance under scaling and rotation transformations.


Model Description

This model (eq-vae-ema) is a regularized version of SD-VAE. We finetune it with EQ-VAE regularization for 44 epochs on Imagenet with EMA weights.

Model Usage

These weights are intended to be used with the EQ-VAE codebase or the CompVis Stable Diffusion codebase.

You can also use this model with the 🧨 diffusers library:

from diffusers import AutoencoderKL
eqvae = AutoencoderKL.from_pretrained("zelaki/eq-vae-ema")

Metrics

Reconstruction performance of eq-vae-ema on Imagenet Validation Set.

Metric Score
FID 0.552
PSNR 26.158
LPIPS 0.133
SSIM 0.725