Improve model card: Add pipeline tag, links, and usage reference

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +39 -3
README.md CHANGED
@@ -1,3 +1,39 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ pipeline_tag: image-to-image
4
+ ---
5
+
6
+ # Flow to the Mode: Mode-Seeking Diffusion Autoencoders for State-of-the-Art Image Tokenization
7
+
8
+ This repository contains **FlowMo**, a transformer-based diffusion autoencoder that achieves state-of-the-art performance for image tokenization at multiple compression rates. It is introduced in the paper [Flow to the Mode: Mode-Seeking Diffusion Autoencoders for State-of-the-Art Image Tokenization](https://huggingface.co/papers/2503.11056).
9
+
10
+ FlowMo operates without using convolutions, adversarial losses, spatially-aligned two-dimensional latent codes, or distilling from other tokenizers. Its key insight is that training should be broken into a mode-matching pre-training stage and a mode-seeking post-training stage.
11
+
12
+ <p align="center">
13
+ <img src="https://github.com/kylesargent/FlowMo/raw/main/demo.gif" alt="FlowMo demo GIF" />
14
+ </p>
15
+
16
+ ## Links
17
+
18
+ * **Project Page:** [https://kylesargent.github.io/flowmo](https://kylesargent.github.io/flowmo)
19
+ * **Code Repository:** [https://github.com/kylesargent/FlowMo](https://github.com/kylesargent/FlowMo)
20
+
21
+ ## Usage
22
+
23
+ The official GitHub repository provides comprehensive instructions for installation, data preparation, training, and evaluation. A Jupyter notebook, `example.ipynb`, is available to demonstrate how to use the FlowMo tokenizer for image reconstruction.
24
+
25
+ ## Citation
26
+
27
+ If you find FlowMo useful, please cite our paper:
28
+
29
+ ```bibtex
30
+ @misc{sargent2025flowmodemodeseekingdiffusion,
31
+ title={Flow to the Mode: Mode-Seeking Diffusion Autoencoders for State-of-the-Art Image Tokenization},
32
+ author={Kyle Sargent and Kyle Hsu and Justin Johnson and Li Fei-Fei and Jiajun Wu},
33
+ year={2025},
34
+ eprint={2503.11056},
35
+ archivePrefix={arXiv},
36
+ primaryClass={cs.CV},
37
+ url={https://arxiv.org/abs/2503.11056},
38
+ }
39
+ ```