Improve model card for Neon project
#5
by
nielsr HF Staff - opened
README.md
CHANGED
|
@@ -1,3 +1,64 @@
|
|
| 1 |
-
---
|
| 2 |
-
license: mit
|
| 3 |
-
--
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
license: mit
|
| 3 |
+
pipeline_tag: unconditional-image-generation
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
+
# Neon: Negative Extrapolation From Self-Training Improves Image Generation
|
| 7 |
+
|
| 8 |
+
This repository contains models improved by **Neon**, a novel method for enhancing generative AI models, as presented in the paper [Neon: Negative Extrapolation From Self-Training Improves Image Generation](https://huggingface.co/papers/2510.03597).
|
| 9 |
+
|
| 10 |
+
Official code: [https://github.com/SinaAlemohammad/Neon](https://github.com/SinaAlemohammad/Neon)
|
| 11 |
+
|
| 12 |
+
## Introduction
|
| 13 |
+
|
| 14 |
+
Scaling generative AI models is bottlenecked by the scarcity of high-quality training data. The ease of synthesizing from a generative model suggests using (unverified) synthetic data to augment a limited corpus of real data for the purpose of fine-tuning in the hope of improving performance. Unfortunately, however, the resulting positive feedback loop leads to model autophagy disorder (MAD, aka model collapse) that results in a rapid degradation in sample quality and/or diversity. In this paper, we introduce Neon (for Negative Extrapolation frOm self-traiNing), a new learning method that turns the degradation from self-training into a powerful signal for self-improvement. Given a base model, Neon first fine-tunes it on its own self-synthesized data but then, counterintuitively, reverses its gradient updates to extrapolate away from the degraded weights. We prove that Neon works because typical inference samplers that favor high-probability regions create a predictable anti-alignment between the synthetic and real data population gradients, which negative extrapolation corrects to better align the model with the true data distribution. Neon is remarkably easy to implement via a simple post-hoc merge that requires no new real data, works effectively with as few as 1k synthetic samples, and typically uses less than 1% additional training compute. We demonstrate Neon’s universality across a range of architectures (diffusion, flow matching, autoregressive, and inductive moment matching models) and datasets (ImageNet, CIFAR-10, and FFHQ). In particular, on ImageNet 256x256, Neon elevates the xAR-L model to a new state-of-the-art FID of 1.02 with only 0.36% additional training compute.
|
| 15 |
+
|
| 16 |
+
## Method
|
| 17 |
+
|
| 18 |
+

|
| 19 |
+
|
| 20 |
+
**In one line:** sample with your usual inference to form a synthetic set $S$; briefly fine-tune the reference model on $S$ to get $\theta_s$; then **reverse** that update with a merge $\theta_{\text{neon}}=(1+w)\,\theta_r - w\,\theta_s$ (small $w>0$), which cancels mode-seeking drift and improves FID.
|
| 21 |
+
|
| 22 |
+
## Benchmark Performance
|
| 23 |
+
|
| 24 |
+
| Model type | Dataset | Base model FID | Neon FID (paper) | Download model |
|
| 25 |
+
| ------------- | ---------------- | -------------: | ---------------: | ------------------------------------------------------------------------------------------------------- |
|
| 26 |
+
| xAR-L | ImageNet-256 | 1.28 | **1.02** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_xARL_imagenet256.pth) |
|
| 27 |
+
| xAR-B | ImageNet-256 | 1.72 | **1.31** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_xARB_imagenet256.pth) |
|
| 28 |
+
| VAR d16 | ImageNet-256 | 3.30 | **2.01** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_VARd16_imagenet256.pth) |
|
| 29 |
+
| VAR d36 | ImageNet-512 | 2.63 | **1.70** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_VARd36_imagenet512.pth) |
|
| 30 |
+
| EDM (cond.) | CIFAR-10 (32×32) | 1.78 | **1.38** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_EDM_conditional_CIFAR10.pkl) |
|
| 31 |
+
| EDM (uncond.) | CIFAR-10 (32×32) | 1.98 | **1.38** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_EDM_unconditional_CIFAR10.pkl) |
|
| 32 |
+
| EDM | FFHQ-64×64 | 2.39 | **1.12** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_EDM_FFHQ.pkl) |
|
| 33 |
+
| IMM | ImageNet-256 | 1.99 | **1.46** | [Download](https://huggingface.co/sinaalemohammad/Neon/resolve/main/Neon_imm_imagenet256.pkl) |
|
| 34 |
+
|
| 35 |
+
## Quickstart
|
| 36 |
+
|
| 37 |
+
Pretrained Neon-improved models can be downloaded directly from this Hugging Face repository via the links in the table above. For detailed instructions on environment setup, downloading models, and evaluating performance, please refer to the [official GitHub repository](https://github.com/SinaAlemohammad/Neon).
|
| 38 |
+
|
| 39 |
+
## Citation
|
| 40 |
+
|
| 41 |
+
If you find Neon useful, please consider citing the paper:
|
| 42 |
+
|
| 43 |
+
```bibtex
|
| 44 |
+
@article{neon2025,
|
| 45 |
+
title={Neon: Negative Extrapolation from Self-Training for Generative Models},
|
| 46 |
+
author={Alemohammad, Sina and collaborators},
|
| 47 |
+
journal={arXiv preprint},
|
| 48 |
+
year={2025}
|
| 49 |
+
}
|
| 50 |
+
```
|
| 51 |
+
|
| 52 |
+
## Acknowledgments
|
| 53 |
+
|
| 54 |
+
This repository builds upon and thanks the following projects:
|
| 55 |
+
|
| 56 |
+
* [VAR — Visual AutoRegressive Modeling](https://github.com/FoundationVision/VAR)
|
| 57 |
+
* [xAR — Beyond Next‑Token: Next‑X Prediction](https://github.com/OliverRensu/xAR)
|
| 58 |
+
* [IMM — Inductive Moment Matching](https://github.com/lumalabs/imm)
|
| 59 |
+
* [EDM — Elucidating the Design Space of Diffusion Models](https://github.com/NVlabs/edm)
|
| 60 |
+
* [MAR VAE (KL‑16) tokenizer](https://huggingface.co/xwen99/mar-vae-kl16)
|
| 61 |
+
|
| 62 |
+
## Contact
|
| 63 |
+
|
| 64 |
+
Questions? Reach out to **Sina Alemohammad** — [sinaalemohammad@gmail.com](mailto:sinaalemohammad@gmail.com).
|