Improve dataset card: Add metadata, paper link, code link, and sample usage

#1
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +49 -1
README.md CHANGED
@@ -1 +1,49 @@
1
- arxiv.org/abs/2509.04394
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ task_categories:
3
+ - text-to-image
4
+ license: apache-2.0
5
+ language:
6
+ - en
7
+ tags:
8
+ - diffusion-model
9
+ - generative-ai
10
+ - image-generation
11
+ - toy-dataset
12
+ ---
13
+
14
+ # TiM-Toy-T2I-Dataset
15
+
16
+ This repository contains the toy dataset for text-to-image (T2I) generation used in the paper [Transition Models: Rethinking the Generative Learning Objective](https://huggingface.co/papers/2509.04394).
17
+
18
+ Transition Models (TiM) introduce an exact, continuous-time dynamics equation that analytically defines state transitions across any finite time interval. This leads to a novel generative paradigm, TiM, which adapts to arbitrary-step transitions, seamlessly traversing the generative trajectory from single leaps to fine-grained refinement with more steps. This specific dataset serves as a toy dataset for training and experimentation with TiM for text-to-image generation.
19
+
20
+ Code: [https://github.com/WZDTHU/TiM](https://github.com/WZDTHU/TiM)
21
+
22
+ ## Sample Usage
23
+
24
+ To download this toy text-to-image dataset for training TiM models, you can follow the instructions from the official GitHub repository under the "Dataset Setup" section:
25
+
26
+ ```bash
27
+ bash tools/download_toy_t2i_dataset.sh
28
+ ```
29
+
30
+ For more details on setting up the environment, downloading models, and other training/sampling scripts, please refer to the [official GitHub repository](https://github.com/WZDTHU/TiM).
31
+
32
+ ## Citation
33
+
34
+ If you find the project useful, please kindly cite:
35
+
36
+ ```bibtex
37
+ @article{wang2025transition,
38
+ title={Transition Models: Rethinking the Generative Learning Objective},
39
+ author={Wang, Zidong and Zhang, Yiyuan and Yue, Xiaoyu and Yue, Xiangyu and Li, Yangguang and Ouyang, Wanli and Bai, Lei},
40
+ year={2025},
41
+ eprint={2509.04394},
42
+ archivePrefix={arXiv},
43
+ primaryClass={cs.LG}
44
+ }
45
+ ```
46
+
47
+ ## License
48
+
49
+ This project is licensed under the Apache-2.0 license.