File size: 4,987 Bytes
f4ebfc5 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 | <div align="center">
<a href="https://arxiv.org/abs/2512.09271"><img src="https://img.shields.io/badge/Arxiv-preprint-red"></a>
<a href="https://welldky.github.io/LongT2IBench-Homepage/"><img src="https://img.shields.io/badge/Homepage-green"></a>
</div>
<h1 align="center">LongT2IBench: A Benchmark for Evaluating Long Text-to-Image Generation with Graph-structured Annotations</h1>
<div align="center">
<a href="https://github.com/yzc-ippl/" target="_blank">Zhichao Yang</a><sup>1</sup>,
<a href="https://github.com/welldky" target="_blank">Tianjiao Gu</a><sup>1</sup>,
<a href="https://github.com/satan-7" target="_blank">Jianjie Wang</a><sup>1</sup>,
<a href="https://github.com/Guapicat0" target="_blank">Feiyu Lin</a><sup>1</sup>,
<a href="https://github.com/sxfly99" target="_blank">Xiangfei Sheng</a><sup>1</sup>,
<a href="https://faculty.xidian.edu.cn/cpf/" target="_blank">Pengfei Chen</a><sup>1*</sup>,
<a href="https://web.xidian.edu.cn/ldli/" target="_blank">Leida Li</a><sup>1,2*</sup>
</div>
<div align="center">
<sup>1</sup>School of Artificial Intelligence, Xidian University
<br>
<sup>2</sup>State Key Laboratory of Electromechanical Integrated Manufacturing of High-Performance Electronic Equipments, Xidian University
</div>
<div align="center">
<sup>*</sup>Corresponding author
</div>
<div align="center">
<img src="LongT2IBench.png" width="800"/>
</div>
<div style="font-family: sans-serif; margin-bottom: 2em;">
<h2 style="border-bottom: 1px solid #eaecef; padding-bottom: 0.3em; margin-bottom: 1em;">News</h2>
<ul style="list-style-type: none; padding-left: 0;">
<li style="margin-bottom: 0.8em;">
<strong>[2025-12-21]</strong> The training code has been released.
</li>
<li style="margin-bottom: 0.8em;">
<strong>[2025-12-09]</strong> The data and pre-trained models have been released.
</li>
<li style="margin-bottom: 0.8em;">
<strong>[2025-11-08]</strong> Our paper, "LongT2IBench: A Benchmark for Evaluating Long Text-to-Image Generation with Graph-structured Annotations", has been accepted for an oral presentation at AAAI 2026!
</li>
</ul>
</div>
## Quick Start
This guide will help you get started with the LongT2IBench inference code.
### 1. Installation
First, clone the repository and install the required dependencies.
```bash
git clone https://github.com/yzc-ippl/LongT2IBench.git
cd LongT2IBench
pip install -r requirements.txt
```
### 2. Download Pre-trained Weights and Dataset
##### Prepare Pre-trained Weights
You can download the pre-trained model weights of <strong>[LongT2IExpert]</strong> from the following link: [**(Baidu Netdisk)**](https://pan.baidu.com/s/1Ltj77l31hyBkn6nLtYctnQ?pwd=i8ug)
Place the downloaded files in the `weights` directory.
- ``./weights/LongT2IBench-checkpoints``: The main model for generation and scoring.
Create the `weights` directory if it doesn't exist and place the files inside.
##### Prepare Datasets
You can download the dataset of <strong>[LongPrompt-3K]</strong> and <strong>[LongT2IBench-14K]</strong> from the following link: [**(Baidu Netdisk)**](https://pan.baidu.com/s/1M_tE9EfA2s0Vn7l9r0GebA?pwd=7b6d)
Place the downloaded files in the `data` directory.
Create the `data` directory if it doesn't exist and place the files inside.
```
LongT2IBench/
|-- weights/
| |-- LongT2IBench-checkpoints
| | |-- config.json
| | |-- ...
| |-- Qwen2.5-VL-7B-Instruct
| | |-- config.json
| | |-- ...
|-- data/
| |-- imgs
| |-- split
| | |-- train.json
| | |-- test.json
| | |-- val.json
|-- config.py
|-- dataset.py
|-- model.py
|-- requirements.txt
|-- README.md
|-- test_generation.py
|-- test_score.py
|-- train.py
```
### 3. Run Inference
The `LongT2IExpert` provides two main inference tasks: Long T2I Alignment Scoring and Long T2I Alignment Interpreting.
##### Long T2I Alignment Scoring
```
python test_score.py
```
##### Long T2I Alignment Interpreting
```
python test_generation.py
```
### 4. Run Training
You can run this code to train <strong>[LongT2IExpert]</strong> from start to finish.
Make sure the initially untrained weights are located at ``./weights/Qwen2.5-VL-7B-Instruct`` :
You can download the untrained weights from the following link [**(Baidu Netdisk)**](https://pan.baidu.com/s/17PcO4CvgB6FDHh6JBgM_Lg?pwd=3h8m)
```bash
python train.py
```
## Citation
If you find this work is useful, pleaes cite our paper!
```bibtex
@misc{yang2025longt2ibenchbenchmarkevaluatinglong,
title={LongT2IBench: A Benchmark for Evaluating Long Text-to-Image Generation with Graph-structured Annotations},
author={Zhichao Yang and Tianjiao Gu and Jianjie Wang and Feiyu Lin and Xiangfei Sheng and Pengfei Chen and Leida Li},
year={2025},
eprint={2512.09271},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2512.09271},
}
```
|