File size: 4,987 Bytes
f4ebfc5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
<div align="center">
    <a href="https://arxiv.org/abs/2512.09271"><img src="https://img.shields.io/badge/Arxiv-preprint-red"></a>
    <a href="https://welldky.github.io/LongT2IBench-Homepage/"><img src="https://img.shields.io/badge/Homepage-green"></a>
</div>

<h1 align="center">LongT2IBench: A Benchmark for Evaluating Long Text-to-Image Generation with Graph-structured Annotations</h1>

<div align="center">
    <a href="https://github.com/yzc-ippl/" target="_blank">Zhichao Yang</a><sup>1</sup>,
    <a href="https://github.com/welldky" target="_blank">Tianjiao Gu</a><sup>1</sup>,
    <a href="https://github.com/satan-7" target="_blank">Jianjie Wang</a><sup>1</sup>,
    <a href="https://github.com/Guapicat0" target="_blank">Feiyu Lin</a><sup>1</sup>,
    <a href="https://github.com/sxfly99" target="_blank">Xiangfei Sheng</a><sup>1</sup>,
    <a href="https://faculty.xidian.edu.cn/cpf/" target="_blank">Pengfei Chen</a><sup>1*</sup>,
    <a href="https://web.xidian.edu.cn/ldli/" target="_blank">Leida Li</a><sup>1,2*</sup>
</div>

<div align="center">
  <sup>1</sup>School of Artificial Intelligence, Xidian University
  <br>
  <sup>2</sup>State Key Laboratory of Electromechanical Integrated Manufacturing of High-Performance Electronic Equipments, Xidian University
</div>

<div align="center">
<sup>*</sup>Corresponding author
</div>

<div align="center">
  <img src="LongT2IBench.png" width="800"/>
</div>

<div style="font-family: sans-serif; margin-bottom: 2em;">
    <h2 style="border-bottom: 1px solid #eaecef; padding-bottom: 0.3em; margin-bottom: 1em;">News</h2>
    <ul style="list-style-type: none; padding-left: 0;">
        <li style="margin-bottom: 0.8em;">
            <strong>[2025-12-21]</strong> The training code has been released.
        </li>
        <li style="margin-bottom: 0.8em;">
            <strong>[2025-12-09]</strong> The data and pre-trained models have been released.
        </li>
        <li style="margin-bottom: 0.8em;">
            <strong>[2025-11-08]</strong> Our paper, "LongT2IBench: A Benchmark for Evaluating Long Text-to-Image Generation with Graph-structured Annotations", has been accepted for an oral presentation at AAAI 2026!
        </li>
    </ul>
</div>

## Quick Start

This guide will help you get started with the LongT2IBench inference code.

### 1. Installation

First, clone the repository and install the required dependencies.

```bash
git clone https://github.com/yzc-ippl/LongT2IBench.git
cd LongT2IBench
pip install -r requirements.txt
```

### 2. Download Pre-trained Weights and Dataset

##### Prepare Pre-trained Weights

You can download the pre-trained model weights of <strong>[LongT2IExpert]</strong> from the following link: [**(Baidu Netdisk)**](https://pan.baidu.com/s/1Ltj77l31hyBkn6nLtYctnQ?pwd=i8ug)

Place the downloaded files in the `weights` directory.

- ``./weights/LongT2IBench-checkpoints``: The main model for generation and scoring.

Create the `weights` directory if it doesn't exist and place the files inside.

##### Prepare Datasets

You can download the dataset of <strong>[LongPrompt-3K]</strong> and <strong>[LongT2IBench-14K]</strong> from the following link: [**(Baidu Netdisk)**](https://pan.baidu.com/s/1M_tE9EfA2s0Vn7l9r0GebA?pwd=7b6d)

Place the downloaded files in the `data` directory.

Create the `data` directory if it doesn't exist and place the files inside.

```
LongT2IBench/
|-- weights/
|   |-- LongT2IBench-checkpoints
|   |   |-- config.json
|   |   |-- ...
|   |-- Qwen2.5-VL-7B-Instruct
|   |   |-- config.json
|   |   |-- ...
|-- data/
|   |-- imgs
|   |-- split
|   |   |-- train.json
|   |   |-- test.json
|   |   |-- val.json
|-- config.py
|-- dataset.py
|-- model.py
|-- requirements.txt
|-- README.md
|-- test_generation.py
|-- test_score.py
|-- train.py
```

### 3. Run Inference

The `LongT2IExpert` provides two main inference tasks: Long T2I Alignment Scoring and Long T2I Alignment Interpreting.

##### Long T2I Alignment Scoring

```
python test_score.py
```

##### Long T2I Alignment Interpreting

```
python test_generation.py
```

### 4. Run Training

You can run this code to train <strong>[LongT2IExpert]</strong> from start to finish. 

Make sure the initially untrained weights are located at ``./weights/Qwen2.5-VL-7B-Instruct`` :

You can download the untrained weights from the following link [**(Baidu Netdisk)**](https://pan.baidu.com/s/17PcO4CvgB6FDHh6JBgM_Lg?pwd=3h8m)

```bash
python train.py
```

## Citation

If you find this work is useful, pleaes cite our paper!

```bibtex
@misc{yang2025longt2ibenchbenchmarkevaluatinglong,
      title={LongT2IBench: A Benchmark for Evaluating Long Text-to-Image Generation with Graph-structured Annotations}, 
      author={Zhichao Yang and Tianjiao Gu and Jianjie Wang and Feiyu Lin and Xiangfei Sheng and Pengfei Chen and Leida Li},
      year={2025},
      eprint={2512.09271},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2512.09271}, 
}
```