File size: 11,238 Bytes
09074af
508ac59
 
 
 
 
 
 
09074af
 
508ac59
 
 
 
 
 
570abc9
508ac59
 
 
09074af
 
 
508ac59
 
 
 
09074af
508ac59
9b94de3
 
 
 
 
 
 
 
508ac59
 
9b94de3
 
 
 
 
 
 
 
 
da26b92
9b94de3
 
 
 
 
 
 
 
 
 
 
da26b92
9b94de3
 
 
42e2761
9b94de3
 
 
 
 
 
508ac59
 
 
 
 
 
 
 
 
 
 
 
9b94de3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fc034ba
9b94de3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
da26b92
 
 
 
 
 
 
 
 
9b94de3
 
 
 
 
 
 
508ac59
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
---
license: cc-by-4.0
task_categories:
- zero-shot-image-classification
- image-to-text
language:
- zh
arxiv_id: 2601.10305
dataset_info:
  features:
  - name: url
    dtype: string
  - name: alt_text
    dtype: string
  - name: recaption
    dtype: string
  splits:
  - name: train
    num_bytes: 19795236
    num_examples: 99892381
  download_size: 19795236
  dataset_size: 99892381
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---

<div align="center">

<img src="Figures/danqing.svg" width="30%">

**100M** Chinese image-text pairs | **12TB** dataset | **2024-2025** web data

<h1 align="center">DanQing: An Up-to-Date Large-Scale Chinese Vision-Language Pre-training Dataset</h1>

[**Project Page**](https://deepglint.github.io/DanQing/) | [**Paper**](https://huggingface.co/papers/2601.10305) | [**Code**](https://github.com/deepglint/DanQing)

</div>


<div align="center">

Hengyu Shen<sup></sup>, [**Tiancheng Gu**](https://scholar.google.com/citations?hl=zh-CN&user=9etrpbYAAAAJ)<sup></sup>, Bin Qin, Lan Wu, Yuling Wu, Shuo Tan, [**Zelong Sun**](https://scholar.google.com/citations?user=mDxuGMgAAAAJ&hl=zh-CN), Jun Wang, Nan Wu, [**Xiang An**](https://anxiangsir.github.io/), [**Weidong Cai**](https://weidong-tom-cai.github.io/), [**Ziyong Feng**](https://scholar.google.com/citations?user=xlKttUEAAAAJ&hl=zh-CN)<sup></sup>, [**Kaicheng Yang**](https://kaicheng-yang0828.github.io)<sup></sup>

<sup></sup> Equal Contribution | <sup></sup> Team Leader | <sup></sup> Project Leader

[![Paper](https://img.shields.io/badge/📄-Paper-red)](https://arxiv.org/abs/2601.10305)
[![Hugging Face](https://img.shields.io/badge/🤗%20Hugging%20Face-Dataset-yellow)](https://huggingface.co/datasets/DeepGlint-AI/DanQing100M)
[![ModelScope](https://img.shields.io/badge/ModelScope-Dataset-blue)](https://www.modelscope.cn/datasets/deepglint/DanQing)
[![License: CC BY 4.0](https://img.shields.io/badge/License-CC%20BY%204.0-lightgrey.svg)](https://creativecommons.org/licenses/by/4.0/)


</div>

## 📣 News

<div align="left">

- [2026/01/16] ✨ We release the [paper](https://arxiv.org/abs/2601.10305) of DanQing.
- [2026/01/15] 🔥 We release the DanQing dataset (images and captions, about 12TB) in [ModelScope](https://www.modelscope.cn/datasets/deepglint/DanQing)
- [2026/01/13] ✨ We release the DanQing dataset (URLs of image and captions) in [🤗 Hugging Face](https://huggingface.co/datasets/DeepGlint-AI/DanQing100M)

> ⚠️ **Note:** Due to the storage and transmission limitations of Hugging Face, we only release the URLs corresponding to the images on Hugging Face. To access the complete dataset, please download it from **ModelScope**. We also provide synthetic short captions (generated by GLM4.1-base-9B) for the Danqing100M dataset in the recaption column.


</div>

---

## 🚀 Sample Usage

You can load the metadata (URLs and captions) directly using the `datasets` library:

```python
from datasets import load_dataset

ds = load_dataset("DeepGlint-AI/DanQing100M")
```

---

## 📑 Table of Contents
- [💡 Highlights](#-highlights)
- [💻 Dataset Information](#-dataset-information)
  - [Data Preview](#data-preview)
  - [Topic Assessment](#topic-assessment)
  - [Image Resolution and Text Length Distribution](#image-resolution-and-text-length-distribution)
  - [Text Quality](#text-quality)
  - [Cosine Similarity and Semantic Distribution](#cosine-similarity-and-semantic-distribution)
- [📊 Performance Comparison](#-performance-comparison)
  - [Zero-Shot Classification](#zero-shot-classification)
  - [Cross-Modal Retrieval (Short Caption)](#cross-modal-retrieval-short-caption)
  - [Cross-Modal Retrieval (Long Caption)](#cross-modal-retrieval-long-caption)
  - [Chinese-Centric Large Multimodal Model Tasks](#chinese-centric-large-multimodal-model-tasks)
- [🧠 Analysis](#-analysis)
  - [Data and Model Scaling](#data-and-model-scaling)
  - [New Concept Understanding](#new-concept-understanding)
- [📥 Download](#-download)
  - [🤗 Hugging Face](#-hugging-face)
    - [Python API](#python-api)
    - [Command Line](#command-line)
  - [ ModelScope](#-modelscope)
    - [Python API](#python-api-1)
    - [Command Line](#command-line-1)
- [📄 License](#-license)
- [📝 Citation](#-citation)

---

## 💡 Highlights

In this paper, we propose **DanQing** dataset, which contains **100 million** image-text pairs collected from Common Crawl. Different from existing datasets, DanQing is curated through a more rigorous selection process, yielding superior data quality. Moreover, DanQing is primarily built from **2024–2025** web data, enabling models to better capture evolving semantic trends and thus offering greater practical utility. 

We compare DanQing with existing datasets by conducting continual pre-training of the SigLIP2 model. Experimental results show that DanQing consistently achieves superior performance across a range of Chinese downstream tasks, including zero-shot classification, cross-modal retrieval, and LMM-based evaluations.

<div align="center">
    <img src="Figures/framework.png" width="100%">
</div>

---

## 💻 Dataset Information

### Data Preview

<div align="center">
    <img src="Figures/case.png" width="100%">
</div>

### Topic Assessment

We implement a topic modeling pipeline based on [BERTopic](https://github.com/MaartenGr/BERTopic). We randomly sample 10M image-text pairs and extract text embeddings using [Chinese-CLIP-L/14](https://github.com/OFA-Sys/Chinese-CLIP). To address high-dimensional clustering, we apply UMAP for dimensionality reduction, followed by HDBSCAN to identify semantic clusters with a minimum cluster size of 1,000 for stability and noise reduction. Finally, we use class-based TF-IDF to extract representative keywords for each topic.

<div align="center">
    <img src="Figures/topic_examples.png" width="100%">
</div>

### Image Resolution and Text Length Distribution

We analyze image resolutions by width, height, and minimum dimension, demonstrating a wide range of visual scales. We also report the distribution of text lengths across **2.2B** Chinese words.

<div align="center">
    <img src="Figures/statistic.png" width="100%">
</div>

### Text Quality

We evaluate the text quality of DanQing using two metrics: **semantic word density** and **perplexity (PPL)**. We randomly sample 10M texts from DanQing, Wukong, and Zero for comparison. Semantic words (nouns, verbs, adjectives) are identified using the jieba toolkit, and their proportion in each sentence is calculated as semantic density. Sentence-level perplexity is computed with a pre-trained Chinese [BERT](https://huggingface.co/google-bert/bert-base-chinese) model.

<div align="center">
    <img src="Figures/quality.png" width="100%">
</div>

### Cosine Similarity and Semantic Distribution

We analyze 10M-sample subsets of DanQing and Wukong by presenting image-text similarity distributions, extracted with [FG-CLIP2-L/16@256](https://huggingface.co/qihoo360/fg-clip2-large). For semantic distribution comparison, 10M images from each dataset are clustered into 10K groups using [FAISS](https://github.com/facebookresearch/faiss), with clusters ranked by sample count.

<div align="center">
    <img src="Figures/distribution.png" width="100%">
</div>

---

## 📊 Performance Comparison

### Zero-Shot Classification

<div align="center">
    <img src="Figures/classification.png" width="80%">
</div>

### Cross-Modal Retrieval (Short Caption)

<div align="center">
    <img src="Figures/short.png" width="100%">
</div>

### Cross-Modal Retrieval (Long Caption)

<div align="center">
    <img src="Figures/long.png" width="100%">
</div>

### Chinese-Centric Large Multimodal Model Tasks

<div align="center">
    <img src="Figures/LMM.png" width="80%">
</div>

---

## 🧠 Analysis

### Data and Model Scaling

We compare the data and model scaling capabilities of DanQing and Wukong, reporting average zero-shot classification and retrieval (long & short caption) performance in the figure below.

<div align="center">
    <img src="Figures/scaling.png" width="100%">
</div>

### New Concept Understanding

We evaluate SigLIP2-L/16 models pre-trained on various Chinese datasets for emergent concept understanding, and find that the model trained on DanQing consistently gives the highest confidence to correct pairs.

<div align="center">
    <img src="Figures/new_concept.png" width="100%">
</div>

---

## 📥 Download

### 🤗 Hugging Face

#### Python API

```python
from datasets import load_dataset

ds = load_dataset("DeepGlint-AI/DanQing100M")
```

#### Command Line

```bash
# Install dependencies
# brew install git-xet  # macOS
# git xet install

# sudo apt update  # Ubuntu/Debian
# sudo apt install aria2

# Install git-lfs
# curl -s https://packagecloud.io/install/repositories/github/git-lfs/script.deb.sh | sudo bash
# sudo apt-get install git-lfs
# git lfs install

# Download dataset URLs and captions
bash hfd.sh DeepGlint-AI/DanQing100M --dataset --tool aria2c -x 10

# Download images using img2dataset
# pip install img2dataset
# For better performance, it's highly recommended to set up a fast dns resolver
# See: https://github.com/rom1504/img2dataset#setting-up-a-high-performance-dns-resolver
img2dataset --url_list DanQing100M/data \
        --input_format "parquet" \
        --url_col "url" \
        --caption_col "alt_text" \
        --output_format webdataset \
        --output_folder DanQing100M-webdataset \
        --processes_count 16 \
        --thread_count 32 \
        --image_size 256 \
        --resize_only_if_bigger=True \
        --resize_mode="keep_ratio" \
        --skip_reencode=True \
        --save_additional_columns '["recaption"]' \
        --enable_wandb False
```

### ModelScope

#### Python API

```python
from modelscope.msdatasets import MsDataset

ds = MsDataset.load('deepglint/DanQing')
```

#### Command Line

```bash
pip install modelscope
modelscope download --dataset deepglint/DanQing
```

---

## 📄 License

The DanQing dataset is licensed under [CC-BY-4.0 License](https://creativecommons.org/licenses/by/4.0/). The full license can be found in the [LICENSE.cc-by-4.0 file](./LICENSE.cc-by-4.0). The dataset is collected from Common Crawl web pages and may contain biased or sensitive content. The collected data is subject to the license to which each content belongs. Users are solely responsible for ensuring compliance with ethical and legal standards in their research or applications.

---

## 📝 Citation

If you find this repository useful, please use the following BibTeX entry for citation.

```bibtex
@misc{danqing,
      title={DanQing: An Up-to-Date Large-Scale Chinese Vision-Language Pre-training Dataset}, 
      author={Hengyu Shen and Tiancheng Gu and Bin Qin and Lan Wu and Yuling Wu and Shuo Tan and Zelong Sun and Jun Wang and Nan Wu and Xiang An and Weidong Cai and Ziyong Feng and Kaicheng Yang},
      year={2026},
      eprint={2601.10305},
      archivePrefix={arXiv},
      primaryClass={cs.CV},
      url={https://arxiv.org/abs/2601.10305}, 
}
```
---

<div align="center">

### ⭐ Don't forget to star this repository if you find it helpful!

</div>