File size: 5,728 Bytes
40a7688
 
225b66f
870e29f
 
225b66f
 
870e29f
 
 
 
40a7688
 
d06b074
40a7688
225b66f
 
 
40a7688
 
 
 
d06b074
40a7688
 
 
 
 
 
 
 
 
 
 
 
d06b074
40a7688
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d06b074
40a7688
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
---
license: cc-by-nc-4.0
task_categories:
- image-classification
- text-to-image
library_name: datasets
tags:
- ai-generated-content
- image-quality-assessment
- real-vs-fake
- multimodal
---

# DANI: Discrepancy Assessing for Natural and AI Images

[Paper: D-Judge: How Far Are We? Evaluating the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal Guidance](https://huggingface.co/papers/2412.17632)
[Code: https://github.com/RenyangLiu/DJudge](https://github.com/RenyangLiu/DJudge)

**A Large-Scale Dataset for Visual Research on AI-Synthesized and Natural Images**

## Overview

DANI (Discrepancy Assessing for Natural and AI Images) is a large-scale, multimodal dataset for benchmarking and broad visual research on both AI-generated images (AIGIs) and natural images.  
The dataset is designed to support a wide range of computer vision and multimodal research tasks, including but not limited to:
- AI-generated vs. real image discrimination
- Representation learning
- Image quality assessment
- Style transfer
- Image reconstruction
- Domain adaptation
- Multimodal understanding and beyond

DANI accompanies the paper:

> Liu, Renyang; Lyu, Ziyu; Zhou, Wei; Ng, See-Kiong.  
> *D-Judge: How Far Are We? Assessing the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal Guidance.*  
> ACM International Conference on Multimedia (MM), 2025.

## Dataset Summary

DANI contains over **445,000 images**, including 5,000 natural images (from COCO, with resolutions 224, 256, 512, 1024) and more than 440,000 AI-generated images produced by diverse state-of-the-art generative models.  
Each sample is annotated with detailed metadata, enabling comprehensive evaluation and flexible use for a broad range of visual and multimodal research.
Images are generated using a wide range of generative models and protocols:

- **Models:** GALIP, DFGAN, SD_V14, SD_V15, Versatile Diffusion (VD), SD_V21, SD_XL, Dalle2, Dalle3, and COCO (real images)
- **Image Sizes:** 224, 256, 512, 768, 1024
- **Generation Types:** Text-to-Image (T2I), Image-to-Image (I2I), Text and Image-to-Image (TI2I)
- **Categories:** indoor, outdoor, etc.

## Data Fields

Each sample in the dataset contains the following fields:

| Field      | Description                                                                  |
|------------|------------------------------------------------------------------------------|
| index      | Unique index for each image                                                  |
| image      | The image itself (as a file, not just path)                                  |
| size       | Image resolution (e.g., 224, 256, 512, 768, 1024)                            |
| category   | Scene category (e.g., `indoor`, `outdoor`, etc.)                             |
| class_id   | COCO class or semantic category ID/name                                      |
| model      | Generative model used (`GALIP`, `DFGAN`, `SD_V14`, `SD_V15`, `VD`, etc.)     |
| gen_type   | Generation method (`T2I`, `I2I`, `TI2I`)                                     |
| reference  | Whether it is a real/natural image (`True` for real, `False` for generated)  |

> *Note:*  
> - **COCO** images have `reference=True`, and may appear at multiple resolutions.  
> - For AI-generated images, the `model` and `gen_type` fields indicate the specific generative model and generation protocol (T2I, I2I, or TI2I) used for each sample.

## Model/Generation Configurations

The dataset covers the following models and settings:

| Model    | Image Size | Generation Types Supported            |
|----------|------------|---------------------------------------|
| GALIP    | 224        | T2I                                   |
| DFGAN    | 256        | T2I                                   |
| SD_V14   | 512        | T2I, I2I, TI2I                        |
| SD_V15   | 512        | T2I, I2I, TI2I                        |
| VD       | 512        | T2I, I2I, TI2I                        |
| SD_V21   | 768        | T2I, I2I, TI2I                        |
| SD_XL    | 1024       | T2I, I2I, TI2I                        |
| Dalle2   | 512        | T2I, I2I                              |
| Dalle3   | 1024       | T2I                                   |
| COCO     | 224,256,512,1024 | Reference/Real Images           |

For each generation type (`T2I`, `I2I`, `TI2I`), a diverse set of models are covered.

## Usage

You can load DANI directly using the 🤗 datasets library:

```python
from datasets import load_dataset

ds = load_dataset("Renyang/DANI")
print(ds)
# Output: DatasetDict({
#     train: Dataset({
#         features: ['index', 'image', 'size', 'category', 'class_id','model', 'gen_type','reference'],
#         num_rows: 540257
#     })
# })
# Access images and metadata
img = ds["train"][0]["image"]
meta = {k: ds["train"][0][k] for k in ds["train"].column_names if k != "image"}

```

*Note:* Images are loaded as PIL Images. Use `.convert("RGB")` if needed.

## Citation

If you use this dataset or the associated benchmark, please cite:

```bibtex
@inproceedings{liu2024djudge,
  title = {D-Judge: How Far Are We? Assessing the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal Guidance},
  author = {Liu, Renyang and Lyu, Ziyu and Zhou, Wei and Ng, See-Kiong},
  booktitle = {ACM International Conference on Multimedia (MM)},
  organization = {ACM},
  year = {2025},
}
```

## License

This dataset is released under the [CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/) license (for non-commercial research use).

## Contact

For questions or collaborations, please visit [Renyang Liu's homepage](https://ryliu68.github.io/).