|
|
--- |
|
|
license: cc-by-nc-4.0 |
|
|
task_categories: |
|
|
- image-classification |
|
|
- text-to-image |
|
|
library_name: datasets |
|
|
tags: |
|
|
- ai-generated-content |
|
|
- image-quality-assessment |
|
|
- real-vs-fake |
|
|
- multimodal |
|
|
--- |
|
|
|
|
|
# DANI: Discrepancy Assessing for Natural and AI Images |
|
|
|
|
|
[Paper: D-Judge: How Far Are We? Evaluating the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal Guidance](https://huggingface.co/papers/2412.17632) |
|
|
[Code: https://github.com/RenyangLiu/DJudge](https://github.com/RenyangLiu/DJudge) |
|
|
|
|
|
**A Large-Scale Dataset for Visual Research on AI-Synthesized and Natural Images** |
|
|
|
|
|
## Overview |
|
|
|
|
|
DANI (Discrepancy Assessing for Natural and AI Images) is a large-scale, multimodal dataset for benchmarking and broad visual research on both AI-generated images (AIGIs) and natural images. |
|
|
The dataset is designed to support a wide range of computer vision and multimodal research tasks, including but not limited to: |
|
|
- AI-generated vs. real image discrimination |
|
|
- Representation learning |
|
|
- Image quality assessment |
|
|
- Style transfer |
|
|
- Image reconstruction |
|
|
- Domain adaptation |
|
|
- Multimodal understanding and beyond |
|
|
|
|
|
DANI accompanies the paper: |
|
|
|
|
|
> Liu, Renyang; Lyu, Ziyu; Zhou, Wei; Ng, See-Kiong. |
|
|
> *D-Judge: How Far Are We? Assessing the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal Guidance.* |
|
|
> ACM International Conference on Multimedia (MM), 2025. |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
DANI contains over **445,000 images**, including 5,000 natural images (from COCO, with resolutions 224, 256, 512, 1024) and more than 440,000 AI-generated images produced by diverse state-of-the-art generative models. |
|
|
Each sample is annotated with detailed metadata, enabling comprehensive evaluation and flexible use for a broad range of visual and multimodal research. |
|
|
Images are generated using a wide range of generative models and protocols: |
|
|
|
|
|
- **Models:** GALIP, DFGAN, SD_V14, SD_V15, Versatile Diffusion (VD), SD_V21, SD_XL, Dalle2, Dalle3, and COCO (real images) |
|
|
- **Image Sizes:** 224, 256, 512, 768, 1024 |
|
|
- **Generation Types:** Text-to-Image (T2I), Image-to-Image (I2I), Text and Image-to-Image (TI2I) |
|
|
- **Categories:** indoor, outdoor, etc. |
|
|
|
|
|
## Data Fields |
|
|
|
|
|
Each sample in the dataset contains the following fields: |
|
|
|
|
|
| Field | Description | |
|
|
|------------|------------------------------------------------------------------------------| |
|
|
| index | Unique index for each image | |
|
|
| image | The image itself (as a file, not just path) | |
|
|
| size | Image resolution (e.g., 224, 256, 512, 768, 1024) | |
|
|
| category | Scene category (e.g., `indoor`, `outdoor`, etc.) | |
|
|
| class_id | COCO class or semantic category ID/name | |
|
|
| model | Generative model used (`GALIP`, `DFGAN`, `SD_V14`, `SD_V15`, `VD`, etc.) | |
|
|
| gen_type | Generation method (`T2I`, `I2I`, `TI2I`) | |
|
|
| reference | Whether it is a real/natural image (`True` for real, `False` for generated) | |
|
|
|
|
|
> *Note:* |
|
|
> - **COCO** images have `reference=True`, and may appear at multiple resolutions. |
|
|
> - For AI-generated images, the `model` and `gen_type` fields indicate the specific generative model and generation protocol (T2I, I2I, or TI2I) used for each sample. |
|
|
|
|
|
## Model/Generation Configurations |
|
|
|
|
|
The dataset covers the following models and settings: |
|
|
|
|
|
| Model | Image Size | Generation Types Supported | |
|
|
|----------|------------|---------------------------------------| |
|
|
| GALIP | 224 | T2I | |
|
|
| DFGAN | 256 | T2I | |
|
|
| SD_V14 | 512 | T2I, I2I, TI2I | |
|
|
| SD_V15 | 512 | T2I, I2I, TI2I | |
|
|
| VD | 512 | T2I, I2I, TI2I | |
|
|
| SD_V21 | 768 | T2I, I2I, TI2I | |
|
|
| SD_XL | 1024 | T2I, I2I, TI2I | |
|
|
| Dalle2 | 512 | T2I, I2I | |
|
|
| Dalle3 | 1024 | T2I | |
|
|
| COCO | 224,256,512,1024 | Reference/Real Images | |
|
|
|
|
|
For each generation type (`T2I`, `I2I`, `TI2I`), a diverse set of models are covered. |
|
|
|
|
|
## Usage |
|
|
|
|
|
You can load DANI directly using the 🤗 datasets library: |
|
|
|
|
|
```python |
|
|
from datasets import load_dataset |
|
|
|
|
|
ds = load_dataset("Renyang/DANI") |
|
|
print(ds) |
|
|
# Output: DatasetDict({ |
|
|
# train: Dataset({ |
|
|
# features: ['index', 'image', 'size', 'category', 'class_id','model', 'gen_type','reference'], |
|
|
# num_rows: 540257 |
|
|
# }) |
|
|
# }) |
|
|
# Access images and metadata |
|
|
img = ds["train"][0]["image"] |
|
|
meta = {k: ds["train"][0][k] for k in ds["train"].column_names if k != "image"} |
|
|
|
|
|
``` |
|
|
|
|
|
*Note:* Images are loaded as PIL Images. Use `.convert("RGB")` if needed. |
|
|
|
|
|
## Citation |
|
|
|
|
|
If you use this dataset or the associated benchmark, please cite: |
|
|
|
|
|
```bibtex |
|
|
@inproceedings{liu2024djudge, |
|
|
title = {D-Judge: How Far Are We? Assessing the Discrepancies Between AI-synthesized Images and Natural Images through Multimodal Guidance}, |
|
|
author = {Liu, Renyang and Lyu, Ziyu and Zhou, Wei and Ng, See-Kiong}, |
|
|
booktitle = {ACM International Conference on Multimedia (MM)}, |
|
|
organization = {ACM}, |
|
|
year = {2025}, |
|
|
} |
|
|
``` |
|
|
|
|
|
## License |
|
|
|
|
|
This dataset is released under the [CC BY-NC 4.0](https://creativecommons.org/licenses/by-nc/4.0/) license (for non-commercial research use). |
|
|
|
|
|
## Contact |
|
|
|
|
|
For questions or collaborations, please visit [Renyang Liu's homepage](https://ryliu68.github.io/). |