Datasets:

Modalities:
Audio
Text
Formats:
webdataset
ArXiv:
Libraries:
Datasets
WebDataset
License:
File size: 2,525 Bytes
35882dd
 
 
 
 
 
 
 
 
9bed078
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
31688d7
9bed078
35882dd
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
license: cc-by-nc-nd-4.0
task_categories:
- text-to-audio
size_categories:
- 1M<n<10M
pretty_name: AudioX-IFcaps
---



# [ICLR 2026] AudioX-IFcaps: Instruction-Following Audio Caption Dataset

<a href="https://zeyuet.github.io/AudioX/" target="_blank"><img src="https://img.shields.io/badge/🌐%20Project%20Page-blue" alt="Project Page"></a>
<a href="https://github.com/ZeyueT/AudioX" target="_blank"><img src="https://img.shields.io/badge/πŸ’»%20GitHub-ffffff?logo=github&logoColor=181717" alt="GitHub"></a>
<a href="https://arxiv.org/pdf/2503.10522" target="_blank"><img src="https://img.shields.io/badge/πŸ“„%20Paper-ICLR%202026-red" alt="Paper"></a>

**AudioX-IFcaps** (Instruction-Following) is a large-scale, high-quality multimodal dataset designed for training unified audio and music generation models. The dataset contains over **7 million samples** with fine-grained, structured annotations that enable precise control over audio generation, including sound event categories, counts, temporal ordering, and timestamps.

## πŸ“Š Dataset Statistics

- **General Audio**: ~1.3m 10-second video-audio clips
- **Music**: ~5.7m 10-second video-music clips
- **Total Duration**: ~16k hours of audio content

## πŸ“ Citation

If you use this dataset in your research, please cite:

```bibtex
@article{tian2025audiox,
  title={Audiox: Diffusion transformer for anything-to-audio generation},
  author={Tian, Zeyue and Jin, Yizhu and Liu, Zhaoyang and Yuan, Ruibin and Tan, Xu and Chen, Qifeng and Xue, Wei and Guo, Yike},
  journal={arXiv preprint arXiv:2503.10522},
  year={2025}
}

@inproceedings{tian2025vidmuse,
  title={Vidmuse: A simple video-to-music generation framework with long-short-term modeling},
  author={Tian, Zeyue and Liu, Zhaoyang and Yuan, Ruibin and Pan, Jiahao and Liu, Qifeng and Tan, Xu and Chen, Qifeng and Xue, Wei and Guo, Yike},
  booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
  pages={18782--18793},
  year={2025}
}
```



## πŸ”— Related Resources

- **Paper**: <a href="https://arxiv.org/pdf/2503.10522" target="_blank">AudioX: Diffusion Transformer for Anything-to-Audio Generation</a> (Accepted to ICLR 2026)
- **Project Page**: <a href="https://zeyuet.github.io/AudioX/" target="_blank">https://zeyuet.github.io/AudioX/</a>
- **Code**: <a href="https://github.com/ZeyueT/AudioX" target="_blank">GitHub Repository</a>

---

**Note**: This dataset is part of the AudioX project. For more information, please refer to the paper and project page.