Datasets:

Modalities:
Audio
Text
Formats:
webdataset
ArXiv:
Libraries:
Datasets
WebDataset
License:
AudioX-IFcaps / README.md
Zeyue7's picture
Update README.md
35882dd verified
metadata
license: cc-by-nc-nd-4.0
task_categories:
  - text-to-audio
size_categories:
  - 1M<n<10M
pretty_name: AudioX-IFcaps

[ICLR 2026] AudioX-IFcaps: Instruction-Following Audio Caption Dataset

Project Page GitHub Paper

AudioX-IFcaps (Instruction-Following) is a large-scale, high-quality multimodal dataset designed for training unified audio and music generation models. The dataset contains over 7 million samples with fine-grained, structured annotations that enable precise control over audio generation, including sound event categories, counts, temporal ordering, and timestamps.

πŸ“Š Dataset Statistics

  • General Audio: ~1.3m 10-second video-audio clips
  • Music: ~5.7m 10-second video-music clips
  • Total Duration: ~16k hours of audio content

πŸ“ Citation

If you use this dataset in your research, please cite:

@article{tian2025audiox,
  title={Audiox: Diffusion transformer for anything-to-audio generation},
  author={Tian, Zeyue and Jin, Yizhu and Liu, Zhaoyang and Yuan, Ruibin and Tan, Xu and Chen, Qifeng and Xue, Wei and Guo, Yike},
  journal={arXiv preprint arXiv:2503.10522},
  year={2025}
}

@inproceedings{tian2025vidmuse,
  title={Vidmuse: A simple video-to-music generation framework with long-short-term modeling},
  author={Tian, Zeyue and Liu, Zhaoyang and Yuan, Ruibin and Pan, Jiahao and Liu, Qifeng and Tan, Xu and Chen, Qifeng and Xue, Wei and Guo, Yike},
  booktitle={Proceedings of the Computer Vision and Pattern Recognition Conference},
  pages={18782--18793},
  year={2025}
}

πŸ”— Related Resources


Note: This dataset is part of the AudioX project. For more information, please refer to the paper and project page.