|
|
--- |
|
|
license: cc-by-nc-4.0 |
|
|
task_categories: |
|
|
- audio-classification |
|
|
- automatic-speech-recognition |
|
|
- question-answering |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- Audio |
|
|
- Multi-modal Large Language Models |
|
|
modalities: |
|
|
- audio |
|
|
- text |
|
|
size_categories: |
|
|
- 1K<n<10K |
|
|
--- |
|
|
# AudioMarathon: A Comprehensive Long-Form Audio Understanding Benchmark |
|
|
|
|
|
## Abstract |
|
|
|
|
|
**AudioMarathon** is a large-scale, multi-task audio understanding benchmark designed to systematically evaluate audio language models' capabilities in processing and comprehending long-form audio content. It provides a diverse set of **10** tasks built upon three pillars: |
|
|
long-context audio inputs with durations ranging from **90.0 to 300.0** seconds, which |
|
|
correspond to encoded sequences of 2,250 to 7,500 audio tokens, respectively, full |
|
|
domain coverage across speech, sound, and music, and complex reasoning that |
|
|
requires multi-hop inference. |
|
|
|
|
|
|
|
|
## π Task Taxonomy & Statistics |
|
|
|
|
|
### Task Categories |
|
|
|
|
|
AudioMarathon organizes tasks into four primary categories: |
|
|
|
|
|
1. **Speech Content Extraction** - |
|
|
2. **Audio Classification** |
|
|
3. **Speaker Information Modeling** |
|
|
|
|
|
### Dataset Statistics |
|
|
|
|
|
| Task ID | Dataset | Task Type | # Samples | Duration | Format | License | Status | |
|
|
| ------- | --------------------------------------- | ---------------------------------- | --------- | ------------ | ---------- | ------------- | ----------- | |
|
|
| 1 | [LibriSpeech-long](#1-librispeech-long) | Automatic Speech Recognition (ASR) | 204 | 1-4min | FLAC 16kHz | CC BY 4.0 | β
Full | |
|
|
| 2 | [RACE](#2-race) | Speech Content Reasoning (SCR) | 820 | 2-4.22min | WAV 16kHz | Apache-2.0 | β
Full | |
|
|
| 3 | [HAD](#3-had) | Speech Detection (SD) | 776 | 3~5min | WAV 16kHz | CC BY 4.0 | β
Full | |
|
|
| 4 | [GTZAN](#4-gtzan) | Music classifier (MC) | 120 | 4min | WAV 22kHz | Research Only | β
Full | |
|
|
| 5 | [TAU](#5-tau) | Audio scene classifier (ASC) | 1145 | 1.5-3.5min | WAV 16kHz | CC BY 4.0 | β
Full | |
|
|
| 6 | [VESUS](#6-vesus) | Emotion Recognition (ER) | 185 | 1.5-2min | WAV 16kHz | Academic Only | β
Full | |
|
|
| 7 | [SLUE](#7-slue) | Speech Entity Recognition (SER) | 490 | 2.75~5min | WAV 16kHz | CC BY 4.0 | β
Full | |
|
|
| 8 | [DESED](#8-desed) | Sound event detection (SED) | 254 | 4.5-5min | WAV 16kHz | Mixed CC* | β
Full | |
|
|
| 9 | [VoxCeleb-Gender](#9-voxceleb-gender) | Speaker Gender Recognition (SGR) | 1614 | 1.5-3.5min | WAV 16kHz | CC BY 4.0 | β
Full | |
|
|
| 10 | [VoxCeleb-Age](#10-voxceleb-age) | Speaker Age Recognition (SAR) | 959 | 1.5-3.5min | WAV 16kHz | CC BY 4.0 | β
Full | |
|
|
|
|
|
**Total**: 6567 samples | ~60GB |
|
|
|
|
|
\* *DESED requires per-clip Freesound attribution (CC0/CC BY 3.0/4.0)* |
|
|
|
|
|
--- |
|
|
|
|
|
## π― Benchmark Objectives |
|
|
|
|
|
AudioMarathon is designed to evaluate: |
|
|
|
|
|
1. **Long-Audio Processing**: Ability to maintain coherence across extended audio sequences |
|
|
2. **Multi-Domain Generalization**: Performance across diverse acoustic environments and tasks |
|
|
3. **Semantic Understanding**: Comprehension of spoken content, not just acoustic patterns |
|
|
4. **Efficiency**: Computational requirements for long-form audio processing |
|
|
|
|
|
--- |
|
|
|
|
|
## π Directory Structure |
|
|
|
|
|
``` |
|
|
Dataset/ |
|
|
βββ librispeech-long/ # Automatic Speech Recognition |
|
|
β βββ README.md |
|
|
β βββ test-clean/ # Clean test set |
|
|
β βββ test-other/ # Noisy test set |
|
|
β βββ dev-clean/ # Clean dev set |
|
|
β βββ dev-other/ # Noisy dev set |
|
|
β |
|
|
βββ race_audio/ # Reading Comprehension |
|
|
β βββ race_benchmark.json # Task metadata |
|
|
β βββ test/ # Audio articles |
|
|
β βββ article_*/ |
|
|
β |
|
|
βββ HAD/ # Half-truth Audio Detection |
|
|
β βββ concatenated_audio/ |
|
|
β βββ had_audio_classification_task.json |
|
|
β βββ real/ # Authentic audio |
|
|
β βββ fake/ # Synthesized audio |
|
|
β |
|
|
βββ GTZAN/ # Music Genre Classification |
|
|
β βββ concatenated_audio/ |
|
|
β βββ music_genre_classification_meta.json |
|
|
β βββ wav/ # Genre-labeled music clips |
|
|
β |
|
|
βββ TAU/ # Acoustic Scene Classification |
|
|
β βββ acoustic_scene_task_meta.json |
|
|
β βββ LICENSE |
|
|
β βββ README.md |
|
|
β βββ concatenated_resampled/ |
|
|
β |
|
|
βββ VESUS/ # Emotion Recognition |
|
|
β βββ audio_emotion_dataset.json |
|
|
β βββ [1-10]/ # Speaker directories |
|
|
β |
|
|
βββ SLUE/ # Named Entity Recognition |
|
|
β βββ merged_audio_data.json |
|
|
β βββ dev/ |
|
|
β βββ test/ |
|
|
β βββ fine-tune/ |
|
|
β |
|
|
βββ DESED/ # Sound Event Detection |
|
|
β βββ DESED_dataset/ |
|
|
β βββ license_public_eval.tsv |
|
|
β βββ concatenated_audio/ |
|
|
β |
|
|
βββ VoxCeleb/ # Speaker Recognition |
|
|
β βββ concatenated_audio/ |
|
|
β β βββ gender_id_task_meta.json |
|
|
β βββ concatenated_audio_age/ |
|
|
β β βββ age_classification_task_meta.json |
|
|
β βββ txt/ |
|
|
β |
|
|
βββ README.md # This file |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## π― Dataset Details |
|
|
|
|
|
### 1. LibriSpeech-long |
|
|
|
|
|
**Task**: Automatic Speech Recognition (ASR) |
|
|
**Description**: Long-form English speech from audiobooks |
|
|
**Format**: FLAC files with `.trans.txt` transcriptions |
|
|
**Splits**: test-clean, test-other, dev-clean, dev-other |
|
|
**License**: CC BY 4.0 |
|
|
**Source**: https://github.com/google-deepmind/librispeech-long |
|
|
|
|
|
**Structure**: |
|
|
``` |
|
|
librispeech-long/ |
|
|
test-clean/ |
|
|
<speaker_id>/ |
|
|
<chapter_id>/ |
|
|
<speaker>-<chapter>-<utterance>.flac |
|
|
<speaker>-<chapter>.trans.txt |
|
|
``` |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@article{kahn2020libri, |
|
|
title={Libri-light: A benchmark for asr with limited or no supervision}, |
|
|
author={Kahn, Jacob and Rivière, Morgane and Zheng, Weiyi and Khudanpur, Sanjeev and others}, |
|
|
journal={ICASSP 2020}, |
|
|
year={2020} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 2. RACE |
|
|
|
|
|
**Task**: Reading Comprehension from Audio |
|
|
**Description**: Multiple-choice questions based on audio passages |
|
|
**Format**: WAV files + JSON metadata |
|
|
**Sample Count**: ~200 articles |
|
|
**License**: Apache-2.0 (verify) |
|
|
**Source**: https://huggingface.co/datasets/ehovy/race |
|
|
|
|
|
**JSON Format**: |
|
|
```json |
|
|
{ |
|
|
"article_id": 7870154, |
|
|
"audio_path": "test/article_7870154/audio.wav", |
|
|
"question": "What did the author do...?", |
|
|
"options": ["A", "B", "C", "D"], |
|
|
"answer": "A" |
|
|
} |
|
|
``` |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@inproceedings{lai2017race, |
|
|
title={RACE: Large-scale reading comprehension dataset from examinations}, |
|
|
author={Lai, Guokun and Xie, Qizhe and Liu, Hanxiao and Yang, Yiming and Hovy, Eduard}, |
|
|
booktitle={EMNLP}, |
|
|
year={2017} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 3. HAD |
|
|
|
|
|
**Task**: Half-truth Audio Detection |
|
|
**Description**: Classify audio as real or containing synthesized segments |
|
|
**License**: CC BY 4.0 |
|
|
**Source**: https://zenodo.org/records/10377492 |
|
|
|
|
|
**JSON Format**: |
|
|
```json |
|
|
{ |
|
|
"path": "real/HAD_train_real_249.wav", |
|
|
"question": "Is this audio authentic or fake?", |
|
|
"choice_a": "Real", |
|
|
"choice_b": "Fake", |
|
|
"answer_gt": "Real", |
|
|
"duration_seconds": 297.78 |
|
|
} |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
### 4. GTZAN |
|
|
|
|
|
**Task**: Music Genre Classification |
|
|
**Description**: 10-genre music classification dataset |
|
|
**Genres**: blues, classical, country, disco, hiphop, jazz, metal, pop, reggae, rock |
|
|
**β οΈ License**: Research Use Only |
|
|
**Source**: https://www.kaggle.com/datasets/andradaolteanu/gtzan-dataset-music-genre-classification |
|
|
|
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@inproceedings{tzanetakis2002musical, |
|
|
title={Musical genre classification of audio signals}, |
|
|
author={Tzanetakis, George and Cook, Perry}, |
|
|
booktitle={IEEE Transactions on Speech and Audio Processing}, |
|
|
year={2002} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 5. TAU |
|
|
|
|
|
**Task**: Acoustic Scene Classification |
|
|
**Description**: Urban sound scene recognition |
|
|
**Scenes**: airport, bus, metro, park, public_square, shopping_mall, street_pedestrian, street_traffic, tram |
|
|
**License**: CC BY 4.0 |
|
|
**Source**: https://zenodo.org/records/7870258 |
|
|
|
|
|
**Files**: |
|
|
- `acoustic_scene_task_meta.json`: Task metadata |
|
|
- `LICENSE`: Original license text |
|
|
- `concatenated_resampled/`: Resampled audio files |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@article{mesaros2018detection, |
|
|
title={Detection and classification of acoustic scenes and events: Outcome of the DCASE 2016 challenge}, |
|
|
author={Mesaros, Annamaria and Heittola, Toni and Virtanen, Tuomas}, |
|
|
journal={IEEE/ACM TASLP}, |
|
|
year={2018} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 6. VESUS |
|
|
|
|
|
**Task**: Emotion Recognition from Speech |
|
|
**Description**: Actors reading neutral script with emotional inflections |
|
|
**Emotions**: neutral, angry, happy, sad, fearful |
|
|
**Actors**: 10 (5 male, 5 female) |
|
|
**β οΈ License**: Academic Use Only (access by request) |
|
|
**Source**: https://engineering.jhu.edu/nsa/vesus/ |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@inproceedings{sager2019vesus, |
|
|
title={VESUS: A crowd-annotated database to study emotion production and perception in spoken English}, |
|
|
author={Sager, Jennifer and Shankar, Raghav and Reinhold, Jacob and Venkataraman, Archana}, |
|
|
booktitle={Interspeech}, |
|
|
year={2019} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 7. SLUE |
|
|
|
|
|
**Task**: Named Entity Recognition (NER) from Speech |
|
|
**Description**: Count named entities in audio segments |
|
|
**Entity Types**: LAW, NORP, ORG, PLACE, QUANT, WHEN |
|
|
**License**: CC BY 4.0 (VoxPopuli-derived) |
|
|
**Source**: https://arxiv.org/abs/2111.10367 |
|
|
|
|
|
**JSON Format**: |
|
|
```json |
|
|
{ |
|
|
"path": "dev/concatenated_audio_with/concatenated_audio_0000.wav", |
|
|
"question": "How many named entities appear?", |
|
|
"options": ["49 entities", "51 entities", "52 entities", "46 entities"], |
|
|
"answer_gt": "D", |
|
|
"entity_count": 49 |
|
|
} |
|
|
``` |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@article{shon2022slue, |
|
|
title={SLUE: New benchmark tasks for spoken language understanding evaluation on natural speech}, |
|
|
author={Shon, Suwon and Pasad, Ankita and Wu, Felix and others}, |
|
|
journal={ICASSP 2022}, |
|
|
year={2022} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 8. DESED |
|
|
|
|
|
**Task**: Sound Event Detection |
|
|
**Description**: Detect domestic sound events |
|
|
**Events**: Alarm bell, Blender, Cat, Dishes, Dog, Electric shaver, Frying, Running water, Speech, Vacuum cleaner |
|
|
**License**: Mixed CC (Freesound sources: CC0, CC BY 3.0/4.0) |
|
|
**Source**: https://github.com/turpaultn/DESED |
|
|
|
|
|
**β οΈ ATTRIBUTION REQUIRED**: |
|
|
- Audio clips sourced from Freesound.org |
|
|
- Each clip has individual CC license |
|
|
- Must maintain attribution when redistributing |
|
|
- See `license_public_eval.tsv` for per-file credits |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@inproceedings{turpault2019sound, |
|
|
title={Sound event detection in domestic environments with weakly labeled data and soundscape synthesis}, |
|
|
author={Turpault, Nicolas and Serizel, Romain and Salamon, Justin and Shah, Ankit Parag}, |
|
|
booktitle={DCASE Workshop}, |
|
|
year={2019} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
### 9. VoxCeleb-Gender |
|
|
|
|
|
**Task**: Speaker Gender Identification |
|
|
**Description**: Binary classification (male/female) |
|
|
**License**: CC BY 4.0 |
|
|
**Source**: https://www.robots.ox.ac.uk/~vgg/data/voxceleb/ |
|
|
|
|
|
**JSON Format**: |
|
|
```json |
|
|
{ |
|
|
"path": "concatenated_audio/speaker_001.wav", |
|
|
"question": "What is the gender of the speaker?", |
|
|
"choice_a": "Male", |
|
|
"choice_b": "Female", |
|
|
"answer_gt": "A" |
|
|
} |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
### 10. VoxCeleb-Age |
|
|
|
|
|
**Task**: Speaker Age Classification |
|
|
**Description**: Multi-class age group classification |
|
|
**Age Groups**: 20s, 30s, 40s, 50s, 60s, 70s |
|
|
**License**: CC BY 4.0 |
|
|
**Source**: VoxCeleb + https://github.com/hechmik/voxceleb_enrichment_age_gender |
|
|
|
|
|
**Note**: Age/gender labels are derivative annotations on VoxCeleb corpus |
|
|
|
|
|
<!-- **Citation**: |
|
|
```bibtex |
|
|
@inproceedings{nagrani2017voxceleb, |
|
|
title={VoxCeleb: a large-scale speaker identification dataset}, |
|
|
author={Nagrani, Arsha and Chung, Joon Son and Zisserman, Andrew}, |
|
|
booktitle={Interspeech}, |
|
|
year={2017} |
|
|
} |
|
|
``` --> |
|
|
|
|
|
--- |
|
|
|
|
|
## π§ Usage Guidelines |
|
|
|
|
|
You can load the dataset via Hugging Face datasets: |
|
|
|
|
|
from datasets import load_dataset |
|
|
ds = load_dataset("Hezep/AudioMarathon") |
|
|
|
|
|
--- |
|
|
|
|
|
|
|
|
|
|
|
### Special Requirements |
|
|
|
|
|
### Disclaimer |
|
|
|
|
|
This benchmark is provided "AS IS" without warranty. Users bear sole responsibility for: |
|
|
- License compliance verification |
|
|
- Obtaining restricted datasets independently |
|
|
- Proper attribution maintenance |
|
|
- Determining fitness for specific use cases |
|
|
|
|
|
--- |
|
|
|
|
|
## π Benchmark Statistics |
|
|
|
|
|
### Overview |
|
|
|
|
|
| Metric | Value | |
|
|
| -------------- | ---------------------------------------- | |
|
|
| Total Tasks | 10 | |
|
|
| Total Samples | 6567 | |
|
|
| Total Duration | 392h | |
|
|
| Total Size | ~60 GB | |
|
|
| Languages | English | |
|
|
| Domains | Speech, Music, Soundscape, Environmental | |
|
|
| Task Types | Classification (7), QA (2), ASR (1) | |
|
|
|
|
|
### Audio Characteristics |
|
|
|
|
|
| Property | Range | Predominant | |
|
|
|----------|-------|-------------| |
|
|
| Sampling Rate | 16 kHz - 22.05 kHz | 16 kHz (90%) | |
|
|
| Duration | 30s - 5+ min | 2-3 min (avg) | |
|
|
| Channels | Mono | Mono (100%) | |
|
|
| Format | FLAC, WAV | WAV (80%) | |
|
|
| Bit Depth | 16-bit | 16-bit (100%) | |
|
|
|
|
|
### Task Distribution |
|
|
|
|
|
| Category | # Tasks | # Samples | % of Total | |
|
|
| ------------------------ | ------- | --------- | ---------- | |
|
|
| Speech Understanding | 3 | 1514 | 23% | |
|
|
| Acoustic Analysis | 3 | 1519 | 23% | |
|
|
| Speaker Characterization | 3 | 2758 | 42% | |
|
|
| Content Authenticity | 1 | 776 | 12% | |
|
|
|
|
|
--- |
|
|
|
|
|
## π Related Resources |
|
|
|
|
|
- **GitHub Repository**: https://github.com/DabDans/AudioMarathon |
|
|
- **Paper**: [] |
|
|
|
|
|
--- |
|
|
|
|
|
## π Citation |
|
|
|
|
|
If you use AudioMarathon in your research, please cite: |
|
|
|
|
|
|
|
|
|
|
|
### Citing Component Datasets |
|
|
|
|
|
When using specific tasks, please also cite the original datasets (see individual task documentation above for BibTeX entries): |
|
|
|
|
|
- **LibriSpeech**: Panayotov et al. (2015) |
|
|
- **RACE**: Lai et al. (2017) |
|
|
- **HAD**: Zenodo record 10377492 |
|
|
- **GTZAN**: Tzanetakis & Cook (2002) |
|
|
- **TAU**: DCASE Challenge (Mesaros et al., 2018) |
|
|
- **VESUS**: Sager et al. (2019) |
|
|
- **SLUE**: Shon et al. (2022) |
|
|
- **DESED**: Turpault et al. (2019) |
|
|
- **VoxCeleb**: Nagrani et al. (2017, 2018) |
|
|
|
|
|
Full BibTeX entries available in individual task sections. |
|
|
|
|
|
--- |
|
|
|
|
|
## π€ Contributing & Support |
|
|
|
|
|
|
|
|
|
|
|
## π§ Contact |
|
|
|
|
|
|
|
|
- **GitHub Issues**: https://github.com/DabDans/AudioMarathon/issues |
|
|
|
|
|
--- |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
## π Acknowledgments |
|
|
|
|
|
AudioMarathon builds upon the pioneering work of numerous research teams. We gratefully acknowledge the creators of: |
|
|
|
|
|
- LibriSpeech (Panayotov et al.) |
|
|
- RACE (Lai et al.) |
|
|
- HAD (Zenodo contributors) |
|
|
- GTZAN (Tzanetakis & Cook) |
|
|
- TAU/DCASE (Mesaros et al.) |
|
|
- VESUS (Sager et al., JHU) |
|
|
- SLUE (Shon et al.) |
|
|
- DESED (Turpault et al. & Freesound community) |
|
|
- VoxCeleb (Nagrani et al., Oxford VGG) |
|
|
|
|
|
Their datasets enable comprehensive audio understanding research. |
|
|
|
|
|
--- |
|
|
|
|
|
<p align="center"> |
|
|
<b>AudioMarathon</b><br> |
|
|
A Comprehensive Long-Form Audio Understanding Benchmark<br> |
|
|
<i>Version 1.0.0 | October 2025</i> |
|
|
</p> |