Hui519 commited on
Commit
5582826
·
verified ·
1 Parent(s): 813b1b0

Update dataset card for AudioEval

Browse files
Files changed (1) hide show
  1. README.md +161 -33
README.md CHANGED
@@ -1,7 +1,20 @@
1
  ---
2
  pretty_name: AudioEval
 
3
  size_categories:
4
  - 1K<n<10K
 
 
 
 
 
 
 
 
 
 
 
 
5
  configs:
6
  - config_name: default
7
  data_files:
@@ -16,50 +29,165 @@ configs:
16
 
17
  # AudioEval
18
 
19
- AudioEval is a text-to-audio evaluation benchmark packaged for a private Hugging Face dataset repository.
20
- This release package keeps the audio files and split metadata in an `AudioFolder` layout so the dataset can be previewed on the Hub and later switched to public visibility without repacking.
21
 
22
- ## What is included
23
 
24
- - `data/train`, `data/validation`, `data/test`: split directories with `.wav` files and `metadata.jsonl`.
25
- - `annotations/ratings.csv`: anonymized per-rater scores with stable `rater_id` values.
26
- - `annotations/system_info.csv`: mapping from `system_id` to the evaluated model name.
27
- - `annotations/prompts.tsv`: prompt metadata used to generate the audio clips.
28
- - `stats/*.csv`: reliability and model summary tables from the local analysis pipeline.
 
 
 
29
 
30
- ## Dataset size
31
 
32
- - 3360 train clips, 420 validation clips, and 420 test clips.
33
- - 4200 total audio clips from 24 text-to-audio systems and 451 prompts.
34
- - 11.712 total hours of audio.
35
- - 25200 total ratings from 12 raters.
36
- - Rating rows by rater type: common=12600, pro=12600.
37
- - Unique raters by type: common=9, pro=3.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
 
39
- ## Main split metadata fields
 
 
 
40
 
41
- - `file_name`: relative path to the audio file inside each split directory.
42
- - `prompt_id`, `prompt_text`, `scene_category`, `sound_event_count`, `audioset_ontology`: prompt-side metadata.
43
- - `system_id`, `system_name`: evaluated generation system metadata.
44
- - `common_*_mean`, `pro_*_mean`: mean scores from common and professional raters.
45
- - `common_*_raw_scores`, `pro_*_raw_scores`: raw score lists per clip for each rating axis.
46
 
47
- The five rating axes are:
48
 
49
- - `production_complexity`
50
- - `content_enjoyment`
51
- - `production_quality`
52
- - `textual_alignment`
53
- - `content_usefulness`
54
 
55
- ## Privacy and release status
56
 
57
- - This package is intended for a private Hugging Face dataset repo first.
58
- - Rater demographic tables are intentionally omitted from this release package.
59
- - Per-rater annotations are anonymized before export.
60
- - Update the license and citation sections before switching the repo to public visibility.
 
 
 
 
 
61
 
62
- ## Notes
63
 
64
  - The original source directory in this workspace remains unchanged.
65
  - The export directory can be regenerated by re-running `scripts/prepare_hf_audioeval_dataset.py`.
 
 
1
  ---
2
  pretty_name: AudioEval
3
+ license: other
4
  size_categories:
5
  - 1K<n<10K
6
+ source_datasets:
7
+ - original
8
+ annotations_creators:
9
+ - expert-generated
10
+ - crowdsourced
11
+ tags:
12
+ - audio
13
+ - text-to-audio
14
+ - benchmark
15
+ - evaluation
16
+ - human-ratings
17
+ - multimodal
18
  configs:
19
  - config_name: default
20
  data_files:
 
29
 
30
  # AudioEval
31
 
32
+ AudioEval is a benchmark for evaluating text-to-audio generation systems from two listener perspectives (`common` and `pro`) and five perceptual dimensions: `production_complexity`, `content_enjoyment`, `production_quality`, `textual_alignment`, and `content_usefulness`.
33
+ This repository is packaged in an `AudioFolder`-compatible layout so it can be previewed directly on the Hugging Face Hub and later switched from private to public visibility without repacking.
34
 
35
+ ## Dataset Summary
36
 
37
+ - 4200 generated audio clips from 24 text-to-audio systems and 451 prompts.
38
+ - 11.712 total hours of audio, with an average clip duration of 10.039 seconds.
39
+ - Official splits: train=3360, validation=420, test=420.
40
+ - 25200 per-rater annotation rows in `annotations/ratings.csv`.
41
+ - 126000 dimension-level human scores overall, because each annotation row contains five dimension scores.
42
+ - 12 raters in total: common=9, pro=3.
43
+ - Rating rows by rater type: common=12600, pro=12600.
44
+ - Prompts span five coarse scene categories: daily life, art and cultural, natural and outdoor, work and production, and transportation and travel.
45
 
46
+ ## What Is In This Repository
47
 
48
+ - `data/train`, `data/validation`, `data/test`: split directories containing `.wav` files and a `metadata.jsonl` file for each split.
49
+ - `annotations/ratings.csv`: anonymized per-rater annotations.
50
+ - `annotations/prompts.tsv`: prompt text plus prompt-side metadata such as `scene_category` and `audioset_ontology`.
51
+ - `annotations/system_info.csv`: mapping from `system_id` to the evaluated system name.
52
+ - `annotations/release_summary.json`: summary counts for this packaged release.
53
+ - `stats/*.csv`: analysis tables such as ICC, Krippendorff's alpha, significance tests, and model result summaries.
54
+
55
+ ## Suggested Uses
56
+
57
+ - Benchmarking automatic evaluators for text-to-audio generation.
58
+ - Training clip-level regression or distribution-prediction models from prompt-audio pairs.
59
+ - Studying rating differences between professional and non-professional listeners.
60
+ - Analyzing inter-rater reliability, disagreement, and cross-dimension correlations.
61
+
62
+ ## Split Sizes
63
+
64
+ | Split | Clips |
65
+ | --- | ---: |
66
+ | train | 3360 |
67
+ | validation | 420 |
68
+ | test | 420 |
69
+
70
+ ## Data Structure
71
+
72
+ ### Split-Level Metadata
73
+
74
+ Each row in `data/*/metadata.jsonl` corresponds to one generated audio clip.
75
+
76
+ | Field | Description |
77
+ | --- | --- |
78
+ | `file_name` | Relative path to the audio file inside the split directory. |
79
+ | `wav_name` | Original clip filename. |
80
+ | `split` | One of `train`, `validation`, or `test`. |
81
+ | `prompt_id` | Prompt identifier such as `P0004` or `N001`. |
82
+ | `prompt_text` | Natural language prompt used to generate the clip. |
83
+ | `scene_category` | Coarse manually assigned prompt category. |
84
+ | `sound_event_count` | Number of sound events recorded in the prompt metadata. |
85
+ | `audioset_ontology` | Prompt-side ontology category. |
86
+ | `system_id` | System identifier such as `S001`. |
87
+ | `system_name` | Human-readable system name. |
88
+ | `num_ratings_common` | Number of common-rater judgments for the clip. |
89
+ | `num_ratings_pro` | Number of professional-rater judgments for the clip. |
90
+ | `num_ratings_overall` | Total number of judgments for the clip. |
91
+ | `common_*_mean` | Mean score from common raters for a given dimension. |
92
+ | `pro_*_mean` | Mean score from professional raters for a given dimension. |
93
+ | `common_*_raw_scores` | List of raw common-rater scores for a given dimension. |
94
+ | `pro_*_raw_scores` | List of raw professional-rater scores for a given dimension. |
95
+
96
+ The five dimensions are `production_complexity`, `content_enjoyment`, `production_quality`, `textual_alignment`, and `content_usefulness`.
97
+
98
+ ### Per-Rater Annotations
99
+
100
+ Each row in `annotations/ratings.csv` contains one rater's five-dimensional judgment for one clip.
101
+
102
+ | Field | Description |
103
+ | --- | --- |
104
+ | `wav_name` | Clip filename. |
105
+ | `split` | Dataset split for the clip. |
106
+ | `prompt_id` | Prompt identifier. |
107
+ | `system_id` | System identifier. |
108
+ | `rater_type` | `common` or `pro`. |
109
+ | `rater_id` | Stable anonymized rater identifier within this release. |
110
+ | `production_complexity` | Integer score from 1 to 10. |
111
+ | `content_enjoyment` | Integer score from 1 to 10. |
112
+ | `production_quality` | Integer score from 1 to 10. |
113
+ | `textual_alignment` | Integer score from 1 to 10. |
114
+ | `content_usefulness` | Integer score from 1 to 10. |
115
+
116
+ ## Loading the Data
117
+
118
+ Once you have access to the repository on the Hub, you can load the split metadata and audio files with `datasets`:
119
+
120
+ ```python
121
+ from datasets import load_dataset
122
+
123
+ train = load_dataset("Hui519/AudioEval", split="train")
124
+ print(train[0]["audio"])
125
+ print(train[0]["prompt_text"])
126
+ ```
127
+
128
+ The anonymized annotation table can be loaded separately from `annotations/ratings.csv`.
129
+
130
+ ## Dataset Creation
131
+
132
+ AudioEval is a benchmark release of machine-generated audio, not a collection of source recordings.
133
+ Each clip is paired with its generation prompt, generating system identifier, and human evaluation results.
134
+ This packaged Hub release was assembled from locally processed benchmark assets by:
135
+
136
+ - standardizing prompt metadata into `annotations/prompts.tsv`,
137
+ - merging raw MOS tables into a unified annotation table,
138
+ - exporting split-level clip metadata into `metadata.jsonl`,
139
+ - and anonymizing rater identities before publication.
140
+
141
+ ## Annotation Process
142
+
143
+ - The release contains two listener groups: `common` and `pro`.
144
+ - There are 9 common raters and 3 professional raters.
145
+ - Each clip receives 3 common-rater and 3 professional-rater judgments for every dimension.
146
+ - The five dimensions are production complexity, content enjoyment, production quality, textual alignment, and content usefulness.
147
+ - Score values are integers from 1 to 10.
148
+
149
+ ## Biases, Risks, and Limitations
150
+
151
+ - Prompt coverage is broad but not uniform across scene categories or sound-event complexity.
152
+ - The prompts are written in English, but generated audio may contain non-speech sounds, speech-like content, music, or multilingual vocal content.
153
+ - The benchmark is designed for evaluation research on text-to-audio systems and should not be treated as a general-purpose audio understanding dataset.
154
+ - Because the audio is machine-generated, outputs may contain artifacts, distorted speech, or unsafe-sounding events that require additional filtering for downstream applications.
155
+ - Anonymized `rater_id` values are release-specific identifiers and should not be linked back to source identities.
156
+
157
+ ## Privacy and Sensitive Information
158
+
159
+ - This release intentionally excludes the original rater demographic tables.
160
+ - No direct personally identifying information is included in the published files.
161
+ - `annotations/ratings.csv` keeps only anonymized rater IDs, rater type, system ID, prompt ID, split, and scores.
162
+ - Although the audio is machine-generated, some clips may depict or imitate human speech, children, alarms, animals, vehicles, or other safety-relevant scenarios.
163
+
164
+ ## License
165
 
166
+ The repository is currently private.
167
+ The YAML metadata uses `license: other` because the final public release license has not been fixed yet.
168
+ Until a public license is selected, access and reuse are governed by the repository owner's access settings and release terms.
169
+ Before switching the repository to public visibility, replace this section and the YAML metadata with the final license identifier if you choose a standard Hugging Face-supported license.
170
 
171
+ ## Citation
 
 
 
 
172
 
173
+ If you use AudioEval, cite the paper below and, when relevant, the Hugging Face dataset repository.
174
 
175
+ Paper:
 
 
 
 
176
 
177
+ - Hui Wang, Jinghua Zhao, Junyang Cheng, Cheng Liu, Yuhang Jia, Haoqin Sun, Jiaming Zhou, and Yong Qin. *AudioEval: Automatic Dual-Perspective and Multi-Dimensional Evaluation of Text-to-Audio-Generation*. arXiv:2510.14570, 2025. DOI: 10.48550/arXiv.2510.14570.
178
 
179
+ ```bibtex
180
+ @article{wang2025audioeval,
181
+ title={AudioEval: Automatic Dual-Perspective and Multi-Dimensional Evaluation of Text-to-Audio-Generation},
182
+ author={Wang, Hui and Zhao, Jinghua and Cheng, Junyang and Liu, Cheng and Jia, Yuhang and Sun, Haoqin and Zhou, Jiaming and Qin, Yong},
183
+ journal={arXiv preprint arXiv:2510.14570},
184
+ year={2025},
185
+ doi={10.48550/arXiv.2510.14570}
186
+ }
187
+ ```
188
 
189
+ ## Release Status
190
 
191
  - The original source directory in this workspace remains unchanged.
192
  - The export directory can be regenerated by re-running `scripts/prepare_hf_audioeval_dataset.py`.
193
+ - This card is written so the same package can remain useful both while private and after a later public release.