Hui519 commited on
Commit
5fe45fb
·
verified ·
1 Parent(s): 5582826

Upload AudioEval dataset package

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +35 -139
  2. data/{test/S001_P0004.wav → S001_P0004.wav} +0 -0
  3. data/{train/S001_P0008.wav → S001_P0008.wav} +0 -0
  4. data/{train/S001_P0013.wav → S001_P0013.wav} +0 -0
  5. data/{train/S001_P0014.wav → S001_P0014.wav} +0 -0
  6. data/{train/S001_P0016.wav → S001_P0016.wav} +0 -0
  7. data/{train/S001_P0017.wav → S001_P0017.wav} +0 -0
  8. data/{train/S001_P0023.wav → S001_P0023.wav} +0 -0
  9. data/{train/S001_P0041.wav → S001_P0041.wav} +0 -0
  10. data/{train/S001_P0045.wav → S001_P0045.wav} +0 -0
  11. data/{train/S001_P0048.wav → S001_P0048.wav} +0 -0
  12. data/{train/S001_P0050.wav → S001_P0050.wav} +0 -0
  13. data/{train/S001_P0053.wav → S001_P0053.wav} +0 -0
  14. data/{train/S001_P0058.wav → S001_P0058.wav} +0 -0
  15. data/{train/S001_P0064.wav → S001_P0064.wav} +0 -0
  16. data/{train/S001_P0072.wav → S001_P0072.wav} +0 -0
  17. data/{validation/S001_P0080.wav → S001_P0080.wav} +0 -0
  18. data/{train/S001_P0082.wav → S001_P0082.wav} +0 -0
  19. data/{train/S001_P0094.wav → S001_P0094.wav} +0 -0
  20. data/{train/S001_P0102.wav → S001_P0102.wav} +0 -0
  21. data/{validation/S001_P0111.wav → S001_P0111.wav} +0 -0
  22. data/{validation/S001_P0112.wav → S001_P0112.wav} +0 -0
  23. data/{train/S001_P0113.wav → S001_P0113.wav} +0 -0
  24. data/{train/S001_P0115.wav → S001_P0115.wav} +0 -0
  25. data/{train/S001_P0119.wav → S001_P0119.wav} +0 -0
  26. data/{train/S001_P0120.wav → S001_P0120.wav} +0 -0
  27. data/{train/S001_P0126.wav → S001_P0126.wav} +0 -0
  28. data/{train/S001_P0136.wav → S001_P0136.wav} +0 -0
  29. data/{train/S001_P0141.wav → S001_P0141.wav} +0 -0
  30. data/{train/S001_P0143.wav → S001_P0143.wav} +0 -0
  31. data/{train/S001_P0151.wav → S001_P0151.wav} +0 -0
  32. data/{train/S001_P0155.wav → S001_P0155.wav} +0 -0
  33. data/{train/S001_P0166.wav → S001_P0166.wav} +0 -0
  34. data/{train/S001_P0173.wav → S001_P0173.wav} +0 -0
  35. data/{train/S001_P0175.wav → S001_P0175.wav} +0 -0
  36. data/{train/S001_P0177.wav → S001_P0177.wav} +0 -0
  37. data/{validation/S001_P0184.wav → S001_P0184.wav} +0 -0
  38. data/{validation/S001_P0186.wav → S001_P0186.wav} +0 -0
  39. data/{test/S001_P0194.wav → S001_P0194.wav} +0 -0
  40. data/{train/S001_P0195.wav → S001_P0195.wav} +0 -0
  41. data/{train/S001_P0213.wav → S001_P0213.wav} +0 -0
  42. data/{validation/S001_P0215.wav → S001_P0215.wav} +0 -0
  43. data/{train/S001_P0217.wav → S001_P0217.wav} +0 -0
  44. data/{test/S001_P0230.wav → S001_P0230.wav} +0 -0
  45. data/{train/S001_P0236.wav → S001_P0236.wav} +0 -0
  46. data/{train/S001_P0259.wav → S001_P0259.wav} +0 -0
  47. data/{test/S001_P0275.wav → S001_P0275.wav} +0 -0
  48. data/{train/S001_P0280.wav → S001_P0280.wav} +0 -0
  49. data/{train/S001_P0283.wav → S001_P0283.wav} +0 -0
  50. data/{test/S001_P0288.wav → S001_P0288.wav} +0 -0
README.md CHANGED
@@ -13,167 +13,69 @@ tags:
13
  - text-to-audio
14
  - benchmark
15
  - evaluation
16
- - human-ratings
17
- - multimodal
18
  configs:
19
  - config_name: default
20
  data_files:
21
- - split: train
22
- path: data/train/**
23
- - split: validation
24
- path: data/validation/**
25
- - split: test
26
- path: data/test/**
27
  drop_labels: true
28
  ---
29
 
30
  # AudioEval
31
 
32
- AudioEval is a benchmark for evaluating text-to-audio generation systems from two listener perspectives (`common` and `pro`) and five perceptual dimensions: `production_complexity`, `content_enjoyment`, `production_quality`, `textual_alignment`, and `content_usefulness`.
33
- This repository is packaged in an `AudioFolder`-compatible layout so it can be previewed directly on the Hugging Face Hub and later switched from private to public visibility without repacking.
34
 
35
- ## Dataset Summary
36
 
37
- - 4200 generated audio clips from 24 text-to-audio systems and 451 prompts.
38
- - 11.712 total hours of audio, with an average clip duration of 10.039 seconds.
39
- - Official splits: train=3360, validation=420, test=420.
40
- - 25200 per-rater annotation rows in `annotations/ratings.csv`.
41
- - 126000 dimension-level human scores overall, because each annotation row contains five dimension scores.
42
- - 12 raters in total: common=9, pro=3.
43
- - Rating rows by rater type: common=12600, pro=12600.
44
- - Prompts span five coarse scene categories: daily life, art and cultural, natural and outdoor, work and production, and transportation and travel.
45
-
46
- ## What Is In This Repository
47
-
48
- - `data/train`, `data/validation`, `data/test`: split directories containing `.wav` files and a `metadata.jsonl` file for each split.
49
  - `annotations/ratings.csv`: anonymized per-rater annotations.
50
- - `annotations/prompts.tsv`: prompt text plus prompt-side metadata such as `scene_category` and `audioset_ontology`.
51
- - `annotations/system_info.csv`: mapping from `system_id` to the evaluated system name.
52
- - `annotations/release_summary.json`: summary counts for this packaged release.
53
- - `stats/*.csv`: analysis tables such as ICC, Krippendorff's alpha, significance tests, and model result summaries.
54
-
55
- ## Suggested Uses
56
-
57
- - Benchmarking automatic evaluators for text-to-audio generation.
58
- - Training clip-level regression or distribution-prediction models from prompt-audio pairs.
59
- - Studying rating differences between professional and non-professional listeners.
60
- - Analyzing inter-rater reliability, disagreement, and cross-dimension correlations.
61
-
62
- ## Split Sizes
63
-
64
- | Split | Clips |
65
- | --- | ---: |
66
- | train | 3360 |
67
- | validation | 420 |
68
- | test | 420 |
69
-
70
- ## Data Structure
71
-
72
- ### Split-Level Metadata
73
-
74
- Each row in `data/*/metadata.jsonl` corresponds to one generated audio clip.
75
-
76
- | Field | Description |
77
- | --- | --- |
78
- | `file_name` | Relative path to the audio file inside the split directory. |
79
- | `wav_name` | Original clip filename. |
80
- | `split` | One of `train`, `validation`, or `test`. |
81
- | `prompt_id` | Prompt identifier such as `P0004` or `N001`. |
82
- | `prompt_text` | Natural language prompt used to generate the clip. |
83
- | `scene_category` | Coarse manually assigned prompt category. |
84
- | `sound_event_count` | Number of sound events recorded in the prompt metadata. |
85
- | `audioset_ontology` | Prompt-side ontology category. |
86
- | `system_id` | System identifier such as `S001`. |
87
- | `system_name` | Human-readable system name. |
88
- | `num_ratings_common` | Number of common-rater judgments for the clip. |
89
- | `num_ratings_pro` | Number of professional-rater judgments for the clip. |
90
- | `num_ratings_overall` | Total number of judgments for the clip. |
91
- | `common_*_mean` | Mean score from common raters for a given dimension. |
92
- | `pro_*_mean` | Mean score from professional raters for a given dimension. |
93
- | `common_*_raw_scores` | List of raw common-rater scores for a given dimension. |
94
- | `pro_*_raw_scores` | List of raw professional-rater scores for a given dimension. |
95
-
96
- The five dimensions are `production_complexity`, `content_enjoyment`, `production_quality`, `textual_alignment`, and `content_usefulness`.
97
-
98
- ### Per-Rater Annotations
99
-
100
- Each row in `annotations/ratings.csv` contains one rater's five-dimensional judgment for one clip.
101
-
102
- | Field | Description |
103
- | --- | --- |
104
- | `wav_name` | Clip filename. |
105
- | `split` | Dataset split for the clip. |
106
- | `prompt_id` | Prompt identifier. |
107
- | `system_id` | System identifier. |
108
- | `rater_type` | `common` or `pro`. |
109
- | `rater_id` | Stable anonymized rater identifier within this release. |
110
- | `production_complexity` | Integer score from 1 to 10. |
111
- | `content_enjoyment` | Integer score from 1 to 10. |
112
- | `production_quality` | Integer score from 1 to 10. |
113
- | `textual_alignment` | Integer score from 1 to 10. |
114
- | `content_usefulness` | Integer score from 1 to 10. |
115
-
116
- ## Loading the Data
117
-
118
- Once you have access to the repository on the Hub, you can load the split metadata and audio files with `datasets`:
119
 
120
- ```python
121
- from datasets import load_dataset
122
 
123
- train = load_dataset("Hui519/AudioEval", split="train")
124
- print(train[0]["audio"])
125
- print(train[0]["prompt_text"])
126
- ```
127
-
128
- The anonymized annotation table can be loaded separately from `annotations/ratings.csv`.
129
-
130
- ## Dataset Creation
131
 
132
- AudioEval is a benchmark release of machine-generated audio, not a collection of source recordings.
133
- Each clip is paired with its generation prompt, generating system identifier, and human evaluation results.
134
- This packaged Hub release was assembled from locally processed benchmark assets by:
135
 
136
- - standardizing prompt metadata into `annotations/prompts.tsv`,
137
- - merging raw MOS tables into a unified annotation table,
138
- - exporting split-level clip metadata into `metadata.jsonl`,
139
- - and anonymizing rater identities before publication.
 
140
 
141
- ## Annotation Process
142
 
143
- - The release contains two listener groups: `common` and `pro`.
144
- - There are 9 common raters and 3 professional raters.
145
- - Each clip receives 3 common-rater and 3 professional-rater judgments for every dimension.
146
- - The five dimensions are production complexity, content enjoyment, production quality, textual alignment, and content usefulness.
147
- - Score values are integers from 1 to 10.
148
 
149
- ## Biases, Risks, and Limitations
150
 
151
- - Prompt coverage is broad but not uniform across scene categories or sound-event complexity.
152
- - The prompts are written in English, but generated audio may contain non-speech sounds, speech-like content, music, or multilingual vocal content.
153
- - The benchmark is designed for evaluation research on text-to-audio systems and should not be treated as a general-purpose audio understanding dataset.
154
- - Because the audio is machine-generated, outputs may contain artifacts, distorted speech, or unsafe-sounding events that require additional filtering for downstream applications.
155
- - Anonymized `rater_id` values are release-specific identifiers and should not be linked back to source identities.
156
 
157
- ## Privacy and Sensitive Information
 
 
 
 
158
 
159
- - This release intentionally excludes the original rater demographic tables.
160
- - No direct personally identifying information is included in the published files.
161
- - `annotations/ratings.csv` keeps only anonymized rater IDs, rater type, system ID, prompt ID, split, and scores.
162
- - Although the audio is machine-generated, some clips may depict or imitate human speech, children, alarms, animals, vehicles, or other safety-relevant scenarios.
163
 
164
  ## License
165
 
166
  The repository is currently private.
167
- The YAML metadata uses `license: other` because the final public release license has not been fixed yet.
168
- Until a public license is selected, access and reuse are governed by the repository owner's access settings and release terms.
169
- Before switching the repository to public visibility, replace this section and the YAML metadata with the final license identifier if you choose a standard Hugging Face-supported license.
170
 
171
  ## Citation
172
 
173
- If you use AudioEval, cite the paper below and, when relevant, the Hugging Face dataset repository.
174
-
175
- Paper:
176
-
177
  - Hui Wang, Jinghua Zhao, Junyang Cheng, Cheng Liu, Yuhang Jia, Haoqin Sun, Jiaming Zhou, and Yong Qin. *AudioEval: Automatic Dual-Perspective and Multi-Dimensional Evaluation of Text-to-Audio-Generation*. arXiv:2510.14570, 2025. DOI: 10.48550/arXiv.2510.14570.
178
 
179
  ```bibtex
@@ -185,9 +87,3 @@ Paper:
185
  doi={10.48550/arXiv.2510.14570}
186
  }
187
  ```
188
-
189
- ## Release Status
190
-
191
- - The original source directory in this workspace remains unchanged.
192
- - The export directory can be regenerated by re-running `scripts/prepare_hf_audioeval_dataset.py`.
193
- - This card is written so the same package can remain useful both while private and after a later public release.
 
13
  - text-to-audio
14
  - benchmark
15
  - evaluation
 
 
16
  configs:
17
  - config_name: default
18
  data_files:
19
+ - split: all
20
+ path: data/**
 
 
 
 
21
  drop_labels: true
22
  ---
23
 
24
  # AudioEval
25
 
26
+ AudioEval is a text-to-audio evaluation benchmark with 4200 generated clips, 451 prompts, 24 systems, and 25200 per-rater annotations.
27
+ This release uses one main clip table in `data/metadata.jsonl` and keeps the original `train` / `validation` / `test` assignment as a `split` column instead of separate split folders.
28
 
29
+ ## Files
30
 
31
+ - `data/metadata.jsonl`: one clip-level table for all 4200 clips.
32
+ - `data/*.wav`: audio files referenced by `file_name`.
 
 
 
 
 
 
 
 
 
 
33
  - `annotations/ratings.csv`: anonymized per-rater annotations.
34
+ - `annotations/prompts.tsv`: prompt metadata.
35
+ - `annotations/system_info.csv`: system name mapping.
36
+ - `stats/*.csv`: reliability and model summary tables.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
 
38
+ ## Summary
 
39
 
40
+ - 11.712 total hours of audio, about 10.039 seconds per clip on average.
41
+ - Original split counts are preserved in the `split` column: train=3360, validation=420, test=420.
42
+ - There are 9 common raters and 3 professional raters.
43
+ - Rating rows by rater type: common=12600, pro=12600.
44
+ - Each rating row contains 5 integer scores from 1 to 10.
 
 
 
45
 
46
+ ## Main Columns
 
 
47
 
48
+ - `file_name`, `wav_name`, `split`, `prompt_id`, `prompt_text`
49
+ - `scene_category`, `sound_event_count`, `audioset_ontology`
50
+ - `system_id`, `system_name`
51
+ - `common_*_mean`, `pro_*_mean`
52
+ - `common_*_raw_scores`, `pro_*_raw_scores`
53
 
54
+ The five evaluation dimensions are `production_complexity`, `content_enjoyment`, `production_quality`, `textual_alignment`, and `content_usefulness`.
55
 
56
+ ## Loading
 
 
 
 
57
 
58
+ Once you have access to the repository on the Hub, you can load the main table like this:
59
 
60
+ ```python
61
+ from datasets import load_dataset
 
 
 
62
 
63
+ data = load_dataset("Hui519/AudioEval", split="all")
64
+ print(data[0]["audio"])
65
+ print(data[0]["split"])
66
+ print(data[0]["prompt_text"])
67
+ ```
68
 
69
+ - `annotations/ratings.csv` also keeps the original split assignment as a column.
70
+ - Rater demographic tables are intentionally excluded from this release.
 
 
71
 
72
  ## License
73
 
74
  The repository is currently private.
75
+ The YAML metadata stays at `license: other` until the final public license is chosen.
 
 
76
 
77
  ## Citation
78
 
 
 
 
 
79
  - Hui Wang, Jinghua Zhao, Junyang Cheng, Cheng Liu, Yuhang Jia, Haoqin Sun, Jiaming Zhou, and Yong Qin. *AudioEval: Automatic Dual-Perspective and Multi-Dimensional Evaluation of Text-to-Audio-Generation*. arXiv:2510.14570, 2025. DOI: 10.48550/arXiv.2510.14570.
80
 
81
  ```bibtex
 
87
  doi={10.48550/arXiv.2510.14570}
88
  }
89
  ```
 
 
 
 
 
 
data/{test/S001_P0004.wav → S001_P0004.wav} RENAMED
File without changes
data/{train/S001_P0008.wav → S001_P0008.wav} RENAMED
File without changes
data/{train/S001_P0013.wav → S001_P0013.wav} RENAMED
File without changes
data/{train/S001_P0014.wav → S001_P0014.wav} RENAMED
File without changes
data/{train/S001_P0016.wav → S001_P0016.wav} RENAMED
File without changes
data/{train/S001_P0017.wav → S001_P0017.wav} RENAMED
File without changes
data/{train/S001_P0023.wav → S001_P0023.wav} RENAMED
File without changes
data/{train/S001_P0041.wav → S001_P0041.wav} RENAMED
File without changes
data/{train/S001_P0045.wav → S001_P0045.wav} RENAMED
File without changes
data/{train/S001_P0048.wav → S001_P0048.wav} RENAMED
File without changes
data/{train/S001_P0050.wav → S001_P0050.wav} RENAMED
File without changes
data/{train/S001_P0053.wav → S001_P0053.wav} RENAMED
File without changes
data/{train/S001_P0058.wav → S001_P0058.wav} RENAMED
File without changes
data/{train/S001_P0064.wav → S001_P0064.wav} RENAMED
File without changes
data/{train/S001_P0072.wav → S001_P0072.wav} RENAMED
File without changes
data/{validation/S001_P0080.wav → S001_P0080.wav} RENAMED
File without changes
data/{train/S001_P0082.wav → S001_P0082.wav} RENAMED
File without changes
data/{train/S001_P0094.wav → S001_P0094.wav} RENAMED
File without changes
data/{train/S001_P0102.wav → S001_P0102.wav} RENAMED
File without changes
data/{validation/S001_P0111.wav → S001_P0111.wav} RENAMED
File without changes
data/{validation/S001_P0112.wav → S001_P0112.wav} RENAMED
File without changes
data/{train/S001_P0113.wav → S001_P0113.wav} RENAMED
File without changes
data/{train/S001_P0115.wav → S001_P0115.wav} RENAMED
File without changes
data/{train/S001_P0119.wav → S001_P0119.wav} RENAMED
File without changes
data/{train/S001_P0120.wav → S001_P0120.wav} RENAMED
File without changes
data/{train/S001_P0126.wav → S001_P0126.wav} RENAMED
File without changes
data/{train/S001_P0136.wav → S001_P0136.wav} RENAMED
File without changes
data/{train/S001_P0141.wav → S001_P0141.wav} RENAMED
File without changes
data/{train/S001_P0143.wav → S001_P0143.wav} RENAMED
File without changes
data/{train/S001_P0151.wav → S001_P0151.wav} RENAMED
File without changes
data/{train/S001_P0155.wav → S001_P0155.wav} RENAMED
File without changes
data/{train/S001_P0166.wav → S001_P0166.wav} RENAMED
File without changes
data/{train/S001_P0173.wav → S001_P0173.wav} RENAMED
File without changes
data/{train/S001_P0175.wav → S001_P0175.wav} RENAMED
File without changes
data/{train/S001_P0177.wav → S001_P0177.wav} RENAMED
File without changes
data/{validation/S001_P0184.wav → S001_P0184.wav} RENAMED
File without changes
data/{validation/S001_P0186.wav → S001_P0186.wav} RENAMED
File without changes
data/{test/S001_P0194.wav → S001_P0194.wav} RENAMED
File without changes
data/{train/S001_P0195.wav → S001_P0195.wav} RENAMED
File without changes
data/{train/S001_P0213.wav → S001_P0213.wav} RENAMED
File without changes
data/{validation/S001_P0215.wav → S001_P0215.wav} RENAMED
File without changes
data/{train/S001_P0217.wav → S001_P0217.wav} RENAMED
File without changes
data/{test/S001_P0230.wav → S001_P0230.wav} RENAMED
File without changes
data/{train/S001_P0236.wav → S001_P0236.wav} RENAMED
File without changes
data/{train/S001_P0259.wav → S001_P0259.wav} RENAMED
File without changes
data/{test/S001_P0275.wav → S001_P0275.wav} RENAMED
File without changes
data/{train/S001_P0280.wav → S001_P0280.wav} RENAMED
File without changes
data/{train/S001_P0283.wav → S001_P0283.wav} RENAMED
File without changes
data/{test/S001_P0288.wav → S001_P0288.wav} RENAMED
File without changes