vadette commited on
Commit
dcb7518
·
verified ·
1 Parent(s): cf9a015

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +235 -0
README.md ADDED
@@ -0,0 +1,235 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc0-1.0
3
+ task_categories:
4
+ - feature-extraction
5
+ language:
6
+ - ko
7
+ - hi
8
+ tags:
9
+ - audio
10
+ - speech
11
+ - prosody
12
+ - acoustics
13
+ - linguistics
14
+ - phonetics
15
+ - voice-analytics
16
+ pretty_name: Alexandria Voice Corpus — Korean & Hindi Macro-Prosody Telemetry
17
+ size_categories:
18
+ - 10K<n<100K
19
+ ---
20
+
21
+ # Alexandria Voice Corpus — Korean & Hindi Macro-Prosody Telemetry
22
+
23
+ A sample release from the **Alexandria Voice Corpus**, a multilingual acoustic telemetry database spanning 60+ languages. This pack contains macro-prosodic feature extractions for **Korean** (6,998 clips) and **Hindi** (18,447 clips), derived from the Mozilla Common Voice CV24 corpus (CC0).
24
+
25
+ No audio is included. This is a structured feature dataset for linguistic research, speech technology development, and cross-linguistic prosody analysis.
26
+
27
+ ---
28
+
29
+ ## Dataset Details
30
+
31
+ ### What is macro-prosody telemetry?
32
+
33
+ Macro-prosody refers to the suprasegmental properties of speech — pitch contour, rhythm, intensity, and voice quality — measured at the clip level rather than the phoneme level. Each row in this dataset represents one spoken utterance with 20+ acoustic features extracted from it.
34
+
35
+ This is distinct from transcription, alignment, or phoneme-level data. It is designed for population-level acoustic analysis, language typology research, and training prosody-aware speech models.
36
+
37
+ ### Dataset Description
38
+
39
+ - **Curated by:** Orator Forge
40
+ - **Language(s):** Korean (`ko`), Hindi (`hi`)
41
+ - **Source corpus:** Mozilla Common Voice CV24 (CC0-1.0)
42
+ - **License:** CC0-1.0
43
+ - **Clips:** 25,445 total (Korean: 6,998 | Hindi: 18,447)
44
+ - **Anonymization standard:** `orator_forge_k5_v1`
45
+
46
+ ### Dataset Sources
47
+
48
+ - **Source project:** [Mozilla Common Voice](https://commonvoice.mozilla.org)
49
+ - **Source license:** [CC0-1.0](https://creativecommons.org/publicdomain/zero/1.0/)
50
+ - **Part of:** Alexandria Voice Corpus (Orator Forge)
51
+
52
+ ---
53
+
54
+ ## Uses
55
+
56
+ ### Direct Use
57
+
58
+ - Cross-linguistic prosody comparison between Korean (language isolate) and Hindi (Indo-Aryan)
59
+ - Training or evaluating prosody-aware TTS and ASR models
60
+ - Rhythm typology research (e.g. mora-timed vs stress-timed speech)
61
+ - Voice quality and breathiness studies
62
+ - Speaker demographic modeling from acoustic features (population level)
63
+ - Feature engineering for downstream speech classification tasks
64
+
65
+ ### Out-of-Scope Use
66
+
67
+ - **Speaker identification or re-identification** — this dataset has been deliberately anonymized to prevent linking acoustic features back to individual speakers. Any attempt to do so violates the terms of use.
68
+ - Direct audio reconstruction — no audio is present in this dataset.
69
+ - Tasks requiring phoneme-level or word-level timing — use a force-aligned corpus instead.
70
+
71
+ ---
72
+
73
+ ## Dataset Structure
74
+
75
+ Each parquet file contains one row per utterance. Files are Snappy-compressed.
76
+
77
+ | Column | Type | Description |
78
+ |---|---|---|
79
+ | `clip_id` | string | Anonymized sequential ID (e.g. `korean_cv24_004521`) |
80
+ | `lang` | string | BCP-47 language code |
81
+ | `lang_name` | string | Language name |
82
+ | `quality_tier` | int | 1 (best) – 2 (good). Only T1/T2 clips included |
83
+ | `duration_ms` | int | Clip duration, bucketed to nearest 100ms |
84
+ | `gender` | string | `male` / `female` / `unknown` |
85
+ | `gender_source` | string | `meta` (self-reported) / `inferred` (pitch-based) / `unknown` |
86
+ | `age` | string | Age bracket (CV metadata where available) |
87
+ | `syllable_count_approx` | int | Approximate syllable count (vowel-count proxy) |
88
+ | `pitch_mean` | float32 | Mean F0 in Hz |
89
+ | `pitch_std` | float32 | F0 standard deviation |
90
+ | `pitch_range` | float32 | F0 range (max – min) in Hz |
91
+ | `pitch_velocity_max` | float32 | Max rate of F0 change (Hz/s) |
92
+ | `intensity_mean` | float32 | Mean RMS intensity (dB) |
93
+ | `intensity_max` | float32 | Peak intensity (dB) |
94
+ | `intensity_range` | float32 | Intensity dynamic range (dB) |
95
+ | `hnr_mean` | float32 | Harmonics-to-noise ratio (dB) |
96
+ | `cpps` | float32 | Cepstral peak prominence smoothed — breathiness indicator |
97
+ | `jitter_local` | float32 | Cycle-to-cycle pitch perturbation |
98
+ | `shimmer_local` | float32 | Cycle-to-cycle amplitude perturbation |
99
+ | `spectral_centroid_mean` | float32 | Mean spectral centroid (Hz) |
100
+ | `spectral_tilt` | float32 | Spectral slope (relates to voice effort) |
101
+ | `mfcc_delta_mean` | float32 | Mean MFCC delta (rate of spectral change) |
102
+ | `zcr_mean` | float32 | Zero-crossing rate |
103
+ | `teo_mean` | float32 | Teager energy operator mean |
104
+ | `npvi` | float32 | Normalized pairwise variability index (rhythm metric) |
105
+ | `articulation_rate` | float32 | Syllables per second (speech only) |
106
+ | `speaking_rate` | float32 | Syllables per second (total duration) |
107
+ | `pause_rate` | float32 | Pauses per second |
108
+ | `speech_ratio` | float32 | Proportion of clip containing voiced speech |
109
+ | `snr_median` | float32 | Signal-to-noise ratio, median (Brouhaha) |
110
+ | `c50_median` | float32 | C50 clarity metric, median (Brouhaha) |
111
+ | `f1_mean` | float32 | First formant mean (Hz) — note: may be 0.0 in this release |
112
+ | `f2_mean` | float32 | Second formant mean (Hz) — note: may be 0.0 in this release |
113
+ | `f3_mean` | float32 | Third formant mean (Hz) — note: may be 0.0 in this release |
114
+
115
+ ### Quality Tiers
116
+
117
+ Clips were graded using [Brouhaha](https://github.com/marianne-m/brouhaha-vad) (SNR + C50 + VAD scoring):
118
+
119
+ | Tier | SNR | C50 | Speech ratio | Description |
120
+ |---|---|---|---|---|
121
+ | T1 | ≥ 20 dB | ≥ 20 dB | ≥ 0.6 | Studio quality |
122
+ | T2 | ≥ 10 dB | ≥ 5 dB | ≥ 0.4 | Clean field recording |
123
+
124
+ Only T1 and T2 clips are included in this release.
125
+
126
+ ### Files
127
+
128
+ ```
129
+ korean_cv24.parquet — 6,998 rows
130
+ hindi_cv24.parquet — 18,447 rows
131
+ ```
132
+
133
+ ---
134
+
135
+ ## Dataset Creation
136
+
137
+ ### Curation Rationale
138
+
139
+ There is a significant gap in publicly available acoustic feature datasets for non-Western and non-European languages. Korean and Hindi together represent over 600 million speakers across two typologically distinct language families — a language isolate and an Indo-Aryan branch of Indo-European. This release provides a free, CC0-licensed baseline for researchers who need structured prosodic features without needing to process raw audio.
140
+
141
+ ### Source Data
142
+
143
+ #### Data Collection and Processing
144
+
145
+ Source audio was drawn from [Mozilla Common Voice CV24](https://commonvoice.mozilla.org), a crowd-sourced corpus of read speech recorded by volunteers under a CC0 license.
146
+
147
+ Processing pipeline:
148
+ 1. MP3 source audio converted to 16kHz mono WAV (ffmpeg, -20 dBFS normalization)
149
+ 2. Quality grading via Brouhaha (SNR, C50, VAD) — only T1/T2 retained
150
+ 3. Acoustic feature extraction via Parselmouth/Praat at 16kHz
151
+ 4. Anonymization and precision degradation applied at export (see below)
152
+
153
+ #### Source Data Producers
154
+
155
+ Recordings were made by volunteer contributors to the Mozilla Common Voice project. Contributors self-reported demographic metadata (age, gender, accent) where willing.
156
+
157
+ ### Anonymization
158
+
159
+ This dataset applies the `orator_forge_k5_v1` anonymization standard:
160
+
161
+ - Original Mozilla filenames replaced with sequential anonymized clip IDs
162
+ - Transcripts removed entirely (approximate syllable count provided as proxy)
163
+ - All continuous acoustic variables truncated to 2 decimal places and stored as float32
164
+ - Duration bucketed to nearest 100ms to prevent cross-referencing with source audio
165
+ - K-anonymity suppression at k=5: rows where the combination of `{gender, age_bucket, duration_bucket}` has fewer than 5 members are excluded
166
+
167
+ ### Personal and Sensitive Information
168
+
169
+ - No names, speaker IDs, or any directly identifying information is present
170
+ - No original audio is included
171
+ - Demographic fields (age, gender) are self-reported by Mozilla Common Voice contributors and are optional — many rows will show `unknown`
172
+ - Formant data (F1/F2/F3) is present but returns 0.0 in this release due to a known extraction issue; this will be corrected in v1.1
173
+
174
+ ---
175
+
176
+ ## Bias, Risks, and Limitations
177
+
178
+ - **Gender balance:** Gender is inferred from pitch for clips lacking self-reported metadata. Pitch-based inference has known limitations for speakers with atypical voices, tonal language speakers, and non-binary individuals. The `gender_source` field distinguishes self-reported from inferred labels.
179
+ - **Recording conditions:** Common Voice is read speech recorded in uncontrolled environments. Acoustic conditions vary significantly across contributors.
180
+ - **Age distribution:** CV contributor demographics skew younger and technically literate. This dataset is not a representative sample of the full speaker population of either language.
181
+ - **Hindi script diversity:** Hindi CV24 clips include speakers from a wide range of regional backgrounds with varying accent profiles. No regional stratification has been applied in this release.
182
+ - **Formant zeros:** F1/F2/F3 return 0.0 across all clips in this release. Do not use formant columns until v1.1.
183
+ - **Prohibited use:** Do not use this dataset to attempt speaker identification or re-linking to source audio. This violates the terms of use regardless of technical feasibility.
184
+
185
+ ### Recommendations
186
+
187
+ Use the `gender_source` field to filter to self-reported gender labels if demographic accuracy is important for your use case. For cross-linguistic rhythm comparisons, nPVI and articulation rate are the most reliable features in this release. Formant-dependent analyses should wait for v1.1.
188
+
189
+ ---
190
+
191
+ ## Citation
192
+
193
+ If you use this dataset, please cite the Mozilla Common Voice project as the source corpus:
194
+
195
+ **BibTeX:**
196
+ ```bibtex
197
+ @dataset{alexandria_korean_hindi_prosody_2026,
198
+ title = {Alexandria Voice Corpus — Korean \& Hindi Macro-Prosody Telemetry},
199
+ author = {Orator Forge},
200
+ year = {2026},
201
+ license = {CC0-1.0},
202
+ note = {Derived from Mozilla Common Voice CV24 (CC0).
203
+ Acoustic features extracted via Parselmouth/Praat.}
204
+ }
205
+
206
+ @misc{mozilla_common_voice,
207
+ title = {Common Voice: A Massively-Multilingual Speech Corpus},
208
+ author = {Ardila, Rosana and others},
209
+ year = {2020},
210
+ url = {https://commonvoice.mozilla.org}
211
+ }
212
+ ```
213
+
214
+ ---
215
+
216
+ ## Glossary
217
+
218
+ | Term | Definition |
219
+ |---|---|
220
+ | **F0 / pitch_mean** | Fundamental frequency — the perceived pitch of the voice, measured in Hz |
221
+ | **HNR** | Harmonics-to-noise ratio — higher values indicate cleaner, more tonal voice quality |
222
+ | **CPPS** | Cepstral peak prominence smoothed — lower values indicate breathier voice |
223
+ | **nPVI** | Normalized pairwise variability index — measures durational variability between adjacent syllables; higher in stress-timed languages |
224
+ | **C50** | Clarity metric from room acoustics; higher = less reverb/echo in the recording |
225
+ | **SNR** | Signal-to-noise ratio — higher = cleaner recording |
226
+ | **Brouhaha** | Quality scoring model used for grading: [github.com/marianne-m/brouhaha-vad](https://github.com/marianne-m/brouhaha-vad) |
227
+ | **T1/T2** | Quality tiers assigned by Brouhaha grading (see Dataset Structure) |
228
+ | **orator_forge_k5_v1** | Anonymization standard: k=5 suppression + sequential IDs + 2dp truncation + 100ms duration bucketing |
229
+
230
+ ---
231
+
232
+ ## Dataset Card Contact
233
+
234
+ c.kleingertner@gmail.com
235
+ ``