Datasets:
File size: 11,106 Bytes
dcb7518 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 | ---
license: cc0-1.0
task_categories:
- feature-extraction
language:
- ko
- hi
tags:
- audio
- speech
- prosody
- acoustics
- linguistics
- phonetics
- voice-analytics
pretty_name: Alexandria Voice Corpus — Korean & Hindi Macro-Prosody Telemetry
size_categories:
- 10K<n<100K
---
# Alexandria Voice Corpus — Korean & Hindi Macro-Prosody Telemetry
A sample release from the **Alexandria Voice Corpus**, a multilingual acoustic telemetry database spanning 60+ languages. This pack contains macro-prosodic feature extractions for **Korean** (6,998 clips) and **Hindi** (18,447 clips), derived from the Mozilla Common Voice CV24 corpus (CC0).
No audio is included. This is a structured feature dataset for linguistic research, speech technology development, and cross-linguistic prosody analysis.
---
## Dataset Details
### What is macro-prosody telemetry?
Macro-prosody refers to the suprasegmental properties of speech — pitch contour, rhythm, intensity, and voice quality — measured at the clip level rather than the phoneme level. Each row in this dataset represents one spoken utterance with 20+ acoustic features extracted from it.
This is distinct from transcription, alignment, or phoneme-level data. It is designed for population-level acoustic analysis, language typology research, and training prosody-aware speech models.
### Dataset Description
- **Curated by:** Orator Forge
- **Language(s):** Korean (`ko`), Hindi (`hi`)
- **Source corpus:** Mozilla Common Voice CV24 (CC0-1.0)
- **License:** CC0-1.0
- **Clips:** 25,445 total (Korean: 6,998 | Hindi: 18,447)
- **Anonymization standard:** `orator_forge_k5_v1`
### Dataset Sources
- **Source project:** [Mozilla Common Voice](https://commonvoice.mozilla.org)
- **Source license:** [CC0-1.0](https://creativecommons.org/publicdomain/zero/1.0/)
- **Part of:** Alexandria Voice Corpus (Orator Forge)
---
## Uses
### Direct Use
- Cross-linguistic prosody comparison between Korean (language isolate) and Hindi (Indo-Aryan)
- Training or evaluating prosody-aware TTS and ASR models
- Rhythm typology research (e.g. mora-timed vs stress-timed speech)
- Voice quality and breathiness studies
- Speaker demographic modeling from acoustic features (population level)
- Feature engineering for downstream speech classification tasks
### Out-of-Scope Use
- **Speaker identification or re-identification** — this dataset has been deliberately anonymized to prevent linking acoustic features back to individual speakers. Any attempt to do so violates the terms of use.
- Direct audio reconstruction — no audio is present in this dataset.
- Tasks requiring phoneme-level or word-level timing — use a force-aligned corpus instead.
---
## Dataset Structure
Each parquet file contains one row per utterance. Files are Snappy-compressed.
| Column | Type | Description |
|---|---|---|
| `clip_id` | string | Anonymized sequential ID (e.g. `korean_cv24_004521`) |
| `lang` | string | BCP-47 language code |
| `lang_name` | string | Language name |
| `quality_tier` | int | 1 (best) – 2 (good). Only T1/T2 clips included |
| `duration_ms` | int | Clip duration, bucketed to nearest 100ms |
| `gender` | string | `male` / `female` / `unknown` |
| `gender_source` | string | `meta` (self-reported) / `inferred` (pitch-based) / `unknown` |
| `age` | string | Age bracket (CV metadata where available) |
| `syllable_count_approx` | int | Approximate syllable count (vowel-count proxy) |
| `pitch_mean` | float32 | Mean F0 in Hz |
| `pitch_std` | float32 | F0 standard deviation |
| `pitch_range` | float32 | F0 range (max – min) in Hz |
| `pitch_velocity_max` | float32 | Max rate of F0 change (Hz/s) |
| `intensity_mean` | float32 | Mean RMS intensity (dB) |
| `intensity_max` | float32 | Peak intensity (dB) |
| `intensity_range` | float32 | Intensity dynamic range (dB) |
| `hnr_mean` | float32 | Harmonics-to-noise ratio (dB) |
| `cpps` | float32 | Cepstral peak prominence smoothed — breathiness indicator |
| `jitter_local` | float32 | Cycle-to-cycle pitch perturbation |
| `shimmer_local` | float32 | Cycle-to-cycle amplitude perturbation |
| `spectral_centroid_mean` | float32 | Mean spectral centroid (Hz) |
| `spectral_tilt` | float32 | Spectral slope (relates to voice effort) |
| `mfcc_delta_mean` | float32 | Mean MFCC delta (rate of spectral change) |
| `zcr_mean` | float32 | Zero-crossing rate |
| `teo_mean` | float32 | Teager energy operator mean |
| `npvi` | float32 | Normalized pairwise variability index (rhythm metric) |
| `articulation_rate` | float32 | Syllables per second (speech only) |
| `speaking_rate` | float32 | Syllables per second (total duration) |
| `pause_rate` | float32 | Pauses per second |
| `speech_ratio` | float32 | Proportion of clip containing voiced speech |
| `snr_median` | float32 | Signal-to-noise ratio, median (Brouhaha) |
| `c50_median` | float32 | C50 clarity metric, median (Brouhaha) |
| `f1_mean` | float32 | First formant mean (Hz) — note: may be 0.0 in this release |
| `f2_mean` | float32 | Second formant mean (Hz) — note: may be 0.0 in this release |
| `f3_mean` | float32 | Third formant mean (Hz) — note: may be 0.0 in this release |
### Quality Tiers
Clips were graded using [Brouhaha](https://github.com/marianne-m/brouhaha-vad) (SNR + C50 + VAD scoring):
| Tier | SNR | C50 | Speech ratio | Description |
|---|---|---|---|---|
| T1 | ≥ 20 dB | ≥ 20 dB | ≥ 0.6 | Studio quality |
| T2 | ≥ 10 dB | ≥ 5 dB | ≥ 0.4 | Clean field recording |
Only T1 and T2 clips are included in this release.
### Files
```
korean_cv24.parquet — 6,998 rows
hindi_cv24.parquet — 18,447 rows
```
---
## Dataset Creation
### Curation Rationale
There is a significant gap in publicly available acoustic feature datasets for non-Western and non-European languages. Korean and Hindi together represent over 600 million speakers across two typologically distinct language families — a language isolate and an Indo-Aryan branch of Indo-European. This release provides a free, CC0-licensed baseline for researchers who need structured prosodic features without needing to process raw audio.
### Source Data
#### Data Collection and Processing
Source audio was drawn from [Mozilla Common Voice CV24](https://commonvoice.mozilla.org), a crowd-sourced corpus of read speech recorded by volunteers under a CC0 license.
Processing pipeline:
1. MP3 source audio converted to 16kHz mono WAV (ffmpeg, -20 dBFS normalization)
2. Quality grading via Brouhaha (SNR, C50, VAD) — only T1/T2 retained
3. Acoustic feature extraction via Parselmouth/Praat at 16kHz
4. Anonymization and precision degradation applied at export (see below)
#### Source Data Producers
Recordings were made by volunteer contributors to the Mozilla Common Voice project. Contributors self-reported demographic metadata (age, gender, accent) where willing.
### Anonymization
This dataset applies the `orator_forge_k5_v1` anonymization standard:
- Original Mozilla filenames replaced with sequential anonymized clip IDs
- Transcripts removed entirely (approximate syllable count provided as proxy)
- All continuous acoustic variables truncated to 2 decimal places and stored as float32
- Duration bucketed to nearest 100ms to prevent cross-referencing with source audio
- K-anonymity suppression at k=5: rows where the combination of `{gender, age_bucket, duration_bucket}` has fewer than 5 members are excluded
### Personal and Sensitive Information
- No names, speaker IDs, or any directly identifying information is present
- No original audio is included
- Demographic fields (age, gender) are self-reported by Mozilla Common Voice contributors and are optional — many rows will show `unknown`
- Formant data (F1/F2/F3) is present but returns 0.0 in this release due to a known extraction issue; this will be corrected in v1.1
---
## Bias, Risks, and Limitations
- **Gender balance:** Gender is inferred from pitch for clips lacking self-reported metadata. Pitch-based inference has known limitations for speakers with atypical voices, tonal language speakers, and non-binary individuals. The `gender_source` field distinguishes self-reported from inferred labels.
- **Recording conditions:** Common Voice is read speech recorded in uncontrolled environments. Acoustic conditions vary significantly across contributors.
- **Age distribution:** CV contributor demographics skew younger and technically literate. This dataset is not a representative sample of the full speaker population of either language.
- **Hindi script diversity:** Hindi CV24 clips include speakers from a wide range of regional backgrounds with varying accent profiles. No regional stratification has been applied in this release.
- **Formant zeros:** F1/F2/F3 return 0.0 across all clips in this release. Do not use formant columns until v1.1.
- **Prohibited use:** Do not use this dataset to attempt speaker identification or re-linking to source audio. This violates the terms of use regardless of technical feasibility.
### Recommendations
Use the `gender_source` field to filter to self-reported gender labels if demographic accuracy is important for your use case. For cross-linguistic rhythm comparisons, nPVI and articulation rate are the most reliable features in this release. Formant-dependent analyses should wait for v1.1.
---
## Citation
If you use this dataset, please cite the Mozilla Common Voice project as the source corpus:
**BibTeX:**
```bibtex
@dataset{alexandria_korean_hindi_prosody_2026,
title = {Alexandria Voice Corpus — Korean \& Hindi Macro-Prosody Telemetry},
author = {Orator Forge},
year = {2026},
license = {CC0-1.0},
note = {Derived from Mozilla Common Voice CV24 (CC0).
Acoustic features extracted via Parselmouth/Praat.}
}
@misc{mozilla_common_voice,
title = {Common Voice: A Massively-Multilingual Speech Corpus},
author = {Ardila, Rosana and others},
year = {2020},
url = {https://commonvoice.mozilla.org}
}
```
---
## Glossary
| Term | Definition |
|---|---|
| **F0 / pitch_mean** | Fundamental frequency — the perceived pitch of the voice, measured in Hz |
| **HNR** | Harmonics-to-noise ratio — higher values indicate cleaner, more tonal voice quality |
| **CPPS** | Cepstral peak prominence smoothed — lower values indicate breathier voice |
| **nPVI** | Normalized pairwise variability index — measures durational variability between adjacent syllables; higher in stress-timed languages |
| **C50** | Clarity metric from room acoustics; higher = less reverb/echo in the recording |
| **SNR** | Signal-to-noise ratio — higher = cleaner recording |
| **Brouhaha** | Quality scoring model used for grading: [github.com/marianne-m/brouhaha-vad](https://github.com/marianne-m/brouhaha-vad) |
| **T1/T2** | Quality tiers assigned by Brouhaha grading (see Dataset Structure) |
| **orator_forge_k5_v1** | Anonymization standard: k=5 suppression + sequential IDs + 2dp truncation + 100ms duration bucketing |
---
## Dataset Card Contact
c.kleingertner@gmail.com
`` |