File size: 4,714 Bytes
7a04089
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f1b481a
7a04089
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
---
license: cc-by-4.0
task_categories:
  - automatic-speech-recognition
  - text-to-speech
language:
  - en
  - de
  - fr
  - es
  - ru
  - ja
  - ko
  - pt
  - tr
  - th
tags:
  - audio
  - speech
  - phoneme-alignment
  - mfa
  - forced-alignment
pretty_name: Multilingual MFA-Aligned Speech Dataset

---

# Multilingual MFA-Aligned Speech Dataset

A large-scale multilingual speech dataset with **word-level and phoneme-level alignments** produced using the Montreal Forced Aligner (MFA).

## Dataset Description

This dataset consolidates multiple speech corpora across various languages, all processed through MFA to provide precise phoneme and word alignments. Each sample includes the original audio, transcript, and detailed timing information for both words and phonemes.

### Features

| Column | Type | Description |
|--------|------|-------------|
| `audio` | Audio | Audio waveform at 16kHz |
| `transcript` | string | Text transcription |
| `phoneme_sequence` | string | Phoneme sequence with spaces between words |
| `words` | list | Word-level alignments: `[{word, start, end}, ...]` |
| `phonemes` | list | Phoneme-level alignments: `[{phoneme, start, end}, ...]` |
| `source` | string | Original dataset source (e.g., voxpopuli, common_voice) |

### Languages & Statistics

| Language | Config | Hours | Samples | Sources |
|----------|--------|-------|---------|---------|
| English | `english` | TBD | TBD | Common Voice, VoxPopuli, GigaSpeech, Emilia, Genshin Voice, Gemini Speech |
| German | `german` | TBD | TBD | Multilingual LibriSpeech, Emilia |
| French | `french` | TBD | TBD | French Game Voice, Multilingual LibriSpeech, Wolof French ASR |
| Spanish | `spanish` | TBD | TBD | CML TTS, LibriVox, TEDx Spanish |
| Russian | `russian` | TBD | TBD | Russian Audio Data, Multilingual LibriSpeech |
| Japanese | `japanese` | TBD | TBD | Combined Japanese Dataset, Japanese Anime Speech |
| Korean | `korean` | TBD | TBD | Zeroth STT Korean, Korea Speech |
| Portuguese | `portuguese` | TBD | TBD | Portuguese TTS, Multilingual LibriSpeech |
| Turkish | `turkish` | TBD | TBD | Turkish Merge Audio, Khan Academy Turkish |
| Thai | `thai` | TBD | TBD | Porjai Thai Voice Dataset |

**Total: ~20,000+ hours** (estimated)

## Usage

### Load a specific language

```python
from datasets import load_dataset

# Load English data
dataset = load_dataset("AAdonis/merged_mfa_alignments", "english", split="train")

# Load German data
dataset = load_dataset("AAdonis/merged_mfa_alignments", "german", split="train")
```

### Access alignments

```python
sample = dataset[0]

# Get audio
audio = sample["audio"]["array"]
sample_rate = sample["audio"]["sampling_rate"]

# Get transcript, phonemes, and source
transcript = sample["transcript"]
phonemes = sample["phoneme_sequence"]  # "h ɛ l oʊ w ɜː l d"
source = sample["source"]  # e.g., "voxpopuli"

# Get word-level alignments
for word_info in sample["words"]:
    print(f"{word_info['word']}: {word_info['start']:.2f}s - {word_info['end']:.2f}s")

# Get phoneme-level alignments
for phon_info in sample["phonemes"]:
    print(f"{phon_info['phoneme']}: {phon_info['start']:.3f}s - {phon_info['end']:.3f}s")
```

### Filter by source

```python
# Get only VoxPopuli samples
voxpopuli_samples = dataset.filter(lambda x: x["source"] == "voxpopuli")

# Get only Common Voice samples
cv_samples = dataset.filter(lambda x: x["source"] == "common_voice")
```

## Processing Details

### MFA Alignment

All samples were aligned using the Montreal Forced Aligner (MFA) with language-specific acoustic models and pronunciation dictionaries.

### Quality Filtering

During processing, samples were filtered and split based on:
- **`<unk>` words**: Samples containing unknown words are split at those boundaries
- **`spn` phonemes**: Spoken noise markers cause sample splits
- **Duration**: Samples are filtered by minimum/maximum duration thresholds
- **Word count**: Minimum word requirements per segment

### Phoneme Sequence Format

The `phoneme_sequence` column contains IPA phonemes with:
- Phonemes within a word are concatenated directly
- Words are separated by spaces
- Example: `"h ɛ l oʊ"` for "hello" (4 phonemes, 1 word)

## Citation

If you use this dataset, please cite the original source datasets and the Montreal Forced Aligner:

```bibtex
@article{mcauliffe2017montreal,
  title={Montreal Forced Aligner: Trainable text-speech alignment using Kaldi},
  author={McAuliffe, Michael and Socolof, Michaela and Mihuc, Sarah and Wagner, Michael and Sonderegger, Morgan},
  journal={Interspeech},
  year={2017}
}
```

## License

This dataset is released under CC-BY-4.0. Please also respect the licenses of the original source datasets.