File size: 5,208 Bytes
607dc7e
 
 
585c979
 
 
 
 
 
 
 
 
 
 
 
607dc7e
 
bbf9987
607dc7e
 
 
bbf9987
 
 
607dc7e
 
 
bbf9987
 
607dc7e
 
 
 
 
 
3f61b79
607dc7e
 
 
 
 
 
 
bbf9987
607dc7e
 
 
 
 
bbf9987
607dc7e
 
 
 
 
 
 
3f61b79
607dc7e
1a953e7
607dc7e
 
 
 
 
 
 
3f61b79
607dc7e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3f61b79
 
bbf9987
 
607dc7e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3f61b79
 
607dc7e
 
 
 
bbf9987
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
---
license: cc-by-nc-4.0
task_categories:
- automatic-speech-recognition
language:
- en
tags:
- non-native
- pronunciation
- speech
- pronunciation assessment
- phoneme
pretty_name: EpaDB
size_categories:
- 1K<n<10K
---

# EpaDB: English Pronunciation by Argentinians

## Dataset Summary

EpaDB is a speech database intended for research in pronunciation scoring. The corpus includes audios from 50 Spanish speakers (25 males and 25 females) from 
Argentina reading phrases in English. Each speaker recorded 64 short phrases containing sounds hard to pronounce for this population adding up to ~3.5 hours of speech.


## Supported Tasks

- **Pronunciation Assessment** – predict utterance-level global scores or phoneme-level correct/incorrect
- **Phone Recognition** - predict phoneme sequences
- **Phone-level Error Detection** – classify each phone as insertion, deletion, distortion, substitution, or correct.
- **Alignment Analysis** – leverage MFA timings to study forced alignment quality or to refine pronunciation models.

## Languages

- L2 utterances: English
- Speaker L1: Spanish

## Dataset Structure

### Data Instances

Each JSON entry describes one utterance:

- Phone sequences for reference transcription (`reference`) and annotators (`annot_1`, optional `annot_2`).
- Phone-level labels (`label_1`, `label_2`) and derived `error_type` categories.
- MFA start/end timestamps per phone (`start_mfa`, `end_mfa`).
- Per-utterance global scores (`global_1`, `global_2`) and propagated speaker levels (`level_1`, `level_2`).
- Speaker metadata (`speaker_id`, `gender`).
- Audio metadata (`duration`, `sample_rate`, `wav_path`) plus the waveform itself.
- Reference sentence orthographic transcription (`transcription`).

### Data Fields

| Field | Type | Description |
|-------|------|-------------|
| `utt_id` | string | Unique utterance identifier (e.g., `spkr28_1`). |
| `speaker_id` | string | Speaker identifier. |
| `sentence_id` | string | Reference sentence ID (matches `reference_transcriptions.txt`). |
| `phone_ids` | sequence[string] | Unique phone identifiers per utterance. |
| `reference` | sequence[string] | reference phones assigned to match the closer aimed pronunciation by the speaker. |
| `annot_1` | sequence[string] | Annotator 1 phones (`-` marks deletions). |
| `annot_2` | sequence[string] | Annotator 3 phones when available, empty otherwise. |
| `label_1` | sequence[string] | Annotator 1 phone labels (`"1"` correct, `"0"` incorrect). |
| `label_2` | sequence[string] | Annotator 3 phone labels when present. |
| `error_type` | sequence[string] | Derived categories: `correct`, `insertion`, `deletion`, `distortion`, `substitution`. |
| `start_mfa` | sequence[float] | Phone start times (seconds). |
| `end_mfa` | sequence[float] | Phone end times (seconds). |
| `global_1` | float or null | Annotator 1 utterance-level score (1–4). |
| `global_2` | float or null | Annotator 3 score when available. |
| `level_1` | string or null | Speaker-level proficiency tier from annotator 1 ("A"/"B"). |
| `level_2` | string or null | Speaker tier from annotator 3. |
| `gender` | string or null | Speaker gender (`"M"`/`"F"`). |
| `duration` | float | Utterance duration in seconds (after resampling to 16 kHz). |
| `sample_rate` | int | Sample rate in Hz (16,000). |
| `wav_path` | string | Waveform filename (`<utt_id>.wav`). |
| `audio` | Audio | Automatically loaded waveform (16 kHz). |
| `transcription` | string or null | Reference sentence text. |

### Data Splits

| Split | # Examples |
|-------|------------|
| train | 1,903 |
| test  | 1,263 |

### Notes

- When annotator 3 did not label an utterance, related fields (`annot_2`, `label_2`, `global_2`, `level_2`) are absent or set to null.
- Error types come from simple heuristics contrasting MFA reference phones with annotator 1 labels.
- Waveforms were resampled to 16 kHz using `ffmpeg` during manifest generation.
- Forced alignments and annotations were merged to produce enriched CSV files per speaker/partition.
- Global scores are averaged per speaker to derive `level_*` tiers (`A` if mean ≥ 3, `B` otherwise).

## Licensing

- Audio and annotations: CC BY-NC 4.0 (non-commercial use allowed with attribution).

## Citation

```
@article{vidal2019epadb,
  title   = {EpaDB: a database for development of pronunciation assessment systems},
  author  = {Vidal, Jazmin and Ferrer, Luciana and Brambilla, Leonardo},
  journal = {Proc. Interspeech},
  pages   = {589--593},
  year    = {2019}
}
```

## Usage

Install dependencies and load the dataset:

```python
from datasets import load_dataset

# Local usage before uploading:
ds = load_dataset(
    "epadb_dataset/epadb.py",
    data_dir="/path/to/epadb",  # folder with train.json, test.json, WAV/
    split="train",
)
print(ds)
print(ds[0]["utt_id"], ds[0]["audio"]["sampling_rate"])  # 16000

# After pushing to the Hugging Face Hub:
# ds = load_dataset("JazminVidal/epadb", split="train")
```

## Acknowledgements

The database is an effort of the Speech Lab at the Laboratorio de Inteligencia Artificial Aplicada from 
the Universidad de Buenos Aires and was partially funded by Google by a Google Latin America Reseach Award in 2018