Datasets:
Update README
#1
by
diarray
- opened
README.md
CHANGED
|
@@ -1,5 +1,5 @@
|
|
| 1 |
---
|
| 2 |
-
license:
|
| 3 |
configs:
|
| 4 |
- config_name: default
|
| 5 |
data_files:
|
|
@@ -35,25 +35,33 @@ dataset_info:
|
|
| 35 |
num_examples: 45
|
| 36 |
download_size: 19975985
|
| 37 |
dataset_size: 20690456
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 38 |
---
|
| 39 |
# Nyana-Eval Dataset
|
| 40 |
|
| 41 |
## Dataset Description
|
| 42 |
|
| 43 |
-
**Nyana-Eval** is a compact, stratified evaluation subset for benchmarking Automatic Speech Recognition (ASR) models in Bambara. It consists of **45 audio recordings** totaling approximately **3.03 minutes
|
| 44 |
|
| 45 |
Nyana-Eval is ideal for:
|
| 46 |
- Rapid evaluation of Bambara ASR models (e.g., WER/CER computation on diverse conditions).
|
| 47 |
- Human-assisted qualitative analysis (e.g., semantic fidelity, code-switching handling).
|
| 48 |
-
-
|
| 49 |
|
| 50 |
**Key Statistics**:
|
| 51 |
- **Total Samples**: 45 (balanced: 15 per source subset).
|
| 52 |
- **Total Duration**: ~3.03 minutes (average ~4.0 seconds per sample).
|
| 53 |
-
- **Audio Format**: Mono-channel WAV files at 16 kHz sampling rate.
|
| 54 |
- **Languages**: Primary: Bambara (Bamana); Secondary: French code-switching (~15% of samples).
|
| 55 |
-
- **License**:
|
| 56 |
-
- **Tags**: `audio`, `speech`, `asr`, `bambara`, `bamanan`, `low-resource`, `evaluation`, `human-evaluation`, `african-languages`.
|
| 57 |
|
| 58 |
Compiled by Robots Mali AI4D Lab, this dataset powers the human-comparative analysis in the [Bambara ASR Models Evaluation Report].
|
| 59 |
|
|
@@ -65,10 +73,10 @@ Nyana-Eval is a single-split dataset (default: `test`), with each entry includin
|
|
| 65 |
|
| 66 |
| Column | Type | Description | Example Value |
|
| 67 |
|-----------------|----------|-----------------------------------------------------------------------------|---------------|
|
| 68 |
-
| `audio` | Audio | Raw audio waveform (array + sampling rate: 16 kHz) or file path.
|
| 69 |
-
| `duration` | Float64 | Length of the audio clip in seconds (range: 0.62s – 15s).
|
| 70 |
-
| `references` | String | Bambara text
|
| 71 |
-
| '8 * models transcriptions' | String |
|
| 72 |
|
| 73 |
### Splits
|
| 74 |
- **Default Split**: Full 45 samples (`test` for evaluation).
|
|
@@ -84,17 +92,17 @@ print(dataset[0]) # Example: {'audio': ..., 'duration': 3.45, 'transcription'
|
|
| 84 |
|
| 85 |
## Sources and Compilation
|
| 86 |
|
| 87 |
-
Nyana-Eval is a **balanced subsample (15 per subset)** from the full 500-sample [RobotsMali/Bam_ASR_Eval_500](https://huggingface.co/datasets/RobotsMali/Bam_ASR_Eval_500) corpus (~36.69 minutes total). Selection criteria ensured diversity: voice variety (age/gender/accents), acoustic challenges (noise/volume/overlaps), and linguistic phenomena (code-switching, proper names
|
| 88 |
|
| 89 |
**Parent Subsets Breakdown** (15 samples each in Nyana-Eval):
|
| 90 |
- **Ref. 1: RobotsMali/kunkado (Hugging Face)** – 15 audios (~1.96 minutes scaled).
|
| 91 |
Semi-supervised interviews and spontaneous discourse. Source: [RobotsMali/kunkado](https://huggingface.co/datasets/RobotsMali/kunkado). Focus: Dialectal variations and natural flow.
|
| 92 |
|
| 93 |
-
- **Ref. 2: jeli-ASR street interviews subset** –
|
| 94 |
-
|
| 95 |
|
| 96 |
-
- **Ref. 3:
|
| 97 |
-
User-generated readings from the Bambara learning
|
| 98 |
|
| 99 |
## Metadata
|
| 100 |
|
|
@@ -106,13 +114,10 @@ Nyana-Eval is a **balanced subsample (15 per subset)** from the full 500-sample
|
|
| 106 |
- **Download Size**: ~25 MB (audios + metadata).
|
| 107 |
- **Ethical Notes**: Ethically sourced/anonymized; focuses on public-domain cultural speech. For research; cite Robots Mali.
|
| 108 |
|
| 109 |
-
###
|
| 110 |
-
- **Dialect Coverage**: Urban Bamana (Bamako-influenced) with rural elements.
|
| 111 |
-
- **Phonetic Coverage**: Tones, nasals, contractions; with OOD proper names (e.g., "Sunjata," "Traoré").
|
| 112 |
-
- **Challenges Represented**:
|
| 113 |
- Code-switching: samples (e.g., "Segou ville").
|
|
|
|
| 114 |
- Noise/Overlaps: (e.g., low-volume interviews, multi-speaker).
|
| 115 |
-
- Human Eval Focus: 45 samples scored 0-3 + bonuses for fidelity, names, switching, robustness.
|
| 116 |
|
| 117 |
### Evaluation Metadata
|
| 118 |
- **WER Baselines**: From report – Soloni-v2: 36.07%; QuartzNet-v0: 65.42% (greedy decoding).
|
|
@@ -127,6 +132,6 @@ Nyana-Eval is a **balanced subsample (15 per subset)** from the full 500-sample
|
|
| 127 |
## Related Resources
|
| 128 |
- **Parent Dataset**: [RobotsMali/Bam_ASR_Eval_500](https://huggingface.co/datasets/RobotsMali/Bam_ASR_Eval_500) (full 500 samples).
|
| 129 |
- **Models**: Test with [RobotsMali ASR models](https://huggingface.co/RobotsMali/models)
|
| 130 |
-
- **App**: Collect similar data via [An Bɛ Kalan](https://play.google.com/store/apps/details?id=
|
| 131 |
|
| 132 |
-
This README is self-contained; explore the attached report PDF for detailed human annotations and model rankings on these exact 45 samples!
|
|
|
|
| 1 |
---
|
| 2 |
+
license: cc-by-4.0
|
| 3 |
configs:
|
| 4 |
- config_name: default
|
| 5 |
data_files:
|
|
|
|
| 35 |
num_examples: 45
|
| 36 |
download_size: 19975985
|
| 37 |
dataset_size: 20690456
|
| 38 |
+
task_categories:
|
| 39 |
+
- automatic-speech-recognition
|
| 40 |
+
language:
|
| 41 |
+
- bm
|
| 42 |
+
tags:
|
| 43 |
+
- speech
|
| 44 |
+
- asr
|
| 45 |
+
- bambara
|
| 46 |
+
- low-resource
|
| 47 |
---
|
| 48 |
# Nyana-Eval Dataset
|
| 49 |
|
| 50 |
## Dataset Description
|
| 51 |
|
| 52 |
+
**Nyana-Eval** is a compact, stratified evaluation subset for benchmarking Automatic Speech Recognition (ASR) models in Bambara. It consists of **45 audio recordings** totaling approximately **3.03 minutes**, carefully selected to represent real-world linguistic and acoustic challenges in low-resource Bambara speech. This dataset is derived from the larger [RobotsMali/Bam_ASR_Eval_500](https://huggingface.co/datasets/RobotsMali/Bam_ASR_Eval_500) corpus and is optimized for quick, reproducible human evaluation.
|
| 53 |
|
| 54 |
Nyana-Eval is ideal for:
|
| 55 |
- Rapid evaluation of Bambara ASR models (e.g., WER/CER computation on diverse conditions).
|
| 56 |
- Human-assisted qualitative analysis (e.g., semantic fidelity, code-switching handling).
|
| 57 |
+
- Testing models on low-resource settings gaps: dialectal variations, noise, proper names, and code-mixing with French.
|
| 58 |
|
| 59 |
**Key Statistics**:
|
| 60 |
- **Total Samples**: 45 (balanced: 15 per source subset).
|
| 61 |
- **Total Duration**: ~3.03 minutes (average ~4.0 seconds per sample).
|
| 62 |
+
- **Audio Format**: Mono-channel WAV files at 16 | 44.1k kHz sampling rate.
|
| 63 |
- **Languages**: Primary: Bambara (Bamana); Secondary: French code-switching (~15% of samples).
|
| 64 |
+
- **License**: CC-BY-4.0 License (open for research, commercial use with attribution).
|
|
|
|
| 65 |
|
| 66 |
Compiled by Robots Mali AI4D Lab, this dataset powers the human-comparative analysis in the [Bambara ASR Models Evaluation Report].
|
| 67 |
|
|
|
|
| 73 |
|
| 74 |
| Column | Type | Description | Example Value |
|
| 75 |
|-----------------|----------|-----------------------------------------------------------------------------|---------------|
|
| 76 |
+
| `audio` | Audio | Raw audio waveform (array + sampling rate: 16 | 44.1k kHz) or file path. | `{"path": "1.1.wav", "array": [...], "sampling_rate": 16000}` |
|
| 77 |
+
| `duration` | Float64 | Length of the audio clip in seconds (range: 0.62s – 15s). | 3.45 |
|
| 78 |
+
| `references` | String | Bambara text | "nɔgɔ ye a ka tɔɔrɔ ye" |
|
| 79 |
+
| '8 * models transcriptions' | String | ASR provised transcriptions | |
|
| 80 |
|
| 81 |
### Splits
|
| 82 |
- **Default Split**: Full 45 samples (`test` for evaluation).
|
|
|
|
| 92 |
|
| 93 |
## Sources and Compilation
|
| 94 |
|
| 95 |
+
Nyana-Eval is a **balanced subsample (15 per subset)** from the full 500-sample [RobotsMali/Bam_ASR_Eval_500](https://huggingface.co/datasets/RobotsMali/Bam_ASR_Eval_500) corpus (~36.69 minutes total). Selection criteria ensured diversity: voice variety (age/gender/accents), acoustic challenges (noise/volume/overlaps), and linguistic phenomena (code-switching, proper names etc.)
|
| 96 |
|
| 97 |
**Parent Subsets Breakdown** (15 samples each in Nyana-Eval):
|
| 98 |
- **Ref. 1: RobotsMali/kunkado (Hugging Face)** – 15 audios (~1.96 minutes scaled).
|
| 99 |
Semi-supervised interviews and spontaneous discourse. Source: [RobotsMali/kunkado](https://huggingface.co/datasets/RobotsMali/kunkado). Focus: Dialectal variations and natural flow.
|
| 100 |
|
| 101 |
+
- **Ref. 2: jeli-ASR street interviews subset** – 30 audios (~1.85 minutes).
|
| 102 |
+
Street interviews Subset from the jeli-asr project. Source: [jeli-asr](https://github.com/robotsmali-ai/jeli-asr/)
|
| 103 |
|
| 104 |
+
- **Ref. 3: Readings of Excerpts from An Bɛ Kalan app (RobotsMali)** – 220 audios (~20.06 minutes).
|
| 105 |
+
User-generated readings and interactions from the mobile app for Bambara learning, captures learner speech with occasional errors or pauses. source: [https://github.com/Robotsmali-ai/an-be-kalan]
|
| 106 |
|
| 107 |
## Metadata
|
| 108 |
|
|
|
|
| 114 |
- **Download Size**: ~25 MB (audios + metadata).
|
| 115 |
- **Ethical Notes**: Ethically sourced/anonymized; focuses on public-domain cultural speech. For research; cite Robots Mali.
|
| 116 |
|
| 117 |
+
### Challenges Represented:
|
|
|
|
|
|
|
|
|
|
| 118 |
- Code-switching: samples (e.g., "Segou ville").
|
| 119 |
+
- Proper names (e.g., "Sunjata," "Traoré").
|
| 120 |
- Noise/Overlaps: (e.g., low-volume interviews, multi-speaker).
|
|
|
|
| 121 |
|
| 122 |
### Evaluation Metadata
|
| 123 |
- **WER Baselines**: From report – Soloni-v2: 36.07%; QuartzNet-v0: 65.42% (greedy decoding).
|
|
|
|
| 132 |
## Related Resources
|
| 133 |
- **Parent Dataset**: [RobotsMali/Bam_ASR_Eval_500](https://huggingface.co/datasets/RobotsMali/Bam_ASR_Eval_500) (full 500 samples).
|
| 134 |
- **Models**: Test with [RobotsMali ASR models](https://huggingface.co/RobotsMali/models)
|
| 135 |
+
- **App**: Collect similar data via [An Bɛ Kalan](https://play.google.com/store/apps/details?id=org.robotsmali.literacy_app).
|
| 136 |
|
| 137 |
+
This README is self-contained; explore the attached report PDF for detailed human annotations and model rankings on these exact 45 samples!
|