Datasets:
id
int32 1
500
| audio
audioduration (s) 0.65
46.1
| text
stringclasses 0
values | duration
float64 0.65
46.1
|
|---|---|---|---|
1
| null | 25.725986
|
|
2
| null | 1.553515
|
|
3
| null | 3.666259
|
|
4
| null | 1.708844
|
|
5
| null | 1.491361
|
|
6
| null | 0.683537
|
|
7
| null | 0.652472
|
|
8
| null | 1.367097
|
|
9
| null | 3.790544
|
|
10
| null | 3.759479
|
|
11
| null | 4.660499
|
|
12
| null | 2.672018
|
|
13
| null | 2.174898
|
|
14
| null | 2.23703
|
|
15
| null | 3.013787
|
|
16
| null | 3.355556
|
|
17
| null | 5.40619
|
|
18
| null | 6.213991
|
|
19
| null | 3.82161
|
|
20
| null | 3.91483
|
|
21
| null | 9.320998
|
|
22
| null | 6.555782
|
|
23
| null | 8.73068
|
|
24
| null | 4.443016
|
|
25
| null | 2.703084
|
|
26
| null | 10.625941
|
|
27
| null | 5.375125
|
|
28
| null | 3.510907
|
|
29
| null | 3.169138
|
|
30
| null | 5.561542
|
|
31
| null | 1.584558
|
|
32
| null | 4.287664
|
|
33
| null | 4.287664
|
|
34
| null | 6.649002
|
|
35
| null | 9.787052
|
|
36
| null | 22.121859
|
|
37
| null | 8.637461
|
|
38
| null | 7.270385
|
|
39
| null | 9.196712
|
|
40
| null | 11.340567
|
|
41
| null | 8.870997
|
|
42
| null | 1.242812
|
|
43
| null | 7.115057
|
|
44
| null | 1.770998
|
|
45
| null | 7.394671
|
|
46
| null | 15.441814
|
|
47
| null | 12.521224
|
|
48
| null | 2.423469
|
|
49
| null | 22.77433
|
|
50
| null | 15.752517
|
|
51
| null | 33.369205
|
|
52
| null | 6.027596
|
|
53
| null | 2.050612
|
|
54
| null | 2.268095
|
|
55
| null | 4.22551
|
|
56
| null | 7.052903
|
|
57
| null | 8.016077
|
|
58
| null | 9.445284
|
|
59
| null | 1.833152
|
|
60
| null | 8.233561
|
|
61
| null | 9.476372
|
|
62
| null | 4.50517
|
|
63
| null | 1.615646
|
|
64
| null | 2.361338
|
|
65
| null | 3.510907
|
|
66
| null | 3.324512
|
|
67
| null | 6.866463
|
|
68
| null | 10.750227
|
|
69
| null | 8.637461
|
|
70
| null | 3.852676
|
|
71
| null | 2.23703
|
|
72
| null | 7.891791
|
|
73
| null | 8.513197
|
|
74
| null | 10.097755
|
|
75
| null | 1.833129
|
|
76
| null | 2.858458
|
|
77
| null | 4.846916
|
|
78
| null | 13.888322
|
|
79
| null | 1.802063
|
|
80
| null | 2.92059
|
|
81
| null | 13.173697
|
|
82
| null | 9.2278
|
|
83
| null | 13.670816
|
|
84
| null | 1.926327
|
|
85
| null | 5.918844
|
|
86
| null | 10.734694
|
|
87
| null | 7.907324
|
|
88
| null | 21.205307
|
|
89
| null | 11.961973
|
|
90
| null | 7.783039
|
|
91
| null | 4.458571
|
|
92
| null | 9.491905
|
|
93
| null | 7.052903
|
|
94
| null | 10.051156
|
|
95
| null | 11.014331
|
|
96
| null | 7.658753
|
|
97
| null | 10.159909
|
|
98
| null | 10.284172
|
|
99
| null | 14.929161
|
|
100
| null | 9.056916
|
Bambara ASR Benchmark
The first standardized evaluation set for Automatic Speech Recognition in Bambara (Bamanankan). One hour of studio-quality constitutional text, transcribed and validated by linguists from Mali's Direction Nationale de l'Éducation Non Formelle et des Langues Nationales (DNENF-LN).
This benchmark accompanies the paper "Where Are We at with Automatic Speech Recognition for the Bambara Language?" and the public leaderboard at MALIBA-AI/bambara-asr-leaderboard.
Purpose: Evaluation only. This dataset is a test set — it is not intended for training.
Overview
| Domain | Malian Constitution (legal/institutional text) |
| Duration | 1.075 hours (64.5 minutes) |
| Segments | 500 |
| Speaker(s) | 1 adult male |
| Segment length | 0.65–46.12s (mean: 7.47s, 97% under 20s) |
| Vocabulary | 1,198 unique words |
| Recording | Single-channel, studio conditions |
| Orthography | Standard Bambara latin script per DNENF-LN |
| Code-switching | None |
| SNR | 99% of utterances ≥ 15 dB (noise-free) |
Why This Benchmark
Bambara ASR has seen rapid growth since 2022, but until now there was no common test set for comparing models. Researchers reported WER/CER on internal splits, making cross-model comparison unreliable.
This benchmark fills that gap. It is deliberately formal and domain-specific: the vocabulary is legal, the orthography is strict, and there is no code-switching. This makes it a stress test for domain robustness and out-of-vocabulary handling.
Because the acoustic conditions are near-optimal (studio recording, clean audio, professional speaker), the error rates reported here should be interpreted as approximate lower bounds — real-world performance on naturalistic Bambara speech will almost certainly be worse.
Vocabulary Isolation
The vocabulary in this dataset has very low overlap with existing Bambara ASR training corpora. To verify that, we have drawn two one-hour subsets from the two publicly available ASR dataset (at the time of compiling this report) by randomly selecting utterances and we observed the following word overlap figures:
| Training corpus | Shared words | Overlap with this benchmark |
|---|---|---|
RobotsMali/bam-asr-early |
297 | ~25% |
RobotsMali/kunkado |
319 | ~27% |
So this benchmark features about 75% unique words when compared with equal sized samples of these two datasets, confirming how narrow the domain is. Words like yuruguyuruguli (disorder/embezzlement) and yamaruyasariya (ordinance/regulation) are central to the constitutional register but virtually absent from conversational Bambara.
Leaderboard
Full results and adjustable metric weights at the public leaderboard.
Dataset Construction
Source. The text and audio originate from the DNENF-LN's official Bambara translation of the Malian Constitution (as of July 2023, 191 articles).
Segmentation. Manual segmentation and forced alignment were performed using Audacity. The resulting segments range from 600ms to 46 seconds.
Transcription validation. An annotator reviewed every segment to align the transcription with the actual spoken audio, correcting cases where the speaker paraphrased or interpreted the written text rather than reading it verbatim. All words were verified against the DNENF-LN dictionary and Bamadaba.
Code-switching protocol. French-derived words already adopted into the official Bambara lexicon (e.g., politiki, afiriki, minisiri) were transcribed in standard Bambara spelling. No French-spelled words remain in the dataset.
Silent/music segments. 8 non-speech segments from the original recording were retained to test model robustness against hallucination on silence.
Known Limitations
- Single speaker, single domain. One male voice reading constitutional text. No speaker or topic diversity.
- Near-ideal acoustics. Studio recording, high SNR. Does not reflect real-world noise, phone-quality audio, or multi-speaker scenarios.
- No code-switching. Urban Bambara frequently mixes French. This benchmark does not test that ability.
- Normalization artifacts. Standard WER penalizes valid Bambara orthographic alternations (e.g., b'a vs. bɛ a, compound segmentation like yɛrɛmahɔrɔnya vs. yɛrɛma hɔrɔnya). A linguistically-aware normalization would yield different error rates.
- 1 hour is small. Consistent patterns across 37 models suggest the findings generalize, but a larger benchmark would strengthen conclusions.
Future versions will incorporate speaker diversity, domain variation, naturalistic speech, varied acoustic conditions, and code-switching.
Usage
from datasets import load_dataset
ds = load_dataset("MALIBA-AI/bambara-asr-benchmark", "eval", split="eval")
Links
- Leaderboard: huggingface.co/spaces/MALIBA-AI/bambara-asr-leaderboard
- Code: github.com/MALIBA-AI/bambara-asr-leaderboard
- Paper: Where Are We at with Automatic Speech Recognition for the Bambara Language?
Citation
@misc{BambaraASRBenchmark2025,
title = {Where Are We at with Automatic Speech Recognition for the Bambara Language?},
author = {Seydou Diallo and Yacouba Diarra and Mamadou K. Keita and Panga Azazia Kamat{\'e} and Adam Bouno Kampo and Aboubacar Ouattara},
year = {2025},
howpublished = {Hugging Face Datasets},
url = {https://huggingface.co/datasets/MALIBA-AI/bambara-asr-benchmark}
}
- Downloads last month
- 17