Dataset Viewer
Auto-converted to Parquet Duplicate
id
int32
1
500
audio
audioduration (s)
0.65
46.1
text
stringclasses
0 values
duration
float64
0.65
46.1
1
null
25.725986
2
null
1.553515
3
null
3.666259
4
null
1.708844
5
null
1.491361
6
null
0.683537
7
null
0.652472
8
null
1.367097
9
null
3.790544
10
null
3.759479
11
null
4.660499
12
null
2.672018
13
null
2.174898
14
null
2.23703
15
null
3.013787
16
null
3.355556
17
null
5.40619
18
null
6.213991
19
null
3.82161
20
null
3.91483
21
null
9.320998
22
null
6.555782
23
null
8.73068
24
null
4.443016
25
null
2.703084
26
null
10.625941
27
null
5.375125
28
null
3.510907
29
null
3.169138
30
null
5.561542
31
null
1.584558
32
null
4.287664
33
null
4.287664
34
null
6.649002
35
null
9.787052
36
null
22.121859
37
null
8.637461
38
null
7.270385
39
null
9.196712
40
null
11.340567
41
null
8.870997
42
null
1.242812
43
null
7.115057
44
null
1.770998
45
null
7.394671
46
null
15.441814
47
null
12.521224
48
null
2.423469
49
null
22.77433
50
null
15.752517
51
null
33.369205
52
null
6.027596
53
null
2.050612
54
null
2.268095
55
null
4.22551
56
null
7.052903
57
null
8.016077
58
null
9.445284
59
null
1.833152
60
null
8.233561
61
null
9.476372
62
null
4.50517
63
null
1.615646
64
null
2.361338
65
null
3.510907
66
null
3.324512
67
null
6.866463
68
null
10.750227
69
null
8.637461
70
null
3.852676
71
null
2.23703
72
null
7.891791
73
null
8.513197
74
null
10.097755
75
null
1.833129
76
null
2.858458
77
null
4.846916
78
null
13.888322
79
null
1.802063
80
null
2.92059
81
null
13.173697
82
null
9.2278
83
null
13.670816
84
null
1.926327
85
null
5.918844
86
null
10.734694
87
null
7.907324
88
null
21.205307
89
null
11.961973
90
null
7.783039
91
null
4.458571
92
null
9.491905
93
null
7.052903
94
null
10.051156
95
null
11.014331
96
null
7.658753
97
null
10.159909
98
null
10.284172
99
null
14.929161
100
null
9.056916
End of preview. Expand in Data Studio

Bambara ASR Benchmark

The first standardized evaluation set for Automatic Speech Recognition in Bambara (Bamanankan). One hour of studio-quality constitutional text, transcribed and validated by linguists from Mali's Direction Nationale de l'Éducation Non Formelle et des Langues Nationales (DNENF-LN).

This benchmark accompanies the paper "Where Are We at with Automatic Speech Recognition for the Bambara Language?" and the public leaderboard at MALIBA-AI/bambara-asr-leaderboard.

Purpose: Evaluation only. This dataset is a test set — it is not intended for training.

Overview

Domain Malian Constitution (legal/institutional text)
Duration 1.075 hours (64.5 minutes)
Segments 500
Speaker(s) 1 adult male
Segment length 0.65–46.12s (mean: 7.47s, 97% under 20s)
Vocabulary 1,198 unique words
Recording Single-channel, studio conditions
Orthography Standard Bambara latin script per DNENF-LN
Code-switching None
SNR 99% of utterances ≥ 15 dB (noise-free)

Why This Benchmark

Bambara ASR has seen rapid growth since 2022, but until now there was no common test set for comparing models. Researchers reported WER/CER on internal splits, making cross-model comparison unreliable.

This benchmark fills that gap. It is deliberately formal and domain-specific: the vocabulary is legal, the orthography is strict, and there is no code-switching. This makes it a stress test for domain robustness and out-of-vocabulary handling.

Because the acoustic conditions are near-optimal (studio recording, clean audio, professional speaker), the error rates reported here should be interpreted as approximate lower bounds — real-world performance on naturalistic Bambara speech will almost certainly be worse.

Vocabulary Isolation

The vocabulary in this dataset has very low overlap with existing Bambara ASR training corpora. To verify that, we have drawn two one-hour subsets from the two publicly available ASR dataset (at the time of compiling this report) by randomly selecting utterances and we observed the following word overlap figures:

Training corpus Shared words Overlap with this benchmark
RobotsMali/bam-asr-early 297 ~25%
RobotsMali/kunkado 319 ~27%

So this benchmark features about 75% unique words when compared with equal sized samples of these two datasets, confirming how narrow the domain is. Words like yuruguyuruguli (disorder/embezzlement) and yamaruyasariya (ordinance/regulation) are central to the constitutional register but virtually absent from conversational Bambara.

Leaderboard

Full results and adjustable metric weights at the public leaderboard.

Dataset Construction

Source. The text and audio originate from the DNENF-LN's official Bambara translation of the Malian Constitution (as of July 2023, 191 articles).

Segmentation. Manual segmentation and forced alignment were performed using Audacity. The resulting segments range from 600ms to 46 seconds.

Transcription validation. An annotator reviewed every segment to align the transcription with the actual spoken audio, correcting cases where the speaker paraphrased or interpreted the written text rather than reading it verbatim. All words were verified against the DNENF-LN dictionary and Bamadaba.

Code-switching protocol. French-derived words already adopted into the official Bambara lexicon (e.g., politiki, afiriki, minisiri) were transcribed in standard Bambara spelling. No French-spelled words remain in the dataset.

Silent/music segments. 8 non-speech segments from the original recording were retained to test model robustness against hallucination on silence.

Known Limitations

  • Single speaker, single domain. One male voice reading constitutional text. No speaker or topic diversity.
  • Near-ideal acoustics. Studio recording, high SNR. Does not reflect real-world noise, phone-quality audio, or multi-speaker scenarios.
  • No code-switching. Urban Bambara frequently mixes French. This benchmark does not test that ability.
  • Normalization artifacts. Standard WER penalizes valid Bambara orthographic alternations (e.g., b'a vs. bɛ a, compound segmentation like yɛrɛmahɔrɔnya vs. yɛrɛma hɔrɔnya). A linguistically-aware normalization would yield different error rates.
  • 1 hour is small. Consistent patterns across 37 models suggest the findings generalize, but a larger benchmark would strengthen conclusions.

Future versions will incorporate speaker diversity, domain variation, naturalistic speech, varied acoustic conditions, and code-switching.

Usage

from datasets import load_dataset

ds = load_dataset("MALIBA-AI/bambara-asr-benchmark", "eval", split="eval")

Links

Citation

@misc{BambaraASRBenchmark2025,
  title        = {Where Are We at with Automatic Speech Recognition for the Bambara Language?},
  author       = {Seydou Diallo and Yacouba Diarra and Mamadou K. Keita and Panga Azazia Kamat{\'e} and Adam Bouno Kampo and Aboubacar Ouattara},
  year         = {2025},
  howpublished = {Hugging Face Datasets},
  url          = {https://huggingface.co/datasets/MALIBA-AI/bambara-asr-benchmark}
}
Downloads last month
17