File size: 8,806 Bytes
3959697
 
0cf9eb1
3959697
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
0cf9eb1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
382cd21
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3959697
 
 
 
 
 
 
0cf9eb1
 
 
 
382cd21
 
 
 
a88276c
 
 
 
 
 
 
 
 
 
 
 
8679523
 
 
 
 
fa4694d
8679523
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
fa4694d
 
 
 
 
 
 
 
8679523
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
---
dataset_info:
- config_name: human-corrected
  features:
  - name: text
    dtype: string
  - name: duration
    dtype: float64
  - name: audio
    dtype: audio
  - name: label-v1
    dtype: string
  - name: label-v2
    dtype: string
  splits:
  - name: train
    num_bytes: 62771143761
    num_examples: 253290
  - name: test
    num_bytes: 1515394591
    num_examples: 6718
  download_size: 59319505964
  dataset_size: 64286538352
- config_name: model-annotated
  features:
  - name: duration
    dtype: float64
  - name: audio
    dtype: audio
  - name: label-v1
    dtype: string
  - name: label-v2
    dtype: string
  splits:
  - name: train
    num_bytes: 55616591334
    num_examples: 355571
  download_size: 66321575877
  dataset_size: 55616591334
- config_name: short
  features:
  - name: audio
    dtype: audio
  - name: duration
    dtype: float64
  - name: label-v1
    dtype: string
  - name: label-v2
    dtype: string
  splits:
  - name: train
    num_bytes: 16345361845
    num_examples: 259183
  download_size: 16319374978
  dataset_size: 16345361845
configs:
- config_name: human-corrected
  data_files:
  - split: train
    path: human-corrected/train-*
  - split: test
    path: human-corrected/test-*
- config_name: model-annotated
  data_files:
  - split: train
    path: model-annotated/train-*
- config_name: short
  data_files:
  - split: train
    path: short/train-*
license: cc-by-4.0
task_categories:
- automatic-speech-recognition
language:
- bm
tags:
- bambara
- African-Next-Voices
- ANV
- RobotsMali
- afvoices
- asr
pretty_name: Robots
---

# 📘 **African Next Voices – Bambara (AfVoices)**

The **AfVoices** dataset is the largest open corpus of spontaneous Bambara speech at its release in late 2025. It contains **423 hours** of segmented audio and **612 hours** of original raw recordings collected across southern Mali. Speech was recorded in natural, conversational settings and annotated using a semi-automated transcription pipeline combining ASR pre-labels and human corrections. We release all the data processing code on [GitHub](https://github.com/RobotsMali-AI/afvoices).

---

## 🔎 **Quick Facts**

| Category                                 | Value                                                                                             |
| ---------------------------------------- | ------------------------------------------------------------------------------------------------- |
| **Total raw hours**                      | 612 h (1,777 raw recordings; publicly available on GCS)                                           |
| **Total segmented hours**                | 423 h (874,762 segments)                                                                          |
| **Speakers**                             | 512                                                                                               |
| **Regions**                              | Bamako, Ségou, Sikasso, Bagineda, Bougouni                                                        |
| **Avg. segment duration**                | ~2 seconds                                                                                        |
| **Subsets**                              | 159 h human-corrected, 212 h model-annotated, 52 h short (<1s)                                    |
| **Age distribution**                     | Broad, across young to elderly speakers (90% between 18 and 45)                                   |
| **Topics**                               | Health, agriculture, Miscellaneous (art, education, history etc.)                                 |
| **SNR distribution (raw recordings)**    | 71.75% High or Very High SNR                                                                      |
| **Train / Test split**                   | 155 h / 4 h                                                                                       |

---

## **Motivation**

The **African Next Voices (ANV)** project is a multi-country effort aiming to gather over **9,000 hours of speech** across 18 African languages. Its goal is to build high-quality datasets that empower local communities, support inclusive AI research, and provide strong foundations for ASR in underrepresented languages.

As part of this initiative, **RobotsMali** led the Bambara data collection for Mali. This dataset reflects RobotsMali’s broader mission to advance AI and NLP research malian languages, with a long-term focus on improving education, access, and technology across Mali and the wider Manding linguistic region.

---

## 🎙️ **Characteristics of the Dataset**

### **Data Collection**

* Speech was collected through trained **facilitators** who guided participants, ensured audio quality, and encouraged natural, topic-focused conversations.
* All recordings are **spontaneous speech**, not read text.
* A custom **Flutter mobile app** ([open-source](https://github.com/RobotsMali-AI/Africa-Voice-App)) was used to simplify the process and reduce training time.
* Geographic focus: **Southern Mali**, to limit extreme accent variation and build a clean baseline corpus.

### **Segmentation and Preprocessing**

* Raw audio was segmented using **Silero VAD**, retaining ~70% of the original duration.
* Segments range from **240 ms to 30 s**.
* Voice activity detection helped remove long silences and improve data usability.

### **Transcriptions**

* Pre-transcribed using the ASR model **soloni-114m-tdt-ctc-v0**.
* Human annotators corrected the transcripts.
* A second model (**soloni-114m-tdt-ctc-v2**) was trained using the corrected transcripts and used to regenerate improved labels.
* Two automatic transcription variants exist for each sample: **v1** (from soloni-v0) and **v2** (from soloni-v2).

### **Acoustic Event Tags**

The following tags appear in transcriptions:

| Tag       | Meaning                                                       |
| --------- | ------------------------------------------------------------- |
| `[um]`    | Vocalized pauses, filler sounds                               |
| `[cs]`    | Code-switched or foreign word                                 |
| `[noise]` | Background noise (applause, coughing, children, etc.)         |
| `[?]`     | Inaudible or overlapped speech                                |
| `[pause]` | Long silence (>5 seconds or >3 seconds at segment boundaries); due to VAD segmentation this tag is rarely used |

---

## 📂 **Subsets**

### **1. Human-corrected (159 h, 260k samples)**

* Fully reviewed and corrected by annotators.
* Only subset with a definitive `text` field containing the validated transcription.

### **2. Model-annotated (212 h, 355k samples)**

* Includes automatic labels: `v1` (soloni-v0) and `v2` (soloni-v2).
* No human review.

### **3. Short subset (52 h, 259k samples)**

* Segments <1 second (formulaic expressions, discourse markers).
* Excluded from human annotation for optimization purposes.
* Automatically labeled (v1 & v2).

---

## ⚠️ **Limitations**

* **Clean dataset vs real-world noise:**
  Over 70% of recordings can be categorized as relatively clean speech. Models trained solely on this dataset may underperform in noisy street or radio environments typical in Mali. See this [report](https://zenodo.org/records/17672774) if you are interested in learning more about the strengths and weaknesses of RobotsMali's ASR models.

* **Reduced code-switching:**
  French terms were often replaced by `[cs]` or normalized into Bambara phonology. This improves model stability but reduces realism for natural bilingual speech.

* **Geographic homogeneity:**
  Focused on the southern region to control accent variability. Broader dialectal coverage might require additional data.

* **Simplified linguistic conditions:**
  Overlaps, multi-speaker settings, and conversational chaos are minimized—again improving training stability at the cost of deployment realism.

---

## 📑 **Citation**

```bibtex
@misc{diarra2025dealinghardfactslowresource,
      title={Dealing with the Hard Facts of Low-Resource African NLP}, 
      author={Yacouba Diarra and Nouhoum Souleymane Coulibaly and Panga Azazia Kamaté and Madani Amadou Tall and Emmanuel Élisé Koné and Aymane Dembélé and Michael Leventhal},
      year={2025},
      eprint={2511.18557},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2511.18557}, 
}
```

---

You may want to download the original 612 hours dataset with its associated metadata for research purposes or to create a derivative. You will find the codes and manifest files to download those files from Google Cloud Storage in this repository: [RobotsMali-AI/afvoices](https://github.com/RobotsMali-AI/afvoices). Do not hesitate to open an issue for Help or suggestions 🤗