Update README.md
Browse files
README.md
CHANGED
|
@@ -102,15 +102,18 @@ The **Simba** family consists of state-of-the-art models fine-tuned using SimbaB
|
|
| 102 |
- **Simba-M** (MMS-1b-all)
|
| 103 |
- **Simba-H** (AfriHuBERT)
|
| 104 |
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
|
|
| 108 |
-
|
| 109 |
-
| π₯**Simba-
|
| 110 |
-
| π₯**Simba-
|
| 111 |
-
| π₯**Simba-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
|
| 113 |
-
* **Simba-S** (based on SeamlessM4T-v2-MT) emerged as the best-performing ASR model overall.
|
| 114 |
|
| 115 |
**π§© Usage Example**
|
| 116 |
|
|
@@ -125,7 +128,9 @@ asr_pipeline = pipeline(
|
|
| 125 |
model="UBC-NLP/Simba-S" #Simba mdoels `UBC-NLP/Simba-S`, `UBC-NLP/Simba-W`, `UBC-NLP/Simba-X`, `UBC-NLP/Simba-H`, `UBC-NLP/Simba-M`
|
| 126 |
)
|
| 127 |
|
|
|
|
| 128 |
asr_pipeline.model.load_adapter("multilingual_african") # Only for `UBC-NLP/Simba-M`
|
|
|
|
| 129 |
|
| 130 |
# Transcribe audio from file
|
| 131 |
result = asr_pipeline("https://africa.dlnlp.ai/simba/audio/afr_Lwazi_afr_test_idx3889.wav")
|
|
@@ -140,122 +145,36 @@ result = asr_pipeline({
|
|
| 140 |
print(result["text"])
|
| 141 |
|
| 142 |
```
|
| 143 |
-
Get started with Simba models in minutes using our interactive Colab notebook: [](https://github.com/UBC-NLP/simba/edit/main/simba_models.ipynb)
|
| 144 |
-
|
| 145 |
-
|
| 146 |
-
## Citation
|
| 147 |
-
|
| 148 |
-
If you use the Simba models or SimbaBench benchmark for your scientific publication, or if you find the resources in this website useful, please cite our paper.
|
| 149 |
|
| 150 |
-
|
| 151 |
|
| 152 |
-
|
| 153 |
-
title = "Voice of a Continent: Mapping {A}frica{'}s Speech Technology Frontier",
|
| 154 |
-
author = "Elmadany, AbdelRahim A. and
|
| 155 |
-
Kwon, Sang Yun and
|
| 156 |
-
Toyin, Hawau Olamide and
|
| 157 |
-
Alcoba Inciarte, Alcides and
|
| 158 |
-
Aldarmaki, Hanan and
|
| 159 |
-
Abdul-Mageed, Muhammad",
|
| 160 |
-
editor = "Christodoulopoulos, Christos and
|
| 161 |
-
Chakraborty, Tanmoy and
|
| 162 |
-
Rose, Carolyn and
|
| 163 |
-
Peng, Violet",
|
| 164 |
-
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
|
| 165 |
-
month = nov,
|
| 166 |
-
year = "2025",
|
| 167 |
-
address = "Suzhou, China",
|
| 168 |
-
publisher = "Association for Computational Linguistics",
|
| 169 |
-
url = "https://aclanthology.org/2025.emnlp-main.559/",
|
| 170 |
-
doi = "10.18653/v1/2025.emnlp-main.559",
|
| 171 |
-
pages = "11039--11061",
|
| 172 |
-
ISBN = "979-8-89176-332-6",
|
| 173 |
-
}
|
| 174 |
|
|
|
|
|
|
|
|
|
|
| 175 |
```
|
| 176 |
|
| 177 |
-
<div align="center">
|
| 178 |
-
|
| 179 |
-
<img src="https://africa.dlnlp.ai/simba/images/VoC_simba" alt="VoC Simba Models Logo">
|
| 180 |
-
|
| 181 |
-
|
| 182 |
-
[](https://aclanthology.org/2025.emnlp-main.559/)
|
| 183 |
-
[](https://africa.dlnlp.ai/simba/)
|
| 184 |
-
[](#simbabench)
|
| 185 |
-
[](https://huggingface.co/collections/UBC-NLP/simba-speech-series)
|
| 186 |
-
[](#demo)
|
| 187 |
-
|
| 188 |
-
</div>
|
| 189 |
-
|
| 190 |
-
## *Bridging the Digital Divide for African AI*
|
| 191 |
-
|
| 192 |
-
**Voice of a Continent** is a comprehensive open-source ecosystem designed to bring African languages to the forefront of artificial intelligence. By providing a unified suite of benchmarking tools and state-of-the-art models, we ensure that the future of speech technology is inclusive, representative, and accessible to over a billion people.
|
| 193 |
-
|
| 194 |
-
## Best-in-Class Multilingual Models
|
| 195 |
-
|
| 196 |
-
Introduced in our EMNLP 2025 paper *[Voice of a Continent](https://aclanthology.org/2025.emnlp-main.559/)*, the **Simba Series** represents the current state-of-the-art for African speech AI.
|
| 197 |
-
|
| 198 |
-
- **Unified Suite:** Models optimized for African languages.
|
| 199 |
-
- **Superior Accuracy:** Outperforms generic multilingual models by leveraging SimbaBench's high-quality, domain-diverse datasets.
|
| 200 |
-
- **Multitask Capability:** Designed for high performance in ASR (Automatic Speech Recognition) and TTS (Text-to-Speech).
|
| 201 |
-
- **Inclusion-First:** Specifically built to mitigate the "digital divide" by empowering speakers of underrepresented languages.
|
| 202 |
-
|
| 203 |
-
The **Simba** family consists of state-of-the-art models fine-tuned using SimbaBench. These models achieve superior performance by leveraging dataset quality, domain diversity, and language family relationships.
|
| 204 |
-
|
| 205 |
-
### π£οΈβοΈ Simba-ASR
|
| 206 |
-
> **The New Standard for African Speech-to-Text**
|
| 207 |
-
|
| 208 |
-
**π― Task** `Automatic Speech Recognition` β Powering high-accuracy transcription across the continent.
|
| 209 |
-
|
| 210 |
-
**π Language Coverage (43 African languages)**
|
| 211 |
-
> **Amharic** (`amh`), **Arabic** (`ara`), **Asante Twi** (`asanti`), **Bambara** (`bam`), **BaoulΓ©** (`bau`), **Bemba** (`bem`), **Ewe** (`ewe`), **Fanti** (`fat`), **Fon** (`fon`), **French** (`fra`), **Ganda** (`lug`), **Hausa** (`hau`), **Igbo** (`ibo`), **Kabiye** (`kab`), **Kinyarwanda** (`kin`), **Kongo** (`kon`), **Lingala** (`lin`), **Luba-Katanga** (`lub`), **Luo** (`luo`), **Malagasy** (`mlg`), **Mossi** (`mos`), **Northern Sotho** (`nso`), **Nyanja** (`nya`), **Oromo** (`orm`), **Portuguese** (`por`), **Shona** (`sna`), **Somali** (`som`), **Southern Sotho** (`sot`), **Swahili** (`swa`), **Swati** (`ssw`), **Tigrinya** (`tir`), **Tsonga** (`tso`), **Tswana** (`tsn`), **Twi** (`twi`), **Umbundu** (`umb`), **Venda** (`ven`), **Wolof** (`wol`), **Xhosa** (`xho`), **Yoruba** (`yor`), **Zulu** (`zul`), **Tamazight** (`tzm`), **Sango** (`sag`), **Dinka** (`din`).
|
| 212 |
-
|
| 213 |
-
**ποΈ Base Architectures**
|
| 214 |
-
|
| 215 |
-
- **Simba-S** (SeamlessM4T-v2-MT) β *Top Performer*
|
| 216 |
-
- **Simba-W** (Whisper-v3-large)
|
| 217 |
-
- **Simba-X** (Wav2Vec2-XLS-R-2b)
|
| 218 |
-
- **Simba-M** (MMS-1b-all)
|
| 219 |
-
- **Simba-H** (AfriHuBERT)
|
| 220 |
-
|
| 221 |
-
| **ASR Models** | **Architecture** | **π€ Hugging Face Model Card** | **Status** |
|
| 222 |
-
|---------|:------------------:| :------------------:| :------------------:|
|
| 223 |
-
| π₯**Simba-S**π₯| SeamlessM4T-v2 | π€ [https://huggingface.co/UBC-NLP/Simba-S](https://huggingface.co/UBC-NLP/Simba-S) | β
Released |
|
| 224 |
-
| π₯**Simba-W**π₯| Whisper | π€ [https://huggingface.co/UBC-NLP/Simba-W](https://huggingface.co/UBC-NLP/Simba-W) | β
Released |
|
| 225 |
-
| π₯**Simba-X**π₯| Wav2Vec2 | π€ [https://huggingface.co/UBC-NLP/Simba-X](https://huggingface.co/UBC-NLP/Simba-X) | β
Released |
|
| 226 |
-
| π₯**Simba-M**π₯| MMS | π€ [https://huggingface.co/UBC-NLP/Simba-M](https://huggingface.co/UBC-NLP/Simba-M) | β
Released |
|
| 227 |
-
| π₯**Simba-H**π₯| HuBERT | π€ [https://huggingface.co/UBC-NLP/Simba-H](https://huggingface.co/UBC-NLP/Simba-H) | β
Released |
|
| 228 |
-
|
| 229 |
-
* **Simba-S** (based on SeamlessM4T-v2-MT) emerged as the best-performing ASR model overall.
|
| 230 |
-
|
| 231 |
-
**π§© Usage Example**
|
| 232 |
-
|
| 233 |
-
You can easily run inference using the Hugging Face `transformers` library.
|
| 234 |
-
|
| 235 |
```python
|
| 236 |
-
|
| 237 |
-
|
| 238 |
-
|
| 239 |
-
asr_pipeline = pipeline(
|
| 240 |
-
"automatic-speech-recognition",
|
| 241 |
-
model="UBC-NLP/Simba-S" #Simba mdoels `UBC-NLP/Simba-S`, `UBC-NLP/Simba-W`, `UBC-NLP/Simba-X`, `UBC-NLP/Simba-H`, `UBC-NLP/Simba-M`
|
| 242 |
-
)
|
| 243 |
-
|
| 244 |
-
asr_pipeline.model.load_adapter("multilingual_african") # Only for `UBC-NLP/Simba-M`
|
| 245 |
-
|
| 246 |
-
# Transcribe audio from file
|
| 247 |
-
result = asr_pipeline("https://africa.dlnlp.ai/simba/audio/afr_Lwazi_afr_test_idx3889.wav")
|
| 248 |
-
print(result["text"])
|
| 249 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 250 |
|
| 251 |
-
|
| 252 |
-
|
| 253 |
-
|
| 254 |
-
|
| 255 |
-
})
|
| 256 |
-
print(result["text"])
|
| 257 |
|
|
|
|
|
|
|
|
|
|
| 258 |
```
|
|
|
|
| 259 |
Get started with Simba models in minutes using our interactive Colab notebook: [](https://github.com/UBC-NLP/simba/edit/main/simba_models.ipynb)
|
| 260 |
|
| 261 |
|
|
|
|
| 102 |
- **Simba-M** (MMS-1b-all)
|
| 103 |
- **Simba-H** (AfriHuBERT)
|
| 104 |
|
| 105 |
+
π Explore the Frontier
|
| 106 |
+
|
| 107 |
+
| **ASR Models** | **Architecture** | **#Parameters** | **π€ Hugging Face Model Card** | **Status** |
|
| 108 |
+
|---------|:------------------:| :------------------:| :------------------:|:------------------:|
|
| 109 |
+
| π₯**Simba-S**π₯| SeamlessM4T-v2 | 2.3B | π€ [https://huggingface.co/UBC-NLP/Simba-S](https://huggingface.co/UBC-NLP/Simba-S) | β
Released |
|
| 110 |
+
| π₯**Simba-W**π₯| Whisper | 1.5B | π€ [https://huggingface.co/UBC-NLP/Simba-W](https://huggingface.co/UBC-NLP/Simba-W) | β
Released |
|
| 111 |
+
| π₯**Simba-X**π₯| Wav2Vec2 | 1B | π€ [https://huggingface.co/UBC-NLP/Simba-X](https://huggingface.co/UBC-NLP/Simba-X) | β
Released |
|
| 112 |
+
| π₯**Simba-M**π₯| MMS | 1B | π€ [https://huggingface.co/UBC-NLP/Simba-M](https://huggingface.co/UBC-NLP/Simba-M) | β
Released |
|
| 113 |
+
| π₯**Simba-H**π₯| HuBERT | 94M | π€ [https://huggingface.co/UBC-NLP/Simba-H](https://huggingface.co/UBC-NLP/Simba-H) | β
Released |
|
| 114 |
+
|
| 115 |
+
* **Simba-S** emerged as the best-performing ASR model overall.
|
| 116 |
|
|
|
|
| 117 |
|
| 118 |
**π§© Usage Example**
|
| 119 |
|
|
|
|
| 128 |
model="UBC-NLP/Simba-S" #Simba mdoels `UBC-NLP/Simba-S`, `UBC-NLP/Simba-W`, `UBC-NLP/Simba-X`, `UBC-NLP/Simba-H`, `UBC-NLP/Simba-M`
|
| 129 |
)
|
| 130 |
|
| 131 |
+
##### Load the multilingual African adapter (Only for `UBC-NLP/Simba-M`)
|
| 132 |
asr_pipeline.model.load_adapter("multilingual_african") # Only for `UBC-NLP/Simba-M`
|
| 133 |
+
###########################
|
| 134 |
|
| 135 |
# Transcribe audio from file
|
| 136 |
result = asr_pipeline("https://africa.dlnlp.ai/simba/audio/afr_Lwazi_afr_test_idx3889.wav")
|
|
|
|
| 145 |
print(result["text"])
|
| 146 |
|
| 147 |
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 148 |
|
| 149 |
+
#### Example Outputs
|
| 150 |
|
| 151 |
+
Using the same audio file with different Simba models:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 152 |
|
| 153 |
+
```python
|
| 154 |
+
# Simba-S
|
| 155 |
+
{'text': 'watter verontwaardiging sou daar, in ons binneste gewees het.'}
|
| 156 |
```
|
| 157 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 158 |
```python
|
| 159 |
+
# Simba-W
|
| 160 |
+
{'text': 'watter veronwaardigingsel daar, in ons binneste gewees het.'}
|
| 161 |
+
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 162 |
|
| 163 |
+
```python
|
| 164 |
+
# Simba-X
|
| 165 |
+
{'text': 'fator fr on ar taamsodr is'}
|
| 166 |
+
```
|
| 167 |
|
| 168 |
+
```python
|
| 169 |
+
# Simba-M
|
| 170 |
+
{'text': 'watter veronwaardiging sodaar in ons binniste gewees het'}
|
| 171 |
+
```
|
|
|
|
|
|
|
| 172 |
|
| 173 |
+
```python
|
| 174 |
+
# Simba-H
|
| 175 |
+
{'text': 'watter vironwaardiging so daar in ons binneste geweeshet'}
|
| 176 |
```
|
| 177 |
+
|
| 178 |
Get started with Simba models in minutes using our interactive Colab notebook: [](https://github.com/UBC-NLP/simba/edit/main/simba_models.ipynb)
|
| 179 |
|
| 180 |
|