--- license: - cc-by-sa-4.0 - cc-by-4.0 annotation_creators: - human-annotated - crowdsourced language_creators: - creator_1 tags: - audio - automatic-speech-recognition - text-to-speech language: - ach - aka - dag - dga - ewe - fat - ful - hau - ibo - kpo - lin - lug - mas - mlg - nyn - sna - sog - swa - twi - yor multilinguality: - multilingual pretty_name: Waxal NLP Datasets task_categories: - automatic-speech-recognition - text-to-speech source_datasets: - UGSpeechData - DigitalUmuganda/AfriVoice - original configs: - config_name: ach_asr data_files: - split: train path: "data/ASR/ach/ach-train-*" - split: validation path: "data/ASR/ach/ach-validation-*" - split: test path: "data/ASR/ach/ach-test-*" - split: unlabeled path: "data/ASR/ach/ach-unlabeled-*" - config_name: ach_tts data_files: - split: train path: "data/TTS/ach/ach-train-*" - split: validation path: "data/TTS/ach/ach-validation-*" - split: test path: "data/TTS/ach/ach-test-*" - config_name: aka_asr data_files: - split: train path: "data/ASR/aka/aka-train-*" - split: validation path: "data/ASR/aka/aka-validation-*" - split: test path: "data/ASR/aka/aka-test-*" - split: unlabeled path: "data/ASR/aka/aka-unlabeled-*" - config_name: dag_asr data_files: - split: train path: "data/ASR/dag/dag-train-*" - split: validation path: "data/ASR/dag/dag-validation-*" - split: test path: "data/ASR/dag/dag-test-*" - split: unlabeled path: "data/ASR/dag/dag-unlabeled-*" - config_name: dga_asr data_files: - split: train path: "data/ASR/dga/dga-train-*" - split: validation path: "data/ASR/dga/dga-validation-*" - split: test path: "data/ASR/dga/dga-test-*" - split: unlabeled path: "data/ASR/dga/dga-unlabeled-*" - config_name: ewe_asr data_files: - split: train path: "data/ASR/ewe/ewe-train-*" - split: validation path: "data/ASR/ewe/ewe-validation-*" - split: test path: "data/ASR/ewe/ewe-test-*" - split: unlabeled path: "data/ASR/ewe/ewe-unlabeled-*" - config_name: fat_tts data_files: - split: train path: "data/TTS/fat/fat-train-*" - split: validation path: "data/TTS/fat/fat-validation-*" - split: test path: "data/TTS/fat/fat-test-*" - config_name: ful_asr data_files: - split: train path: "data/ASR/ful/ful-train-*" - split: validation path: "data/ASR/ful/ful-validation-*" - split: test path: "data/ASR/ful/ful-test-*" - split: unlabeled path: "data/ASR/ful/ful-unlabeled-*" - config_name: ful_tts data_files: - split: train path: "data/TTS/ful/ful-train-*" - split: validation path: "data/TTS/ful/ful-validation-*" - split: test path: "data/TTS/ful/ful-test-*" - config_name: hau_tts data_files: - split: train path: "data/TTS/hau/hau-train-*" - split: validation path: "data/TTS/hau/hau-validation-*" - split: test path: "data/TTS/hau/hau-test-*" - config_name: ibo_tts data_files: - split: train path: "data/TTS/ibo/ibo-train-*" - split: validation path: "data/TTS/ibo/ibo-validation-*" - split: test path: "data/TTS/ibo/ibo-test-*" - config_name: kpo_asr data_files: - split: train path: "data/ASR/kpo/kpo-train-*" - split: validation path: "data/ASR/kpo/kpo-validation-*" - split: test path: "data/ASR/kpo/kpo-test-*" - split: unlabeled path: "data/ASR/kpo/kpo-unlabeled-*" - config_name: lin_asr data_files: - split: train path: "data/ASR/lin/lin-train-*" - split: validation path: "data/ASR/lin/lin-validation-*" - split: test path: "data/ASR/lin/lin-test-*" - split: unlabeled path: "data/ASR/lin/lin-unlabeled-*" - config_name: lug_asr data_files: - split: train path: "data/ASR/lug/lug-train-*" - split: validation path: "data/ASR/lug/lug-validation-*" - split: test path: "data/ASR/lug/lug-test-*" - split: unlabeled path: "data/ASR/lug/lug-unlabeled-*" - config_name: lug_tts data_files: - split: train path: "data/TTS/lug/lug-train-*" - split: validation path: "data/TTS/lug/lug-validation-*" - split: test path: "data/TTS/lug/lug-test-*" - config_name: mas_asr data_files: - split: train path: "data/ASR/mas/mas-train-*" - split: validation path: "data/ASR/mas/mas-validation-*" - split: test path: "data/ASR/mas/mas-test-*" - split: unlabeled path: "data/ASR/mas/mas-unlabeled-*" - config_name: mlg_asr data_files: - split: train path: "data/ASR/mlg/mlg-train-*" - split: validation path: "data/ASR/mlg/mlg-validation-*" - split: test path: "data/ASR/mlg/mlg-test-*" - split: unlabeled path: "data/ASR/mlg/mlg-unlabeled-*" - config_name: nyn_asr data_files: - split: train path: "data/ASR/nyn/nyn-train-*" - split: validation path: "data/ASR/nyn/nyn-validation-*" - split: test path: "data/ASR/nyn/nyn-test-*" - split: unlabeled path: "data/ASR/nyn/nyn-unlabeled-*" - config_name: nyn_tts data_files: - split: train path: "data/TTS/nyn/nyn-train-*" - split: validation path: "data/TTS/nyn/nyn-validation-*" - split: test path: "data/TTS/nyn/nyn-test-*" - config_name: sna_asr data_files: - split: train path: "data/ASR/sna/sna-train-*" - split: validation path: "data/ASR/sna/sna-validation-*" - split: test path: "data/ASR/sna/sna-test-*" - split: unlabeled path: "data/ASR/sna/sna-unlabeled-*" - config_name: sog_asr data_files: - split: train path: "data/ASR/sog/sog-train-*" - split: validation path: "data/ASR/sog/sog-validation-*" - split: test path: "data/ASR/sog/sog-test-*" - split: unlabeled path: "data/ASR/sog/sog-unlabeled-*" - config_name: swa_tts data_files: - split: train path: "data/TTS/swa/swa-train-*" - split: validation path: "data/TTS/swa/swa-validation-*" - split: test path: "data/TTS/swa/swa-test-*" - config_name: twi_tts data_files: - split: train path: "data/TTS/twi/twi-train-*" - split: validation path: "data/TTS/twi/twi-validation-*" - split: test path: "data/TTS/twi/twi-test-*" - config_name: yor_tts data_files: - split: train path: "data/TTS/yor/yor-train-*" - split: validation path: "data/TTS/yor/yor-validation-*" - split: test path: "data/TTS/yor/yor-test-*" dataset_info: - config_name: ach_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: ach_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: aka_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: dag_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: dga_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: ewe_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: fat_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: ful_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: ful_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: hau_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: ibo_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: kpo_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: lin_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: lug_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: lug_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: mas_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: mlg_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: nyn_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: nyn_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: sna_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: sog_asr features: - name: id dtype: string - name: speaker_id dtype: string - name: transcription dtype: string - name: language dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: swa_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: twi_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio - config_name: yor_tts features: - name: id dtype: string - name: speaker_id dtype: string - name: text dtype: string - name: locale dtype: string - name: gender dtype: string - name: audio dtype: audio --- # Waxal Datasets ## Table of Contents - [Dataset Description](#dataset-description) - [ASR Dataset](#asr-dataset) - [TTS Dataset](#tts-dataset) - [How to Use](#how-to-use) - [Dataset Structure](#dataset-structure) - [ASR Data Fields](#asr-data-fields) - [TTS Data Fields](#tts-data-fields) - [Data Splits](#data-splits) - [Dataset Curation](#dataset-curation) - [Considerations for Using the Data](#considerations-for-using-the-data) - [Additional Information](#additional-information) ## Dataset Description The Waxal project provides datasets for both Automated Speech Recognition (ASR) and Text-to-Speech (TTS) for African languages. The goal of this dataset's creation and release is to facilitate research that improves the accuracy and fluency of speech and language technology for these underserved languages, and to serve as a repository for digital preservation. The Waxal datasets are collections acquired through partnerships with Makerere University, The University of Ghana, Digital Umuganda, and Media Trust. Acquisition was funded by Google and the Gates Foundation under an agreement to make the dataset openly accessible. ### ASR Dataset The Waxal ASR dataset is a collection of data in 14 African languages. It consists of approximately 1,250 hours of transcribed natural speech from a wide variety of voices. The 14 languages in this dataset represent over 100 million speakers across 40 Sub-Saharan African countries. Provider | Languages | License :------------------ | :--------------------------------------- | :------------: Makerere University | Acholi, Luganda, Masaaba, Nyankole, Soga | `CC-BY-4.0` University of Ghana | Akan, Ewe, Dagbani, Dagaare, Ikposo | `CC-BY-4.0` Digital Umuganda | Fula, Lingala, Shona, Malagasy | `CC-BY-4.0` ### TTS Dataset The Waxal TTS dataset is a collection of text-to-speech data in 10 African languages. It consists of approximately 240 hours of scripted natural speech from a wide variety of voices. Provider | Languages | License :------------------ | :----------------------------------- | :------------: Makerere University | Acholi, Luganda, Kiswahili, Nyankole | `CC-BY-4.0` University of Ghana | Akan (Fante, Twi) | `CC-BY-4.0` Media Trust | Fula, Igbo, Hausa, Yoruba | `CC-BY-4.0` ### How to Use The `datasets` library allows you to load and pre-process your dataset in pure Python, at scale. First, ensure you have the necessary dependencies installed to handle audio data. You will need `ffmpeg` installed on your system. **Google Colab / Ubuntu** ```bash sudo apt-get install ffmpeg pip install datasets[audio] ``` **macOS** ```bash brew install ffmpeg pip install datasets[audio] ``` **Windows** Download and install from [ffmpeg.org](https://ffmpeg.org/download.html) and ensure it's in your PATH. ```bash pip install datasets[audio] ``` If you encounter `RuntimeError: Could not load libtorchcodec`, please ensure `ffmpeg` is correctly installed or check for compatibility between your `torch`, `torchaudio`, and `torchcodec` versions. **Loading ASR Data** To load ASR data for a specific language, specify the configuration name, e.g. `sna_asr` for Shona ASR data. ```python from datasets import load_dataset, Audio # Load Shona (sna) ASR dataset asr_data = load_dataset("google/WaxalNLP", "sna_asr") # Access splits train = asr_data['train'] val = asr_data['validation'] test = asr_data['test'] # Example: Accessing audio bytes and other fields example = train[0] print(f"Transcription: {example['transcription']}") print(f"Sampling Rate: {example['audio']['sampling_rate']}") # 'array' contains the decoded audio bytes as a numpy array print(f"Audio Array Shape: {example['audio']['array'].shape}") ``` **Loading TTS Data** To load TTS data for a specific language, specify the configuration name, e.g. `swa_tts` for Swahili TTS data. ```python from datasets import load_dataset # Load Swahili (swa) TTS dataset tts_data = load_dataset("google/WaxalNLP", "swa_tts") # Access splits train = tts_data['train'] ``` ## Dataset Structure ### ASR Data Fields ```python { 'id': 'sna_0', 'speaker_id': '...', 'audio': { 'array': [...], 'sample_rate': 16_000 }, 'transcription': '...', 'language': 'sna', 'gender': 'Female', } ``` * **id**: Unique identifier. * **speaker_id**: Unique identifier for the speaker. * **audio**: Audio data. * **transcription**: Transcription of the audio. * **language**: ISO 639-2 language code. * **gender**: Speaker gender ('Male', 'Female', or empty). ### TTS Data Fields ```python { 'id': 'swa_0', 'speaker_id': '...', 'audio': { 'array': [...], 'sample_rate': 16_000 }, 'text': '...', 'locale': 'swa', 'gender': 'Female', } ``` * **id**: Unique identifier. * **speaker_id**: Unique identifier for the speaker. * **audio**: Audio data. * **text**: Text script. * **locale**: ISO 639-2 language code. * **gender**: Speaker gender. ### Data Splits For the **ASR Dataset**, the data with transcriptions is split as follows: * **train**: 80% of labeled data. * **validation**: 10% of labeled data. * **test**: 10% of labeled data. The **unlabeled** split contains all samples that do not have a corresponding transcription. The **TTS Dataset** follows a similar structure, with data split into `train`, `validation`, and `test` sets. ## Dataset Curation The data was gathered by multiple partners: Provider | Dataset | License :------------------ | :------------------------------------------------------- | :------ University of Ghana | [UGSpeechData](https://doi.org/10.57760/sciencedb.22298) | `CC BY 4.0` Digital Umuganda | [AfriVoice](DigitalUmuganda/AfriVoice) | `CC-BY 4.0` Makerere University | [Yogera Dataset](https://doi.org/10.7910/DVN/BEROE0) | `CC-BY 4.0` Media Trust | | `CC-BY 4.0` ## Considerations for Using the Data Please check the license for the specific languages you are using, as they may differ between providers. **Affiliation:** Google Research ## Version and Maintenance - **Current Version:** 1.0.0 - **Last Updated:** 01/2026