Automatic Speech Recognition
Transformers
PyTorch
Igbo
wav2vec2
How to use from the
Use from the
Transformers library
# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("automatic-speech-recognition", model="AstralZander/igbo_ASR")
# Load model directly
from transformers import AutoProcessor, AutoModelForCTC

processor = AutoProcessor.from_pretrained("AstralZander/igbo_ASR")
model = AutoModelForCTC.from_pretrained("AstralZander/igbo_ASR")
Quick Links

facebook/wav2vec2-xls-r-300m fine-tuned on google/fleurs and mozilla-foundation/common_voice_13_0 for Igbo language.

WER: 0.51

Code for running:

from huggingsound import SpeechRecognitionModel

model = SpeechRecognitionModel("AstralZander/igbo_ASR")
audio_paths = [audio_path] # List with paths to audio
transcriptions = model.transcribe(audio_paths)

transcriptions # List of transcriptions, timestamps and probabilities
transcriptions[ind_audio]['transcription'] # Transcription of audio with the ind_audio index from the audio_paths list
Downloads last month
137
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Datasets used to train AstralZander/igbo_ASR

Space using AstralZander/igbo_ASR 1