google/fleurs
Viewer • Updated • 768k • 57.7k • 402
How to use steja/whisper-small-telugu-large-data with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("automatic-speech-recognition", model="steja/whisper-small-telugu-large-data") # Load model directly
from transformers import AutoProcessor, AutoModelForSpeechSeq2Seq
processor = AutoProcessor.from_pretrained("steja/whisper-small-telugu-large-data")
model = AutoModelForSpeechSeq2Seq.from_pretrained("steja/whisper-small-telugu-large-data")This model is a fine-tuned version of openai/whisper-small on the google/fleurs and openslr dataset in telugu. It achieves the following results on the evaluation set (google/fleurs, test set):
openai/whisper-small has the following zero shot performance on google/fleurs test set:
More information needed
More information needed
More information needed
The following hyperparameters were used during training:
| Training Loss | Epoch | Step | Validation Loss | Wer |
|---|---|---|---|---|
| 0.128 | 2.27 | 500 | 0.2015 | 45.1692 |
| 0.0462 | 4.55 | 1000 | 0.1877 | 41.1050 |
| 0.0184 | 6.82 | 1500 | 0.2241 | 40.5153 |
| 0.0045 | 9.09 | 2000 | 0.2590 | 39.7260 |
| 0.0019 | 11.36 | 2500 | 0.2824 | 39.0819 |
| 0.0006 | 13.64 | 3000 | 0.3002 | 38.9096 |
| 0.0002 | 15.91 | 3500 | 0.3141 | 38.5920 |
| 0.0001 | 18.18 | 4000 | 0.3232 | 38.7463 |
| 0.0001 | 20.45 | 4500 | 0.3289 | 38.8370 |
| 0.0001 | 22.73 | 5000 | 0.3310 | 38.8460 |