Description

This model is a fine-tuned version of openai/whisper-small on Bangla speech data.

Environment:

  • Python version: 3.12.12
  • PyTorch version: 2.8.0+cu126
  • NumPy version: 1.26.4

Training Parameters:

  • Total epochs: 8
  • Learning Rate: 2e-5
  • Batch size per device: 4
  • Gradient accumulation: 4
  • Warmup Steps: 200
  • LR Scheduler: cosine_with_restarts

Validation Set Evaluation:

Epoch Training Loss Validation Loss WER (%) Normalized Levenshtein Similarity (%)
0 1.436900 1.500822 23.480663 84.353535
2 1.433300 1.479616 20.883978 85.525253
4 1.418600 1.467377 16.022099 88.111111
6 1.413600 1.469696 14.751381 89.313131
8 1.413300 1.470008 14.530387 89.444444
Downloads last month
1
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ishmamzarif/finetuned-modelv4

Finetuned
(3268)
this model
Finetunes
1 model