You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

whisper-small-lug-4src

This model is a fine-tuned version of sulaimank/whisper-small-CV-Fleurs-lg-300 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6772
  • Wer: 0.2145
  • Cer: 0.1136

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 1e-05
  • train_batch_size: 8
  • eval_batch_size: 4
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 16
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 500
  • num_epochs: 30

Training results

Training Loss Epoch Step Cer Validation Loss Wer
0.6768 1.1508 1000 0.2882 0.4067 0.3768
0.3794 2.3017 2000 0.2152 0.3882 0.3322
0.3018 3.4525 3000 0.4216 0.2716 0.1417
0.2216 4.6033 4000 0.4807 0.2945 0.1581
0.1689 5.7542 5000 0.5392 0.2358 0.1309
0.0748 6.9050 6000 0.5808 0.2331 0.1226
0.0453 8.0553 7000 0.6295 0.2382 0.1290
0.0276 9.2061 8000 0.6217 0.2413 0.1291
0.0166 10.3569 9000 0.6351 0.2300 0.1183
0.0115 11.5078 10000 0.6329 0.2366 0.1237
0.0096 12.6586 11000 0.6458 0.2428 0.1259
0.0059 13.8094 12000 0.6414 0.2323 0.1182
0.0045 14.9603 13000 0.6418 0.2638 0.1463
0.0043 16.1105 14000 0.6474 0.2362 0.1269
0.0026 17.2614 15000 0.6542 0.2284 0.1259
0.0026 18.4122 16000 0.6498 0.2253 0.1164
0.0013 19.5630 17000 0.6504 0.2758 0.1572
0.0015 20.7139 18000 0.6608 0.2187 0.1161
0.0015 21.8647 19000 0.6602 0.2234 0.1173
0.0015 23.0150 20000 0.6661 0.2455 0.1296
0.0008 24.1658 21000 0.6625 0.2265 0.1214
0.0003 25.3166 22000 0.6698 0.2498 0.1384
0.0003 26.4675 23000 0.6725 0.2347 0.1234
0.0001 27.6183 24000 0.6738 0.2187 0.1194
0.0001 28.7691 25000 0.6766 0.2133 0.1144
0.0001 29.9200 26000 0.6772 0.2145 0.1136

Framework versions

  • Transformers 5.3.0
  • Pytorch 2.10.0+cu130
  • Datasets 4.0.0
  • Tokenizers 0.22.2
Downloads last month
240
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for sulaimank/whisper-small-lug-4src

Finetuned
(2)
this model
Finetunes
1 model