deepdml commited on
Commit
50f7501
·
verified ·
1 Parent(s): 5b856b2

End of training

Browse files
Files changed (1) hide show
  1. README.md +34 -32
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  language:
3
  - ar
4
  license: apache-2.0
@@ -6,11 +7,11 @@ base_model: openai/whisper-tiny
6
  tags:
7
  - generated_from_trainer
8
  datasets:
 
9
  - ymoslem/MediaSpeech
10
  - deepdml/Tunisian_MSA
11
  - UBC-NLP/Casablanca
12
  - fixie-ai/common_voice_17_0
13
- - google/fleurs
14
  metrics:
15
  - wer
16
  model-index:
@@ -21,12 +22,13 @@ model-index:
21
  type: automatic-speech-recognition
22
  dataset:
23
  name: Common Voice 17.0
24
- type: ymoslem/MediaSpeech
25
  metrics:
26
  - name: Wer
27
  type: wer
28
- value: 60.61585354657461
29
  ---
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
31
  should probably proofread and complete it, then remove this comment. -->
32
 
@@ -34,9 +36,9 @@ should probably proofread and complete it, then remove this comment. -->
34
 
35
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 17.0 dataset.
36
  It achieves the following results on the evaluation set:
37
- - Loss: 0.7162
38
- - Wer: 60.6159
39
- - Cer: 21.8903
40
 
41
  ## Model description
42
 
@@ -59,38 +61,38 @@ The following hyperparameters were used during training:
59
  - train_batch_size: 64
60
  - eval_batch_size: 64
61
  - seed: 42
62
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
63
  - lr_scheduler_type: linear
64
  - lr_scheduler_warmup_ratio: 0.04
65
- - training_steps: 5000
66
 
67
  ### Training results
68
 
69
- | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
70
- |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
71
- | 0.8928 | 0.2 | 1000 | 0.7965 | 66.8809 | 25.5809 |
72
- | 0.6731 | 0.4 | 2000 | 0.7496 | 63.1479 | 23.1687 |
73
- | 0.5235 | 0.6 | 3000 | 0.7214 | 61.6845 | 22.3557 |
74
- | 0.4641 | 0.8 | 4000 | 0.7161 | 60.8490 | 21.9854 |
75
- | 0.4296 | 1.0 | 5000 | 0.7162 | 60.6159 | 21.8903 |
 
 
 
 
 
 
 
 
 
 
 
 
 
76
 
77
 
78
  ### Framework versions
79
 
80
- - Transformers 4.42.0.dev0
81
- - Pytorch 2.3.0+cu121
82
- - Datasets 2.19.1
83
- - Tokenizers 0.19.1
84
-
85
- ## Citation
86
-
87
- Please cite the model using the following BibTeX entry:
88
-
89
- ```bibtex
90
- @misc{deepdml/whisper-tiny-ar-mix-norm,
91
- title={Fine-tuned Whisper tiny ASR model for speech recognition in Arabic},
92
- author={Jimenez, David},
93
- howpublished={\url{https://huggingface.co/deepdml/whisper-tiny-ar-mix-norm}},
94
- year={2026}
95
- }
96
- ```
 
1
  ---
2
+ library_name: transformers
3
  language:
4
  - ar
5
  license: apache-2.0
 
7
  tags:
8
  - generated_from_trainer
9
  datasets:
10
+ - google/fleurs
11
  - ymoslem/MediaSpeech
12
  - deepdml/Tunisian_MSA
13
  - UBC-NLP/Casablanca
14
  - fixie-ai/common_voice_17_0
 
15
  metrics:
16
  - wer
17
  model-index:
 
22
  type: automatic-speech-recognition
23
  dataset:
24
  name: Common Voice 17.0
25
+ type: google/fleurs
26
  metrics:
27
  - name: Wer
28
  type: wer
29
+ value: 52.17678705862912
30
  ---
31
+
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
  should probably proofread and complete it, then remove this comment. -->
34
 
 
36
 
37
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 17.0 dataset.
38
  It achieves the following results on the evaluation set:
39
+ - Loss: 0.6321
40
+ - Wer: 52.1768
41
+ - Cer: 18.3597
42
 
43
  ## Model description
44
 
 
61
  - train_batch_size: 64
62
  - eval_batch_size: 64
63
  - seed: 42
64
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
65
  - lr_scheduler_type: linear
66
  - lr_scheduler_warmup_ratio: 0.04
67
+ - training_steps: 18000
68
 
69
  ### Training results
70
 
71
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
72
+ |:-------------:|:------:|:-----:|:---------------:|:-------:|:-------:|
73
+ | 0.9731 | 0.0556 | 1000 | 0.8246 | 68.5849 | 26.4636 |
74
+ | 0.6833 | 0.1111 | 2000 | 0.7503 | 63.3554 | 23.5338 |
75
+ | 0.4756 | 0.1667 | 3000 | 0.7112 | 60.5773 | 21.8069 |
76
+ | 0.3473 | 0.2222 | 4000 | 0.7019 | 59.5509 | 21.6413 |
77
+ | 0.2547 | 0.2778 | 5000 | 0.6910 | 59.1212 | 21.5653 |
78
+ | 0.1777 | 0.3333 | 6000 | 0.6924 | 57.6816 | 20.6340 |
79
+ | 0.128 | 1.0197 | 7000 | 0.6828 | 57.0996 | 20.5314 |
80
+ | 0.11 | 1.0752 | 8000 | 0.6706 | 56.0768 | 20.2707 |
81
+ | 0.0869 | 1.1308 | 9000 | 0.6622 | 55.4654 | 20.0036 |
82
+ | 0.0714 | 1.1863 | 10000 | 0.6506 | 54.8448 | 19.6163 |
83
+ | 0.0594 | 1.2419 | 11000 | 0.6427 | 54.9714 | 19.4470 |
84
+ | 0.0541 | 1.2974 | 12000 | 0.6365 | 53.4089 | 19.0258 |
85
+ | 0.0484 | 1.353 | 13000 | 0.6371 | 53.7100 | 19.1604 |
86
+ | 0.0445 | 2.0393 | 14000 | 0.6359 | 53.8697 | 19.4603 |
87
+ | 0.042 | 2.0949 | 15000 | 0.6348 | 52.5403 | 18.6839 |
88
+ | 0.0346 | 2.1504 | 16000 | 0.6317 | 52.7809 | 18.6861 |
89
+ | 0.0339 | 2.206 | 17000 | 0.6436 | 52.5128 | 18.5581 |
90
+ | 0.0404 | 2.2616 | 18000 | 0.6321 | 52.1768 | 18.3597 |
91
 
92
 
93
  ### Framework versions
94
 
95
+ - Transformers 4.48.0.dev0
96
+ - Pytorch 2.5.1+cu121
97
+ - Datasets 3.6.0
98
+ - Tokenizers 0.21.0