deepdml commited on
Commit
7983137
·
1 Parent(s): 5592952

End of training

Browse files
Files changed (1) hide show
  1. README.md +30 -29
README.md CHANGED
@@ -8,8 +8,6 @@ tags:
8
  datasets:
9
  - voice-biomarkers/openslr-32-hq-SA-languages-Afrikaans
10
  - google/fleurs
11
- - dsfsi-anv/multilingual-nchlt-dataset
12
- - andreoosthuizen/afrikaans-30s
13
  metrics:
14
  - wer
15
  model-index:
@@ -27,8 +25,9 @@ model-index:
27
  metrics:
28
  - name: Wer
29
  type: wer
30
- value: 49.05627705627706
31
  ---
 
32
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
33
  should probably proofread and complete it, then remove this comment. -->
34
 
@@ -36,9 +35,9 @@ should probably proofread and complete it, then remove this comment. -->
36
 
37
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 17.0 dataset.
38
  It achieves the following results on the evaluation set:
39
- - Loss: 1.3265
40
- - Wer: 49.0563
41
- - Cer: 20.0715
42
 
43
  ## Model description
44
 
@@ -58,23 +57,38 @@ More information needed
58
 
59
  The following hyperparameters were used during training:
60
  - learning_rate: 1e-05
61
- - train_batch_size: 128
62
- - eval_batch_size: 128
63
  - seed: 42
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
  - lr_scheduler_warmup_ratio: 0.04
67
- - training_steps: 5000
68
 
69
  ### Training results
70
 
71
- | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
72
- |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|
73
- | 0.0873 | 0.2 | 1000 | 1.1737 | 43.8442 | 18.5673 |
74
- | 0.0323 | 1.014 | 2000 | 1.2562 | 46.8571 | 19.0335 |
75
- | 0.0184 | 1.214 | 3000 | 1.2972 | 42.9957 | 17.3357 |
76
- | 0.0164 | 2.028 | 4000 | 1.3082 | 45.8528 | 18.4676 |
77
- | 0.0154 | 2.228 | 5000 | 1.3265 | 49.0563 | 20.0715 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
 
79
 
80
  ### Framework versions
@@ -83,16 +97,3 @@ The following hyperparameters were used during training:
83
  - Pytorch 2.3.0+cu121
84
  - Datasets 2.19.1
85
  - Tokenizers 0.19.1
86
-
87
- ## Citation
88
-
89
- Please cite the model using the following BibTeX entry:
90
-
91
- ```bibtex
92
- @misc{deepdml/whisper-tiny-af-mix-norm,
93
- title={Fine-tuned Whisper tiny ASR model for speech recognition in Afrikaans},
94
- author={Jimenez, David},
95
- howpublished={\url{https://huggingface.co/deepdml/whisper-tiny-af-mix-norm}},
96
- year={2026}
97
- }
98
- ```
 
8
  datasets:
9
  - voice-biomarkers/openslr-32-hq-SA-languages-Afrikaans
10
  - google/fleurs
 
 
11
  metrics:
12
  - wer
13
  model-index:
 
25
  metrics:
26
  - name: Wer
27
  type: wer
28
+ value: 52.17316017316017
29
  ---
30
+
31
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
32
  should probably proofread and complete it, then remove this comment. -->
33
 
 
35
 
36
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the Common Voice 17.0 dataset.
37
  It achieves the following results on the evaluation set:
38
+ - Loss: 1.3668
39
+ - Wer: 52.1732
40
+ - Cer: 20.9395
41
 
42
  ## Model description
43
 
 
57
 
58
  The following hyperparameters were used during training:
59
  - learning_rate: 1e-05
60
+ - train_batch_size: 64
61
+ - eval_batch_size: 64
62
  - seed: 42
63
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
64
  - lr_scheduler_type: linear
65
  - lr_scheduler_warmup_ratio: 0.04
66
+ - training_steps: 2000
67
 
68
  ### Training results
69
 
70
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
71
+ |:-------------:|:-------:|:----:|:---------------:|:-------:|:-------:|
72
+ | 1.3199 | 0.05 | 100 | 1.4770 | 59.2208 | 24.5022 |
73
+ | 0.6393 | 1.0315 | 200 | 1.2510 | 51.2554 | 20.9454 |
74
+ | 0.3811 | 2.013 | 300 | 1.2197 | 49.4545 | 20.1155 |
75
+ | 0.261 | 2.063 | 400 | 1.2089 | 48.3290 | 19.2036 |
76
+ | 0.186 | 3.0445 | 500 | 1.2141 | 48.0693 | 19.8575 |
77
+ | 0.1459 | 4.026 | 600 | 1.2341 | 49.8701 | 20.2621 |
78
+ | 0.0963 | 5.0075 | 700 | 1.2517 | 48.4675 | 19.5437 |
79
+ | 0.0809 | 5.0575 | 800 | 1.2674 | 51.0823 | 21.0715 |
80
+ | 0.0536 | 6.039 | 900 | 1.2812 | 48.2597 | 19.5408 |
81
+ | 0.0432 | 7.0205 | 1000 | 1.3003 | 48.5022 | 19.4910 |
82
+ | 0.0379 | 8.002 | 1100 | 1.3117 | 51.6190 | 21.2298 |
83
+ | 0.0333 | 8.052 | 1200 | 1.3314 | 52.3463 | 21.7078 |
84
+ | 0.0247 | 9.0335 | 1300 | 1.3389 | 52.0 | 21.4644 |
85
+ | 0.0201 | 10.015 | 1400 | 1.3484 | 51.4113 | 22.1769 |
86
+ | 0.0194 | 10.065 | 1500 | 1.3469 | 51.8442 | 21.0685 |
87
+ | 0.0191 | 11.0465 | 1600 | 1.3536 | 52.4502 | 21.3471 |
88
+ | 0.0179 | 12.028 | 1700 | 1.3611 | 51.7229 | 21.1272 |
89
+ | 0.0155 | 13.0095 | 1800 | 1.3637 | 52.4329 | 20.9512 |
90
+ | 0.0159 | 13.0595 | 1900 | 1.3651 | 52.0346 | 20.9483 |
91
+ | 0.0152 | 14.041 | 2000 | 1.3668 | 52.1732 | 20.9395 |
92
 
93
 
94
  ### Framework versions
 
97
  - Pytorch 2.3.0+cu121
98
  - Datasets 2.19.1
99
  - Tokenizers 0.19.1