ales commited on
Commit
c54ad8c
·
1 Parent(s): a7afa35

update model card README.md

Browse files
Files changed (2) hide show
  1. README.md +18 -13
  2. train.log +2 -0
README.md CHANGED
@@ -1,41 +1,38 @@
1
  ---
2
- language:
3
- - be
4
  license: apache-2.0
5
  tags:
6
- - whisper-event
7
  - generated_from_trainer
8
  datasets:
9
- - mozilla-foundation/common_voice_11_0
10
  metrics:
11
  - wer
12
  model-index:
13
- - name: Whisper Small Belarusian
14
  results:
15
  - task:
16
  name: Automatic Speech Recognition
17
  type: automatic-speech-recognition
18
  dataset:
19
- name: mozilla-foundation/common_voice_11_0 be
20
- type: mozilla-foundation/common_voice_11_0
21
  config: be
22
  split: validation
23
  args: be
24
  metrics:
25
  - name: Wer
26
  type: wer
27
- value: 75.27472527472527
28
  ---
29
 
30
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
31
  should probably proofread and complete it, then remove this comment. -->
32
 
33
- # Whisper Small Belarusian
34
 
35
- This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the mozilla-foundation/common_voice_11_0 be dataset.
36
  It achieves the following results on the evaluation set:
37
- - Loss: 0.9745
38
- - Wer: 75.2747
39
 
40
  ## Model description
41
 
@@ -61,7 +58,7 @@ The following hyperparameters were used during training:
61
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
  - lr_scheduler_type: linear
63
  - lr_scheduler_warmup_steps: 5
64
- - training_steps: 20
65
  - mixed_precision_training: Native AMP
66
 
67
  ### Training results
@@ -70,6 +67,14 @@ The following hyperparameters were used during training:
70
  |:-------------:|:-----:|:----:|:---------------:|:-------:|
71
  | 2.4473 | 0.5 | 10 | 1.3675 | 95.4212 |
72
  | 1.256 | 1.0 | 20 | 0.9745 | 75.2747 |
 
 
 
 
 
 
 
 
73
 
74
 
75
  ### Framework versions
 
1
  ---
 
 
2
  license: apache-2.0
3
  tags:
 
4
  - generated_from_trainer
5
  datasets:
6
+ - common_voice_11_0
7
  metrics:
8
  - wer
9
  model-index:
10
+ - name: whisper-tiny-be-test
11
  results:
12
  - task:
13
  name: Automatic Speech Recognition
14
  type: automatic-speech-recognition
15
  dataset:
16
+ name: common_voice_11_0
17
+ type: common_voice_11_0
18
  config: be
19
  split: validation
20
  args: be
21
  metrics:
22
  - name: Wer
23
  type: wer
24
+ value: 61.53846153846154
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
  should probably proofread and complete it, then remove this comment. -->
29
 
30
+ # whisper-tiny-be-test
31
 
32
+ This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the common_voice_11_0 dataset.
33
  It achieves the following results on the evaluation set:
34
+ - Loss: 0.5759
35
+ - Wer: 61.5385
36
 
37
  ## Model description
38
 
 
58
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
  - lr_scheduler_type: linear
60
  - lr_scheduler_warmup_steps: 5
61
+ - training_steps: 100
62
  - mixed_precision_training: Native AMP
63
 
64
  ### Training results
 
67
  |:-------------:|:-----:|:----:|:---------------:|:-------:|
68
  | 2.4473 | 0.5 | 10 | 1.3675 | 95.4212 |
69
  | 1.256 | 1.0 | 20 | 0.9745 | 75.2747 |
70
+ | 0.9934 | 0.3 | 30 | 0.8114 | 72.1612 |
71
+ | 0.9568 | 0.4 | 40 | 0.7814 | 72.7106 |
72
+ | 0.6856 | 0.5 | 50 | 0.7517 | 76.9231 |
73
+ | 0.7808 | 0.6 | 60 | 0.6514 | 63.5531 |
74
+ | 0.6826 | 0.7 | 70 | 0.6197 | 60.4396 |
75
+ | 0.7832 | 0.8 | 80 | 0.6129 | 65.9341 |
76
+ | 0.6031 | 0.9 | 90 | 0.5877 | 61.3553 |
77
+ | 0.6678 | 1.0 | 100 | 0.5759 | 61.5385 |
78
 
79
 
80
  ### Framework versions
train.log CHANGED
@@ -31,3 +31,5 @@
31
  {'loss': 0.6031, 'learning_rate': 1.3684210526315791e-05, 'epoch': 0.9}
32
  {'eval_loss': 0.5876654982566833, 'eval_wer': 61.35531135531136, 'eval_runtime': 20.4075, 'eval_samples_per_second': 3.136, 'eval_steps_per_second': 0.098, 'epoch': 0.9}
33
  {'loss': 0.6678, 'learning_rate': 3.1578947368421056e-06, 'epoch': 1.0}
 
 
 
31
  {'loss': 0.6031, 'learning_rate': 1.3684210526315791e-05, 'epoch': 0.9}
32
  {'eval_loss': 0.5876654982566833, 'eval_wer': 61.35531135531136, 'eval_runtime': 20.4075, 'eval_samples_per_second': 3.136, 'eval_steps_per_second': 0.098, 'epoch': 0.9}
33
  {'loss': 0.6678, 'learning_rate': 3.1578947368421056e-06, 'epoch': 1.0}
34
+ {'eval_loss': 0.5758526921272278, 'eval_wer': 61.53846153846154, 'eval_runtime': 19.5593, 'eval_samples_per_second': 3.272, 'eval_steps_per_second': 0.102, 'epoch': 1.0}
35
+ {'train_runtime': 782.3972, 'train_samples_per_second': 4.09, 'train_steps_per_second': 0.128, 'train_loss': 0.6153274965286255, 'epoch': 1.0}