jethrowang commited on
Commit
3d8f36f
·
verified ·
1 Parent(s): d7663a3

End of training

Browse files
Files changed (1) hide show
  1. README.md +18 -41
README.md CHANGED
@@ -7,22 +7,9 @@ tags:
7
  - generated_from_trainer
8
  datasets:
9
  - formospeech/tat_asr_aligned
10
- metrics:
11
- - wer
12
  model-index:
13
  - name: Whisper Tiny Taiwanese Android
14
- results:
15
- - task:
16
- name: Automatic Speech Recognition
17
- type: automatic-speech-recognition
18
- dataset:
19
- name: TAT ASR Aligned
20
- type: formospeech/tat_asr_aligned
21
- args: 'config: taiwanese, split: test'
22
- metrics:
23
- - name: Wer
24
- type: wer
25
- value: 46.86953383124636
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -32,8 +19,8 @@ should probably proofread and complete it, then remove this comment. -->
32
 
33
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the TAT ASR Aligned dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 0.9412
36
- - Wer: 46.8695
37
 
38
  ## Model description
39
 
@@ -52,40 +39,30 @@ More information needed
52
  ### Training hyperparameters
53
 
54
  The following hyperparameters were used during training:
55
- - learning_rate: 0.001
56
  - train_batch_size: 64
57
  - eval_batch_size: 32
58
  - seed: 42
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
- - lr_scheduler_warmup_steps: 1362
62
- - training_steps: 13620
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
66
 
67
- | Training Loss | Epoch | Step | Validation Loss | Wer |
68
- |:-------------:|:-------:|:-----:|:---------------:|:-------:|
69
- | 0.4657 | 0.9985 | 681 | 0.5472 | 63.0785 |
70
- | 0.4955 | 1.9971 | 1362 | 0.6142 | 66.7046 |
71
- | 0.3604 | 2.9956 | 2043 | 0.5341 | 59.9550 |
72
- | 0.2486 | 3.9941 | 2724 | 0.4831 | 56.0985 |
73
- | 0.1743 | 4.9927 | 3405 | 0.4750 | 54.1050 |
74
- | 0.1358 | 5.9912 | 4086 | 0.4915 | 51.4590 |
75
- | 0.1023 | 6.9897 | 4767 | 0.5325 | 52.2309 |
76
- | 0.0769 | 7.9883 | 5448 | 0.5405 | 50.7483 |
77
- | 0.0636 | 8.9868 | 6129 | 0.5518 | 50.9815 |
78
- | 0.0473 | 9.9853 | 6810 | 0.5817 | 49.4905 |
79
- | 0.0373 | 10.9839 | 7491 | 0.5983 | 49.9986 |
80
- | 0.0275 | 11.9824 | 8172 | 0.6474 | 50.0847 |
81
- | 0.0206 | 12.9809 | 8853 | 0.7030 | 50.2652 |
82
- | 0.0156 | 13.9795 | 9534 | 0.7327 | 49.8459 |
83
- | 0.009 | 14.9780 | 10215 | 0.7506 | 48.4244 |
84
- | 0.0071 | 15.9765 | 10896 | 0.7994 | 48.5271 |
85
- | 0.003 | 16.9751 | 11577 | 0.8406 | 47.6747 |
86
- | 0.0014 | 17.9736 | 12258 | 0.8839 | 47.4665 |
87
- | 0.0003 | 18.9721 | 12939 | 0.9245 | 46.7890 |
88
- | 0.0001 | 19.9707 | 13620 | 0.9412 | 46.8695 |
89
 
90
 
91
  ### Framework versions
 
7
  - generated_from_trainer
8
  datasets:
9
  - formospeech/tat_asr_aligned
 
 
10
  model-index:
11
  - name: Whisper Tiny Taiwanese Android
12
+ results: []
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
19
 
20
  This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the TAT ASR Aligned dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.5581
23
+ - Cer: 10.2914
24
 
25
  ## Model description
26
 
 
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
+ - learning_rate: 0.0001
43
  - train_batch_size: 64
44
  - eval_batch_size: 32
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
+ - lr_scheduler_warmup_steps: 681
49
+ - training_steps: 6810
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
+ | Training Loss | Epoch | Step | Validation Loss | Cer |
55
+ |:-------------:|:------:|:----:|:---------------:|:-------:|
56
+ | 0.3523 | 0.9985 | 681 | 0.4273 | 15.0771 |
57
+ | 0.2104 | 1.9971 | 1362 | 0.3778 | 12.2058 |
58
+ | 0.1224 | 2.9956 | 2043 | 0.3942 | 11.9977 |
59
+ | 0.0738 | 3.9941 | 2724 | 0.4164 | 11.9422 |
60
+ | 0.0423 | 4.9927 | 3405 | 0.4579 | 11.4839 |
61
+ | 0.0216 | 5.9912 | 4086 | 0.4818 | 11.3165 |
62
+ | 0.0107 | 6.9897 | 4767 | 0.5189 | 10.9872 |
63
+ | 0.0036 | 7.9883 | 5448 | 0.5398 | 10.8198 |
64
+ | 0.0012 | 8.9868 | 6129 | 0.5509 | 10.3179 |
65
+ | 0.0006 | 9.9853 | 6810 | 0.5581 | 10.2914 |
 
 
 
 
 
 
 
 
 
 
66
 
67
 
68
  ### Framework versions