sant commited on
Commit
a2aa996
·
verified ·
1 Parent(s): 8c842ac

End of training

Browse files
README.md CHANGED
@@ -6,19 +6,19 @@ tags:
6
  metrics:
7
  - wer
8
  model-index:
9
- - name: small-Cotrsc
10
  results: []
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
- # small-Cotrsc
17
 
18
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.0196
21
- - Wer: 19.7923
22
 
23
  ## Model description
24
 
@@ -37,23 +37,24 @@ More information needed
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
- - learning_rate: 1e-05
41
- - train_batch_size: 12
42
  - eval_batch_size: 8
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 500
47
- - training_steps: 1200
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Wer |
53
  |:-------------:|:------:|:----:|:---------------:|:-------:|
54
- | 0.0186 | 0.1634 | 400 | 0.0212 | 21.7574 |
55
- | 0.0206 | 0.3268 | 800 | 0.0209 | 21.5890 |
56
- | 0.0206 | 0.4902 | 1200 | 0.0196 | 19.7923 |
 
57
 
58
 
59
  ### Framework versions
 
6
  metrics:
7
  - wer
8
  model-index:
9
+ - name: Whisper-opus
10
  results: []
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # Whisper-opus
17
 
18
  This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.0113
21
+ - Wer: 74.0448
22
 
23
  ## Model description
24
 
 
37
  ### Training hyperparameters
38
 
39
  The following hyperparameters were used during training:
40
+ - learning_rate: 2e-05
41
+ - train_batch_size: 24
42
  - eval_batch_size: 8
43
  - seed: 42
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 500
47
+ - training_steps: 2000
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Wer |
53
  |:-------------:|:------:|:----:|:---------------:|:-------:|
54
+ | 0.0173 | 0.3671 | 500 | 0.0139 | 70.8788 |
55
+ | 0.0194 | 0.7342 | 1000 | 0.0130 | 48.4443 |
56
+ | 0.0172 | 1.1013 | 1500 | 0.0119 | 56.8231 |
57
+ | 0.0117 | 1.4684 | 2000 | 0.0113 | 74.0448 |
58
 
59
 
60
  ### Framework versions
generation_config.json CHANGED
@@ -1,38 +1,4 @@
1
  {
2
- "alignment_heads": [
3
- [
4
- 3,
5
- 1
6
- ],
7
- [
8
- 4,
9
- 2
10
- ],
11
- [
12
- 4,
13
- 3
14
- ],
15
- [
16
- 4,
17
- 7
18
- ],
19
- [
20
- 5,
21
- 1
22
- ],
23
- [
24
- 5,
25
- 2
26
- ],
27
- [
28
- 5,
29
- 4
30
- ],
31
- [
32
- 5,
33
- 6
34
- ]
35
- ],
36
  "begin_suppress_tokens": [
37
  220,
38
  50257
@@ -43,121 +9,19 @@
43
  "forced_decoder_ids": [
44
  [
45
  1,
46
- null
47
  ],
48
  [
49
  2,
50
  50359
 
 
 
 
51
  ]
52
  ],
53
- "is_multilingual": true,
54
- "lang_to_id": {
55
- "<|af|>": 50327,
56
- "<|am|>": 50334,
57
- "<|ar|>": 50272,
58
- "<|as|>": 50350,
59
- "<|az|>": 50304,
60
- "<|ba|>": 50355,
61
- "<|be|>": 50330,
62
- "<|bg|>": 50292,
63
- "<|bn|>": 50302,
64
- "<|bo|>": 50347,
65
- "<|br|>": 50309,
66
- "<|bs|>": 50315,
67
- "<|ca|>": 50270,
68
- "<|cs|>": 50283,
69
- "<|cy|>": 50297,
70
- "<|da|>": 50285,
71
- "<|de|>": 50261,
72
- "<|el|>": 50281,
73
- "<|en|>": 50259,
74
- "<|es|>": 50262,
75
- "<|et|>": 50307,
76
- "<|eu|>": 50310,
77
- "<|fa|>": 50300,
78
- "<|fi|>": 50277,
79
- "<|fo|>": 50338,
80
- "<|fr|>": 50265,
81
- "<|gl|>": 50319,
82
- "<|gu|>": 50333,
83
- "<|haw|>": 50352,
84
- "<|ha|>": 50354,
85
- "<|he|>": 50279,
86
- "<|hi|>": 50276,
87
- "<|hr|>": 50291,
88
- "<|ht|>": 50339,
89
- "<|hu|>": 50286,
90
- "<|hy|>": 50312,
91
- "<|id|>": 50275,
92
- "<|is|>": 50311,
93
- "<|it|>": 50274,
94
- "<|ja|>": 50266,
95
- "<|jw|>": 50356,
96
- "<|ka|>": 50329,
97
- "<|kk|>": 50316,
98
- "<|km|>": 50323,
99
- "<|kn|>": 50306,
100
- "<|ko|>": 50264,
101
- "<|la|>": 50294,
102
- "<|lb|>": 50345,
103
- "<|ln|>": 50353,
104
- "<|lo|>": 50336,
105
- "<|lt|>": 50293,
106
- "<|lv|>": 50301,
107
- "<|mg|>": 50349,
108
- "<|mi|>": 50295,
109
- "<|mk|>": 50308,
110
- "<|ml|>": 50296,
111
- "<|mn|>": 50314,
112
- "<|mr|>": 50320,
113
- "<|ms|>": 50282,
114
- "<|mt|>": 50343,
115
- "<|my|>": 50346,
116
- "<|ne|>": 50313,
117
- "<|nl|>": 50271,
118
- "<|nn|>": 50342,
119
- "<|no|>": 50288,
120
- "<|oc|>": 50328,
121
- "<|pa|>": 50321,
122
- "<|pl|>": 50269,
123
- "<|ps|>": 50340,
124
- "<|pt|>": 50267,
125
- "<|ro|>": 50284,
126
- "<|ru|>": 50263,
127
- "<|sa|>": 50344,
128
- "<|sd|>": 50332,
129
- "<|si|>": 50322,
130
- "<|sk|>": 50298,
131
- "<|sl|>": 50305,
132
- "<|sn|>": 50324,
133
- "<|so|>": 50326,
134
- "<|sq|>": 50317,
135
- "<|sr|>": 50303,
136
- "<|su|>": 50357,
137
- "<|sv|>": 50273,
138
- "<|sw|>": 50318,
139
- "<|ta|>": 50287,
140
- "<|te|>": 50299,
141
- "<|tg|>": 50331,
142
- "<|th|>": 50289,
143
- "<|tk|>": 50341,
144
- "<|tl|>": 50348,
145
- "<|tr|>": 50268,
146
- "<|tt|>": 50351,
147
- "<|uk|>": 50280,
148
- "<|ur|>": 50290,
149
- "<|uz|>": 50337,
150
- "<|vi|>": 50278,
151
- "<|yi|>": 50335,
152
- "<|yo|>": 50325,
153
- "<|zh|>": 50260
154
- },
155
- "max_initial_timestamp_index": 50,
156
  "max_length": 448,
157
- "no_timestamps_token_id": 50363,
158
  "pad_token_id": 50257,
159
- "prev_sot_token_id": 50361,
160
- "return_timestamps": false,
161
  "suppress_tokens": [
162
  1,
163
  2,
@@ -248,9 +112,6 @@
248
  50361,
249
  50362
250
  ],
251
- "task_to_id": {
252
- "transcribe": 50359,
253
- "translate": 50358
254
- },
255
- "transformers_version": "4.41.2"
256
  }
 
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "begin_suppress_tokens": [
3
  220,
4
  50257
 
9
  "forced_decoder_ids": [
10
  [
11
  1,
12
+ 50259
13
  ],
14
  [
15
  2,
16
  50359
17
+ ],
18
+ [
19
+ 3,
20
+ 50363
21
  ]
22
  ],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "max_length": 448,
 
24
  "pad_token_id": 50257,
 
 
25
  "suppress_tokens": [
26
  1,
27
  2,
 
112
  50361,
113
  50362
114
  ],
115
+ "transformers_version": "4.41.2",
116
+ "use_cache": false
 
 
 
117
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0d7aae63d783c9621b7c551b2bd6443db6cdd8ede1567126c2c6f13b17f33b7
3
  size 290403936
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5784c3e46e361db977c49763a7d67278e7339189638352851d4c21472378f34e
3
  size 290403936
runs/Jun22_14-05-49_7c59a2b76305/events.out.tfevents.1719065152.7c59a2b76305.34.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bcd0dcd2fca0493a7f41239962a820630277179ed17441bd02fbe618a5ac9b0c
3
- size 8918
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f82e502edb3564478c31da10a1e3aacf2608d615986ea380cbf67f9614434266
3
+ size 9272