abdouaziiz commited on
Commit
b41186d
·
verified ·
1 Parent(s): 11567d2

End of training

Browse files
README.md CHANGED
@@ -4,11 +4,23 @@ license: apache-2.0
4
  base_model: openai/whisper-small
5
  tags:
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - wer
9
  model-index:
10
  - name: MODELING_23
11
- results: []
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -16,10 +28,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # MODELING_23
18
 
19
- This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.5140
22
- - Wer: 0.2730
23
 
24
  ## Model description
25
 
 
4
  base_model: openai/whisper-small
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - abdouaziiz/full_wolof_normalized
9
  metrics:
10
  - wer
11
  model-index:
12
  - name: MODELING_23
13
+ results:
14
+ - task:
15
+ name: Automatic Speech Recognition
16
+ type: automatic-speech-recognition
17
+ dataset:
18
+ name: abdouaziiz/full_wolof_normalized
19
+ type: abdouaziiz/full_wolof_normalized
20
+ metrics:
21
+ - name: Wer
22
+ type: wer
23
+ value: 0.26790202840942307
24
  ---
25
 
26
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
28
 
29
  # MODELING_23
30
 
31
+ This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the abdouaziiz/full_wolof_normalized dataset.
32
  It achieves the following results on the evaluation set:
33
+ - Loss: 0.4923
34
+ - Wer: 0.2679
35
 
36
  ## Model description
37
 
all_results.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.356112169888375,
3
+ "eval_loss": 0.49226129055023193,
4
+ "eval_runtime": 608.3552,
5
+ "eval_samples": 1544,
6
+ "eval_samples_per_second": 2.538,
7
+ "eval_steps_per_second": 0.159,
8
+ "eval_wer": 0.26790202840942307,
9
+ "total_flos": 7.38663190659072e+19,
10
+ "train_loss": 0.6509292511940002,
11
+ "train_runtime": 136739.1848,
12
+ "train_samples": 58758,
13
+ "train_samples_per_second": 11.233,
14
+ "train_steps_per_second": 0.351
15
+ }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.356112169888375,
3
+ "eval_loss": 0.49226129055023193,
4
+ "eval_runtime": 608.3552,
5
+ "eval_samples": 1544,
6
+ "eval_samples_per_second": 2.538,
7
+ "eval_steps_per_second": 0.159,
8
+ "eval_wer": 0.26790202840942307
9
+ }
runs/Aug03_21-13-12_bambaratts/events.out.tfevents.1754393008.bambaratts.10974.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c44f79eccf5e0426b80acf59ba2c7cfa694aa1a3b5511bc5fe9f07b5a788169d
3
+ size 406
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 4.356112169888375,
3
+ "total_flos": 7.38663190659072e+19,
4
+ "train_loss": 0.6509292511940002,
5
+ "train_runtime": 136739.1848,
6
+ "train_samples": 58758,
7
+ "train_samples_per_second": 11.233,
8
+ "train_steps_per_second": 0.351
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,307 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.26790202840942307,
3
+ "best_model_checkpoint": "MODELING_23/checkpoint-6500",
4
+ "epoch": 4.356112169888375,
5
+ "eval_steps": 500,
6
+ "global_step": 8000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.27225701061802343,
13
+ "grad_norm": 11.537489891052246,
14
+ "learning_rate": 2.972033368091762e-05,
15
+ "loss": 1.9983,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.27225701061802343,
20
+ "eval_loss": 0.711998701095581,
21
+ "eval_runtime": 794.9584,
22
+ "eval_samples_per_second": 1.942,
23
+ "eval_steps_per_second": 0.122,
24
+ "eval_wer": 0.6396212076927573,
25
+ "step": 500
26
+ },
27
+ {
28
+ "epoch": 0.5445140212360469,
29
+ "grad_norm": 10.056964874267578,
30
+ "learning_rate": 2.9407507820646507e-05,
31
+ "loss": 1.2974,
32
+ "step": 1000
33
+ },
34
+ {
35
+ "epoch": 0.5445140212360469,
36
+ "eval_loss": 0.6104683876037598,
37
+ "eval_runtime": 757.7831,
38
+ "eval_samples_per_second": 2.038,
39
+ "eval_steps_per_second": 0.128,
40
+ "eval_wer": 0.578301280177705,
41
+ "step": 1000
42
+ },
43
+ {
44
+ "epoch": 0.8167710318540703,
45
+ "grad_norm": 9.043569564819336,
46
+ "learning_rate": 2.909468196037539e-05,
47
+ "loss": 1.1407,
48
+ "step": 1500
49
+ },
50
+ {
51
+ "epoch": 0.8167710318540703,
52
+ "eval_loss": 0.5547229051589966,
53
+ "eval_runtime": 785.9187,
54
+ "eval_samples_per_second": 1.965,
55
+ "eval_steps_per_second": 0.123,
56
+ "eval_wer": 0.572397264277781,
57
+ "step": 1500
58
+ },
59
+ {
60
+ "epoch": 1.0890280424720937,
61
+ "grad_norm": 8.64743423461914,
62
+ "learning_rate": 2.8781856100104274e-05,
63
+ "loss": 0.9514,
64
+ "step": 2000
65
+ },
66
+ {
67
+ "epoch": 1.0890280424720937,
68
+ "eval_loss": 0.5316600799560547,
69
+ "eval_runtime": 748.1024,
70
+ "eval_samples_per_second": 2.064,
71
+ "eval_steps_per_second": 0.13,
72
+ "eval_wer": 0.46518968843163616,
73
+ "step": 2000
74
+ },
75
+ {
76
+ "epoch": 1.361285053090117,
77
+ "grad_norm": 8.403170585632324,
78
+ "learning_rate": 2.846903023983316e-05,
79
+ "loss": 0.7178,
80
+ "step": 2500
81
+ },
82
+ {
83
+ "epoch": 1.361285053090117,
84
+ "eval_loss": 0.5095750689506531,
85
+ "eval_runtime": 752.8094,
86
+ "eval_samples_per_second": 2.051,
87
+ "eval_steps_per_second": 0.129,
88
+ "eval_wer": 0.4634944759455194,
89
+ "step": 2500
90
+ },
91
+ {
92
+ "epoch": 1.6335420637081404,
93
+ "grad_norm": 8.097471237182617,
94
+ "learning_rate": 2.8156204379562044e-05,
95
+ "loss": 0.7264,
96
+ "step": 3000
97
+ },
98
+ {
99
+ "epoch": 1.6335420637081404,
100
+ "eval_loss": 0.49745145440101624,
101
+ "eval_runtime": 664.6083,
102
+ "eval_samples_per_second": 2.323,
103
+ "eval_steps_per_second": 0.146,
104
+ "eval_wer": 0.3532472087449582,
105
+ "step": 3000
106
+ },
107
+ {
108
+ "epoch": 1.905799074326164,
109
+ "grad_norm": 7.794954776763916,
110
+ "learning_rate": 2.784337851929093e-05,
111
+ "loss": 0.7091,
112
+ "step": 3500
113
+ },
114
+ {
115
+ "epoch": 1.905799074326164,
116
+ "eval_loss": 0.4781028628349304,
117
+ "eval_runtime": 631.8344,
118
+ "eval_samples_per_second": 2.444,
119
+ "eval_steps_per_second": 0.154,
120
+ "eval_wer": 0.31595253405038876,
121
+ "step": 3500
122
+ },
123
+ {
124
+ "epoch": 2.1780560849441875,
125
+ "grad_norm": 6.23008918762207,
126
+ "learning_rate": 2.7530552659019814e-05,
127
+ "loss": 0.5207,
128
+ "step": 4000
129
+ },
130
+ {
131
+ "epoch": 2.1780560849441875,
132
+ "eval_loss": 0.47789445519447327,
133
+ "eval_runtime": 658.2478,
134
+ "eval_samples_per_second": 2.346,
135
+ "eval_steps_per_second": 0.147,
136
+ "eval_wer": 0.3010463552931548,
137
+ "step": 4000
138
+ },
139
+ {
140
+ "epoch": 2.4503130955622106,
141
+ "grad_norm": 6.356353282928467,
142
+ "learning_rate": 2.7217726798748698e-05,
143
+ "loss": 0.4213,
144
+ "step": 4500
145
+ },
146
+ {
147
+ "epoch": 2.4503130955622106,
148
+ "eval_loss": 0.4791910946369171,
149
+ "eval_runtime": 671.3319,
150
+ "eval_samples_per_second": 2.3,
151
+ "eval_steps_per_second": 0.144,
152
+ "eval_wer": 0.41065061086105104,
153
+ "step": 4500
154
+ },
155
+ {
156
+ "epoch": 2.722570106180234,
157
+ "grad_norm": 7.0236334800720215,
158
+ "learning_rate": 2.690490093847758e-05,
159
+ "loss": 0.442,
160
+ "step": 5000
161
+ },
162
+ {
163
+ "epoch": 2.722570106180234,
164
+ "eval_loss": 0.47505632042884827,
165
+ "eval_runtime": 644.8379,
166
+ "eval_samples_per_second": 2.394,
167
+ "eval_steps_per_second": 0.15,
168
+ "eval_wer": 0.3007540772783071,
169
+ "step": 5000
170
+ },
171
+ {
172
+ "epoch": 2.9948271167982576,
173
+ "grad_norm": 8.722136497497559,
174
+ "learning_rate": 2.6592700729927007e-05,
175
+ "loss": 0.4337,
176
+ "step": 5500
177
+ },
178
+ {
179
+ "epoch": 2.9948271167982576,
180
+ "eval_loss": 0.4598409831523895,
181
+ "eval_runtime": 635.9124,
182
+ "eval_samples_per_second": 2.428,
183
+ "eval_steps_per_second": 0.153,
184
+ "eval_wer": 0.3061904483544748,
185
+ "step": 5500
186
+ },
187
+ {
188
+ "epoch": 3.2670841274162807,
189
+ "grad_norm": 6.036080360412598,
190
+ "learning_rate": 2.627987486965589e-05,
191
+ "loss": 0.2341,
192
+ "step": 6000
193
+ },
194
+ {
195
+ "epoch": 3.2670841274162807,
196
+ "eval_loss": 0.49184465408325195,
197
+ "eval_runtime": 640.365,
198
+ "eval_samples_per_second": 2.411,
199
+ "eval_steps_per_second": 0.151,
200
+ "eval_wer": 0.3112760858128252,
201
+ "step": 6000
202
+ },
203
+ {
204
+ "epoch": 3.5393411380343043,
205
+ "grad_norm": 5.870471000671387,
206
+ "learning_rate": 2.5967049009384774e-05,
207
+ "loss": 0.2406,
208
+ "step": 6500
209
+ },
210
+ {
211
+ "epoch": 3.5393411380343043,
212
+ "eval_loss": 0.49226129055023193,
213
+ "eval_runtime": 613.8807,
214
+ "eval_samples_per_second": 2.515,
215
+ "eval_steps_per_second": 0.158,
216
+ "eval_wer": 0.26790202840942307,
217
+ "step": 6500
218
+ },
219
+ {
220
+ "epoch": 3.811598148652328,
221
+ "grad_norm": 6.475042343139648,
222
+ "learning_rate": 2.565422314911366e-05,
223
+ "loss": 0.2479,
224
+ "step": 7000
225
+ },
226
+ {
227
+ "epoch": 3.811598148652328,
228
+ "eval_loss": 0.4866102933883667,
229
+ "eval_runtime": 623.7701,
230
+ "eval_samples_per_second": 2.475,
231
+ "eval_steps_per_second": 0.156,
232
+ "eval_wer": 0.2886537674636114,
233
+ "step": 7000
234
+ },
235
+ {
236
+ "epoch": 4.083855159270351,
237
+ "grad_norm": 4.934613227844238,
238
+ "learning_rate": 2.5341397288842544e-05,
239
+ "loss": 0.2078,
240
+ "step": 7500
241
+ },
242
+ {
243
+ "epoch": 4.083855159270351,
244
+ "eval_loss": 0.501932680606842,
245
+ "eval_runtime": 633.874,
246
+ "eval_samples_per_second": 2.436,
247
+ "eval_steps_per_second": 0.153,
248
+ "eval_wer": 0.29134272520021043,
249
+ "step": 7500
250
+ },
251
+ {
252
+ "epoch": 4.356112169888375,
253
+ "grad_norm": 4.6739068031311035,
254
+ "learning_rate": 2.502857142857143e-05,
255
+ "loss": 0.1258,
256
+ "step": 8000
257
+ },
258
+ {
259
+ "epoch": 4.356112169888375,
260
+ "eval_loss": 0.5139926075935364,
261
+ "eval_runtime": 635.4969,
262
+ "eval_samples_per_second": 2.43,
263
+ "eval_steps_per_second": 0.153,
264
+ "eval_wer": 0.2729876658677734,
265
+ "step": 8000
266
+ },
267
+ {
268
+ "epoch": 4.356112169888375,
269
+ "step": 8000,
270
+ "total_flos": 7.38663190659072e+19,
271
+ "train_loss": 0.6509292511940002,
272
+ "train_runtime": 136739.1848,
273
+ "train_samples_per_second": 11.233,
274
+ "train_steps_per_second": 0.351
275
+ }
276
+ ],
277
+ "logging_steps": 500,
278
+ "max_steps": 48000,
279
+ "num_input_tokens_seen": 0,
280
+ "num_train_epochs": 27,
281
+ "save_steps": 500,
282
+ "stateful_callbacks": {
283
+ "EarlyStoppingCallback": {
284
+ "args": {
285
+ "early_stopping_patience": 3,
286
+ "early_stopping_threshold": 0.0
287
+ },
288
+ "attributes": {
289
+ "early_stopping_patience_counter": 3
290
+ }
291
+ },
292
+ "TrainerControl": {
293
+ "args": {
294
+ "should_epoch_stop": false,
295
+ "should_evaluate": false,
296
+ "should_log": false,
297
+ "should_save": true,
298
+ "should_training_stop": true
299
+ },
300
+ "attributes": {}
301
+ }
302
+ },
303
+ "total_flos": 7.38663190659072e+19,
304
+ "train_batch_size": 16,
305
+ "trial_name": null,
306
+ "trial_params": null
307
+ }