ZipeiZhang commited on
Commit
e04e7af
·
1 Parent(s): 17c635f

model updated

Browse files
checkpoint-8000/config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-small",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
+ "architectures": [
7
+ "WhisperForConditionalGeneration"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": [
11
+ 220,
12
+ 50257
13
+ ],
14
+ "bos_token_id": 50257,
15
+ "classifier_proj_size": 256,
16
+ "d_model": 768,
17
+ "decoder_attention_heads": 12,
18
+ "decoder_ffn_dim": 3072,
19
+ "decoder_layerdrop": 0.0,
20
+ "decoder_layers": 12,
21
+ "decoder_start_token_id": 50258,
22
+ "dropout": 0.0,
23
+ "encoder_attention_heads": 12,
24
+ "encoder_ffn_dim": 3072,
25
+ "encoder_layerdrop": 0.0,
26
+ "encoder_layers": 12,
27
+ "eos_token_id": 50257,
28
+ "forced_decoder_ids": null,
29
+ "init_std": 0.02,
30
+ "is_encoder_decoder": true,
31
+ "mask_feature_length": 10,
32
+ "mask_feature_min_masks": 0,
33
+ "mask_feature_prob": 0.0,
34
+ "mask_time_length": 10,
35
+ "mask_time_min_masks": 2,
36
+ "mask_time_prob": 0.05,
37
+ "max_length": 448,
38
+ "max_source_positions": 1500,
39
+ "max_target_positions": 448,
40
+ "median_filter_width": 7,
41
+ "model_type": "whisper",
42
+ "num_hidden_layers": 12,
43
+ "num_mel_bins": 80,
44
+ "pad_token_id": 50257,
45
+ "scale_embedding": false,
46
+ "suppress_tokens": [],
47
+ "torch_dtype": "float32",
48
+ "transformers_version": "4.35.2",
49
+ "use_cache": true,
50
+ "use_weighted_layer_sum": false,
51
+ "vocab_size": 51865
52
+ }
checkpoint-8000/generation_config.json ADDED
@@ -0,0 +1,263 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alignment_heads": [
3
+ [
4
+ 5,
5
+ 3
6
+ ],
7
+ [
8
+ 5,
9
+ 9
10
+ ],
11
+ [
12
+ 8,
13
+ 0
14
+ ],
15
+ [
16
+ 8,
17
+ 4
18
+ ],
19
+ [
20
+ 8,
21
+ 7
22
+ ],
23
+ [
24
+ 8,
25
+ 8
26
+ ],
27
+ [
28
+ 9,
29
+ 0
30
+ ],
31
+ [
32
+ 9,
33
+ 7
34
+ ],
35
+ [
36
+ 9,
37
+ 9
38
+ ],
39
+ [
40
+ 10,
41
+ 5
42
+ ]
43
+ ],
44
+ "begin_suppress_tokens": [
45
+ 220,
46
+ 50257
47
+ ],
48
+ "bos_token_id": 50257,
49
+ "decoder_start_token_id": 50258,
50
+ "eos_token_id": 50257,
51
+ "forced_decoder_ids": [
52
+ [
53
+ 1,
54
+ null
55
+ ],
56
+ [
57
+ 2,
58
+ 50359
59
+ ]
60
+ ],
61
+ "is_multilingual": true,
62
+ "lang_to_id": {
63
+ "<|af|>": 50327,
64
+ "<|am|>": 50334,
65
+ "<|ar|>": 50272,
66
+ "<|as|>": 50350,
67
+ "<|az|>": 50304,
68
+ "<|ba|>": 50355,
69
+ "<|be|>": 50330,
70
+ "<|bg|>": 50292,
71
+ "<|bn|>": 50302,
72
+ "<|bo|>": 50347,
73
+ "<|br|>": 50309,
74
+ "<|bs|>": 50315,
75
+ "<|ca|>": 50270,
76
+ "<|cs|>": 50283,
77
+ "<|cy|>": 50297,
78
+ "<|da|>": 50285,
79
+ "<|de|>": 50261,
80
+ "<|el|>": 50281,
81
+ "<|en|>": 50259,
82
+ "<|es|>": 50262,
83
+ "<|et|>": 50307,
84
+ "<|eu|>": 50310,
85
+ "<|fa|>": 50300,
86
+ "<|fi|>": 50277,
87
+ "<|fo|>": 50338,
88
+ "<|fr|>": 50265,
89
+ "<|gl|>": 50319,
90
+ "<|gu|>": 50333,
91
+ "<|haw|>": 50352,
92
+ "<|ha|>": 50354,
93
+ "<|he|>": 50279,
94
+ "<|hi|>": 50276,
95
+ "<|hr|>": 50291,
96
+ "<|ht|>": 50339,
97
+ "<|hu|>": 50286,
98
+ "<|hy|>": 50312,
99
+ "<|id|>": 50275,
100
+ "<|is|>": 50311,
101
+ "<|it|>": 50274,
102
+ "<|ja|>": 50266,
103
+ "<|jw|>": 50356,
104
+ "<|ka|>": 50329,
105
+ "<|kk|>": 50316,
106
+ "<|km|>": 50323,
107
+ "<|kn|>": 50306,
108
+ "<|ko|>": 50264,
109
+ "<|la|>": 50294,
110
+ "<|lb|>": 50345,
111
+ "<|ln|>": 50353,
112
+ "<|lo|>": 50336,
113
+ "<|lt|>": 50293,
114
+ "<|lv|>": 50301,
115
+ "<|mg|>": 50349,
116
+ "<|mi|>": 50295,
117
+ "<|mk|>": 50308,
118
+ "<|ml|>": 50296,
119
+ "<|mn|>": 50314,
120
+ "<|mr|>": 50320,
121
+ "<|ms|>": 50282,
122
+ "<|mt|>": 50343,
123
+ "<|my|>": 50346,
124
+ "<|ne|>": 50313,
125
+ "<|nl|>": 50271,
126
+ "<|nn|>": 50342,
127
+ "<|no|>": 50288,
128
+ "<|oc|>": 50328,
129
+ "<|pa|>": 50321,
130
+ "<|pl|>": 50269,
131
+ "<|ps|>": 50340,
132
+ "<|pt|>": 50267,
133
+ "<|ro|>": 50284,
134
+ "<|ru|>": 50263,
135
+ "<|sa|>": 50344,
136
+ "<|sd|>": 50332,
137
+ "<|si|>": 50322,
138
+ "<|sk|>": 50298,
139
+ "<|sl|>": 50305,
140
+ "<|sn|>": 50324,
141
+ "<|so|>": 50326,
142
+ "<|sq|>": 50317,
143
+ "<|sr|>": 50303,
144
+ "<|su|>": 50357,
145
+ "<|sv|>": 50273,
146
+ "<|sw|>": 50318,
147
+ "<|ta|>": 50287,
148
+ "<|te|>": 50299,
149
+ "<|tg|>": 50331,
150
+ "<|th|>": 50289,
151
+ "<|tk|>": 50341,
152
+ "<|tl|>": 50348,
153
+ "<|tr|>": 50268,
154
+ "<|tt|>": 50351,
155
+ "<|uk|>": 50280,
156
+ "<|ur|>": 50290,
157
+ "<|uz|>": 50337,
158
+ "<|vi|>": 50278,
159
+ "<|yi|>": 50335,
160
+ "<|yo|>": 50325,
161
+ "<|zh|>": 50260
162
+ },
163
+ "max_initial_timestamp_index": 1,
164
+ "max_length": 448,
165
+ "no_timestamps_token_id": 50363,
166
+ "pad_token_id": 50257,
167
+ "return_timestamps": false,
168
+ "suppress_tokens": [
169
+ 1,
170
+ 2,
171
+ 7,
172
+ 8,
173
+ 9,
174
+ 10,
175
+ 14,
176
+ 25,
177
+ 26,
178
+ 27,
179
+ 28,
180
+ 29,
181
+ 31,
182
+ 58,
183
+ 59,
184
+ 60,
185
+ 61,
186
+ 62,
187
+ 63,
188
+ 90,
189
+ 91,
190
+ 92,
191
+ 93,
192
+ 359,
193
+ 503,
194
+ 522,
195
+ 542,
196
+ 873,
197
+ 893,
198
+ 902,
199
+ 918,
200
+ 922,
201
+ 931,
202
+ 1350,
203
+ 1853,
204
+ 1982,
205
+ 2460,
206
+ 2627,
207
+ 3246,
208
+ 3253,
209
+ 3268,
210
+ 3536,
211
+ 3846,
212
+ 3961,
213
+ 4183,
214
+ 4667,
215
+ 6585,
216
+ 6647,
217
+ 7273,
218
+ 9061,
219
+ 9383,
220
+ 10428,
221
+ 10929,
222
+ 11938,
223
+ 12033,
224
+ 12331,
225
+ 12562,
226
+ 13793,
227
+ 14157,
228
+ 14635,
229
+ 15265,
230
+ 15618,
231
+ 16553,
232
+ 16604,
233
+ 18362,
234
+ 18956,
235
+ 20075,
236
+ 21675,
237
+ 22520,
238
+ 26130,
239
+ 26161,
240
+ 26435,
241
+ 28279,
242
+ 29464,
243
+ 31650,
244
+ 32302,
245
+ 32470,
246
+ 36865,
247
+ 42863,
248
+ 47425,
249
+ 49870,
250
+ 50254,
251
+ 50258,
252
+ 50358,
253
+ 50359,
254
+ 50360,
255
+ 50361,
256
+ 50362
257
+ ],
258
+ "task_to_id": {
259
+ "transcribe": 50359,
260
+ "translate": 50358
261
+ },
262
+ "transformers_version": "4.35.2"
263
+ }
checkpoint-8000/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daaaf0bebfd66b090622504ba6169a69368bb067d853fff6af7bc80100a4ab0f
3
+ size 966995080
checkpoint-8000/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745bd8b2b44d2c841e140658015bc5613c15a54ed94010775f6b4645629858e2
3
+ size 1925064044
checkpoint-8000/preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
checkpoint-8000/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:784b8146198311f72f710aa91e95ab12d1c38fb99b2cf7054df1d54f7a08bcc6
3
+ size 14244
checkpoint-8000/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cee471c36a070fe31d714bec29767e262f3969de51c929a44322fd212536ece1
3
+ size 1064
checkpoint-8000/trainer_state.json ADDED
@@ -0,0 +1,1975 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 160.24633680186878,
3
+ "best_model_checkpoint": "whisper_hi_test/checkpoint-6000",
4
+ "epoch": 11.347517730496454,
5
+ "eval_steps": 2000,
6
+ "global_step": 8000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.04,
13
+ "learning_rate": 4.4e-07,
14
+ "loss": 5.185,
15
+ "step": 25
16
+ },
17
+ {
18
+ "epoch": 0.07,
19
+ "learning_rate": 9.400000000000001e-07,
20
+ "loss": 3.641,
21
+ "step": 50
22
+ },
23
+ {
24
+ "epoch": 0.11,
25
+ "learning_rate": 1.44e-06,
26
+ "loss": 2.3655,
27
+ "step": 75
28
+ },
29
+ {
30
+ "epoch": 0.14,
31
+ "learning_rate": 1.94e-06,
32
+ "loss": 1.1676,
33
+ "step": 100
34
+ },
35
+ {
36
+ "epoch": 0.18,
37
+ "learning_rate": 2.4400000000000004e-06,
38
+ "loss": 0.823,
39
+ "step": 125
40
+ },
41
+ {
42
+ "epoch": 0.21,
43
+ "learning_rate": 2.9400000000000002e-06,
44
+ "loss": 0.6644,
45
+ "step": 150
46
+ },
47
+ {
48
+ "epoch": 0.25,
49
+ "learning_rate": 3.44e-06,
50
+ "loss": 0.5072,
51
+ "step": 175
52
+ },
53
+ {
54
+ "epoch": 0.28,
55
+ "learning_rate": 3.94e-06,
56
+ "loss": 0.3801,
57
+ "step": 200
58
+ },
59
+ {
60
+ "epoch": 0.32,
61
+ "learning_rate": 4.440000000000001e-06,
62
+ "loss": 0.2505,
63
+ "step": 225
64
+ },
65
+ {
66
+ "epoch": 0.35,
67
+ "learning_rate": 4.94e-06,
68
+ "loss": 0.2115,
69
+ "step": 250
70
+ },
71
+ {
72
+ "epoch": 0.39,
73
+ "learning_rate": 5.4400000000000004e-06,
74
+ "loss": 0.2107,
75
+ "step": 275
76
+ },
77
+ {
78
+ "epoch": 0.43,
79
+ "learning_rate": 5.94e-06,
80
+ "loss": 0.2276,
81
+ "step": 300
82
+ },
83
+ {
84
+ "epoch": 0.46,
85
+ "learning_rate": 6.440000000000001e-06,
86
+ "loss": 0.1958,
87
+ "step": 325
88
+ },
89
+ {
90
+ "epoch": 0.5,
91
+ "learning_rate": 6.9400000000000005e-06,
92
+ "loss": 0.2054,
93
+ "step": 350
94
+ },
95
+ {
96
+ "epoch": 0.53,
97
+ "learning_rate": 7.440000000000001e-06,
98
+ "loss": 0.2074,
99
+ "step": 375
100
+ },
101
+ {
102
+ "epoch": 0.57,
103
+ "learning_rate": 7.94e-06,
104
+ "loss": 0.2117,
105
+ "step": 400
106
+ },
107
+ {
108
+ "epoch": 0.6,
109
+ "learning_rate": 8.44e-06,
110
+ "loss": 0.2116,
111
+ "step": 425
112
+ },
113
+ {
114
+ "epoch": 0.64,
115
+ "learning_rate": 8.94e-06,
116
+ "loss": 0.1729,
117
+ "step": 450
118
+ },
119
+ {
120
+ "epoch": 0.67,
121
+ "learning_rate": 9.440000000000001e-06,
122
+ "loss": 0.2045,
123
+ "step": 475
124
+ },
125
+ {
126
+ "epoch": 0.71,
127
+ "learning_rate": 9.940000000000001e-06,
128
+ "loss": 0.1927,
129
+ "step": 500
130
+ },
131
+ {
132
+ "epoch": 0.74,
133
+ "learning_rate": 9.960000000000001e-06,
134
+ "loss": 0.1793,
135
+ "step": 525
136
+ },
137
+ {
138
+ "epoch": 0.78,
139
+ "learning_rate": 9.914545454545455e-06,
140
+ "loss": 0.1943,
141
+ "step": 550
142
+ },
143
+ {
144
+ "epoch": 0.82,
145
+ "learning_rate": 9.86909090909091e-06,
146
+ "loss": 0.2225,
147
+ "step": 575
148
+ },
149
+ {
150
+ "epoch": 0.85,
151
+ "learning_rate": 9.823636363636364e-06,
152
+ "loss": 0.2155,
153
+ "step": 600
154
+ },
155
+ {
156
+ "epoch": 0.89,
157
+ "learning_rate": 9.77818181818182e-06,
158
+ "loss": 0.1918,
159
+ "step": 625
160
+ },
161
+ {
162
+ "epoch": 0.92,
163
+ "learning_rate": 9.732727272727273e-06,
164
+ "loss": 0.2165,
165
+ "step": 650
166
+ },
167
+ {
168
+ "epoch": 0.96,
169
+ "learning_rate": 9.687272727272729e-06,
170
+ "loss": 0.1858,
171
+ "step": 675
172
+ },
173
+ {
174
+ "epoch": 0.99,
175
+ "learning_rate": 9.641818181818183e-06,
176
+ "loss": 0.2113,
177
+ "step": 700
178
+ },
179
+ {
180
+ "epoch": 1.03,
181
+ "learning_rate": 9.596363636363636e-06,
182
+ "loss": 0.1174,
183
+ "step": 725
184
+ },
185
+ {
186
+ "epoch": 1.06,
187
+ "learning_rate": 9.550909090909092e-06,
188
+ "loss": 0.0893,
189
+ "step": 750
190
+ },
191
+ {
192
+ "epoch": 1.1,
193
+ "learning_rate": 9.505454545454547e-06,
194
+ "loss": 0.0886,
195
+ "step": 775
196
+ },
197
+ {
198
+ "epoch": 1.13,
199
+ "learning_rate": 9.460000000000001e-06,
200
+ "loss": 0.0984,
201
+ "step": 800
202
+ },
203
+ {
204
+ "epoch": 1.17,
205
+ "learning_rate": 9.414545454545455e-06,
206
+ "loss": 0.0979,
207
+ "step": 825
208
+ },
209
+ {
210
+ "epoch": 1.21,
211
+ "learning_rate": 9.36909090909091e-06,
212
+ "loss": 0.0912,
213
+ "step": 850
214
+ },
215
+ {
216
+ "epoch": 1.24,
217
+ "learning_rate": 9.323636363636366e-06,
218
+ "loss": 0.0901,
219
+ "step": 875
220
+ },
221
+ {
222
+ "epoch": 1.28,
223
+ "learning_rate": 9.278181818181818e-06,
224
+ "loss": 0.0973,
225
+ "step": 900
226
+ },
227
+ {
228
+ "epoch": 1.31,
229
+ "learning_rate": 9.232727272727273e-06,
230
+ "loss": 0.0915,
231
+ "step": 925
232
+ },
233
+ {
234
+ "epoch": 1.35,
235
+ "learning_rate": 9.187272727272728e-06,
236
+ "loss": 0.1084,
237
+ "step": 950
238
+ },
239
+ {
240
+ "epoch": 1.38,
241
+ "learning_rate": 9.141818181818182e-06,
242
+ "loss": 0.116,
243
+ "step": 975
244
+ },
245
+ {
246
+ "epoch": 1.42,
247
+ "learning_rate": 9.096363636363638e-06,
248
+ "loss": 0.0895,
249
+ "step": 1000
250
+ },
251
+ {
252
+ "epoch": 1.45,
253
+ "learning_rate": 9.050909090909091e-06,
254
+ "loss": 0.0877,
255
+ "step": 1025
256
+ },
257
+ {
258
+ "epoch": 1.49,
259
+ "learning_rate": 9.005454545454547e-06,
260
+ "loss": 0.0968,
261
+ "step": 1050
262
+ },
263
+ {
264
+ "epoch": 1.52,
265
+ "learning_rate": 8.96e-06,
266
+ "loss": 0.0855,
267
+ "step": 1075
268
+ },
269
+ {
270
+ "epoch": 1.56,
271
+ "learning_rate": 8.914545454545456e-06,
272
+ "loss": 0.0797,
273
+ "step": 1100
274
+ },
275
+ {
276
+ "epoch": 1.6,
277
+ "learning_rate": 8.86909090909091e-06,
278
+ "loss": 0.0899,
279
+ "step": 1125
280
+ },
281
+ {
282
+ "epoch": 1.63,
283
+ "learning_rate": 8.823636363636364e-06,
284
+ "loss": 0.0971,
285
+ "step": 1150
286
+ },
287
+ {
288
+ "epoch": 1.67,
289
+ "learning_rate": 8.778181818181819e-06,
290
+ "loss": 0.1113,
291
+ "step": 1175
292
+ },
293
+ {
294
+ "epoch": 1.7,
295
+ "learning_rate": 8.732727272727274e-06,
296
+ "loss": 0.0933,
297
+ "step": 1200
298
+ },
299
+ {
300
+ "epoch": 1.74,
301
+ "learning_rate": 8.687272727272728e-06,
302
+ "loss": 0.0958,
303
+ "step": 1225
304
+ },
305
+ {
306
+ "epoch": 1.77,
307
+ "learning_rate": 8.641818181818182e-06,
308
+ "loss": 0.1102,
309
+ "step": 1250
310
+ },
311
+ {
312
+ "epoch": 1.81,
313
+ "learning_rate": 8.596363636363637e-06,
314
+ "loss": 0.0917,
315
+ "step": 1275
316
+ },
317
+ {
318
+ "epoch": 1.84,
319
+ "learning_rate": 8.550909090909091e-06,
320
+ "loss": 0.098,
321
+ "step": 1300
322
+ },
323
+ {
324
+ "epoch": 1.88,
325
+ "learning_rate": 8.505454545454545e-06,
326
+ "loss": 0.0854,
327
+ "step": 1325
328
+ },
329
+ {
330
+ "epoch": 1.91,
331
+ "learning_rate": 8.46e-06,
332
+ "loss": 0.1084,
333
+ "step": 1350
334
+ },
335
+ {
336
+ "epoch": 1.95,
337
+ "learning_rate": 8.414545454545456e-06,
338
+ "loss": 0.0937,
339
+ "step": 1375
340
+ },
341
+ {
342
+ "epoch": 1.99,
343
+ "learning_rate": 8.36909090909091e-06,
344
+ "loss": 0.0851,
345
+ "step": 1400
346
+ },
347
+ {
348
+ "epoch": 2.02,
349
+ "learning_rate": 8.323636363636365e-06,
350
+ "loss": 0.0584,
351
+ "step": 1425
352
+ },
353
+ {
354
+ "epoch": 2.06,
355
+ "learning_rate": 8.278181818181819e-06,
356
+ "loss": 0.034,
357
+ "step": 1450
358
+ },
359
+ {
360
+ "epoch": 2.09,
361
+ "learning_rate": 8.232727272727274e-06,
362
+ "loss": 0.0318,
363
+ "step": 1475
364
+ },
365
+ {
366
+ "epoch": 2.13,
367
+ "learning_rate": 8.187272727272728e-06,
368
+ "loss": 0.0383,
369
+ "step": 1500
370
+ },
371
+ {
372
+ "epoch": 2.16,
373
+ "learning_rate": 8.141818181818183e-06,
374
+ "loss": 0.0314,
375
+ "step": 1525
376
+ },
377
+ {
378
+ "epoch": 2.2,
379
+ "learning_rate": 8.096363636363637e-06,
380
+ "loss": 0.0291,
381
+ "step": 1550
382
+ },
383
+ {
384
+ "epoch": 2.23,
385
+ "learning_rate": 8.05090909090909e-06,
386
+ "loss": 0.0344,
387
+ "step": 1575
388
+ },
389
+ {
390
+ "epoch": 2.27,
391
+ "learning_rate": 8.005454545454546e-06,
392
+ "loss": 0.0363,
393
+ "step": 1600
394
+ },
395
+ {
396
+ "epoch": 2.3,
397
+ "learning_rate": 7.960000000000002e-06,
398
+ "loss": 0.0335,
399
+ "step": 1625
400
+ },
401
+ {
402
+ "epoch": 2.34,
403
+ "learning_rate": 7.914545454545455e-06,
404
+ "loss": 0.0351,
405
+ "step": 1650
406
+ },
407
+ {
408
+ "epoch": 2.38,
409
+ "learning_rate": 7.86909090909091e-06,
410
+ "loss": 0.027,
411
+ "step": 1675
412
+ },
413
+ {
414
+ "epoch": 2.41,
415
+ "learning_rate": 7.823636363636365e-06,
416
+ "loss": 0.0251,
417
+ "step": 1700
418
+ },
419
+ {
420
+ "epoch": 2.45,
421
+ "learning_rate": 7.778181818181818e-06,
422
+ "loss": 0.0295,
423
+ "step": 1725
424
+ },
425
+ {
426
+ "epoch": 2.48,
427
+ "learning_rate": 7.732727272727272e-06,
428
+ "loss": 0.037,
429
+ "step": 1750
430
+ },
431
+ {
432
+ "epoch": 2.52,
433
+ "learning_rate": 7.687272727272728e-06,
434
+ "loss": 0.0329,
435
+ "step": 1775
436
+ },
437
+ {
438
+ "epoch": 2.55,
439
+ "learning_rate": 7.641818181818183e-06,
440
+ "loss": 0.0271,
441
+ "step": 1800
442
+ },
443
+ {
444
+ "epoch": 2.59,
445
+ "learning_rate": 7.596363636363638e-06,
446
+ "loss": 0.0331,
447
+ "step": 1825
448
+ },
449
+ {
450
+ "epoch": 2.62,
451
+ "learning_rate": 7.550909090909092e-06,
452
+ "loss": 0.0308,
453
+ "step": 1850
454
+ },
455
+ {
456
+ "epoch": 2.66,
457
+ "learning_rate": 7.505454545454546e-06,
458
+ "loss": 0.034,
459
+ "step": 1875
460
+ },
461
+ {
462
+ "epoch": 2.7,
463
+ "learning_rate": 7.4600000000000006e-06,
464
+ "loss": 0.0313,
465
+ "step": 1900
466
+ },
467
+ {
468
+ "epoch": 2.73,
469
+ "learning_rate": 7.414545454545455e-06,
470
+ "loss": 0.0355,
471
+ "step": 1925
472
+ },
473
+ {
474
+ "epoch": 2.77,
475
+ "learning_rate": 7.36909090909091e-06,
476
+ "loss": 0.0363,
477
+ "step": 1950
478
+ },
479
+ {
480
+ "epoch": 2.8,
481
+ "learning_rate": 7.3236363636363635e-06,
482
+ "loss": 0.0386,
483
+ "step": 1975
484
+ },
485
+ {
486
+ "epoch": 2.84,
487
+ "learning_rate": 7.278181818181819e-06,
488
+ "loss": 0.0378,
489
+ "step": 2000
490
+ },
491
+ {
492
+ "epoch": 2.84,
493
+ "eval_loss": 0.18704022467136383,
494
+ "eval_runtime": 812.0155,
495
+ "eval_samples_per_second": 5.799,
496
+ "eval_steps_per_second": 2.9,
497
+ "eval_wer": 195.09449989382034,
498
+ "step": 2000
499
+ },
500
+ {
501
+ "epoch": 2.87,
502
+ "learning_rate": 7.2327272727272736e-06,
503
+ "loss": 0.0357,
504
+ "step": 2025
505
+ },
506
+ {
507
+ "epoch": 2.91,
508
+ "learning_rate": 7.187272727272728e-06,
509
+ "loss": 0.0344,
510
+ "step": 2050
511
+ },
512
+ {
513
+ "epoch": 2.94,
514
+ "learning_rate": 7.141818181818182e-06,
515
+ "loss": 0.0288,
516
+ "step": 2075
517
+ },
518
+ {
519
+ "epoch": 2.98,
520
+ "learning_rate": 7.0963636363636365e-06,
521
+ "loss": 0.0274,
522
+ "step": 2100
523
+ },
524
+ {
525
+ "epoch": 3.01,
526
+ "learning_rate": 7.050909090909092e-06,
527
+ "loss": 0.0191,
528
+ "step": 2125
529
+ },
530
+ {
531
+ "epoch": 3.05,
532
+ "learning_rate": 7.0054545454545465e-06,
533
+ "loss": 0.0134,
534
+ "step": 2150
535
+ },
536
+ {
537
+ "epoch": 3.09,
538
+ "learning_rate": 6.96e-06,
539
+ "loss": 0.0161,
540
+ "step": 2175
541
+ },
542
+ {
543
+ "epoch": 3.12,
544
+ "learning_rate": 6.914545454545455e-06,
545
+ "loss": 0.0133,
546
+ "step": 2200
547
+ },
548
+ {
549
+ "epoch": 3.16,
550
+ "learning_rate": 6.8690909090909095e-06,
551
+ "loss": 0.0138,
552
+ "step": 2225
553
+ },
554
+ {
555
+ "epoch": 3.19,
556
+ "learning_rate": 6.823636363636365e-06,
557
+ "loss": 0.0148,
558
+ "step": 2250
559
+ },
560
+ {
561
+ "epoch": 3.23,
562
+ "learning_rate": 6.778181818181818e-06,
563
+ "loss": 0.011,
564
+ "step": 2275
565
+ },
566
+ {
567
+ "epoch": 3.26,
568
+ "learning_rate": 6.732727272727273e-06,
569
+ "loss": 0.0166,
570
+ "step": 2300
571
+ },
572
+ {
573
+ "epoch": 3.3,
574
+ "learning_rate": 6.687272727272728e-06,
575
+ "loss": 0.0138,
576
+ "step": 2325
577
+ },
578
+ {
579
+ "epoch": 3.33,
580
+ "learning_rate": 6.6418181818181824e-06,
581
+ "loss": 0.0116,
582
+ "step": 2350
583
+ },
584
+ {
585
+ "epoch": 3.37,
586
+ "learning_rate": 6.596363636363637e-06,
587
+ "loss": 0.0146,
588
+ "step": 2375
589
+ },
590
+ {
591
+ "epoch": 3.4,
592
+ "learning_rate": 6.550909090909091e-06,
593
+ "loss": 0.0129,
594
+ "step": 2400
595
+ },
596
+ {
597
+ "epoch": 3.44,
598
+ "learning_rate": 6.505454545454546e-06,
599
+ "loss": 0.0126,
600
+ "step": 2425
601
+ },
602
+ {
603
+ "epoch": 3.48,
604
+ "learning_rate": 6.460000000000001e-06,
605
+ "loss": 0.0143,
606
+ "step": 2450
607
+ },
608
+ {
609
+ "epoch": 3.51,
610
+ "learning_rate": 6.4145454545454554e-06,
611
+ "loss": 0.0139,
612
+ "step": 2475
613
+ },
614
+ {
615
+ "epoch": 3.55,
616
+ "learning_rate": 6.369090909090909e-06,
617
+ "loss": 0.0157,
618
+ "step": 2500
619
+ },
620
+ {
621
+ "epoch": 3.58,
622
+ "learning_rate": 6.323636363636364e-06,
623
+ "loss": 0.0152,
624
+ "step": 2525
625
+ },
626
+ {
627
+ "epoch": 3.62,
628
+ "learning_rate": 6.278181818181819e-06,
629
+ "loss": 0.015,
630
+ "step": 2550
631
+ },
632
+ {
633
+ "epoch": 3.65,
634
+ "learning_rate": 6.232727272727274e-06,
635
+ "loss": 0.0134,
636
+ "step": 2575
637
+ },
638
+ {
639
+ "epoch": 3.69,
640
+ "learning_rate": 6.1872727272727276e-06,
641
+ "loss": 0.0135,
642
+ "step": 2600
643
+ },
644
+ {
645
+ "epoch": 3.72,
646
+ "learning_rate": 6.141818181818182e-06,
647
+ "loss": 0.0146,
648
+ "step": 2625
649
+ },
650
+ {
651
+ "epoch": 3.76,
652
+ "learning_rate": 6.096363636363637e-06,
653
+ "loss": 0.0138,
654
+ "step": 2650
655
+ },
656
+ {
657
+ "epoch": 3.79,
658
+ "learning_rate": 6.050909090909091e-06,
659
+ "loss": 0.0142,
660
+ "step": 2675
661
+ },
662
+ {
663
+ "epoch": 3.83,
664
+ "learning_rate": 6.005454545454545e-06,
665
+ "loss": 0.0121,
666
+ "step": 2700
667
+ },
668
+ {
669
+ "epoch": 3.87,
670
+ "learning_rate": 5.9600000000000005e-06,
671
+ "loss": 0.0138,
672
+ "step": 2725
673
+ },
674
+ {
675
+ "epoch": 3.9,
676
+ "learning_rate": 5.914545454545455e-06,
677
+ "loss": 0.0131,
678
+ "step": 2750
679
+ },
680
+ {
681
+ "epoch": 3.94,
682
+ "learning_rate": 5.86909090909091e-06,
683
+ "loss": 0.0148,
684
+ "step": 2775
685
+ },
686
+ {
687
+ "epoch": 3.97,
688
+ "learning_rate": 5.823636363636364e-06,
689
+ "loss": 0.0093,
690
+ "step": 2800
691
+ },
692
+ {
693
+ "epoch": 4.01,
694
+ "learning_rate": 5.778181818181818e-06,
695
+ "loss": 0.0098,
696
+ "step": 2825
697
+ },
698
+ {
699
+ "epoch": 4.04,
700
+ "learning_rate": 5.7327272727272735e-06,
701
+ "loss": 0.0057,
702
+ "step": 2850
703
+ },
704
+ {
705
+ "epoch": 4.08,
706
+ "learning_rate": 5.687272727272728e-06,
707
+ "loss": 0.0072,
708
+ "step": 2875
709
+ },
710
+ {
711
+ "epoch": 4.11,
712
+ "learning_rate": 5.641818181818183e-06,
713
+ "loss": 0.0035,
714
+ "step": 2900
715
+ },
716
+ {
717
+ "epoch": 4.15,
718
+ "learning_rate": 5.5963636363636365e-06,
719
+ "loss": 0.0057,
720
+ "step": 2925
721
+ },
722
+ {
723
+ "epoch": 4.18,
724
+ "learning_rate": 5.550909090909091e-06,
725
+ "loss": 0.0041,
726
+ "step": 2950
727
+ },
728
+ {
729
+ "epoch": 4.22,
730
+ "learning_rate": 5.505454545454546e-06,
731
+ "loss": 0.0038,
732
+ "step": 2975
733
+ },
734
+ {
735
+ "epoch": 4.26,
736
+ "learning_rate": 5.460000000000001e-06,
737
+ "loss": 0.0037,
738
+ "step": 3000
739
+ },
740
+ {
741
+ "epoch": 4.29,
742
+ "learning_rate": 5.414545454545455e-06,
743
+ "loss": 0.0054,
744
+ "step": 3025
745
+ },
746
+ {
747
+ "epoch": 4.33,
748
+ "learning_rate": 5.3690909090909094e-06,
749
+ "loss": 0.0073,
750
+ "step": 3050
751
+ },
752
+ {
753
+ "epoch": 4.36,
754
+ "learning_rate": 5.323636363636364e-06,
755
+ "loss": 0.0048,
756
+ "step": 3075
757
+ },
758
+ {
759
+ "epoch": 4.4,
760
+ "learning_rate": 5.278181818181819e-06,
761
+ "loss": 0.0048,
762
+ "step": 3100
763
+ },
764
+ {
765
+ "epoch": 4.43,
766
+ "learning_rate": 5.232727272727272e-06,
767
+ "loss": 0.004,
768
+ "step": 3125
769
+ },
770
+ {
771
+ "epoch": 4.47,
772
+ "learning_rate": 5.187272727272728e-06,
773
+ "loss": 0.0032,
774
+ "step": 3150
775
+ },
776
+ {
777
+ "epoch": 4.5,
778
+ "learning_rate": 5.141818181818182e-06,
779
+ "loss": 0.0074,
780
+ "step": 3175
781
+ },
782
+ {
783
+ "epoch": 4.54,
784
+ "learning_rate": 5.096363636363637e-06,
785
+ "loss": 0.003,
786
+ "step": 3200
787
+ },
788
+ {
789
+ "epoch": 4.57,
790
+ "learning_rate": 5.050909090909092e-06,
791
+ "loss": 0.0036,
792
+ "step": 3225
793
+ },
794
+ {
795
+ "epoch": 4.61,
796
+ "learning_rate": 5.005454545454545e-06,
797
+ "loss": 0.0072,
798
+ "step": 3250
799
+ },
800
+ {
801
+ "epoch": 4.65,
802
+ "learning_rate": 4.960000000000001e-06,
803
+ "loss": 0.0053,
804
+ "step": 3275
805
+ },
806
+ {
807
+ "epoch": 4.68,
808
+ "learning_rate": 4.914545454545455e-06,
809
+ "loss": 0.0036,
810
+ "step": 3300
811
+ },
812
+ {
813
+ "epoch": 4.72,
814
+ "learning_rate": 4.869090909090909e-06,
815
+ "loss": 0.0049,
816
+ "step": 3325
817
+ },
818
+ {
819
+ "epoch": 4.75,
820
+ "learning_rate": 4.823636363636365e-06,
821
+ "loss": 0.0078,
822
+ "step": 3350
823
+ },
824
+ {
825
+ "epoch": 4.79,
826
+ "learning_rate": 4.778181818181818e-06,
827
+ "loss": 0.0042,
828
+ "step": 3375
829
+ },
830
+ {
831
+ "epoch": 4.82,
832
+ "learning_rate": 4.732727272727273e-06,
833
+ "loss": 0.0038,
834
+ "step": 3400
835
+ },
836
+ {
837
+ "epoch": 4.86,
838
+ "learning_rate": 4.6872727272727275e-06,
839
+ "loss": 0.0045,
840
+ "step": 3425
841
+ },
842
+ {
843
+ "epoch": 4.89,
844
+ "learning_rate": 4.641818181818182e-06,
845
+ "loss": 0.0057,
846
+ "step": 3450
847
+ },
848
+ {
849
+ "epoch": 4.93,
850
+ "learning_rate": 4.596363636363637e-06,
851
+ "loss": 0.0058,
852
+ "step": 3475
853
+ },
854
+ {
855
+ "epoch": 4.96,
856
+ "learning_rate": 4.550909090909091e-06,
857
+ "loss": 0.0036,
858
+ "step": 3500
859
+ },
860
+ {
861
+ "epoch": 5.0,
862
+ "learning_rate": 4.505454545454546e-06,
863
+ "loss": 0.0045,
864
+ "step": 3525
865
+ },
866
+ {
867
+ "epoch": 5.04,
868
+ "learning_rate": 4.4600000000000005e-06,
869
+ "loss": 0.0019,
870
+ "step": 3550
871
+ },
872
+ {
873
+ "epoch": 5.07,
874
+ "learning_rate": 4.414545454545455e-06,
875
+ "loss": 0.0019,
876
+ "step": 3575
877
+ },
878
+ {
879
+ "epoch": 5.11,
880
+ "learning_rate": 4.36909090909091e-06,
881
+ "loss": 0.0027,
882
+ "step": 3600
883
+ },
884
+ {
885
+ "epoch": 5.14,
886
+ "learning_rate": 4.3236363636363634e-06,
887
+ "loss": 0.0022,
888
+ "step": 3625
889
+ },
890
+ {
891
+ "epoch": 5.18,
892
+ "learning_rate": 4.278181818181819e-06,
893
+ "loss": 0.0031,
894
+ "step": 3650
895
+ },
896
+ {
897
+ "epoch": 5.21,
898
+ "learning_rate": 4.2327272727272735e-06,
899
+ "loss": 0.003,
900
+ "step": 3675
901
+ },
902
+ {
903
+ "epoch": 5.25,
904
+ "learning_rate": 4.187272727272727e-06,
905
+ "loss": 0.0017,
906
+ "step": 3700
907
+ },
908
+ {
909
+ "epoch": 5.28,
910
+ "learning_rate": 4.141818181818183e-06,
911
+ "loss": 0.0021,
912
+ "step": 3725
913
+ },
914
+ {
915
+ "epoch": 5.32,
916
+ "learning_rate": 4.096363636363636e-06,
917
+ "loss": 0.0025,
918
+ "step": 3750
919
+ },
920
+ {
921
+ "epoch": 5.35,
922
+ "learning_rate": 4.050909090909092e-06,
923
+ "loss": 0.0012,
924
+ "step": 3775
925
+ },
926
+ {
927
+ "epoch": 5.39,
928
+ "learning_rate": 4.005454545454546e-06,
929
+ "loss": 0.0014,
930
+ "step": 3800
931
+ },
932
+ {
933
+ "epoch": 5.43,
934
+ "learning_rate": 3.96e-06,
935
+ "loss": 0.0014,
936
+ "step": 3825
937
+ },
938
+ {
939
+ "epoch": 5.46,
940
+ "learning_rate": 3.914545454545455e-06,
941
+ "loss": 0.0027,
942
+ "step": 3850
943
+ },
944
+ {
945
+ "epoch": 5.5,
946
+ "learning_rate": 3.869090909090909e-06,
947
+ "loss": 0.0013,
948
+ "step": 3875
949
+ },
950
+ {
951
+ "epoch": 5.53,
952
+ "learning_rate": 3.823636363636364e-06,
953
+ "loss": 0.0026,
954
+ "step": 3900
955
+ },
956
+ {
957
+ "epoch": 5.57,
958
+ "learning_rate": 3.7781818181818186e-06,
959
+ "loss": 0.0015,
960
+ "step": 3925
961
+ },
962
+ {
963
+ "epoch": 5.6,
964
+ "learning_rate": 3.7327272727272728e-06,
965
+ "loss": 0.002,
966
+ "step": 3950
967
+ },
968
+ {
969
+ "epoch": 5.64,
970
+ "learning_rate": 3.6872727272727278e-06,
971
+ "loss": 0.0013,
972
+ "step": 3975
973
+ },
974
+ {
975
+ "epoch": 5.67,
976
+ "learning_rate": 3.641818181818182e-06,
977
+ "loss": 0.0023,
978
+ "step": 4000
979
+ },
980
+ {
981
+ "epoch": 5.67,
982
+ "eval_loss": 0.20714430510997772,
983
+ "eval_runtime": 827.3621,
984
+ "eval_samples_per_second": 5.692,
985
+ "eval_steps_per_second": 2.846,
986
+ "eval_wer": 240.6880441707369,
987
+ "step": 4000
988
+ },
989
+ {
990
+ "epoch": 5.71,
991
+ "learning_rate": 3.5963636363636365e-06,
992
+ "loss": 0.0016,
993
+ "step": 4025
994
+ },
995
+ {
996
+ "epoch": 5.74,
997
+ "learning_rate": 3.550909090909091e-06,
998
+ "loss": 0.0018,
999
+ "step": 4050
1000
+ },
1001
+ {
1002
+ "epoch": 5.78,
1003
+ "learning_rate": 3.5054545454545457e-06,
1004
+ "loss": 0.0019,
1005
+ "step": 4075
1006
+ },
1007
+ {
1008
+ "epoch": 5.82,
1009
+ "learning_rate": 3.46e-06,
1010
+ "loss": 0.0014,
1011
+ "step": 4100
1012
+ },
1013
+ {
1014
+ "epoch": 5.85,
1015
+ "learning_rate": 3.414545454545455e-06,
1016
+ "loss": 0.0022,
1017
+ "step": 4125
1018
+ },
1019
+ {
1020
+ "epoch": 5.89,
1021
+ "learning_rate": 3.3690909090909095e-06,
1022
+ "loss": 0.0017,
1023
+ "step": 4150
1024
+ },
1025
+ {
1026
+ "epoch": 5.92,
1027
+ "learning_rate": 3.3236363636363637e-06,
1028
+ "loss": 0.0011,
1029
+ "step": 4175
1030
+ },
1031
+ {
1032
+ "epoch": 5.96,
1033
+ "learning_rate": 3.2781818181818187e-06,
1034
+ "loss": 0.0011,
1035
+ "step": 4200
1036
+ },
1037
+ {
1038
+ "epoch": 5.99,
1039
+ "learning_rate": 3.232727272727273e-06,
1040
+ "loss": 0.0027,
1041
+ "step": 4225
1042
+ },
1043
+ {
1044
+ "epoch": 6.03,
1045
+ "learning_rate": 3.187272727272728e-06,
1046
+ "loss": 0.0009,
1047
+ "step": 4250
1048
+ },
1049
+ {
1050
+ "epoch": 6.06,
1051
+ "learning_rate": 3.141818181818182e-06,
1052
+ "loss": 0.0009,
1053
+ "step": 4275
1054
+ },
1055
+ {
1056
+ "epoch": 6.1,
1057
+ "learning_rate": 3.0963636363636367e-06,
1058
+ "loss": 0.0008,
1059
+ "step": 4300
1060
+ },
1061
+ {
1062
+ "epoch": 6.13,
1063
+ "learning_rate": 3.050909090909091e-06,
1064
+ "loss": 0.0009,
1065
+ "step": 4325
1066
+ },
1067
+ {
1068
+ "epoch": 6.17,
1069
+ "learning_rate": 3.005454545454546e-06,
1070
+ "loss": 0.0009,
1071
+ "step": 4350
1072
+ },
1073
+ {
1074
+ "epoch": 6.21,
1075
+ "learning_rate": 2.96e-06,
1076
+ "loss": 0.0008,
1077
+ "step": 4375
1078
+ },
1079
+ {
1080
+ "epoch": 6.24,
1081
+ "learning_rate": 2.914545454545455e-06,
1082
+ "loss": 0.0008,
1083
+ "step": 4400
1084
+ },
1085
+ {
1086
+ "epoch": 6.28,
1087
+ "learning_rate": 2.8690909090909092e-06,
1088
+ "loss": 0.0008,
1089
+ "step": 4425
1090
+ },
1091
+ {
1092
+ "epoch": 6.31,
1093
+ "learning_rate": 2.823636363636364e-06,
1094
+ "loss": 0.001,
1095
+ "step": 4450
1096
+ },
1097
+ {
1098
+ "epoch": 6.35,
1099
+ "learning_rate": 2.778181818181818e-06,
1100
+ "loss": 0.0008,
1101
+ "step": 4475
1102
+ },
1103
+ {
1104
+ "epoch": 6.38,
1105
+ "learning_rate": 2.732727272727273e-06,
1106
+ "loss": 0.0008,
1107
+ "step": 4500
1108
+ },
1109
+ {
1110
+ "epoch": 6.42,
1111
+ "learning_rate": 2.687272727272727e-06,
1112
+ "loss": 0.0009,
1113
+ "step": 4525
1114
+ },
1115
+ {
1116
+ "epoch": 6.45,
1117
+ "learning_rate": 2.641818181818182e-06,
1118
+ "loss": 0.0009,
1119
+ "step": 4550
1120
+ },
1121
+ {
1122
+ "epoch": 6.49,
1123
+ "learning_rate": 2.5963636363636364e-06,
1124
+ "loss": 0.0012,
1125
+ "step": 4575
1126
+ },
1127
+ {
1128
+ "epoch": 6.52,
1129
+ "learning_rate": 2.550909090909091e-06,
1130
+ "loss": 0.0009,
1131
+ "step": 4600
1132
+ },
1133
+ {
1134
+ "epoch": 6.56,
1135
+ "learning_rate": 2.505454545454546e-06,
1136
+ "loss": 0.0014,
1137
+ "step": 4625
1138
+ },
1139
+ {
1140
+ "epoch": 6.6,
1141
+ "learning_rate": 2.46e-06,
1142
+ "loss": 0.0008,
1143
+ "step": 4650
1144
+ },
1145
+ {
1146
+ "epoch": 6.63,
1147
+ "learning_rate": 2.4145454545454548e-06,
1148
+ "loss": 0.0008,
1149
+ "step": 4675
1150
+ },
1151
+ {
1152
+ "epoch": 6.67,
1153
+ "learning_rate": 2.3690909090909094e-06,
1154
+ "loss": 0.001,
1155
+ "step": 4700
1156
+ },
1157
+ {
1158
+ "epoch": 6.7,
1159
+ "learning_rate": 2.323636363636364e-06,
1160
+ "loss": 0.0008,
1161
+ "step": 4725
1162
+ },
1163
+ {
1164
+ "epoch": 6.74,
1165
+ "learning_rate": 2.278181818181818e-06,
1166
+ "loss": 0.0015,
1167
+ "step": 4750
1168
+ },
1169
+ {
1170
+ "epoch": 6.77,
1171
+ "learning_rate": 2.2327272727272727e-06,
1172
+ "loss": 0.0008,
1173
+ "step": 4775
1174
+ },
1175
+ {
1176
+ "epoch": 6.81,
1177
+ "learning_rate": 2.1872727272727273e-06,
1178
+ "loss": 0.0009,
1179
+ "step": 4800
1180
+ },
1181
+ {
1182
+ "epoch": 6.84,
1183
+ "learning_rate": 2.1418181818181823e-06,
1184
+ "loss": 0.0008,
1185
+ "step": 4825
1186
+ },
1187
+ {
1188
+ "epoch": 6.88,
1189
+ "learning_rate": 2.0963636363636365e-06,
1190
+ "loss": 0.0008,
1191
+ "step": 4850
1192
+ },
1193
+ {
1194
+ "epoch": 6.91,
1195
+ "learning_rate": 2.050909090909091e-06,
1196
+ "loss": 0.0008,
1197
+ "step": 4875
1198
+ },
1199
+ {
1200
+ "epoch": 6.95,
1201
+ "learning_rate": 2.0054545454545457e-06,
1202
+ "loss": 0.0009,
1203
+ "step": 4900
1204
+ },
1205
+ {
1206
+ "epoch": 6.99,
1207
+ "learning_rate": 1.9600000000000003e-06,
1208
+ "loss": 0.0008,
1209
+ "step": 4925
1210
+ },
1211
+ {
1212
+ "epoch": 7.02,
1213
+ "learning_rate": 1.914545454545455e-06,
1214
+ "loss": 0.001,
1215
+ "step": 4950
1216
+ },
1217
+ {
1218
+ "epoch": 7.06,
1219
+ "learning_rate": 1.8690909090909093e-06,
1220
+ "loss": 0.0007,
1221
+ "step": 4975
1222
+ },
1223
+ {
1224
+ "epoch": 7.09,
1225
+ "learning_rate": 1.8236363636363639e-06,
1226
+ "loss": 0.0007,
1227
+ "step": 5000
1228
+ },
1229
+ {
1230
+ "epoch": 7.13,
1231
+ "learning_rate": 1.7781818181818183e-06,
1232
+ "loss": 0.0006,
1233
+ "step": 5025
1234
+ },
1235
+ {
1236
+ "epoch": 7.16,
1237
+ "learning_rate": 1.7327272727272728e-06,
1238
+ "loss": 0.0006,
1239
+ "step": 5050
1240
+ },
1241
+ {
1242
+ "epoch": 7.2,
1243
+ "learning_rate": 1.6872727272727274e-06,
1244
+ "loss": 0.0006,
1245
+ "step": 5075
1246
+ },
1247
+ {
1248
+ "epoch": 7.23,
1249
+ "learning_rate": 1.6418181818181818e-06,
1250
+ "loss": 0.0011,
1251
+ "step": 5100
1252
+ },
1253
+ {
1254
+ "epoch": 7.27,
1255
+ "learning_rate": 1.5963636363636364e-06,
1256
+ "loss": 0.0007,
1257
+ "step": 5125
1258
+ },
1259
+ {
1260
+ "epoch": 7.3,
1261
+ "learning_rate": 1.550909090909091e-06,
1262
+ "loss": 0.0008,
1263
+ "step": 5150
1264
+ },
1265
+ {
1266
+ "epoch": 7.34,
1267
+ "learning_rate": 1.5054545454545454e-06,
1268
+ "loss": 0.0006,
1269
+ "step": 5175
1270
+ },
1271
+ {
1272
+ "epoch": 7.38,
1273
+ "learning_rate": 1.46e-06,
1274
+ "loss": 0.0007,
1275
+ "step": 5200
1276
+ },
1277
+ {
1278
+ "epoch": 7.41,
1279
+ "learning_rate": 1.4145454545454546e-06,
1280
+ "loss": 0.0006,
1281
+ "step": 5225
1282
+ },
1283
+ {
1284
+ "epoch": 7.45,
1285
+ "learning_rate": 1.369090909090909e-06,
1286
+ "loss": 0.0006,
1287
+ "step": 5250
1288
+ },
1289
+ {
1290
+ "epoch": 7.48,
1291
+ "learning_rate": 1.3236363636363636e-06,
1292
+ "loss": 0.0006,
1293
+ "step": 5275
1294
+ },
1295
+ {
1296
+ "epoch": 7.52,
1297
+ "learning_rate": 1.2781818181818184e-06,
1298
+ "loss": 0.0007,
1299
+ "step": 5300
1300
+ },
1301
+ {
1302
+ "epoch": 7.55,
1303
+ "learning_rate": 1.2327272727272728e-06,
1304
+ "loss": 0.0006,
1305
+ "step": 5325
1306
+ },
1307
+ {
1308
+ "epoch": 7.59,
1309
+ "learning_rate": 1.1872727272727274e-06,
1310
+ "loss": 0.0006,
1311
+ "step": 5350
1312
+ },
1313
+ {
1314
+ "epoch": 7.62,
1315
+ "learning_rate": 1.141818181818182e-06,
1316
+ "loss": 0.0006,
1317
+ "step": 5375
1318
+ },
1319
+ {
1320
+ "epoch": 7.66,
1321
+ "learning_rate": 1.0963636363636366e-06,
1322
+ "loss": 0.0006,
1323
+ "step": 5400
1324
+ },
1325
+ {
1326
+ "epoch": 7.7,
1327
+ "learning_rate": 1.0509090909090911e-06,
1328
+ "loss": 0.0006,
1329
+ "step": 5425
1330
+ },
1331
+ {
1332
+ "epoch": 7.73,
1333
+ "learning_rate": 1.0054545454545455e-06,
1334
+ "loss": 0.0007,
1335
+ "step": 5450
1336
+ },
1337
+ {
1338
+ "epoch": 7.77,
1339
+ "learning_rate": 9.600000000000001e-07,
1340
+ "loss": 0.0006,
1341
+ "step": 5475
1342
+ },
1343
+ {
1344
+ "epoch": 7.8,
1345
+ "learning_rate": 9.145454545454546e-07,
1346
+ "loss": 0.0006,
1347
+ "step": 5500
1348
+ },
1349
+ {
1350
+ "epoch": 7.84,
1351
+ "learning_rate": 8.690909090909091e-07,
1352
+ "loss": 0.0006,
1353
+ "step": 5525
1354
+ },
1355
+ {
1356
+ "epoch": 7.87,
1357
+ "learning_rate": 8.236363636363637e-07,
1358
+ "loss": 0.0006,
1359
+ "step": 5550
1360
+ },
1361
+ {
1362
+ "epoch": 7.91,
1363
+ "learning_rate": 7.781818181818182e-07,
1364
+ "loss": 0.0006,
1365
+ "step": 5575
1366
+ },
1367
+ {
1368
+ "epoch": 7.94,
1369
+ "learning_rate": 7.327272727272727e-07,
1370
+ "loss": 0.0006,
1371
+ "step": 5600
1372
+ },
1373
+ {
1374
+ "epoch": 7.98,
1375
+ "learning_rate": 6.872727272727274e-07,
1376
+ "loss": 0.0006,
1377
+ "step": 5625
1378
+ },
1379
+ {
1380
+ "epoch": 8.01,
1381
+ "learning_rate": 6.418181818181819e-07,
1382
+ "loss": 0.0006,
1383
+ "step": 5650
1384
+ },
1385
+ {
1386
+ "epoch": 8.05,
1387
+ "learning_rate": 5.963636363636364e-07,
1388
+ "loss": 0.0006,
1389
+ "step": 5675
1390
+ },
1391
+ {
1392
+ "epoch": 8.09,
1393
+ "learning_rate": 5.50909090909091e-07,
1394
+ "loss": 0.0005,
1395
+ "step": 5700
1396
+ },
1397
+ {
1398
+ "epoch": 8.12,
1399
+ "learning_rate": 5.054545454545456e-07,
1400
+ "loss": 0.0005,
1401
+ "step": 5725
1402
+ },
1403
+ {
1404
+ "epoch": 8.16,
1405
+ "learning_rate": 4.6000000000000004e-07,
1406
+ "loss": 0.0006,
1407
+ "step": 5750
1408
+ },
1409
+ {
1410
+ "epoch": 8.19,
1411
+ "learning_rate": 4.1454545454545453e-07,
1412
+ "loss": 0.0006,
1413
+ "step": 5775
1414
+ },
1415
+ {
1416
+ "epoch": 8.23,
1417
+ "learning_rate": 3.6909090909090913e-07,
1418
+ "loss": 0.0005,
1419
+ "step": 5800
1420
+ },
1421
+ {
1422
+ "epoch": 8.26,
1423
+ "learning_rate": 3.2363636363636367e-07,
1424
+ "loss": 0.0006,
1425
+ "step": 5825
1426
+ },
1427
+ {
1428
+ "epoch": 8.3,
1429
+ "learning_rate": 2.781818181818182e-07,
1430
+ "loss": 0.0006,
1431
+ "step": 5850
1432
+ },
1433
+ {
1434
+ "epoch": 8.33,
1435
+ "learning_rate": 2.3272727272727273e-07,
1436
+ "loss": 0.0005,
1437
+ "step": 5875
1438
+ },
1439
+ {
1440
+ "epoch": 8.37,
1441
+ "learning_rate": 1.872727272727273e-07,
1442
+ "loss": 0.0006,
1443
+ "step": 5900
1444
+ },
1445
+ {
1446
+ "epoch": 8.4,
1447
+ "learning_rate": 1.4181818181818184e-07,
1448
+ "loss": 0.0006,
1449
+ "step": 5925
1450
+ },
1451
+ {
1452
+ "epoch": 8.44,
1453
+ "learning_rate": 9.636363636363636e-08,
1454
+ "loss": 0.0006,
1455
+ "step": 5950
1456
+ },
1457
+ {
1458
+ "epoch": 8.48,
1459
+ "learning_rate": 5.0909090909090914e-08,
1460
+ "loss": 0.0005,
1461
+ "step": 5975
1462
+ },
1463
+ {
1464
+ "epoch": 8.51,
1465
+ "learning_rate": 5.454545454545455e-09,
1466
+ "loss": 0.0006,
1467
+ "step": 6000
1468
+ },
1469
+ {
1470
+ "epoch": 8.51,
1471
+ "eval_loss": 0.2148810774087906,
1472
+ "eval_runtime": 818.6792,
1473
+ "eval_samples_per_second": 5.752,
1474
+ "eval_steps_per_second": 2.877,
1475
+ "eval_wer": 160.24633680186878,
1476
+ "step": 6000
1477
+ },
1478
+ {
1479
+ "epoch": 8.55,
1480
+ "learning_rate": 2.535897435897436e-06,
1481
+ "loss": 0.0006,
1482
+ "step": 6025
1483
+ },
1484
+ {
1485
+ "epoch": 8.58,
1486
+ "learning_rate": 2.503846153846154e-06,
1487
+ "loss": 0.0006,
1488
+ "step": 6050
1489
+ },
1490
+ {
1491
+ "epoch": 8.62,
1492
+ "learning_rate": 2.471794871794872e-06,
1493
+ "loss": 0.0005,
1494
+ "step": 6075
1495
+ },
1496
+ {
1497
+ "epoch": 8.65,
1498
+ "learning_rate": 2.43974358974359e-06,
1499
+ "loss": 0.0006,
1500
+ "step": 6100
1501
+ },
1502
+ {
1503
+ "epoch": 8.69,
1504
+ "learning_rate": 2.407692307692308e-06,
1505
+ "loss": 0.0006,
1506
+ "step": 6125
1507
+ },
1508
+ {
1509
+ "epoch": 8.72,
1510
+ "learning_rate": 2.375641025641026e-06,
1511
+ "loss": 0.0006,
1512
+ "step": 6150
1513
+ },
1514
+ {
1515
+ "epoch": 8.76,
1516
+ "learning_rate": 2.3435897435897438e-06,
1517
+ "loss": 0.0006,
1518
+ "step": 6175
1519
+ },
1520
+ {
1521
+ "epoch": 8.79,
1522
+ "learning_rate": 2.3115384615384617e-06,
1523
+ "loss": 0.0006,
1524
+ "step": 6200
1525
+ },
1526
+ {
1527
+ "epoch": 8.83,
1528
+ "learning_rate": 2.2794871794871796e-06,
1529
+ "loss": 0.0006,
1530
+ "step": 6225
1531
+ },
1532
+ {
1533
+ "epoch": 8.87,
1534
+ "learning_rate": 2.2474358974358975e-06,
1535
+ "loss": 0.0005,
1536
+ "step": 6250
1537
+ },
1538
+ {
1539
+ "epoch": 8.9,
1540
+ "learning_rate": 2.215384615384616e-06,
1541
+ "loss": 0.0005,
1542
+ "step": 6275
1543
+ },
1544
+ {
1545
+ "epoch": 8.94,
1546
+ "learning_rate": 2.1833333333333333e-06,
1547
+ "loss": 0.0005,
1548
+ "step": 6300
1549
+ },
1550
+ {
1551
+ "epoch": 8.97,
1552
+ "learning_rate": 2.1512820512820517e-06,
1553
+ "loss": 0.0005,
1554
+ "step": 6325
1555
+ },
1556
+ {
1557
+ "epoch": 9.01,
1558
+ "learning_rate": 2.1192307692307696e-06,
1559
+ "loss": 0.0005,
1560
+ "step": 6350
1561
+ },
1562
+ {
1563
+ "epoch": 9.04,
1564
+ "learning_rate": 2.0871794871794875e-06,
1565
+ "loss": 0.0005,
1566
+ "step": 6375
1567
+ },
1568
+ {
1569
+ "epoch": 9.08,
1570
+ "learning_rate": 2.0551282051282054e-06,
1571
+ "loss": 0.0005,
1572
+ "step": 6400
1573
+ },
1574
+ {
1575
+ "epoch": 9.11,
1576
+ "learning_rate": 2.0230769230769233e-06,
1577
+ "loss": 0.0005,
1578
+ "step": 6425
1579
+ },
1580
+ {
1581
+ "epoch": 9.15,
1582
+ "learning_rate": 1.991025641025641e-06,
1583
+ "loss": 0.0005,
1584
+ "step": 6450
1585
+ },
1586
+ {
1587
+ "epoch": 9.18,
1588
+ "learning_rate": 1.958974358974359e-06,
1589
+ "loss": 0.0005,
1590
+ "step": 6475
1591
+ },
1592
+ {
1593
+ "epoch": 9.22,
1594
+ "learning_rate": 1.926923076923077e-06,
1595
+ "loss": 0.0005,
1596
+ "step": 6500
1597
+ },
1598
+ {
1599
+ "epoch": 9.26,
1600
+ "learning_rate": 1.8948717948717951e-06,
1601
+ "loss": 0.0005,
1602
+ "step": 6525
1603
+ },
1604
+ {
1605
+ "epoch": 9.29,
1606
+ "learning_rate": 1.8628205128205128e-06,
1607
+ "loss": 0.0005,
1608
+ "step": 6550
1609
+ },
1610
+ {
1611
+ "epoch": 9.33,
1612
+ "learning_rate": 1.830769230769231e-06,
1613
+ "loss": 0.0005,
1614
+ "step": 6575
1615
+ },
1616
+ {
1617
+ "epoch": 9.36,
1618
+ "learning_rate": 1.7987179487179486e-06,
1619
+ "loss": 0.0005,
1620
+ "step": 6600
1621
+ },
1622
+ {
1623
+ "epoch": 9.4,
1624
+ "learning_rate": 1.7666666666666668e-06,
1625
+ "loss": 0.0005,
1626
+ "step": 6625
1627
+ },
1628
+ {
1629
+ "epoch": 9.43,
1630
+ "learning_rate": 1.7346153846153849e-06,
1631
+ "loss": 0.0005,
1632
+ "step": 6650
1633
+ },
1634
+ {
1635
+ "epoch": 9.47,
1636
+ "learning_rate": 1.7025641025641026e-06,
1637
+ "loss": 0.0004,
1638
+ "step": 6675
1639
+ },
1640
+ {
1641
+ "epoch": 9.5,
1642
+ "learning_rate": 1.6705128205128207e-06,
1643
+ "loss": 0.0005,
1644
+ "step": 6700
1645
+ },
1646
+ {
1647
+ "epoch": 9.54,
1648
+ "learning_rate": 1.6384615384615384e-06,
1649
+ "loss": 0.0005,
1650
+ "step": 6725
1651
+ },
1652
+ {
1653
+ "epoch": 9.57,
1654
+ "learning_rate": 1.6064102564102565e-06,
1655
+ "loss": 0.0005,
1656
+ "step": 6750
1657
+ },
1658
+ {
1659
+ "epoch": 9.61,
1660
+ "learning_rate": 1.5743589743589746e-06,
1661
+ "loss": 0.0005,
1662
+ "step": 6775
1663
+ },
1664
+ {
1665
+ "epoch": 9.65,
1666
+ "learning_rate": 1.5423076923076923e-06,
1667
+ "loss": 0.0005,
1668
+ "step": 6800
1669
+ },
1670
+ {
1671
+ "epoch": 9.68,
1672
+ "learning_rate": 1.5102564102564105e-06,
1673
+ "loss": 0.0005,
1674
+ "step": 6825
1675
+ },
1676
+ {
1677
+ "epoch": 9.72,
1678
+ "learning_rate": 1.4782051282051281e-06,
1679
+ "loss": 0.0005,
1680
+ "step": 6850
1681
+ },
1682
+ {
1683
+ "epoch": 9.75,
1684
+ "learning_rate": 1.4461538461538463e-06,
1685
+ "loss": 0.0005,
1686
+ "step": 6875
1687
+ },
1688
+ {
1689
+ "epoch": 9.79,
1690
+ "learning_rate": 1.4141025641025644e-06,
1691
+ "loss": 0.0005,
1692
+ "step": 6900
1693
+ },
1694
+ {
1695
+ "epoch": 9.82,
1696
+ "learning_rate": 1.382051282051282e-06,
1697
+ "loss": 0.0005,
1698
+ "step": 6925
1699
+ },
1700
+ {
1701
+ "epoch": 9.86,
1702
+ "learning_rate": 1.3500000000000002e-06,
1703
+ "loss": 0.0005,
1704
+ "step": 6950
1705
+ },
1706
+ {
1707
+ "epoch": 9.89,
1708
+ "learning_rate": 1.317948717948718e-06,
1709
+ "loss": 0.0004,
1710
+ "step": 6975
1711
+ },
1712
+ {
1713
+ "epoch": 9.93,
1714
+ "learning_rate": 1.285897435897436e-06,
1715
+ "loss": 0.0005,
1716
+ "step": 7000
1717
+ },
1718
+ {
1719
+ "epoch": 9.96,
1720
+ "learning_rate": 1.253846153846154e-06,
1721
+ "loss": 0.0005,
1722
+ "step": 7025
1723
+ },
1724
+ {
1725
+ "epoch": 10.0,
1726
+ "learning_rate": 1.2217948717948718e-06,
1727
+ "loss": 0.0005,
1728
+ "step": 7050
1729
+ },
1730
+ {
1731
+ "epoch": 10.04,
1732
+ "learning_rate": 1.1897435897435897e-06,
1733
+ "loss": 0.0004,
1734
+ "step": 7075
1735
+ },
1736
+ {
1737
+ "epoch": 10.07,
1738
+ "learning_rate": 1.1576923076923077e-06,
1739
+ "loss": 0.0004,
1740
+ "step": 7100
1741
+ },
1742
+ {
1743
+ "epoch": 10.11,
1744
+ "learning_rate": 1.1256410256410258e-06,
1745
+ "loss": 0.0004,
1746
+ "step": 7125
1747
+ },
1748
+ {
1749
+ "epoch": 10.14,
1750
+ "learning_rate": 1.0935897435897437e-06,
1751
+ "loss": 0.0004,
1752
+ "step": 7150
1753
+ },
1754
+ {
1755
+ "epoch": 10.18,
1756
+ "learning_rate": 1.0615384615384616e-06,
1757
+ "loss": 0.0004,
1758
+ "step": 7175
1759
+ },
1760
+ {
1761
+ "epoch": 10.21,
1762
+ "learning_rate": 1.0294871794871795e-06,
1763
+ "loss": 0.0004,
1764
+ "step": 7200
1765
+ },
1766
+ {
1767
+ "epoch": 10.25,
1768
+ "learning_rate": 9.974358974358974e-07,
1769
+ "loss": 0.0004,
1770
+ "step": 7225
1771
+ },
1772
+ {
1773
+ "epoch": 10.28,
1774
+ "learning_rate": 9.653846153846155e-07,
1775
+ "loss": 0.0004,
1776
+ "step": 7250
1777
+ },
1778
+ {
1779
+ "epoch": 10.32,
1780
+ "learning_rate": 9.333333333333334e-07,
1781
+ "loss": 0.0004,
1782
+ "step": 7275
1783
+ },
1784
+ {
1785
+ "epoch": 10.35,
1786
+ "learning_rate": 9.012820512820513e-07,
1787
+ "loss": 0.0004,
1788
+ "step": 7300
1789
+ },
1790
+ {
1791
+ "epoch": 10.39,
1792
+ "learning_rate": 8.692307692307692e-07,
1793
+ "loss": 0.0004,
1794
+ "step": 7325
1795
+ },
1796
+ {
1797
+ "epoch": 10.43,
1798
+ "learning_rate": 8.371794871794873e-07,
1799
+ "loss": 0.0004,
1800
+ "step": 7350
1801
+ },
1802
+ {
1803
+ "epoch": 10.46,
1804
+ "learning_rate": 8.051282051282052e-07,
1805
+ "loss": 0.0004,
1806
+ "step": 7375
1807
+ },
1808
+ {
1809
+ "epoch": 10.5,
1810
+ "learning_rate": 7.730769230769232e-07,
1811
+ "loss": 0.0004,
1812
+ "step": 7400
1813
+ },
1814
+ {
1815
+ "epoch": 10.53,
1816
+ "learning_rate": 7.410256410256411e-07,
1817
+ "loss": 0.0004,
1818
+ "step": 7425
1819
+ },
1820
+ {
1821
+ "epoch": 10.57,
1822
+ "learning_rate": 7.08974358974359e-07,
1823
+ "loss": 0.0004,
1824
+ "step": 7450
1825
+ },
1826
+ {
1827
+ "epoch": 10.6,
1828
+ "learning_rate": 6.769230769230769e-07,
1829
+ "loss": 0.0004,
1830
+ "step": 7475
1831
+ },
1832
+ {
1833
+ "epoch": 10.64,
1834
+ "learning_rate": 6.448717948717949e-07,
1835
+ "loss": 0.0004,
1836
+ "step": 7500
1837
+ },
1838
+ {
1839
+ "epoch": 10.67,
1840
+ "learning_rate": 6.128205128205128e-07,
1841
+ "loss": 0.0004,
1842
+ "step": 7525
1843
+ },
1844
+ {
1845
+ "epoch": 10.71,
1846
+ "learning_rate": 5.807692307692308e-07,
1847
+ "loss": 0.0004,
1848
+ "step": 7550
1849
+ },
1850
+ {
1851
+ "epoch": 10.74,
1852
+ "learning_rate": 5.487179487179488e-07,
1853
+ "loss": 0.0004,
1854
+ "step": 7575
1855
+ },
1856
+ {
1857
+ "epoch": 10.78,
1858
+ "learning_rate": 5.166666666666667e-07,
1859
+ "loss": 0.0004,
1860
+ "step": 7600
1861
+ },
1862
+ {
1863
+ "epoch": 10.82,
1864
+ "learning_rate": 4.846153846153847e-07,
1865
+ "loss": 0.0004,
1866
+ "step": 7625
1867
+ },
1868
+ {
1869
+ "epoch": 10.85,
1870
+ "learning_rate": 4.5256410256410264e-07,
1871
+ "loss": 0.0004,
1872
+ "step": 7650
1873
+ },
1874
+ {
1875
+ "epoch": 10.89,
1876
+ "learning_rate": 4.2051282051282054e-07,
1877
+ "loss": 0.0004,
1878
+ "step": 7675
1879
+ },
1880
+ {
1881
+ "epoch": 10.92,
1882
+ "learning_rate": 3.8846153846153845e-07,
1883
+ "loss": 0.0004,
1884
+ "step": 7700
1885
+ },
1886
+ {
1887
+ "epoch": 10.96,
1888
+ "learning_rate": 3.5641025641025647e-07,
1889
+ "loss": 0.0004,
1890
+ "step": 7725
1891
+ },
1892
+ {
1893
+ "epoch": 10.99,
1894
+ "learning_rate": 3.243589743589744e-07,
1895
+ "loss": 0.0004,
1896
+ "step": 7750
1897
+ },
1898
+ {
1899
+ "epoch": 11.03,
1900
+ "learning_rate": 2.9230769230769234e-07,
1901
+ "loss": 0.0004,
1902
+ "step": 7775
1903
+ },
1904
+ {
1905
+ "epoch": 11.06,
1906
+ "learning_rate": 2.602564102564103e-07,
1907
+ "loss": 0.0004,
1908
+ "step": 7800
1909
+ },
1910
+ {
1911
+ "epoch": 11.1,
1912
+ "learning_rate": 2.2820512820512823e-07,
1913
+ "loss": 0.0003,
1914
+ "step": 7825
1915
+ },
1916
+ {
1917
+ "epoch": 11.13,
1918
+ "learning_rate": 1.9615384615384617e-07,
1919
+ "loss": 0.0004,
1920
+ "step": 7850
1921
+ },
1922
+ {
1923
+ "epoch": 11.17,
1924
+ "learning_rate": 1.6410256410256413e-07,
1925
+ "loss": 0.0004,
1926
+ "step": 7875
1927
+ },
1928
+ {
1929
+ "epoch": 11.21,
1930
+ "learning_rate": 1.3205128205128206e-07,
1931
+ "loss": 0.0003,
1932
+ "step": 7900
1933
+ },
1934
+ {
1935
+ "epoch": 11.24,
1936
+ "learning_rate": 1.0000000000000001e-07,
1937
+ "loss": 0.0004,
1938
+ "step": 7925
1939
+ },
1940
+ {
1941
+ "epoch": 11.28,
1942
+ "learning_rate": 6.794871794871796e-08,
1943
+ "loss": 0.0004,
1944
+ "step": 7950
1945
+ },
1946
+ {
1947
+ "epoch": 11.31,
1948
+ "learning_rate": 3.58974358974359e-08,
1949
+ "loss": 0.0004,
1950
+ "step": 7975
1951
+ },
1952
+ {
1953
+ "epoch": 11.35,
1954
+ "learning_rate": 3.846153846153846e-09,
1955
+ "loss": 0.0003,
1956
+ "step": 8000
1957
+ },
1958
+ {
1959
+ "epoch": 11.35,
1960
+ "eval_loss": 0.2225753217935562,
1961
+ "eval_runtime": 823.0719,
1962
+ "eval_samples_per_second": 5.721,
1963
+ "eval_steps_per_second": 2.861,
1964
+ "eval_wer": 164.51475897218094,
1965
+ "step": 8000
1966
+ }
1967
+ ],
1968
+ "logging_steps": 25,
1969
+ "max_steps": 8000,
1970
+ "num_train_epochs": 12,
1971
+ "save_steps": 2000,
1972
+ "total_flos": 3.692940792274944e+19,
1973
+ "trial_name": null,
1974
+ "trial_params": null
1975
+ }
checkpoint-8000/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea53e5816e3fe2ed9c22847eb9625b55b24b1e5aee0fd886c31f082828bc4d5a
3
+ size 4728
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d9e989154852e3245bd3067e84320c7a1dd3b97a88da32adbc1e2da790d9da0f
3
  size 966995080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:daaaf0bebfd66b090622504ba6169a69368bb067d853fff6af7bc80100a4ab0f
3
  size 966995080
runs/Dec07_12-01-44_smurf/events.out.tfevents.1701943333.smurf.1063018.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73f8a54f408fc5cd764be254e4d96b47eb957c69bfd3779b412542aa680c80c0
3
- size 5328
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c42f9e41c5a2117ded483c37ef5e13ce14c0c2e808a07e3b2c346451cb13e7db
3
+ size 18246
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e29181f470382fce52e35a2769c9ed58dabf92a7ebfe37f86e4a495bbe2dd96a
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea53e5816e3fe2ed9c22847eb9625b55b24b1e5aee0fd886c31f082828bc4d5a
3
  size 4728