Basma2423 commited on
Commit
449d6d1
·
verified ·
1 Parent(s): d30e3a9

Reset to have only the last checkpoint

Browse files
config.json CHANGED
@@ -1,32 +1,58 @@
1
  {
 
 
 
 
2
  "architectures": [
3
- "T5ForConditionalGeneration"
4
  ],
 
 
 
5
  "classifier_dropout": 0.0,
6
- "d_ff": 2048,
7
- "d_kv": 64,
8
  "d_model": 768,
9
- "decoder_start_token_id": 0,
10
- "dense_act_fn": "gelu_new",
11
- "dropout_rate": 0.1,
12
- "eos_token_id": 1,
13
- "feed_forward_proj": "gated-gelu",
14
- "initializer_factor": 1.0,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  "is_encoder_decoder": true,
16
- "is_gated_act": true,
17
- "layer_norm_epsilon": 1e-06,
18
- "model_type": "t5",
19
- "num_decoder_layers": 12,
20
- "num_heads": 12,
21
- "num_layers": 12,
22
- "output_past": true,
23
- "pad_token_id": 0,
24
- "relative_attention_max_distance": 128,
25
- "relative_attention_num_buckets": 32,
26
- "tie_word_embeddings": false,
27
- "tokenizer_class": "T5Tokenizer",
 
 
 
 
28
  "torch_dtype": "float32",
29
- "transformers_version": "4.53.0",
30
- "use_cache": true,
31
- "vocab_size": 110208
32
  }
 
1
  {
2
+ "activation_dropout": 0.1,
3
+ "activation_function": "gelu",
4
+ "add_bias_logits": false,
5
+ "add_final_layer_norm": true,
6
  "architectures": [
7
+ "MBartForConditionalGeneration"
8
  ],
9
+ "attention_dropout": 0.1,
10
+ "bos_token_id": 0,
11
+ "classif_dropout": 0.1,
12
  "classifier_dropout": 0.0,
 
 
13
  "d_model": 768,
14
+ "decoder_attention_heads": 12,
15
+ "decoder_ffn_dim": 3072,
16
+ "decoder_layerdrop": 0.0,
17
+ "decoder_layers": 6,
18
+ "decoder_start_token_id": 2,
19
+ "do_blenderbot_90_layernorm": false,
20
+ "dropout": 0.1,
21
+ "early_stopping": null,
22
+ "encoder_attention_heads": 12,
23
+ "encoder_ffn_dim": 3072,
24
+ "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
+ "eos_token_id": 2,
27
+ "extra_pos_embeddings": 2,
28
+ "force_bos_token_to_be_generated": false,
29
+ "forced_eos_token_id": 2,
30
+ "gradient_checkpointing": false,
31
+ "id2label": {
32
+ "0": "LABEL_0",
33
+ "1": "LABEL_1",
34
+ "2": "LABEL_2"
35
+ },
36
+ "init_std": 0.02,
37
  "is_encoder_decoder": true,
38
+ "label2id": {
39
+ "LABEL_0": 0,
40
+ "LABEL_1": 1,
41
+ "LABEL_2": 2
42
+ },
43
+ "max_position_embeddings": 1024,
44
+ "model_type": "mbart",
45
+ "no_repeat_ngram_size": null,
46
+ "normalize_before": true,
47
+ "normalize_embedding": true,
48
+ "num_beams": null,
49
+ "num_hidden_layers": 6,
50
+ "pad_token_id": 1,
51
+ "scale_embedding": false,
52
+ "static_position_embeddings": false,
53
+ "tokenizer_class": "BarthezTokenizer",
54
  "torch_dtype": "float32",
55
+ "transformers_version": "4.51.3",
56
+ "use_cache": false,
57
+ "vocab_size": 50594
58
  }
generation_config.json CHANGED
@@ -1,6 +1,9 @@
1
  {
2
- "decoder_start_token_id": 0,
3
- "eos_token_id": 1,
4
- "pad_token_id": 0,
5
- "transformers_version": "4.53.0"
 
 
 
6
  }
 
1
  {
2
+ "bos_token_id": 0,
3
+ "decoder_start_token_id": 2,
4
+ "eos_token_id": 2,
5
+ "forced_eos_token_id": 2,
6
+ "pad_token_id": 1,
7
+ "transformers_version": "4.51.3",
8
+ "use_cache": false
9
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:030be25143ccd1d78f909c41318fb2ecc301e9c81c4835b77323d13e3a68f361
3
- size 1470068592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2139aa0df60063c5f2a89b26029f1e46a73d98ca4786df7e9f5e8f8aa2ce3fd8
3
+ size 558937304
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:38b1742741bbc7df44be28b94d10d9ab05ad1cce2d8e70ac745824b916791397
3
- size 2996811
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e0ac63a97ee73869f6f93c12515905f081226b53257dc0915438424c133a795
3
+ size 1823179
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd54aa42eef0dd482320f74af4126bc2997a856ba65d8619476c5b9c42dc0d7e
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b876e459300ff0cd15dd802956a4ebbc05ffef47b7f642fe15ce092c4b74dca
3
  size 14645
scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9380964b4b2206fff8b4f1535e87c70866b61a1d77c927a05f1aaa0e184890d1
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fc8c60b513e06adf874c662773dc09a90c9b37987241c9c4c0214e207bd2ab9
3
  size 1383
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6218a7bb0fe5f214de8ffed80e034797b70aa538fcf9b5e6aedd7d40c0317f31
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85ab216a2a3026e9bbcd8baba7003d57f9145c6e87bcb81b605155e4178e8c69
3
  size 1465
special_tokens_map.json CHANGED
@@ -1,106 +1,18 @@
1
  {
2
- "additional_special_tokens": [
3
- "<extra_id_0>",
4
- "<extra_id_1>",
5
- "<extra_id_2>",
6
- "<extra_id_3>",
7
- "<extra_id_4>",
8
- "<extra_id_5>",
9
- "<extra_id_6>",
10
- "<extra_id_7>",
11
- "<extra_id_8>",
12
- "<extra_id_9>",
13
- "<extra_id_10>",
14
- "<extra_id_11>",
15
- "<extra_id_12>",
16
- "<extra_id_13>",
17
- "<extra_id_14>",
18
- "<extra_id_15>",
19
- "<extra_id_16>",
20
- "<extra_id_17>",
21
- "<extra_id_18>",
22
- "<extra_id_19>",
23
- "<extra_id_20>",
24
- "<extra_id_21>",
25
- "<extra_id_22>",
26
- "<extra_id_23>",
27
- "<extra_id_24>",
28
- "<extra_id_25>",
29
- "<extra_id_26>",
30
- "<extra_id_27>",
31
- "<extra_id_28>",
32
- "<extra_id_29>",
33
- "<extra_id_30>",
34
- "<extra_id_31>",
35
- "<extra_id_32>",
36
- "<extra_id_33>",
37
- "<extra_id_34>",
38
- "<extra_id_35>",
39
- "<extra_id_36>",
40
- "<extra_id_37>",
41
- "<extra_id_38>",
42
- "<extra_id_39>",
43
- "<extra_id_40>",
44
- "<extra_id_41>",
45
- "<extra_id_42>",
46
- "<extra_id_43>",
47
- "<extra_id_44>",
48
- "<extra_id_45>",
49
- "<extra_id_46>",
50
- "<extra_id_47>",
51
- "<extra_id_48>",
52
- "<extra_id_49>",
53
- "<extra_id_50>",
54
- "<extra_id_51>",
55
- "<extra_id_52>",
56
- "<extra_id_53>",
57
- "<extra_id_54>",
58
- "<extra_id_55>",
59
- "<extra_id_56>",
60
- "<extra_id_57>",
61
- "<extra_id_58>",
62
- "<extra_id_59>",
63
- "<extra_id_60>",
64
- "<extra_id_61>",
65
- "<extra_id_62>",
66
- "<extra_id_63>",
67
- "<extra_id_64>",
68
- "<extra_id_65>",
69
- "<extra_id_66>",
70
- "<extra_id_67>",
71
- "<extra_id_68>",
72
- "<extra_id_69>",
73
- "<extra_id_70>",
74
- "<extra_id_71>",
75
- "<extra_id_72>",
76
- "<extra_id_73>",
77
- "<extra_id_74>",
78
- "<extra_id_75>",
79
- "<extra_id_76>",
80
- "<extra_id_77>",
81
- "<extra_id_78>",
82
- "<extra_id_79>",
83
- "<extra_id_80>",
84
- "<extra_id_81>",
85
- "<extra_id_82>",
86
- "<extra_id_83>",
87
- "<extra_id_84>",
88
- "<extra_id_85>",
89
- "<extra_id_86>",
90
- "<extra_id_87>",
91
- "<extra_id_88>",
92
- "<extra_id_89>",
93
- "<extra_id_90>",
94
- "<extra_id_91>",
95
- "<extra_id_92>",
96
- "<extra_id_93>",
97
- "<extra_id_94>",
98
- "<extra_id_95>",
99
- "<extra_id_96>",
100
- "<extra_id_97>",
101
- "<extra_id_98>",
102
- "<extra_id_99>"
103
- ],
104
  "eos_token": {
105
  "content": "</s>",
106
  "lstrip": false,
@@ -108,6 +20,13 @@
108
  "rstrip": false,
109
  "single_word": false
110
  },
 
 
 
 
 
 
 
111
  "pad_token": {
112
  "content": "<pad>",
113
  "lstrip": false,
@@ -115,6 +34,13 @@
115
  "rstrip": false,
116
  "single_word": false
117
  },
 
 
 
 
 
 
 
118
  "unk_token": {
119
  "content": "<unk>",
120
  "lstrip": false,
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
  "eos_token": {
17
  "content": "</s>",
18
  "lstrip": false,
 
20
  "rstrip": false,
21
  "single_word": false
22
  },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
  "pad_token": {
31
  "content": "<pad>",
32
  "lstrip": false,
 
34
  "rstrip": false,
35
  "single_word": false
36
  },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
  "unk_token": {
45
  "content": "<unk>",
46
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34e5c5c453959763bceb14ae9fcaa635cfc2cbdfae4174fa14d7591a52b7fb0a
3
- size 15315418
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3ead0d4351c24a85331ce4bc6928a28880e1f478202f380cd75ac6dc68878f5
3
+ size 3888340
tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
trainer_state.json CHANGED
@@ -1,364 +1,142 @@
1
  {
2
- "best_global_step": 16280,
3
- "best_metric": 0.05104936846864332,
4
- "best_model_checkpoint": "checkpoints/checkpoint-16280",
5
- "epoch": 1.9992631708215645,
6
- "eval_steps": 1628,
7
- "global_step": 16280,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
- "epoch": 0.06140243153628884,
14
- "grad_norm": 0.31788545846939087,
15
- "learning_rate": 4.99487482593161e-05,
16
- "loss": 0.639,
17
  "step": 500
18
  },
19
  {
20
- "epoch": 0.12280486307257768,
21
- "grad_norm": 0.21716584265232086,
22
- "learning_rate": 4.9794381003990576e-05,
23
- "loss": 0.3961,
24
  "step": 1000
25
  },
26
  {
27
- "epoch": 0.1842072946088665,
28
- "grad_norm": 0.32310155034065247,
29
- "learning_rate": 4.9537535426562465e-05,
30
- "loss": 0.3303,
31
- "step": 1500
 
 
32
  },
33
  {
34
- "epoch": 0.19992631708215644,
35
- "eval_avg": 0.07433640047160056,
36
- "eval_cer": 0.048444434926464684,
37
- "eval_der": 0.054747223027151605,
38
- "eval_loss": 0.2203781008720398,
39
- "eval_runtime": 1149.449,
40
- "eval_samples_per_second": 45.035,
41
- "eval_steps_per_second": 0.352,
42
- "eval_wer": 0.1198175434611854,
43
- "step": 1628
44
  },
45
  {
46
- "epoch": 0.24560972614515536,
47
- "grad_norm": 0.1829567700624466,
48
- "learning_rate": 4.997155120906588e-05,
49
- "loss": 0.287,
50
  "step": 2000
51
  },
52
  {
53
- "epoch": 0.3070121576814442,
54
- "grad_norm": 0.1720687448978424,
55
- "learning_rate": 4.984333190353011e-05,
56
- "loss": 0.2682,
 
 
 
 
 
 
 
 
 
57
  "step": 2500
58
  },
59
  {
60
- "epoch": 0.368414589217733,
61
- "grad_norm": 0.19322410225868225,
62
- "learning_rate": 4.961243195561407e-05,
63
- "loss": 0.2506,
64
  "step": 3000
65
  },
66
  {
67
- "epoch": 0.3998526341643129,
68
- "eval_avg": 0.061321180853411116,
69
- "eval_cer": 0.04388846969480041,
70
- "eval_der": 0.03895160822949064,
71
- "eval_loss": 0.18064290285110474,
72
- "eval_runtime": 1155.5329,
73
- "eval_samples_per_second": 44.798,
74
- "eval_steps_per_second": 0.35,
75
- "eval_wer": 0.10112346463594228,
76
- "step": 3256
77
  },
78
  {
79
- "epoch": 0.42981702075402184,
80
- "grad_norm": 0.16168585419654846,
81
- "learning_rate": 4.998779393767303e-05,
82
- "loss": 0.2339,
83
  "step": 3500
84
  },
85
  {
86
- "epoch": 0.4912194522903107,
87
- "grad_norm": 0.12988826632499695,
88
- "learning_rate": 4.988596305882391e-05,
89
- "loss": 0.2271,
90
  "step": 4000
91
  },
92
  {
93
- "epoch": 0.5526218838265995,
94
- "grad_norm": 0.11692152917385101,
95
- "learning_rate": 4.9681275337621505e-05,
96
- "loss": 0.2181,
97
- "step": 4500
 
 
98
  },
99
  {
100
- "epoch": 0.5997789512464694,
101
- "eval_avg": 0.06223662921729023,
102
- "eval_cer": 0.03700322837976026,
103
- "eval_der": 0.055956727518601486,
104
- "eval_loss": 0.19135934114456177,
105
- "eval_runtime": 1153.4306,
106
- "eval_samples_per_second": 44.879,
107
- "eval_steps_per_second": 0.351,
108
- "eval_wer": 0.09374993175350896,
109
- "step": 4884
110
  },
111
  {
112
- "epoch": 0.6140243153628884,
113
- "grad_norm": 0.12457627058029175,
114
- "learning_rate": 4.9374576774384186e-05,
115
- "loss": 0.2112,
116
  "step": 5000
117
  },
118
  {
119
- "epoch": 0.6754267468991773,
120
- "grad_norm": 0.13500584661960602,
121
- "learning_rate": 4.8967134993181585e-05,
122
- "loss": 0.2042,
123
  "step": 5500
124
  },
125
  {
126
- "epoch": 0.736829178435466,
127
- "grad_norm": 0.14300259947776794,
128
- "learning_rate": 4.846063400258344e-05,
129
- "loss": 0.1994,
130
- "step": 6000
131
- },
132
- {
133
- "epoch": 0.7982316099717549,
134
- "grad_norm": 0.15229490399360657,
135
- "learning_rate": 4.785716723543856e-05,
136
- "loss": 0.1941,
137
- "step": 6500
138
- },
139
- {
140
- "epoch": 0.7997052683286258,
141
- "eval_avg": 0.05848093903880644,
142
- "eval_cer": 0.034985685323039174,
143
- "eval_der": 0.05257239447503411,
144
- "eval_loss": 0.17743617296218872,
145
- "eval_runtime": 1154.8064,
146
- "eval_samples_per_second": 44.826,
147
- "eval_steps_per_second": 0.351,
148
- "eval_wer": 0.08788473731834603,
149
- "step": 6512
150
- },
151
- {
152
- "epoch": 0.8596340415080437,
153
- "grad_norm": 0.10620284080505371,
154
- "learning_rate": 4.9950986655011213e-05,
155
- "loss": 0.1892,
156
- "step": 7000
157
- },
158
- {
159
- "epoch": 0.9210364730443326,
160
- "grad_norm": 0.10538509488105774,
161
- "learning_rate": 4.9798882653846754e-05,
162
- "loss": 0.1843,
163
- "step": 7500
164
- },
165
- {
166
- "epoch": 0.9824389045806214,
167
- "grad_norm": 0.11998545378446579,
168
- "learning_rate": 4.954428172468978e-05,
169
- "loss": 0.1814,
170
- "step": 8000
171
- },
172
- {
173
- "epoch": 0.9996315854107822,
174
- "eval_avg": 0.05574338139826937,
175
- "eval_cer": 0.0337392713322537,
176
- "eval_der": 0.05011513570945588,
177
- "eval_loss": 0.16507230699062347,
178
- "eval_runtime": 1158.6086,
179
- "eval_samples_per_second": 44.679,
180
- "eval_steps_per_second": 0.35,
181
- "eval_wer": 0.08337573715309854,
182
- "step": 8140
183
- },
184
- {
185
- "epoch": 1.0438413361169103,
186
- "grad_norm": 0.12560783326625824,
187
- "learning_rate": 4.9189048589485774e-05,
188
- "loss": 0.1764,
189
- "step": 8500
190
- },
191
- {
192
- "epoch": 1.105243767653199,
193
- "grad_norm": 0.10523095726966858,
194
- "learning_rate": 4.8733228183318665e-05,
195
- "loss": 0.1735,
196
- "step": 9000
197
- },
198
- {
199
- "epoch": 1.1666461991894879,
200
- "grad_norm": 0.11041487008333206,
201
- "learning_rate": 4.817931533427457e-05,
202
- "loss": 0.1715,
203
- "step": 9500
204
- },
205
- {
206
- "epoch": 1.1995579024929388,
207
- "eval_avg": 0.0544335846636687,
208
- "eval_cer": 0.03342891994782228,
209
- "eval_der": 0.048513281174538105,
210
- "eval_loss": 0.1576041877269745,
211
- "eval_runtime": 1159.9625,
212
- "eval_samples_per_second": 44.626,
213
- "eval_steps_per_second": 0.349,
214
- "eval_wer": 0.08135855286864571,
215
- "step": 9768
216
- },
217
- {
218
- "epoch": 1.2280486307257767,
219
- "grad_norm": 0.1073341891169548,
220
- "learning_rate": 4.998896962187747e-05,
221
- "loss": 0.1674,
222
- "step": 10000
223
- },
224
- {
225
- "epoch": 1.2894510622620656,
226
- "grad_norm": 0.11093363165855408,
227
- "learning_rate": 4.988961418344719e-05,
228
- "loss": 0.167,
229
- "step": 10500
230
- },
231
- {
232
- "epoch": 1.3508534937983545,
233
- "grad_norm": 0.09265608340501785,
234
- "learning_rate": 4.968738681210237e-05,
235
- "loss": 0.165,
236
- "step": 11000
237
- },
238
- {
239
- "epoch": 1.3994842195750952,
240
- "eval_avg": 0.05450751736133735,
241
- "eval_cer": 0.03384634860456726,
242
- "eval_der": 0.04814585144402502,
243
- "eval_loss": 0.15283821523189545,
244
- "eval_runtime": 1150.9092,
245
- "eval_samples_per_second": 44.977,
246
- "eval_steps_per_second": 0.352,
247
- "eval_wer": 0.08153035203541975,
248
- "step": 11396
249
- },
250
- {
251
- "epoch": 1.4122559253346432,
252
- "grad_norm": 0.10535308718681335,
253
- "learning_rate": 4.999780685937908e-05,
254
- "loss": 0.1617,
255
- "step": 11500
256
- },
257
- {
258
- "epoch": 1.473658356870932,
259
- "grad_norm": 0.09392343461513519,
260
- "learning_rate": 4.992486965915939e-05,
261
- "loss": 0.161,
262
- "step": 12000
263
- },
264
- {
265
- "epoch": 1.535060788407221,
266
- "grad_norm": 0.0898473933339119,
267
- "learning_rate": 4.974891481067152e-05,
268
- "loss": 0.1593,
269
- "step": 12500
270
- },
271
- {
272
- "epoch": 1.5964632199435098,
273
- "grad_norm": 0.10352090746164322,
274
- "learning_rate": 4.94706695576259e-05,
275
- "loss": 0.1581,
276
- "step": 13000
277
- },
278
- {
279
- "epoch": 1.5994105366572517,
280
- "eval_avg": 0.052648695144827624,
281
- "eval_cer": 0.03269155476372104,
282
- "eval_der": 0.04647078141601352,
283
- "eval_loss": 0.1520785242319107,
284
- "eval_runtime": 1154.3982,
285
- "eval_samples_per_second": 44.842,
286
- "eval_steps_per_second": 0.351,
287
- "eval_wer": 0.07878374925474832,
288
- "step": 13024
289
- },
290
- {
291
- "epoch": 1.6578656514797987,
292
- "grad_norm": 0.08162818104028702,
293
- "learning_rate": 4.995337159613658e-05,
294
- "loss": 0.1563,
295
- "step": 13500
296
- },
297
- {
298
- "epoch": 1.7192680830160874,
299
- "grad_norm": 0.10093415528535843,
300
- "learning_rate": 4.9803736946497604e-05,
301
- "loss": 0.1549,
302
- "step": 14000
303
- },
304
- {
305
- "epoch": 1.7806705145523762,
306
- "grad_norm": 0.0779104083776474,
307
- "learning_rate": 4.9551585305458784e-05,
308
- "loss": 0.1531,
309
- "step": 14500
310
- },
311
- {
312
- "epoch": 1.799336853739408,
313
- "eval_avg": 0.05215597602151043,
314
- "eval_cer": 0.032804504015484064,
315
- "eval_der": 0.046157249106538095,
316
- "eval_loss": 0.1437014639377594,
317
- "eval_runtime": 1158.5459,
318
- "eval_samples_per_second": 44.681,
319
- "eval_steps_per_second": 0.35,
320
- "eval_wer": 0.07750617494250915,
321
- "step": 14652
322
- },
323
- {
324
- "epoch": 1.8420729460886651,
325
- "grad_norm": 0.09926008433103561,
326
- "learning_rate": 4.919795884774044e-05,
327
- "loss": 0.1519,
328
- "step": 15000
329
- },
330
- {
331
- "epoch": 1.903475377624954,
332
- "grad_norm": 0.0932171419262886,
333
- "learning_rate": 4.874431915635338e-05,
334
- "loss": 0.1508,
335
- "step": 15500
336
- },
337
- {
338
- "epoch": 1.9648778091612429,
339
- "grad_norm": 0.08440640568733215,
340
- "learning_rate": 4.819254118169078e-05,
341
- "loss": 0.1493,
342
- "step": 16000
343
- },
344
- {
345
- "epoch": 1.9992631708215645,
346
- "eval_avg": 0.05104936846864332,
347
- "eval_cer": 0.03196083873281993,
348
- "eval_der": 0.04519598184101095,
349
- "eval_loss": 0.1432473361492157,
350
- "eval_runtime": 1163.7456,
351
- "eval_samples_per_second": 44.481,
352
- "eval_steps_per_second": 0.348,
353
- "eval_wer": 0.07599128483209908,
354
- "step": 16280
355
  }
356
  ],
357
  "logging_steps": 500,
358
- "max_steps": 24429,
359
  "num_input_tokens_seen": 0,
360
- "num_train_epochs": 3,
361
- "save_steps": 1628,
362
  "stateful_callbacks": {
363
  "TrainerControl": {
364
  "args": {
@@ -366,13 +144,13 @@
366
  "should_evaluate": false,
367
  "should_log": false,
368
  "should_save": true,
369
- "should_training_stop": false
370
  },
371
  "attributes": {}
372
  }
373
  },
374
- "total_flos": 1.8107754052964844e+18,
375
- "train_batch_size": 128,
376
  "trial_name": null,
377
  "trial_params": null
378
  }
 
1
  {
2
+ "best_global_step": 5570,
3
+ "best_metric": 0.7717538371980056,
4
+ "best_model_checkpoint": "./checkpoints/checkpoint-5570",
5
+ "epoch": 0.9998653923812088,
6
+ "eval_steps": 1114,
7
+ "global_step": 5571,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "epoch": 0.0897384125274824,
14
+ "grad_norm": 0.10910310596227646,
15
+ "learning_rate": 4.901672070547205e-05,
16
+ "loss": 2.0607,
17
  "step": 500
18
  },
19
  {
20
+ "epoch": 0.1794768250549648,
21
+ "grad_norm": 0.11283092200756073,
22
+ "learning_rate": 4.6136704554635066e-05,
23
+ "loss": 0.1219,
24
  "step": 1000
25
  },
26
  {
27
+ "epoch": 0.19993718311123077,
28
+ "eval_cer": 0.7744026050691318,
29
+ "eval_loss": 0.06905069202184677,
30
+ "eval_runtime": 2200.6581,
31
+ "eval_samples_per_second": 80.463,
32
+ "eval_steps_per_second": 0.21,
33
+ "step": 1114
34
  },
35
  {
36
+ "epoch": 0.26921523758244714,
37
+ "grad_norm": 0.07702942937612534,
38
+ "learning_rate": 4.9413107833705604e-05,
39
+ "loss": 0.0709,
40
+ "step": 1500
 
 
 
 
 
41
  },
42
  {
43
+ "epoch": 0.3589536501099296,
44
+ "grad_norm": 0.048379454761743546,
45
+ "learning_rate": 4.695072154439477e-05,
46
+ "loss": 0.0526,
47
  "step": 2000
48
  },
49
  {
50
+ "epoch": 0.39987436622246153,
51
+ "eval_cer": 0.7726060406484112,
52
+ "eval_loss": 0.03332927078008652,
53
+ "eval_runtime": 436.8886,
54
+ "eval_samples_per_second": 405.3,
55
+ "eval_steps_per_second": 1.057,
56
+ "step": 2228
57
+ },
58
+ {
59
+ "epoch": 0.44869206263741196,
60
+ "grad_norm": 0.04356146231293678,
61
+ "learning_rate": 4.970863550514242e-05,
62
+ "loss": 0.0398,
63
  "step": 2500
64
  },
65
  {
66
+ "epoch": 0.5384304751648943,
67
+ "grad_norm": 0.03718848526477814,
68
+ "learning_rate": 4.767405209364289e-05,
69
+ "loss": 0.0348,
70
  "step": 3000
71
  },
72
  {
73
+ "epoch": 0.5998115493336923,
74
+ "eval_cer": 0.7720533296293449,
75
+ "eval_loss": 0.022723019123077393,
76
+ "eval_runtime": 439.0491,
77
+ "eval_samples_per_second": 403.306,
78
+ "eval_steps_per_second": 1.052,
79
+ "step": 3342
 
 
 
80
  },
81
  {
82
+ "epoch": 0.6281688876923768,
83
+ "grad_norm": 0.03279432654380798,
84
+ "learning_rate": 4.384876624004233e-05,
85
+ "loss": 0.03,
86
  "step": 3500
87
  },
88
  {
89
+ "epoch": 0.7179073002198592,
90
+ "grad_norm": 0.032072387635707855,
91
+ "learning_rate": 3.8534883130104934e-05,
92
+ "loss": 0.027,
93
  "step": 4000
94
  },
95
  {
96
+ "epoch": 0.7997487324449231,
97
+ "eval_cer": 0.7718523934618883,
98
+ "eval_loss": 0.01841309294104576,
99
+ "eval_runtime": 436.0956,
100
+ "eval_samples_per_second": 406.037,
101
+ "eval_steps_per_second": 1.059,
102
+ "step": 4456
103
  },
104
  {
105
+ "epoch": 0.8076457127473415,
106
+ "grad_norm": 0.0274814460426569,
107
+ "learning_rate": 3.2152071176988856e-05,
108
+ "loss": 0.0248,
109
+ "step": 4500
 
 
 
 
 
110
  },
111
  {
112
+ "epoch": 0.8973841252748239,
113
+ "grad_norm": 0.025461740791797638,
114
+ "learning_rate": 2.520441835335607e-05,
115
+ "loss": 0.0232,
116
  "step": 5000
117
  },
118
  {
119
+ "epoch": 0.9871225378023063,
120
+ "grad_norm": 0.026277432218194008,
121
+ "learning_rate": 1.8240621417367316e-05,
122
+ "loss": 0.0222,
123
  "step": 5500
124
  },
125
  {
126
+ "epoch": 0.9996859155561538,
127
+ "eval_cer": 0.7717538371980056,
128
+ "eval_loss": 0.016575267538428307,
129
+ "eval_runtime": 437.5161,
130
+ "eval_samples_per_second": 404.719,
131
+ "eval_steps_per_second": 1.056,
132
+ "step": 5570
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
133
  }
134
  ],
135
  "logging_steps": 500,
136
+ "max_steps": 5571,
137
  "num_input_tokens_seen": 0,
138
+ "num_train_epochs": 1,
139
+ "save_steps": 1114,
140
  "stateful_callbacks": {
141
  "TrainerControl": {
142
  "args": {
 
144
  "should_evaluate": false,
145
  "should_log": false,
146
  "should_save": true,
147
+ "should_training_stop": true
148
  },
149
  "attributes": {}
150
  }
151
  },
152
+ "total_flos": 6.522138159281603e+17,
153
+ "train_batch_size": 384,
154
  "trial_name": null,
155
  "trial_params": null
156
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:994dcd51f10df37a598334a01dc75b4d6c85da6c809dc43933bd3b4a8f34c7ee
3
- size 5905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bce4c2a570e5406234205443c86622d09135dc71fc8963eb2e85ed7f5af2b7c
3
+ size 5841