loris3 commited on
Commit
f476a26
·
verified ·
1 Parent(s): 35b0656

Upload folder using huggingface_hub

Browse files
checkpoints/checkpoint-1942/config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "head_dim": 64,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 2048,
14
+ "max_position_embeddings": 256,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "num_key_value_heads": 12,
20
+ "pad_token_id": 1,
21
+ "pretraining_tp": 1,
22
+ "rms_norm_eps": 1e-06,
23
+ "rope_scaling": null,
24
+ "rope_theta": 10000.0,
25
+ "tie_word_embeddings": true,
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.47.0",
28
+ "use_cache": true,
29
+ "vocab_size": 16000
30
+ }
checkpoints/checkpoint-1942/generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "pad_token_id": 1,
6
+ "transformers_version": "4.47.0"
7
+ }
checkpoints/checkpoint-1942/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6769f208d631f497c88a59ab4a39a1ae3c3e752b1d6693a7ad93e29b1c9ac241
3
+ size 388979624
checkpoints/checkpoint-1942/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac737f53664d2608697791ba68099c672fdae4196d677bf0652350000d4046f1
3
+ size 778027770
checkpoints/checkpoint-1942/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8789b5b5ca08009ccf418dfb9b37e200608d0d80ccbc11521ae7fb336cb08be5
3
+ size 14244
checkpoints/checkpoint-1942/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b2effe62a90ed54917e97dbe71307968ad4d9ebbbaddffddd7c89283361431f
3
+ size 1064
checkpoints/checkpoint-1942/trainer_state.json ADDED
@@ -0,0 +1,349 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 4.998804113848362,
5
+ "eval_steps": 500,
6
+ "global_step": 1942,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.09567089213106912,
13
+ "grad_norm": 19.851242065429688,
14
+ "learning_rate": 0.00011666666666666667,
15
+ "loss": 121.1525,
16
+ "step": 50
17
+ },
18
+ {
19
+ "epoch": 0.19134178426213824,
20
+ "grad_norm": 13.892197608947754,
21
+ "learning_rate": 0.00023333333333333333,
22
+ "loss": 77.1477,
23
+ "step": 100
24
+ },
25
+ {
26
+ "epoch": 0.2870126763932074,
27
+ "grad_norm": 12.60777759552002,
28
+ "learning_rate": 0.00035,
29
+ "loss": 64.8839,
30
+ "step": 150
31
+ },
32
+ {
33
+ "epoch": 0.3826835685242765,
34
+ "grad_norm": 8.890196800231934,
35
+ "learning_rate": 0.00046666666666666666,
36
+ "loss": 59.8849,
37
+ "step": 200
38
+ },
39
+ {
40
+ "epoch": 0.4783544606553456,
41
+ "grad_norm": 8.705379486083984,
42
+ "learning_rate": 0.0005833333333333334,
43
+ "loss": 57.5299,
44
+ "step": 250
45
+ },
46
+ {
47
+ "epoch": 0.5740253527864148,
48
+ "grad_norm": 7.457199573516846,
49
+ "learning_rate": 0.0007,
50
+ "loss": 56.2806,
51
+ "step": 300
52
+ },
53
+ {
54
+ "epoch": 0.6696962449174838,
55
+ "grad_norm": 7.943981647491455,
56
+ "learning_rate": 0.000699821634561209,
57
+ "loss": 55.3251,
58
+ "step": 350
59
+ },
60
+ {
61
+ "epoch": 0.765367137048553,
62
+ "grad_norm": 7.716305732727051,
63
+ "learning_rate": 0.0006992867200404345,
64
+ "loss": 52.4636,
65
+ "step": 400
66
+ },
67
+ {
68
+ "epoch": 0.8610380291796221,
69
+ "grad_norm": 7.123641014099121,
70
+ "learning_rate": 0.0006983958016391807,
71
+ "loss": 52.215,
72
+ "step": 450
73
+ },
74
+ {
75
+ "epoch": 0.9567089213106912,
76
+ "grad_norm": 7.223691463470459,
77
+ "learning_rate": 0.0006971497874091708,
78
+ "loss": 52.1157,
79
+ "step": 500
80
+ },
81
+ {
82
+ "epoch": 0.9988041138483617,
83
+ "eval_accuracy": 0.0,
84
+ "eval_loss": 8.116789817810059,
85
+ "eval_normalizer": 685885.0,
86
+ "eval_runtime": 103.9352,
87
+ "eval_samples_per_second": 514.33,
88
+ "eval_steps_per_second": 1.01,
89
+ "step": 522
90
+ },
91
+ {
92
+ "epoch": 1.0535756995933987,
93
+ "grad_norm": 7.065810203552246,
94
+ "learning_rate": 0.0006955499473268326,
95
+ "loss": 50.0351,
96
+ "step": 550
97
+ },
98
+ {
99
+ "epoch": 1.1492465917244679,
100
+ "grad_norm": 7.027311325073242,
101
+ "learning_rate": 0.0006935979119988993,
102
+ "loss": 48.6995,
103
+ "step": 600
104
+ },
105
+ {
106
+ "epoch": 1.244917483855537,
107
+ "grad_norm": 6.899056911468506,
108
+ "learning_rate": 0.0006912956710004438,
109
+ "loss": 48.9226,
110
+ "step": 650
111
+ },
112
+ {
113
+ "epoch": 1.3405883759866062,
114
+ "grad_norm": 7.2765116691589355,
115
+ "learning_rate": 0.0006886455708470427,
116
+ "loss": 48.9019,
117
+ "step": 700
118
+ },
119
+ {
120
+ "epoch": 1.4362592681176751,
121
+ "grad_norm": 8.293883323669434,
122
+ "learning_rate": 0.0006856503126031346,
123
+ "loss": 44.5473,
124
+ "step": 750
125
+ },
126
+ {
127
+ "epoch": 1.5319301602487443,
128
+ "grad_norm": 7.498640537261963,
129
+ "learning_rate": 0.0006823129491290102,
130
+ "loss": 45.238,
131
+ "step": 800
132
+ },
133
+ {
134
+ "epoch": 1.6276010523798135,
135
+ "grad_norm": 7.582442283630371,
136
+ "learning_rate": 0.0006786368819692442,
137
+ "loss": 45.6905,
138
+ "step": 850
139
+ },
140
+ {
141
+ "epoch": 1.7232719445108826,
142
+ "grad_norm": 8.016205787658691,
143
+ "learning_rate": 0.0006746258578857331,
144
+ "loss": 42.9074,
145
+ "step": 900
146
+ },
147
+ {
148
+ "epoch": 1.8189428366419516,
149
+ "grad_norm": 8.324728012084961,
150
+ "learning_rate": 0.000670283965038881,
151
+ "loss": 40.5995,
152
+ "step": 950
153
+ },
154
+ {
155
+ "epoch": 1.9146137287730207,
156
+ "grad_norm": 8.61531925201416,
157
+ "learning_rate": 0.0006656156288208179,
158
+ "loss": 41.5048,
159
+ "step": 1000
160
+ },
161
+ {
162
+ "epoch": 1.9988041138483617,
163
+ "eval_accuracy": 0.0,
164
+ "eval_loss": 8.638134956359863,
165
+ "eval_normalizer": 685885.0,
166
+ "eval_runtime": 104.5141,
167
+ "eval_samples_per_second": 511.481,
168
+ "eval_steps_per_second": 1.005,
169
+ "step": 1044
170
+ },
171
+ {
172
+ "epoch": 2.0114805070557282,
173
+ "grad_norm": 119.22761535644531,
174
+ "learning_rate": 0.000660625607344904,
175
+ "loss": 52.7863,
176
+ "step": 1050
177
+ },
178
+ {
179
+ "epoch": 2.1071513991867974,
180
+ "grad_norm": 7.394373893737793,
181
+ "learning_rate": 0.0006553189865961112,
182
+ "loss": 87.7065,
183
+ "step": 1100
184
+ },
185
+ {
186
+ "epoch": 2.2028222913178666,
187
+ "grad_norm": 5.8150200843811035,
188
+ "learning_rate": 0.0006497011752472301,
189
+ "loss": 68.8677,
190
+ "step": 1150
191
+ },
192
+ {
193
+ "epoch": 2.2984931834489357,
194
+ "grad_norm": 8.597038269042969,
195
+ "learning_rate": 0.0006437778991461825,
196
+ "loss": 59.723,
197
+ "step": 1200
198
+ },
199
+ {
200
+ "epoch": 2.3597225544128198,
201
+ "eval_accuracy": 0.0,
202
+ "eval_loss": 5.760004997253418,
203
+ "eval_normalizer": 685885.0,
204
+ "eval_runtime": 102.8572,
205
+ "eval_samples_per_second": 519.721,
206
+ "eval_steps_per_second": 1.021,
207
+ "step": 1232
208
+ },
209
+ {
210
+ "epoch": 3.0344415211671847,
211
+ "grad_norm": 9.546870231628418,
212
+ "learning_rate": 0.0006375551954800587,
213
+ "loss": 49.7983,
214
+ "step": 1250
215
+ },
216
+ {
217
+ "epoch": 3.130112413298254,
218
+ "grad_norm": 11.537946701049805,
219
+ "learning_rate": 0.0006310394066218296,
220
+ "loss": 38.324,
221
+ "step": 1300
222
+ },
223
+ {
224
+ "epoch": 3.225783305429323,
225
+ "grad_norm": 11.162705421447754,
226
+ "learning_rate": 0.0006242371736660025,
227
+ "loss": 27.47,
228
+ "step": 1350
229
+ },
230
+ {
231
+ "epoch": 3.321454197560392,
232
+ "grad_norm": 12.32886791229248,
233
+ "learning_rate": 0.000617155429659811,
234
+ "loss": 18.7372,
235
+ "step": 1400
236
+ },
237
+ {
238
+ "epoch": 3.3597225544128198,
239
+ "eval_accuracy": 0.0,
240
+ "eval_loss": 7.427099227905273,
241
+ "eval_normalizer": 685885.0,
242
+ "eval_runtime": 110.0904,
243
+ "eval_samples_per_second": 485.573,
244
+ "eval_steps_per_second": 0.954,
245
+ "step": 1420
246
+ },
247
+ {
248
+ "epoch": 4.057402535278642,
249
+ "grad_norm": 13.549798011779785,
250
+ "learning_rate": 0.0006098013925368385,
251
+ "loss": 57.3773,
252
+ "step": 1450
253
+ },
254
+ {
255
+ "epoch": 4.15307342740971,
256
+ "grad_norm": 10.063131332397461,
257
+ "learning_rate": 0.0006021825577602754,
258
+ "loss": 71.4705,
259
+ "step": 1500
260
+ },
261
+ {
262
+ "epoch": 4.24874431954078,
263
+ "grad_norm": 9.314085006713867,
264
+ "learning_rate": 0.0005943066906833104,
265
+ "loss": 69.1017,
266
+ "step": 1550
267
+ },
268
+ {
269
+ "epoch": 4.344415211671849,
270
+ "grad_norm": 8.48709774017334,
271
+ "learning_rate": 0.0005861818186344407,
272
+ "loss": 66.7627,
273
+ "step": 1600
274
+ },
275
+ {
276
+ "epoch": 4.440086103802918,
277
+ "grad_norm": 9.457844734191895,
278
+ "learning_rate": 0.00057781622273577,
279
+ "loss": 62.5878,
280
+ "step": 1650
281
+ },
282
+ {
283
+ "epoch": 4.535756995933987,
284
+ "grad_norm": 8.216409683227539,
285
+ "learning_rate": 0.0005692184294626307,
286
+ "loss": 63.0275,
287
+ "step": 1700
288
+ },
289
+ {
290
+ "epoch": 4.631427888065057,
291
+ "grad_norm": 8.80875301361084,
292
+ "learning_rate": 0.0005603972019531362,
293
+ "loss": 62.8742,
294
+ "step": 1750
295
+ },
296
+ {
297
+ "epoch": 4.727098780196125,
298
+ "grad_norm": 9.469932556152344,
299
+ "learning_rate": 0.0005513615310765172,
300
+ "loss": 56.6002,
301
+ "step": 1800
302
+ },
303
+ {
304
+ "epoch": 4.822769672327194,
305
+ "grad_norm": 9.208184242248535,
306
+ "learning_rate": 0.0005421206262693491,
307
+ "loss": 55.5842,
308
+ "step": 1850
309
+ },
310
+ {
311
+ "epoch": 4.918440564458264,
312
+ "grad_norm": 9.542802810668945,
313
+ "learning_rate": 0.0005326839061490078,
314
+ "loss": 56.4509,
315
+ "step": 1900
316
+ },
317
+ {
318
+ "epoch": 4.998804113848362,
319
+ "eval_accuracy": 0.0,
320
+ "eval_loss": 5.560527801513672,
321
+ "eval_normalizer": 685885.0,
322
+ "eval_runtime": 102.9429,
323
+ "eval_samples_per_second": 519.288,
324
+ "eval_steps_per_second": 1.02,
325
+ "step": 1942
326
+ }
327
+ ],
328
+ "logging_steps": 50,
329
+ "max_steps": 5220,
330
+ "num_input_tokens_seen": 0,
331
+ "num_train_epochs": 10,
332
+ "save_steps": 500,
333
+ "stateful_callbacks": {
334
+ "TrainerControl": {
335
+ "args": {
336
+ "should_epoch_stop": false,
337
+ "should_evaluate": false,
338
+ "should_log": false,
339
+ "should_save": true,
340
+ "should_training_stop": false
341
+ },
342
+ "attributes": {}
343
+ }
344
+ },
345
+ "total_flos": 1.4697852813063014e+17,
346
+ "train_batch_size": 128,
347
+ "trial_name": null,
348
+ "trial_params": null
349
+ }
checkpoints/checkpoint-1942/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112439a8bf818313a9112c0345790e09c2dd4279f702a6944e87378fe04ce172
3
+ size 5432