rbelanec commited on
Commit
f2db9a5
·
verified ·
1 Parent(s): 2c0860b

End of training

Browse files
README.md CHANGED
@@ -17,10 +17,10 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # test
19
 
20
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 0.3559
23
- - Num Input Tokens Seen: 43600
24
 
25
  ## Model description
26
 
 
17
 
18
  # test
19
 
20
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the wsc dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.3497
23
+ - Num Input Tokens Seen: 43904
24
 
25
  ## Model description
26
 
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.35078784823417664,
4
- "eval_runtime": 1.0329,
5
- "eval_samples_per_second": 54.219,
6
- "eval_steps_per_second": 27.109,
7
  "num_input_tokens_seen": 43904,
8
  "total_flos": 278458437992448.0,
9
- "train_loss": 0.39900598229174633,
10
- "train_runtime": 65.0582,
11
- "train_samples_per_second": 7.655,
12
- "train_steps_per_second": 3.827
13
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 0.3496658504009247,
4
+ "eval_runtime": 1.2462,
5
+ "eval_samples_per_second": 44.938,
6
+ "eval_steps_per_second": 22.469,
7
  "num_input_tokens_seen": 43904,
8
  "total_flos": 278458437992448.0,
9
+ "train_loss": 0.3966633373474979,
10
+ "train_runtime": 65.6097,
11
+ "train_samples_per_second": 7.59,
12
+ "train_steps_per_second": 3.795
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.35078784823417664,
4
- "eval_runtime": 1.0329,
5
- "eval_samples_per_second": 54.219,
6
- "eval_steps_per_second": 27.109,
7
  "num_input_tokens_seen": 43904
8
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_loss": 0.3496658504009247,
4
+ "eval_runtime": 1.2462,
5
+ "eval_samples_per_second": 44.938,
6
+ "eval_steps_per_second": 22.469,
7
  "num_input_tokens_seen": 43904
8
  }
train_results.json CHANGED
@@ -2,8 +2,8 @@
2
  "epoch": 1.0,
3
  "num_input_tokens_seen": 43904,
4
  "total_flos": 278458437992448.0,
5
- "train_loss": 0.39900598229174633,
6
- "train_runtime": 65.0582,
7
- "train_samples_per_second": 7.655,
8
- "train_steps_per_second": 3.827
9
  }
 
2
  "epoch": 1.0,
3
  "num_input_tokens_seen": 43904,
4
  "total_flos": 278458437992448.0,
5
+ "train_loss": 0.3966633373474979,
6
+ "train_runtime": 65.6097,
7
+ "train_samples_per_second": 7.59,
8
+ "train_steps_per_second": 3.795
9
  }
trainer_state.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "best_global_step": 208,
3
- "best_metric": 0.35078784823417664,
4
- "best_model_checkpoint": "saves/test/checkpoint-208",
5
  "epoch": 1.0,
6
  "eval_steps": 13,
7
  "global_step": 249,
@@ -11,564 +11,564 @@
11
  "log_history": [
12
  {
13
  "epoch": 0.020080321285140562,
14
- "grad_norm": 30.61184310913086,
15
  "learning_rate": 8.000000000000001e-06,
16
- "loss": 0.8482,
17
  "num_input_tokens_seen": 832,
18
  "step": 5
19
  },
20
  {
21
  "epoch": 0.040160642570281124,
22
- "grad_norm": 19.559144973754883,
23
  "learning_rate": 1.8e-05,
24
- "loss": 0.7661,
25
  "num_input_tokens_seen": 1760,
26
  "step": 10
27
  },
28
  {
29
  "epoch": 0.05220883534136546,
30
- "eval_loss": 0.6881787180900574,
31
- "eval_runtime": 0.7795,
32
- "eval_samples_per_second": 71.84,
33
- "eval_steps_per_second": 35.92,
34
  "num_input_tokens_seen": 2288,
35
  "step": 13
36
  },
37
  {
38
  "epoch": 0.060240963855421686,
39
- "grad_norm": 9.706626892089844,
40
  "learning_rate": 2.8000000000000003e-05,
41
- "loss": 0.7028,
42
  "num_input_tokens_seen": 2608,
43
  "step": 15
44
  },
45
  {
46
  "epoch": 0.08032128514056225,
47
- "grad_norm": 17.458187103271484,
48
  "learning_rate": 3.8e-05,
49
- "loss": 0.5538,
50
  "num_input_tokens_seen": 3536,
51
  "step": 20
52
  },
53
  {
54
  "epoch": 0.10040160642570281,
55
- "grad_norm": 11.384882926940918,
56
  "learning_rate": 4.8e-05,
57
- "loss": 0.6839,
58
  "num_input_tokens_seen": 4496,
59
  "step": 25
60
  },
61
  {
62
  "epoch": 0.10441767068273092,
63
- "eval_loss": 0.4647722542285919,
64
- "eval_runtime": 0.76,
65
- "eval_samples_per_second": 73.686,
66
- "eval_steps_per_second": 36.843,
67
  "num_input_tokens_seen": 4656,
68
  "step": 26
69
  },
70
  {
71
  "epoch": 0.12048192771084337,
72
- "grad_norm": 8.721712112426758,
73
  "learning_rate": 4.996067037544542e-05,
74
- "loss": 0.5189,
75
  "num_input_tokens_seen": 5424,
76
  "step": 30
77
  },
78
  {
79
  "epoch": 0.14056224899598393,
80
- "grad_norm": 7.760402202606201,
81
  "learning_rate": 4.980110583549062e-05,
82
- "loss": 0.374,
83
  "num_input_tokens_seen": 6304,
84
  "step": 35
85
  },
86
  {
87
  "epoch": 0.1566265060240964,
88
- "eval_loss": 0.38417601585388184,
89
- "eval_runtime": 0.8123,
90
- "eval_samples_per_second": 68.942,
91
- "eval_steps_per_second": 34.471,
92
  "num_input_tokens_seen": 6944,
93
  "step": 39
94
  },
95
  {
96
  "epoch": 0.1606425702811245,
97
- "grad_norm": 1.9442764520645142,
98
  "learning_rate": 4.951963201008076e-05,
99
- "loss": 0.4006,
100
  "num_input_tokens_seen": 7072,
101
  "step": 40
102
  },
103
  {
104
  "epoch": 0.18072289156626506,
105
- "grad_norm": 6.7856879234313965,
106
  "learning_rate": 4.91176324775594e-05,
107
- "loss": 0.3861,
108
  "num_input_tokens_seen": 7856,
109
  "step": 45
110
  },
111
  {
112
  "epoch": 0.20080321285140562,
113
- "grad_norm": 9.397907257080078,
114
  "learning_rate": 4.8597083257709194e-05,
115
- "loss": 0.3624,
116
  "num_input_tokens_seen": 8880,
117
  "step": 50
118
  },
119
  {
120
  "epoch": 0.20883534136546184,
121
- "eval_loss": 0.3785439133644104,
122
- "eval_runtime": 0.7026,
123
- "eval_samples_per_second": 79.709,
124
- "eval_steps_per_second": 39.854,
125
  "num_input_tokens_seen": 9232,
126
  "step": 52
127
  },
128
  {
129
  "epoch": 0.22088353413654618,
130
- "grad_norm": 2.288628101348877,
131
  "learning_rate": 4.796054309867053e-05,
132
- "loss": 0.4093,
133
  "num_input_tokens_seen": 9680,
134
  "step": 55
135
  },
136
  {
137
  "epoch": 0.24096385542168675,
138
- "grad_norm": 6.740528583526611,
139
  "learning_rate": 4.721114089947181e-05,
140
- "loss": 0.3391,
141
  "num_input_tokens_seen": 10576,
142
  "step": 60
143
  },
144
  {
145
  "epoch": 0.26104417670682734,
146
- "grad_norm": 2.4798903465270996,
147
  "learning_rate": 4.6352560329995686e-05,
148
- "loss": 0.3164,
149
  "num_input_tokens_seen": 11424,
150
  "step": 65
151
  },
152
  {
153
  "epoch": 0.26104417670682734,
154
- "eval_loss": 0.36688488721847534,
155
- "eval_runtime": 0.9363,
156
- "eval_samples_per_second": 59.811,
157
- "eval_steps_per_second": 29.905,
158
  "num_input_tokens_seen": 11424,
159
  "step": 65
160
  },
161
  {
162
  "epoch": 0.28112449799196787,
163
- "grad_norm": 2.8193576335906982,
164
  "learning_rate": 4.538902172398151e-05,
165
- "loss": 0.3673,
166
  "num_input_tokens_seen": 12224,
167
  "step": 70
168
  },
169
  {
170
  "epoch": 0.30120481927710846,
171
- "grad_norm": 1.4678492546081543,
172
  "learning_rate": 4.4325261334068426e-05,
173
- "loss": 0.3623,
174
  "num_input_tokens_seen": 13168,
175
  "step": 75
176
  },
177
  {
178
  "epoch": 0.3132530120481928,
179
- "eval_loss": 0.3628059923648834,
180
- "eval_runtime": 0.7816,
181
- "eval_samples_per_second": 71.644,
182
- "eval_steps_per_second": 35.822,
183
  "num_input_tokens_seen": 13760,
184
  "step": 78
185
  },
186
  {
187
  "epoch": 0.321285140562249,
188
- "grad_norm": 7.307011127471924,
189
  "learning_rate": 4.316650805085068e-05,
190
- "loss": 0.3654,
191
  "num_input_tokens_seen": 14080,
192
  "step": 80
193
  },
194
  {
195
  "epoch": 0.3413654618473896,
196
- "grad_norm": 6.259345531463623,
197
  "learning_rate": 4.1918457700381855e-05,
198
- "loss": 0.3651,
199
  "num_input_tokens_seen": 15056,
200
  "step": 85
201
  },
202
  {
203
  "epoch": 0.3614457831325301,
204
- "grad_norm": 7.0643415451049805,
205
  "learning_rate": 4.058724504646834e-05,
206
- "loss": 0.3656,
207
  "num_input_tokens_seen": 15904,
208
  "step": 90
209
  },
210
  {
211
  "epoch": 0.3654618473895582,
212
- "eval_loss": 0.358058899641037,
213
- "eval_runtime": 0.762,
214
- "eval_samples_per_second": 73.49,
215
- "eval_steps_per_second": 36.745,
216
  "num_input_tokens_seen": 16048,
217
  "step": 91
218
  },
219
  {
220
  "epoch": 0.3815261044176707,
221
- "grad_norm": 6.522783279418945,
222
  "learning_rate": 3.9179413635373897e-05,
223
- "loss": 0.3311,
224
  "num_input_tokens_seen": 16688,
225
  "step": 95
226
  },
227
  {
228
  "epoch": 0.40160642570281124,
229
- "grad_norm": 6.159587860107422,
230
  "learning_rate": 3.770188363116324e-05,
231
- "loss": 0.2954,
232
  "num_input_tokens_seen": 17552,
233
  "step": 100
234
  },
235
  {
236
  "epoch": 0.41767068273092367,
237
- "eval_loss": 0.3805600702762604,
238
- "eval_runtime": 0.8411,
239
- "eval_samples_per_second": 66.579,
240
- "eval_steps_per_second": 33.289,
241
  "num_input_tokens_seen": 18272,
242
  "step": 104
243
  },
244
  {
245
  "epoch": 0.42168674698795183,
246
- "grad_norm": 2.2846226692199707,
247
  "learning_rate": 3.616191779978907e-05,
248
- "loss": 0.329,
249
  "num_input_tokens_seen": 18400,
250
  "step": 105
251
  },
252
  {
253
  "epoch": 0.44176706827309237,
254
- "grad_norm": 5.145074844360352,
255
  "learning_rate": 3.456708580912725e-05,
256
- "loss": 0.3768,
257
  "num_input_tokens_seen": 19456,
258
  "step": 110
259
  },
260
  {
261
  "epoch": 0.46184738955823296,
262
- "grad_norm": 8.501044273376465,
263
  "learning_rate": 3.292522702044221e-05,
264
- "loss": 0.4359,
265
  "num_input_tokens_seen": 20288,
266
  "step": 115
267
  },
268
  {
269
  "epoch": 0.46987951807228917,
270
- "eval_loss": 0.37039586901664734,
271
- "eval_runtime": 0.8018,
272
- "eval_samples_per_second": 69.844,
273
- "eval_steps_per_second": 34.922,
274
  "num_input_tokens_seen": 20656,
275
  "step": 117
276
  },
277
  {
278
  "epoch": 0.4819277108433735,
279
- "grad_norm": 2.1846377849578857,
280
  "learning_rate": 3.1244411954180676e-05,
281
- "loss": 0.3823,
282
  "num_input_tokens_seen": 21328,
283
  "step": 120
284
  },
285
  {
286
  "epoch": 0.5020080321285141,
287
- "grad_norm": 1.8924623727798462,
288
  "learning_rate": 2.9532902619507462e-05,
289
- "loss": 0.3669,
290
  "num_input_tokens_seen": 22304,
291
  "step": 125
292
  },
293
  {
294
  "epoch": 0.5220883534136547,
295
- "grad_norm": 6.834183692932129,
296
  "learning_rate": 2.7799111902582696e-05,
297
- "loss": 0.356,
298
  "num_input_tokens_seen": 23056,
299
  "step": 130
300
  },
301
  {
302
  "epoch": 0.5220883534136547,
303
- "eval_loss": 0.3524550795555115,
304
- "eval_runtime": 0.8184,
305
- "eval_samples_per_second": 68.423,
306
- "eval_steps_per_second": 34.211,
307
  "num_input_tokens_seen": 23056,
308
  "step": 130
309
  },
310
  {
311
  "epoch": 0.5421686746987951,
312
- "grad_norm": 8.209941864013672,
313
  "learning_rate": 2.6051562213206632e-05,
314
- "loss": 0.3505,
315
  "num_input_tokens_seen": 23840,
316
  "step": 135
317
  },
318
  {
319
  "epoch": 0.5622489959839357,
320
- "grad_norm": 1.8095901012420654,
321
  "learning_rate": 2.429884359310328e-05,
322
- "loss": 0.3685,
323
  "num_input_tokens_seen": 24832,
324
  "step": 140
325
  },
326
  {
327
  "epoch": 0.5742971887550201,
328
- "eval_loss": 0.35455432534217834,
329
- "eval_runtime": 0.8626,
330
- "eval_samples_per_second": 64.92,
331
- "eval_steps_per_second": 32.46,
332
  "num_input_tokens_seen": 25312,
333
  "step": 143
334
  },
335
  {
336
  "epoch": 0.5823293172690763,
337
- "grad_norm": 1.6747307777404785,
338
  "learning_rate": 2.2549571491760986e-05,
339
- "loss": 0.3842,
340
  "num_input_tokens_seen": 25648,
341
  "step": 145
342
  },
343
  {
344
  "epoch": 0.6024096385542169,
345
- "grad_norm": 7.220715522766113,
346
  "learning_rate": 2.0812344417381595e-05,
347
- "loss": 0.3393,
348
  "num_input_tokens_seen": 26496,
349
  "step": 150
350
  },
351
  {
352
  "epoch": 0.6224899598393574,
353
- "grad_norm": 6.922730922698975,
354
  "learning_rate": 1.909570167110415e-05,
355
- "loss": 0.3832,
356
  "num_input_tokens_seen": 27392,
357
  "step": 155
358
  },
359
  {
360
  "epoch": 0.6265060240963856,
361
- "eval_loss": 0.3515147566795349,
362
- "eval_runtime": 1.4586,
363
- "eval_samples_per_second": 38.393,
364
- "eval_steps_per_second": 19.197,
365
  "num_input_tokens_seen": 27552,
366
  "step": 156
367
  },
368
  {
369
  "epoch": 0.642570281124498,
370
- "grad_norm": 1.6281641721725464,
371
  "learning_rate": 1.7408081372259632e-05,
372
- "loss": 0.3618,
373
  "num_input_tokens_seen": 28272,
374
  "step": 160
375
  },
376
  {
377
  "epoch": 0.6626506024096386,
378
- "grad_norm": 5.65364933013916,
379
  "learning_rate": 1.5757778980982626e-05,
380
- "loss": 0.3202,
381
  "num_input_tokens_seen": 29184,
382
  "step": 165
383
  },
384
  {
385
  "epoch": 0.678714859437751,
386
- "eval_loss": 0.3523559272289276,
387
- "eval_runtime": 0.8084,
388
- "eval_samples_per_second": 69.268,
389
- "eval_steps_per_second": 34.634,
390
  "num_input_tokens_seen": 29984,
391
  "step": 169
392
  },
393
  {
394
  "epoch": 0.6827309236947792,
395
- "grad_norm": 2.0312561988830566,
396
  "learning_rate": 1.4152906522061048e-05,
397
- "loss": 0.3334,
398
  "num_input_tokens_seen": 30128,
399
  "step": 170
400
  },
401
  {
402
  "epoch": 0.7028112449799196,
403
- "grad_norm": 1.9656075239181519,
404
  "learning_rate": 1.2601352710458313e-05,
405
- "loss": 0.3209,
406
  "num_input_tokens_seen": 30976,
407
  "step": 175
408
  },
409
  {
410
  "epoch": 0.7228915662650602,
411
- "grad_norm": 5.087824821472168,
412
  "learning_rate": 1.1110744174509952e-05,
413
- "loss": 0.3678,
414
  "num_input_tokens_seen": 31776,
415
  "step": 180
416
  },
417
  {
418
  "epoch": 0.7309236947791165,
419
- "eval_loss": 0.3510570526123047,
420
- "eval_runtime": 0.7299,
421
- "eval_samples_per_second": 76.72,
422
- "eval_steps_per_second": 38.36,
423
  "num_input_tokens_seen": 32080,
424
  "step": 182
425
  },
426
  {
427
  "epoch": 0.7429718875502008,
428
- "grad_norm": 7.287527084350586,
429
  "learning_rate": 9.688407967401248e-06,
430
- "loss": 0.369,
431
  "num_input_tokens_seen": 32608,
432
  "step": 185
433
  },
434
  {
435
  "epoch": 0.7630522088353414,
436
- "grad_norm": 6.0406174659729,
437
  "learning_rate": 8.341335551199902e-06,
438
- "loss": 0.3996,
439
  "num_input_tokens_seen": 33360,
440
  "step": 190
441
  },
442
  {
443
  "epoch": 0.7831325301204819,
444
- "grad_norm": 7.134504795074463,
445
  "learning_rate": 7.076148430479321e-06,
446
- "loss": 0.3704,
447
  "num_input_tokens_seen": 34176,
448
  "step": 195
449
  },
450
  {
451
  "epoch": 0.7831325301204819,
452
- "eval_loss": 0.35649141669273376,
453
- "eval_runtime": 0.7576,
454
- "eval_samples_per_second": 73.922,
455
- "eval_steps_per_second": 36.961,
456
  "num_input_tokens_seen": 34176,
457
  "step": 195
458
  },
459
  {
460
  "epoch": 0.8032128514056225,
461
- "grad_norm": 1.6589281558990479,
462
  "learning_rate": 5.899065604459814e-06,
463
- "loss": 0.3519,
464
  "num_input_tokens_seen": 34992,
465
  "step": 200
466
  },
467
  {
468
  "epoch": 0.8232931726907631,
469
- "grad_norm": 1.7492696046829224,
470
  "learning_rate": 4.81587299765594e-06,
471
- "loss": 0.3651,
472
  "num_input_tokens_seen": 35888,
473
  "step": 205
474
  },
475
  {
476
  "epoch": 0.8353413654618473,
477
- "eval_loss": 0.35078784823417664,
478
- "eval_runtime": 0.7392,
479
- "eval_samples_per_second": 75.761,
480
- "eval_steps_per_second": 37.88,
481
  "num_input_tokens_seen": 36512,
482
  "step": 208
483
  },
484
  {
485
  "epoch": 0.8433734939759037,
486
- "grad_norm": 1.6302882432937622,
487
  "learning_rate": 3.831895019292897e-06,
488
- "loss": 0.3677,
489
  "num_input_tokens_seen": 36848,
490
  "step": 210
491
  },
492
  {
493
  "epoch": 0.8634538152610441,
494
- "grad_norm": 2.224395990371704,
495
  "learning_rate": 2.9519683912911266e-06,
496
- "loss": 0.372,
497
  "num_input_tokens_seen": 37888,
498
  "step": 215
499
  },
500
  {
501
  "epoch": 0.8835341365461847,
502
- "grad_norm": 1.8795971870422363,
503
  "learning_rate": 2.1804183734670277e-06,
504
- "loss": 0.3666,
505
  "num_input_tokens_seen": 38768,
506
  "step": 220
507
  },
508
  {
509
  "epoch": 0.8875502008032129,
510
- "eval_loss": 0.3531496226787567,
511
- "eval_runtime": 0.8251,
512
- "eval_samples_per_second": 67.869,
513
- "eval_steps_per_second": 33.935,
514
  "num_input_tokens_seen": 38912,
515
  "step": 221
516
  },
517
  {
518
  "epoch": 0.9036144578313253,
519
- "grad_norm": 1.104848861694336,
520
  "learning_rate": 1.5210375028143097e-06,
521
- "loss": 0.3642,
522
  "num_input_tokens_seen": 39488,
523
  "step": 225
524
  },
525
  {
526
  "epoch": 0.9236947791164659,
527
- "grad_norm": 1.8672505617141724,
528
  "learning_rate": 9.770669513725128e-07,
529
- "loss": 0.3489,
530
  "num_input_tokens_seen": 40336,
531
  "step": 230
532
  },
533
  {
534
  "epoch": 0.9397590361445783,
535
- "eval_loss": 0.35156872868537903,
536
- "eval_runtime": 0.9032,
537
- "eval_samples_per_second": 62.003,
538
- "eval_steps_per_second": 31.001,
539
  "num_input_tokens_seen": 41120,
540
  "step": 234
541
  },
542
  {
543
  "epoch": 0.9437751004016064,
544
- "grad_norm": 1.5349013805389404,
545
  "learning_rate": 5.5118059431781e-07,
546
- "loss": 0.3569,
547
  "num_input_tokens_seen": 41328,
548
  "step": 235
549
  },
550
  {
551
  "epoch": 0.963855421686747,
552
- "grad_norm": 1.8111923933029175,
553
  "learning_rate": 2.454718665888589e-07,
554
- "loss": 0.3432,
555
  "num_input_tokens_seen": 42176,
556
  "step": 240
557
  },
558
  {
559
  "epoch": 0.9839357429718876,
560
- "grad_norm": 1.2632478475570679,
561
  "learning_rate": 6.14434726538493e-08,
562
- "loss": 0.3405,
563
  "num_input_tokens_seen": 43312,
564
  "step": 245
565
  },
566
  {
567
  "epoch": 0.9919678714859438,
568
- "eval_loss": 0.3515389859676361,
569
- "eval_runtime": 0.7167,
570
- "eval_samples_per_second": 78.134,
571
- "eval_steps_per_second": 39.067,
572
  "num_input_tokens_seen": 43600,
573
  "step": 247
574
  },
@@ -577,10 +577,10 @@
577
  "num_input_tokens_seen": 43904,
578
  "step": 249,
579
  "total_flos": 278458437992448.0,
580
- "train_loss": 0.39900598229174633,
581
- "train_runtime": 65.0582,
582
- "train_samples_per_second": 7.655,
583
- "train_steps_per_second": 3.827
584
  }
585
  ],
586
  "logging_steps": 5,
 
1
  {
2
+ "best_global_step": 234,
3
+ "best_metric": 0.3496658504009247,
4
+ "best_model_checkpoint": "saves/test/checkpoint-234",
5
  "epoch": 1.0,
6
  "eval_steps": 13,
7
  "global_step": 249,
 
11
  "log_history": [
12
  {
13
  "epoch": 0.020080321285140562,
14
+ "grad_norm": 30.13371467590332,
15
  "learning_rate": 8.000000000000001e-06,
16
+ "loss": 0.8454,
17
  "num_input_tokens_seen": 832,
18
  "step": 5
19
  },
20
  {
21
  "epoch": 0.040160642570281124,
22
+ "grad_norm": 20.238059997558594,
23
  "learning_rate": 1.8e-05,
24
+ "loss": 0.7689,
25
  "num_input_tokens_seen": 1760,
26
  "step": 10
27
  },
28
  {
29
  "epoch": 0.05220883534136546,
30
+ "eval_loss": 0.6837943196296692,
31
+ "eval_runtime": 0.9982,
32
+ "eval_samples_per_second": 56.101,
33
+ "eval_steps_per_second": 28.051,
34
  "num_input_tokens_seen": 2288,
35
  "step": 13
36
  },
37
  {
38
  "epoch": 0.060240963855421686,
39
+ "grad_norm": 9.776243209838867,
40
  "learning_rate": 2.8000000000000003e-05,
41
+ "loss": 0.6956,
42
  "num_input_tokens_seen": 2608,
43
  "step": 15
44
  },
45
  {
46
  "epoch": 0.08032128514056225,
47
+ "grad_norm": 17.104515075683594,
48
  "learning_rate": 3.8e-05,
49
+ "loss": 0.5552,
50
  "num_input_tokens_seen": 3536,
51
  "step": 20
52
  },
53
  {
54
  "epoch": 0.10040160642570281,
55
+ "grad_norm": 11.06977653503418,
56
  "learning_rate": 4.8e-05,
57
+ "loss": 0.6557,
58
  "num_input_tokens_seen": 4496,
59
  "step": 25
60
  },
61
  {
62
  "epoch": 0.10441767068273092,
63
+ "eval_loss": 0.46041956543922424,
64
+ "eval_runtime": 0.7224,
65
+ "eval_samples_per_second": 77.516,
66
+ "eval_steps_per_second": 38.758,
67
  "num_input_tokens_seen": 4656,
68
  "step": 26
69
  },
70
  {
71
  "epoch": 0.12048192771084337,
72
+ "grad_norm": 8.536067008972168,
73
  "learning_rate": 4.996067037544542e-05,
74
+ "loss": 0.4952,
75
  "num_input_tokens_seen": 5424,
76
  "step": 30
77
  },
78
  {
79
  "epoch": 0.14056224899598393,
80
+ "grad_norm": 7.301750659942627,
81
  "learning_rate": 4.980110583549062e-05,
82
+ "loss": 0.3647,
83
  "num_input_tokens_seen": 6304,
84
  "step": 35
85
  },
86
  {
87
  "epoch": 0.1566265060240964,
88
+ "eval_loss": 0.3834764063358307,
89
+ "eval_runtime": 0.753,
90
+ "eval_samples_per_second": 74.371,
91
+ "eval_steps_per_second": 37.185,
92
  "num_input_tokens_seen": 6944,
93
  "step": 39
94
  },
95
  {
96
  "epoch": 0.1606425702811245,
97
+ "grad_norm": 2.0650792121887207,
98
  "learning_rate": 4.951963201008076e-05,
99
+ "loss": 0.4023,
100
  "num_input_tokens_seen": 7072,
101
  "step": 40
102
  },
103
  {
104
  "epoch": 0.18072289156626506,
105
+ "grad_norm": 6.442427158355713,
106
  "learning_rate": 4.91176324775594e-05,
107
+ "loss": 0.3811,
108
  "num_input_tokens_seen": 7856,
109
  "step": 45
110
  },
111
  {
112
  "epoch": 0.20080321285140562,
113
+ "grad_norm": 9.193267822265625,
114
  "learning_rate": 4.8597083257709194e-05,
115
+ "loss": 0.3506,
116
  "num_input_tokens_seen": 8880,
117
  "step": 50
118
  },
119
  {
120
  "epoch": 0.20883534136546184,
121
+ "eval_loss": 0.38357535004615784,
122
+ "eval_runtime": 0.8575,
123
+ "eval_samples_per_second": 65.305,
124
+ "eval_steps_per_second": 32.653,
125
  "num_input_tokens_seen": 9232,
126
  "step": 52
127
  },
128
  {
129
  "epoch": 0.22088353413654618,
130
+ "grad_norm": 2.6214115619659424,
131
  "learning_rate": 4.796054309867053e-05,
132
+ "loss": 0.4133,
133
  "num_input_tokens_seen": 9680,
134
  "step": 55
135
  },
136
  {
137
  "epoch": 0.24096385542168675,
138
+ "grad_norm": 6.665003776550293,
139
  "learning_rate": 4.721114089947181e-05,
140
+ "loss": 0.3351,
141
  "num_input_tokens_seen": 10576,
142
  "step": 60
143
  },
144
  {
145
  "epoch": 0.26104417670682734,
146
+ "grad_norm": 2.606006383895874,
147
  "learning_rate": 4.6352560329995686e-05,
148
+ "loss": 0.3084,
149
  "num_input_tokens_seen": 11424,
150
  "step": 65
151
  },
152
  {
153
  "epoch": 0.26104417670682734,
154
+ "eval_loss": 0.36912769079208374,
155
+ "eval_runtime": 0.9476,
156
+ "eval_samples_per_second": 59.094,
157
+ "eval_steps_per_second": 29.547,
158
  "num_input_tokens_seen": 11424,
159
  "step": 65
160
  },
161
  {
162
  "epoch": 0.28112449799196787,
163
+ "grad_norm": 2.763232946395874,
164
  "learning_rate": 4.538902172398151e-05,
165
+ "loss": 0.374,
166
  "num_input_tokens_seen": 12224,
167
  "step": 70
168
  },
169
  {
170
  "epoch": 0.30120481927710846,
171
+ "grad_norm": 1.5414751768112183,
172
  "learning_rate": 4.4325261334068426e-05,
173
+ "loss": 0.3649,
174
  "num_input_tokens_seen": 13168,
175
  "step": 75
176
  },
177
  {
178
  "epoch": 0.3132530120481928,
179
+ "eval_loss": 0.3668709397315979,
180
+ "eval_runtime": 0.8423,
181
+ "eval_samples_per_second": 66.484,
182
+ "eval_steps_per_second": 33.242,
183
  "num_input_tokens_seen": 13760,
184
  "step": 78
185
  },
186
  {
187
  "epoch": 0.321285140562249,
188
+ "grad_norm": 7.335175514221191,
189
  "learning_rate": 4.316650805085068e-05,
190
+ "loss": 0.3613,
191
  "num_input_tokens_seen": 14080,
192
  "step": 80
193
  },
194
  {
195
  "epoch": 0.3413654618473896,
196
+ "grad_norm": 6.502185344696045,
197
  "learning_rate": 4.1918457700381855e-05,
198
+ "loss": 0.3601,
199
  "num_input_tokens_seen": 15056,
200
  "step": 85
201
  },
202
  {
203
  "epoch": 0.3614457831325301,
204
+ "grad_norm": 6.903041362762451,
205
  "learning_rate": 4.058724504646834e-05,
206
+ "loss": 0.3612,
207
  "num_input_tokens_seen": 15904,
208
  "step": 90
209
  },
210
  {
211
  "epoch": 0.3654618473895582,
212
+ "eval_loss": 0.36205965280532837,
213
+ "eval_runtime": 0.7854,
214
+ "eval_samples_per_second": 71.302,
215
+ "eval_steps_per_second": 35.651,
216
  "num_input_tokens_seen": 16048,
217
  "step": 91
218
  },
219
  {
220
  "epoch": 0.3815261044176707,
221
+ "grad_norm": 6.267019748687744,
222
  "learning_rate": 3.9179413635373897e-05,
223
+ "loss": 0.3307,
224
  "num_input_tokens_seen": 16688,
225
  "step": 95
226
  },
227
  {
228
  "epoch": 0.40160642570281124,
229
+ "grad_norm": 6.203659534454346,
230
  "learning_rate": 3.770188363116324e-05,
231
+ "loss": 0.2896,
232
  "num_input_tokens_seen": 17552,
233
  "step": 100
234
  },
235
  {
236
  "epoch": 0.41767068273092367,
237
+ "eval_loss": 0.37523433566093445,
238
+ "eval_runtime": 0.7387,
239
+ "eval_samples_per_second": 75.808,
240
+ "eval_steps_per_second": 37.904,
241
  "num_input_tokens_seen": 18272,
242
  "step": 104
243
  },
244
  {
245
  "epoch": 0.42168674698795183,
246
+ "grad_norm": 2.0911366939544678,
247
  "learning_rate": 3.616191779978907e-05,
248
+ "loss": 0.3154,
249
  "num_input_tokens_seen": 18400,
250
  "step": 105
251
  },
252
  {
253
  "epoch": 0.44176706827309237,
254
+ "grad_norm": 5.106855392456055,
255
  "learning_rate": 3.456708580912725e-05,
256
+ "loss": 0.3694,
257
  "num_input_tokens_seen": 19456,
258
  "step": 110
259
  },
260
  {
261
  "epoch": 0.46184738955823296,
262
+ "grad_norm": 8.49660587310791,
263
  "learning_rate": 3.292522702044221e-05,
264
+ "loss": 0.4278,
265
  "num_input_tokens_seen": 20288,
266
  "step": 115
267
  },
268
  {
269
  "epoch": 0.46987951807228917,
270
+ "eval_loss": 0.3691355884075165,
271
+ "eval_runtime": 0.7901,
272
+ "eval_samples_per_second": 70.875,
273
+ "eval_steps_per_second": 35.437,
274
  "num_input_tokens_seen": 20656,
275
  "step": 117
276
  },
277
  {
278
  "epoch": 0.4819277108433735,
279
+ "grad_norm": 2.1965019702911377,
280
  "learning_rate": 3.1244411954180676e-05,
281
+ "loss": 0.3821,
282
  "num_input_tokens_seen": 21328,
283
  "step": 120
284
  },
285
  {
286
  "epoch": 0.5020080321285141,
287
+ "grad_norm": 1.8442065715789795,
288
  "learning_rate": 2.9532902619507462e-05,
289
+ "loss": 0.3707,
290
  "num_input_tokens_seen": 22304,
291
  "step": 125
292
  },
293
  {
294
  "epoch": 0.5220883534136547,
295
+ "grad_norm": 6.632619857788086,
296
  "learning_rate": 2.7799111902582696e-05,
297
+ "loss": 0.3591,
298
  "num_input_tokens_seen": 23056,
299
  "step": 130
300
  },
301
  {
302
  "epoch": 0.5220883534136547,
303
+ "eval_loss": 0.3583144545555115,
304
+ "eval_runtime": 0.9332,
305
+ "eval_samples_per_second": 60.01,
306
+ "eval_steps_per_second": 30.005,
307
  "num_input_tokens_seen": 23056,
308
  "step": 130
309
  },
310
  {
311
  "epoch": 0.5421686746987951,
312
+ "grad_norm": 7.6188459396362305,
313
  "learning_rate": 2.6051562213206632e-05,
314
+ "loss": 0.3438,
315
  "num_input_tokens_seen": 23840,
316
  "step": 135
317
  },
318
  {
319
  "epoch": 0.5622489959839357,
320
+ "grad_norm": 1.9679665565490723,
321
  "learning_rate": 2.429884359310328e-05,
322
+ "loss": 0.3726,
323
  "num_input_tokens_seen": 24832,
324
  "step": 140
325
  },
326
  {
327
  "epoch": 0.5742971887550201,
328
+ "eval_loss": 0.3531239330768585,
329
+ "eval_runtime": 0.8476,
330
+ "eval_samples_per_second": 66.071,
331
+ "eval_steps_per_second": 33.036,
332
  "num_input_tokens_seen": 25312,
333
  "step": 143
334
  },
335
  {
336
  "epoch": 0.5823293172690763,
337
+ "grad_norm": 1.664465069770813,
338
  "learning_rate": 2.2549571491760986e-05,
339
+ "loss": 0.3791,
340
  "num_input_tokens_seen": 25648,
341
  "step": 145
342
  },
343
  {
344
  "epoch": 0.6024096385542169,
345
+ "grad_norm": 7.226130962371826,
346
  "learning_rate": 2.0812344417381595e-05,
347
+ "loss": 0.3411,
348
  "num_input_tokens_seen": 26496,
349
  "step": 150
350
  },
351
  {
352
  "epoch": 0.6224899598393574,
353
+ "grad_norm": 7.071829319000244,
354
  "learning_rate": 1.909570167110415e-05,
355
+ "loss": 0.3829,
356
  "num_input_tokens_seen": 27392,
357
  "step": 155
358
  },
359
  {
360
  "epoch": 0.6265060240963856,
361
+ "eval_loss": 0.35204586386680603,
362
+ "eval_runtime": 0.805,
363
+ "eval_samples_per_second": 69.565,
364
+ "eval_steps_per_second": 34.782,
365
  "num_input_tokens_seen": 27552,
366
  "step": 156
367
  },
368
  {
369
  "epoch": 0.642570281124498,
370
+ "grad_norm": 1.6194074153900146,
371
  "learning_rate": 1.7408081372259632e-05,
372
+ "loss": 0.3653,
373
  "num_input_tokens_seen": 28272,
374
  "step": 160
375
  },
376
  {
377
  "epoch": 0.6626506024096386,
378
+ "grad_norm": 5.642275810241699,
379
  "learning_rate": 1.5757778980982626e-05,
380
+ "loss": 0.3318,
381
  "num_input_tokens_seen": 29184,
382
  "step": 165
383
  },
384
  {
385
  "epoch": 0.678714859437751,
386
+ "eval_loss": 0.35015299916267395,
387
+ "eval_runtime": 0.7303,
388
+ "eval_samples_per_second": 76.682,
389
+ "eval_steps_per_second": 38.341,
390
  "num_input_tokens_seen": 29984,
391
  "step": 169
392
  },
393
  {
394
  "epoch": 0.6827309236947792,
395
+ "grad_norm": 1.980563759803772,
396
  "learning_rate": 1.4152906522061048e-05,
397
+ "loss": 0.337,
398
  "num_input_tokens_seen": 30128,
399
  "step": 170
400
  },
401
  {
402
  "epoch": 0.7028112449799196,
403
+ "grad_norm": 2.214860200881958,
404
  "learning_rate": 1.2601352710458313e-05,
405
+ "loss": 0.3303,
406
  "num_input_tokens_seen": 30976,
407
  "step": 175
408
  },
409
  {
410
  "epoch": 0.7228915662650602,
411
+ "grad_norm": 5.042296886444092,
412
  "learning_rate": 1.1110744174509952e-05,
413
+ "loss": 0.3655,
414
  "num_input_tokens_seen": 31776,
415
  "step": 180
416
  },
417
  {
418
  "epoch": 0.7309236947791165,
419
+ "eval_loss": 0.3543336093425751,
420
+ "eval_runtime": 0.8576,
421
+ "eval_samples_per_second": 65.301,
422
+ "eval_steps_per_second": 32.65,
423
  "num_input_tokens_seen": 32080,
424
  "step": 182
425
  },
426
  {
427
  "epoch": 0.7429718875502008,
428
+ "grad_norm": 7.1150078773498535,
429
  "learning_rate": 9.688407967401248e-06,
430
+ "loss": 0.3755,
431
  "num_input_tokens_seen": 32608,
432
  "step": 185
433
  },
434
  {
435
  "epoch": 0.7630522088353414,
436
+ "grad_norm": 6.07709264755249,
437
  "learning_rate": 8.341335551199902e-06,
438
+ "loss": 0.3967,
439
  "num_input_tokens_seen": 33360,
440
  "step": 190
441
  },
442
  {
443
  "epoch": 0.7831325301204819,
444
+ "grad_norm": 7.137906074523926,
445
  "learning_rate": 7.076148430479321e-06,
446
+ "loss": 0.3703,
447
  "num_input_tokens_seen": 34176,
448
  "step": 195
449
  },
450
  {
451
  "epoch": 0.7831325301204819,
452
+ "eval_loss": 0.35257503390312195,
453
+ "eval_runtime": 0.7915,
454
+ "eval_samples_per_second": 70.751,
455
+ "eval_steps_per_second": 35.376,
456
  "num_input_tokens_seen": 34176,
457
  "step": 195
458
  },
459
  {
460
  "epoch": 0.8032128514056225,
461
+ "grad_norm": 1.6914492845535278,
462
  "learning_rate": 5.899065604459814e-06,
463
+ "loss": 0.355,
464
  "num_input_tokens_seen": 34992,
465
  "step": 200
466
  },
467
  {
468
  "epoch": 0.8232931726907631,
469
+ "grad_norm": 1.7085988521575928,
470
  "learning_rate": 4.81587299765594e-06,
471
+ "loss": 0.3585,
472
  "num_input_tokens_seen": 35888,
473
  "step": 205
474
  },
475
  {
476
  "epoch": 0.8353413654618473,
477
+ "eval_loss": 0.3534963130950928,
478
+ "eval_runtime": 0.8554,
479
+ "eval_samples_per_second": 65.464,
480
+ "eval_steps_per_second": 32.732,
481
  "num_input_tokens_seen": 36512,
482
  "step": 208
483
  },
484
  {
485
  "epoch": 0.8433734939759037,
486
+ "grad_norm": 1.570613145828247,
487
  "learning_rate": 3.831895019292897e-06,
488
+ "loss": 0.3715,
489
  "num_input_tokens_seen": 36848,
490
  "step": 210
491
  },
492
  {
493
  "epoch": 0.8634538152610441,
494
+ "grad_norm": 2.1737060546875,
495
  "learning_rate": 2.9519683912911266e-06,
496
+ "loss": 0.3719,
497
  "num_input_tokens_seen": 37888,
498
  "step": 215
499
  },
500
  {
501
  "epoch": 0.8835341365461847,
502
+ "grad_norm": 1.8593710660934448,
503
  "learning_rate": 2.1804183734670277e-06,
504
+ "loss": 0.3626,
505
  "num_input_tokens_seen": 38768,
506
  "step": 220
507
  },
508
  {
509
  "epoch": 0.8875502008032129,
510
+ "eval_loss": 0.35169586539268494,
511
+ "eval_runtime": 0.8105,
512
+ "eval_samples_per_second": 69.096,
513
+ "eval_steps_per_second": 34.548,
514
  "num_input_tokens_seen": 38912,
515
  "step": 221
516
  },
517
  {
518
  "epoch": 0.9036144578313253,
519
+ "grad_norm": 1.1120434999465942,
520
  "learning_rate": 1.5210375028143097e-06,
521
+ "loss": 0.3549,
522
  "num_input_tokens_seen": 39488,
523
  "step": 225
524
  },
525
  {
526
  "epoch": 0.9236947791164659,
527
+ "grad_norm": 1.7197054624557495,
528
  "learning_rate": 9.770669513725128e-07,
529
+ "loss": 0.3419,
530
  "num_input_tokens_seen": 40336,
531
  "step": 230
532
  },
533
  {
534
  "epoch": 0.9397590361445783,
535
+ "eval_loss": 0.3496658504009247,
536
+ "eval_runtime": 0.8315,
537
+ "eval_samples_per_second": 67.352,
538
+ "eval_steps_per_second": 33.676,
539
  "num_input_tokens_seen": 41120,
540
  "step": 234
541
  },
542
  {
543
  "epoch": 0.9437751004016064,
544
+ "grad_norm": 1.5268943309783936,
545
  "learning_rate": 5.5118059431781e-07,
546
+ "loss": 0.3635,
547
  "num_input_tokens_seen": 41328,
548
  "step": 235
549
  },
550
  {
551
  "epoch": 0.963855421686747,
552
+ "grad_norm": 1.8964738845825195,
553
  "learning_rate": 2.454718665888589e-07,
554
+ "loss": 0.3463,
555
  "num_input_tokens_seen": 42176,
556
  "step": 240
557
  },
558
  {
559
  "epoch": 0.9839357429718876,
560
+ "grad_norm": 1.2437193393707275,
561
  "learning_rate": 6.14434726538493e-08,
562
+ "loss": 0.3311,
563
  "num_input_tokens_seen": 43312,
564
  "step": 245
565
  },
566
  {
567
  "epoch": 0.9919678714859438,
568
+ "eval_loss": 0.3558541238307953,
569
+ "eval_runtime": 0.7677,
570
+ "eval_samples_per_second": 72.948,
571
+ "eval_steps_per_second": 36.474,
572
  "num_input_tokens_seen": 43600,
573
  "step": 247
574
  },
 
577
  "num_input_tokens_seen": 43904,
578
  "step": 249,
579
  "total_flos": 278458437992448.0,
580
+ "train_loss": 0.3966633373474979,
581
+ "train_runtime": 65.6097,
582
+ "train_samples_per_second": 7.59,
583
+ "train_steps_per_second": 3.795
584
  }
585
  ],
586
  "logging_steps": 5,
training_eval_loss.png CHANGED
training_loss.png CHANGED