File size: 18,195 Bytes
0f3163f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
{
  "best_global_step": null,
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.18467220683287167,
  "eval_steps": 500,
  "global_step": 300,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0030778701138811943,
      "grad_norm": 0.47482776641845703,
      "learning_rate": 4.9999916942231176e-05,
      "loss": 1.1992,
      "num_input_tokens_seen": 42384,
      "step": 5,
      "train_runtime": 19.849,
      "train_tokens_per_second": 2135.325
    },
    {
      "epoch": 0.006155740227762389,
      "grad_norm": 0.24747319519519806,
      "learning_rate": 4.999957952099119e-05,
      "loss": 1.0247,
      "num_input_tokens_seen": 86000,
      "step": 10,
      "train_runtime": 39.0811,
      "train_tokens_per_second": 2200.554
    },
    {
      "epoch": 0.009233610341643583,
      "grad_norm": 0.10441822558641434,
      "learning_rate": 4.9998982548669973e-05,
      "loss": 0.8631,
      "num_input_tokens_seen": 124784,
      "step": 15,
      "train_runtime": 57.1682,
      "train_tokens_per_second": 2182.751
    },
    {
      "epoch": 0.012311480455524777,
      "grad_norm": 0.11058320850133896,
      "learning_rate": 4.999812603146542e-05,
      "loss": 0.9173,
      "num_input_tokens_seen": 165952,
      "step": 20,
      "train_runtime": 75.5733,
      "train_tokens_per_second": 2195.909
    },
    {
      "epoch": 0.015389350569405972,
      "grad_norm": 0.09879708290100098,
      "learning_rate": 4.9997009978270085e-05,
      "loss": 0.8722,
      "num_input_tokens_seen": 201584,
      "step": 25,
      "train_runtime": 92.5588,
      "train_tokens_per_second": 2177.903
    },
    {
      "epoch": 0.018467220683287166,
      "grad_norm": 0.11391397565603256,
      "learning_rate": 4.999563440067108e-05,
      "loss": 0.866,
      "num_input_tokens_seen": 240656,
      "step": 30,
      "train_runtime": 111.1373,
      "train_tokens_per_second": 2165.393
    },
    {
      "epoch": 0.02154509079716836,
      "grad_norm": 0.09722574800252914,
      "learning_rate": 4.9993999312949936e-05,
      "loss": 0.853,
      "num_input_tokens_seen": 287248,
      "step": 35,
      "train_runtime": 130.6684,
      "train_tokens_per_second": 2198.297
    },
    {
      "epoch": 0.024622960911049555,
      "grad_norm": 0.12133017182350159,
      "learning_rate": 4.99921047320825e-05,
      "loss": 0.88,
      "num_input_tokens_seen": 328448,
      "step": 40,
      "train_runtime": 149.193,
      "train_tokens_per_second": 2201.498
    },
    {
      "epoch": 0.027700831024930747,
      "grad_norm": 0.12174953520298004,
      "learning_rate": 4.9989950677738726e-05,
      "loss": 0.8503,
      "num_input_tokens_seen": 370512,
      "step": 45,
      "train_runtime": 168.2236,
      "train_tokens_per_second": 2202.497
    },
    {
      "epoch": 0.030778701138811943,
      "grad_norm": 0.12440948933362961,
      "learning_rate": 4.9987537172282466e-05,
      "loss": 0.8707,
      "num_input_tokens_seen": 409936,
      "step": 50,
      "train_runtime": 186.7223,
      "train_tokens_per_second": 2195.432
    },
    {
      "epoch": 0.03385657125269314,
      "grad_norm": 0.12797962129116058,
      "learning_rate": 4.998486424077127e-05,
      "loss": 0.8475,
      "num_input_tokens_seen": 450928,
      "step": 55,
      "train_runtime": 205.1107,
      "train_tokens_per_second": 2198.462
    },
    {
      "epoch": 0.03693444136657433,
      "grad_norm": 0.09793120622634888,
      "learning_rate": 4.9981931910956086e-05,
      "loss": 0.8409,
      "num_input_tokens_seen": 491152,
      "step": 60,
      "train_runtime": 223.3671,
      "train_tokens_per_second": 2198.856
    },
    {
      "epoch": 0.040012311480455524,
      "grad_norm": 0.10957301408052444,
      "learning_rate": 4.9978740213281005e-05,
      "loss": 0.8524,
      "num_input_tokens_seen": 532160,
      "step": 65,
      "train_runtime": 241.5682,
      "train_tokens_per_second": 2202.939
    },
    {
      "epoch": 0.04309018159433672,
      "grad_norm": 0.12283050268888474,
      "learning_rate": 4.9975289180882936e-05,
      "loss": 0.8944,
      "num_input_tokens_seen": 573392,
      "step": 70,
      "train_runtime": 259.8567,
      "train_tokens_per_second": 2206.57
    },
    {
      "epoch": 0.046168051708217916,
      "grad_norm": 0.1338459849357605,
      "learning_rate": 4.9971578849591225e-05,
      "loss": 0.8478,
      "num_input_tokens_seen": 613488,
      "step": 75,
      "train_runtime": 278.0456,
      "train_tokens_per_second": 2206.429
    },
    {
      "epoch": 0.04924592182209911,
      "grad_norm": 0.10969562828540802,
      "learning_rate": 4.996760925792737e-05,
      "loss": 0.8331,
      "num_input_tokens_seen": 656480,
      "step": 80,
      "train_runtime": 296.3508,
      "train_tokens_per_second": 2215.212
    },
    {
      "epoch": 0.0523237919359803,
      "grad_norm": 0.1411936730146408,
      "learning_rate": 4.996338044710452e-05,
      "loss": 0.8306,
      "num_input_tokens_seen": 698272,
      "step": 85,
      "train_runtime": 315.096,
      "train_tokens_per_second": 2216.061
    },
    {
      "epoch": 0.055401662049861494,
      "grad_norm": 0.1435951143503189,
      "learning_rate": 4.995889246102711e-05,
      "loss": 0.8318,
      "num_input_tokens_seen": 735376,
      "step": 90,
      "train_runtime": 332.4774,
      "train_tokens_per_second": 2211.807
    },
    {
      "epoch": 0.05847953216374269,
      "grad_norm": 0.15133805572986603,
      "learning_rate": 4.995414534629039e-05,
      "loss": 0.8674,
      "num_input_tokens_seen": 772544,
      "step": 95,
      "train_runtime": 350.9252,
      "train_tokens_per_second": 2201.449
    },
    {
      "epoch": 0.061557402277623886,
      "grad_norm": 0.10615513473749161,
      "learning_rate": 4.9949139152179944e-05,
      "loss": 0.901,
      "num_input_tokens_seen": 814928,
      "step": 100,
      "train_runtime": 369.5452,
      "train_tokens_per_second": 2205.219
    },
    {
      "epoch": 0.06463527239150507,
      "grad_norm": 0.10441838204860687,
      "learning_rate": 4.994387393067117e-05,
      "loss": 0.8522,
      "num_input_tokens_seen": 855104,
      "step": 105,
      "train_runtime": 388.7482,
      "train_tokens_per_second": 2199.635
    },
    {
      "epoch": 0.06771314250538628,
      "grad_norm": 0.1405644714832306,
      "learning_rate": 4.993834973642874e-05,
      "loss": 0.8614,
      "num_input_tokens_seen": 896272,
      "step": 110,
      "train_runtime": 406.8129,
      "train_tokens_per_second": 2203.155
    },
    {
      "epoch": 0.07079101261926747,
      "grad_norm": 0.11736899614334106,
      "learning_rate": 4.993256662680604e-05,
      "loss": 0.8371,
      "num_input_tokens_seen": 940864,
      "step": 115,
      "train_runtime": 425.836,
      "train_tokens_per_second": 2209.452
    },
    {
      "epoch": 0.07386888273314866,
      "grad_norm": 0.1269819736480713,
      "learning_rate": 4.992652466184458e-05,
      "loss": 0.9,
      "num_input_tokens_seen": 983360,
      "step": 120,
      "train_runtime": 444.5802,
      "train_tokens_per_second": 2211.884
    },
    {
      "epoch": 0.07694675284702986,
      "grad_norm": 0.10944321751594543,
      "learning_rate": 4.992022390427335e-05,
      "loss": 0.8629,
      "num_input_tokens_seen": 1025120,
      "step": 125,
      "train_runtime": 463.3805,
      "train_tokens_per_second": 2212.264
    },
    {
      "epoch": 0.08002462296091105,
      "grad_norm": 0.1438760608434677,
      "learning_rate": 4.99136644195082e-05,
      "loss": 0.8797,
      "num_input_tokens_seen": 1065360,
      "step": 130,
      "train_runtime": 480.9152,
      "train_tokens_per_second": 2215.276
    },
    {
      "epoch": 0.08310249307479224,
      "grad_norm": 0.13859188556671143,
      "learning_rate": 4.9906846275651125e-05,
      "loss": 0.8317,
      "num_input_tokens_seen": 1106112,
      "step": 135,
      "train_runtime": 498.8557,
      "train_tokens_per_second": 2217.299
    },
    {
      "epoch": 0.08618036318867343,
      "grad_norm": 0.1412304937839508,
      "learning_rate": 4.989976954348958e-05,
      "loss": 0.8422,
      "num_input_tokens_seen": 1149632,
      "step": 140,
      "train_runtime": 517.2481,
      "train_tokens_per_second": 2222.593
    },
    {
      "epoch": 0.08925823330255463,
      "grad_norm": 0.1513959765434265,
      "learning_rate": 4.989243429649573e-05,
      "loss": 0.8403,
      "num_input_tokens_seen": 1187824,
      "step": 145,
      "train_runtime": 535.2196,
      "train_tokens_per_second": 2219.321
    },
    {
      "epoch": 0.09233610341643583,
      "grad_norm": 0.13685107231140137,
      "learning_rate": 4.988484061082572e-05,
      "loss": 0.8961,
      "num_input_tokens_seen": 1229088,
      "step": 150,
      "train_runtime": 553.4777,
      "train_tokens_per_second": 2220.664
    },
    {
      "epoch": 0.09541397353031703,
      "grad_norm": 0.12999333441257477,
      "learning_rate": 4.987698856531884e-05,
      "loss": 0.814,
      "num_input_tokens_seen": 1273056,
      "step": 155,
      "train_runtime": 572.9621,
      "train_tokens_per_second": 2221.885
    },
    {
      "epoch": 0.09849184364419822,
      "grad_norm": 0.1481194943189621,
      "learning_rate": 4.986887824149674e-05,
      "loss": 0.8542,
      "num_input_tokens_seen": 1310288,
      "step": 160,
      "train_runtime": 589.8405,
      "train_tokens_per_second": 2221.427
    },
    {
      "epoch": 0.10156971375807941,
      "grad_norm": 0.1396579146385193,
      "learning_rate": 4.9860509723562573e-05,
      "loss": 0.8722,
      "num_input_tokens_seen": 1351312,
      "step": 165,
      "train_runtime": 607.8908,
      "train_tokens_per_second": 2222.952
    },
    {
      "epoch": 0.1046475838719606,
      "grad_norm": 0.12688206136226654,
      "learning_rate": 4.985188309840012e-05,
      "loss": 0.8587,
      "num_input_tokens_seen": 1396336,
      "step": 170,
      "train_runtime": 626.8902,
      "train_tokens_per_second": 2227.401
    },
    {
      "epoch": 0.1077254539858418,
      "grad_norm": 0.13825534284114838,
      "learning_rate": 4.984299845557287e-05,
      "loss": 0.8014,
      "num_input_tokens_seen": 1437984,
      "step": 175,
      "train_runtime": 645.2717,
      "train_tokens_per_second": 2228.494
    },
    {
      "epoch": 0.11080332409972299,
      "grad_norm": 0.12385495752096176,
      "learning_rate": 4.983385588732312e-05,
      "loss": 0.8385,
      "num_input_tokens_seen": 1475360,
      "step": 180,
      "train_runtime": 662.6866,
      "train_tokens_per_second": 2226.332
    },
    {
      "epoch": 0.11388119421360418,
      "grad_norm": 0.12246419489383698,
      "learning_rate": 4.982445548857102e-05,
      "loss": 0.8372,
      "num_input_tokens_seen": 1513152,
      "step": 185,
      "train_runtime": 680.3316,
      "train_tokens_per_second": 2224.139
    },
    {
      "epoch": 0.11695906432748537,
      "grad_norm": 0.12281708419322968,
      "learning_rate": 4.981479735691354e-05,
      "loss": 0.8446,
      "num_input_tokens_seen": 1556256,
      "step": 190,
      "train_runtime": 698.7652,
      "train_tokens_per_second": 2227.152
    },
    {
      "epoch": 0.12003693444136658,
      "grad_norm": 0.13810615241527557,
      "learning_rate": 4.980488159262353e-05,
      "loss": 0.8141,
      "num_input_tokens_seen": 1598832,
      "step": 195,
      "train_runtime": 717.4389,
      "train_tokens_per_second": 2228.527
    },
    {
      "epoch": 0.12311480455524777,
      "grad_norm": 0.15435761213302612,
      "learning_rate": 4.97947082986486e-05,
      "loss": 0.8398,
      "num_input_tokens_seen": 1641712,
      "step": 200,
      "train_runtime": 736.4476,
      "train_tokens_per_second": 2229.231
    },
    {
      "epoch": 0.12619267466912895,
      "grad_norm": 0.15266934037208557,
      "learning_rate": 4.978427758061009e-05,
      "loss": 0.8714,
      "num_input_tokens_seen": 1684544,
      "step": 205,
      "train_runtime": 756.7894,
      "train_tokens_per_second": 2225.908
    },
    {
      "epoch": 0.12927054478301014,
      "grad_norm": 0.13250574469566345,
      "learning_rate": 4.977358954680202e-05,
      "loss": 0.8116,
      "num_input_tokens_seen": 1723120,
      "step": 210,
      "train_runtime": 774.2541,
      "train_tokens_per_second": 2225.523
    },
    {
      "epoch": 0.13234841489689136,
      "grad_norm": 0.11579591035842896,
      "learning_rate": 4.9762644308189865e-05,
      "loss": 0.8204,
      "num_input_tokens_seen": 1768096,
      "step": 215,
      "train_runtime": 793.0535,
      "train_tokens_per_second": 2229.479
    },
    {
      "epoch": 0.13542628501077256,
      "grad_norm": 0.12654677033424377,
      "learning_rate": 4.9751441978409484e-05,
      "loss": 0.8334,
      "num_input_tokens_seen": 1808176,
      "step": 220,
      "train_runtime": 810.905,
      "train_tokens_per_second": 2229.825
    },
    {
      "epoch": 0.13850415512465375,
      "grad_norm": 0.15607228875160217,
      "learning_rate": 4.97399826737659e-05,
      "loss": 0.8177,
      "num_input_tokens_seen": 1848480,
      "step": 225,
      "train_runtime": 829.3126,
      "train_tokens_per_second": 2228.93
    },
    {
      "epoch": 0.14158202523853494,
      "grad_norm": 0.1535269021987915,
      "learning_rate": 4.972826651323211e-05,
      "loss": 0.807,
      "num_input_tokens_seen": 1889168,
      "step": 230,
      "train_runtime": 847.0449,
      "train_tokens_per_second": 2230.304
    },
    {
      "epoch": 0.14465989535241613,
      "grad_norm": 0.13663041591644287,
      "learning_rate": 4.971629361844785e-05,
      "loss": 0.8557,
      "num_input_tokens_seen": 1930784,
      "step": 235,
      "train_runtime": 865.4457,
      "train_tokens_per_second": 2230.971
    },
    {
      "epoch": 0.14773776546629733,
      "grad_norm": 0.1719619184732437,
      "learning_rate": 4.9704064113718326e-05,
      "loss": 0.8436,
      "num_input_tokens_seen": 1974176,
      "step": 240,
      "train_runtime": 883.9962,
      "train_tokens_per_second": 2233.24
    },
    {
      "epoch": 0.15081563558017852,
      "grad_norm": 0.15830345451831818,
      "learning_rate": 4.9691578126012905e-05,
      "loss": 0.853,
      "num_input_tokens_seen": 2013008,
      "step": 245,
      "train_runtime": 902.1681,
      "train_tokens_per_second": 2231.3
    },
    {
      "epoch": 0.1538935056940597,
      "grad_norm": 0.14472956955432892,
      "learning_rate": 4.967883578496385e-05,
      "loss": 0.874,
      "num_input_tokens_seen": 2053328,
      "step": 250,
      "train_runtime": 919.7822,
      "train_tokens_per_second": 2232.407
    },
    {
      "epoch": 0.1569713758079409,
      "grad_norm": 0.13329900801181793,
      "learning_rate": 4.966583722286491e-05,
      "loss": 0.8333,
      "num_input_tokens_seen": 2094992,
      "step": 255,
      "train_runtime": 937.9537,
      "train_tokens_per_second": 2233.577
    },
    {
      "epoch": 0.1600492459218221,
      "grad_norm": 0.1551639437675476,
      "learning_rate": 4.9652582574669995e-05,
      "loss": 0.864,
      "num_input_tokens_seen": 2135376,
      "step": 260,
      "train_runtime": 955.8955,
      "train_tokens_per_second": 2233.901
    },
    {
      "epoch": 0.1631271160357033,
      "grad_norm": 0.12460612505674362,
      "learning_rate": 4.963907197799174e-05,
      "loss": 0.8321,
      "num_input_tokens_seen": 2176624,
      "step": 265,
      "train_runtime": 974.3283,
      "train_tokens_per_second": 2233.974
    },
    {
      "epoch": 0.16620498614958448,
      "grad_norm": 0.13981439173221588,
      "learning_rate": 4.9625305573100115e-05,
      "loss": 0.8349,
      "num_input_tokens_seen": 2220496,
      "step": 270,
      "train_runtime": 992.5551,
      "train_tokens_per_second": 2237.151
    },
    {
      "epoch": 0.16928285626346568,
      "grad_norm": 0.15424658358097076,
      "learning_rate": 4.961128350292094e-05,
      "loss": 0.8248,
      "num_input_tokens_seen": 2263808,
      "step": 275,
      "train_runtime": 1010.9218,
      "train_tokens_per_second": 2239.35
    },
    {
      "epoch": 0.17236072637734687,
      "grad_norm": 0.14288848638534546,
      "learning_rate": 4.959700591303438e-05,
      "loss": 0.8488,
      "num_input_tokens_seen": 2311696,
      "step": 280,
      "train_runtime": 1030.8143,
      "train_tokens_per_second": 2242.592
    },
    {
      "epoch": 0.17543859649122806,
      "grad_norm": 0.15639737248420715,
      "learning_rate": 4.95824729516735e-05,
      "loss": 0.8602,
      "num_input_tokens_seen": 2353200,
      "step": 285,
      "train_runtime": 1049.2401,
      "train_tokens_per_second": 2242.766
    },
    {
      "epoch": 0.17851646660510925,
      "grad_norm": 0.1268603354692459,
      "learning_rate": 4.9567684769722664e-05,
      "loss": 0.8987,
      "num_input_tokens_seen": 2396000,
      "step": 290,
      "train_runtime": 1067.7324,
      "train_tokens_per_second": 2244.008
    },
    {
      "epoch": 0.18159433671899045,
      "grad_norm": 0.14782211184501648,
      "learning_rate": 4.9552641520716014e-05,
      "loss": 0.7794,
      "num_input_tokens_seen": 2438032,
      "step": 295,
      "train_runtime": 1085.9257,
      "train_tokens_per_second": 2245.119
    },
    {
      "epoch": 0.18467220683287167,
      "grad_norm": 0.15259359776973724,
      "learning_rate": 4.953734336083583e-05,
      "loss": 0.8648,
      "num_input_tokens_seen": 2478048,
      "step": 300,
      "train_runtime": 1103.5538,
      "train_tokens_per_second": 2245.516
    }
  ],
  "logging_steps": 5,
  "max_steps": 4875,
  "num_input_tokens_seen": 2478048,
  "num_train_epochs": 3,
  "save_steps": 100,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 4.765645728959693e+17,
  "train_batch_size": 2,
  "trial_name": null,
  "trial_params": null
}