| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.2917653815780028, | |
| "eval_steps": 500, | |
| "global_step": 2250, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00574341302318903, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 1.9920000000000002e-05, | |
| "loss": 1.3417, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.01148682604637806, | |
| "grad_norm": 2.609375, | |
| "learning_rate": 1.9831111111111113e-05, | |
| "loss": 0.654, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.01723023906956709, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 1.9742222222222223e-05, | |
| "loss": 0.6043, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.02297365209275612, | |
| "grad_norm": 2.296875, | |
| "learning_rate": 1.9653333333333334e-05, | |
| "loss": 0.5744, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.02871706511594515, | |
| "grad_norm": 1.890625, | |
| "learning_rate": 1.9564444444444444e-05, | |
| "loss": 0.563, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03446047813913418, | |
| "grad_norm": 2.3125, | |
| "learning_rate": 1.9475555555555558e-05, | |
| "loss": 0.5506, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.04020389116232321, | |
| "grad_norm": 1.984375, | |
| "learning_rate": 1.938666666666667e-05, | |
| "loss": 0.563, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.04594730418551224, | |
| "grad_norm": 2.125, | |
| "learning_rate": 1.929777777777778e-05, | |
| "loss": 0.561, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.05169071720870127, | |
| "grad_norm": 1.9765625, | |
| "learning_rate": 1.920888888888889e-05, | |
| "loss": 0.5288, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.0574341302318903, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 1.912e-05, | |
| "loss": 0.5693, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.06317754325507934, | |
| "grad_norm": 2.21875, | |
| "learning_rate": 1.9031111111111114e-05, | |
| "loss": 0.5353, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06892095627826836, | |
| "grad_norm": 1.9140625, | |
| "learning_rate": 1.8942222222222224e-05, | |
| "loss": 0.5383, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.0746643693014574, | |
| "grad_norm": 2.40625, | |
| "learning_rate": 1.8853333333333335e-05, | |
| "loss": 0.4849, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.08040778232464642, | |
| "grad_norm": 1.8671875, | |
| "learning_rate": 1.8764444444444445e-05, | |
| "loss": 0.5155, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.08615119534783545, | |
| "grad_norm": 2.265625, | |
| "learning_rate": 1.867555555555556e-05, | |
| "loss": 0.5071, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.09189460837102448, | |
| "grad_norm": 2.03125, | |
| "learning_rate": 1.858666666666667e-05, | |
| "loss": 0.5065, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.09763802139421351, | |
| "grad_norm": 1.84375, | |
| "learning_rate": 1.849777777777778e-05, | |
| "loss": 0.509, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.10338143441740254, | |
| "grad_norm": 1.703125, | |
| "learning_rate": 1.840888888888889e-05, | |
| "loss": 0.5076, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.10912484744059157, | |
| "grad_norm": 2.0, | |
| "learning_rate": 1.832e-05, | |
| "loss": 0.4804, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.1148682604637806, | |
| "grad_norm": 1.8828125, | |
| "learning_rate": 1.823111111111111e-05, | |
| "loss": 0.4584, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.12061167348696963, | |
| "grad_norm": 1.8046875, | |
| "learning_rate": 1.814222222222222e-05, | |
| "loss": 0.4951, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.12635508651015867, | |
| "grad_norm": 1.7265625, | |
| "learning_rate": 1.8053333333333332e-05, | |
| "loss": 0.4758, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.13209849953334768, | |
| "grad_norm": 2.1875, | |
| "learning_rate": 1.7964444444444446e-05, | |
| "loss": 0.5146, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.13784191255653672, | |
| "grad_norm": 2.140625, | |
| "learning_rate": 1.7875555555555556e-05, | |
| "loss": 0.5004, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.14358532557972575, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 1.7786666666666667e-05, | |
| "loss": 0.4744, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1493287386029148, | |
| "grad_norm": 2.078125, | |
| "learning_rate": 1.7697777777777777e-05, | |
| "loss": 0.4943, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.1550721516261038, | |
| "grad_norm": 2.046875, | |
| "learning_rate": 1.760888888888889e-05, | |
| "loss": 0.4751, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.16081556464929284, | |
| "grad_norm": 2.09375, | |
| "learning_rate": 1.752e-05, | |
| "loss": 0.4599, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.16655897767248187, | |
| "grad_norm": 1.8984375, | |
| "learning_rate": 1.7431111111111112e-05, | |
| "loss": 0.4656, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1723023906956709, | |
| "grad_norm": 1.9609375, | |
| "learning_rate": 1.7342222222222226e-05, | |
| "loss": 0.4242, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.17804580371885995, | |
| "grad_norm": 2.0625, | |
| "learning_rate": 1.7253333333333336e-05, | |
| "loss": 0.4804, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.18378921674204896, | |
| "grad_norm": 1.8046875, | |
| "learning_rate": 1.7164444444444447e-05, | |
| "loss": 0.4736, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.189532629765238, | |
| "grad_norm": 2.0625, | |
| "learning_rate": 1.7075555555555557e-05, | |
| "loss": 0.4317, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.19527604278842703, | |
| "grad_norm": 2.46875, | |
| "learning_rate": 1.6986666666666668e-05, | |
| "loss": 0.4752, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.20101945581161607, | |
| "grad_norm": 1.8359375, | |
| "learning_rate": 1.6897777777777778e-05, | |
| "loss": 0.3932, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.20676286883480507, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 1.680888888888889e-05, | |
| "loss": 0.4742, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.2125062818579941, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 1.672e-05, | |
| "loss": 0.4542, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.21824969488118315, | |
| "grad_norm": 1.9140625, | |
| "learning_rate": 1.6631111111111113e-05, | |
| "loss": 0.4284, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.22399310790437219, | |
| "grad_norm": 1.90625, | |
| "learning_rate": 1.6542222222222223e-05, | |
| "loss": 0.4232, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2297365209275612, | |
| "grad_norm": 1.890625, | |
| "learning_rate": 1.6453333333333334e-05, | |
| "loss": 0.4299, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.23547993395075023, | |
| "grad_norm": 2.375, | |
| "learning_rate": 1.6364444444444444e-05, | |
| "loss": 0.4246, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.24122334697393927, | |
| "grad_norm": 2.015625, | |
| "learning_rate": 1.6275555555555558e-05, | |
| "loss": 0.4215, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.2469667599971283, | |
| "grad_norm": 1.9375, | |
| "learning_rate": 1.618666666666667e-05, | |
| "loss": 0.4398, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.25271017302031734, | |
| "grad_norm": 2.1875, | |
| "learning_rate": 1.609777777777778e-05, | |
| "loss": 0.4319, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.2584535860435064, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 1.600888888888889e-05, | |
| "loss": 0.4227, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.26419699906669536, | |
| "grad_norm": 2.390625, | |
| "learning_rate": 1.5920000000000003e-05, | |
| "loss": 0.431, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.2699404120898844, | |
| "grad_norm": 2.421875, | |
| "learning_rate": 1.5831111111111114e-05, | |
| "loss": 0.4323, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.27568382511307343, | |
| "grad_norm": 2.15625, | |
| "learning_rate": 1.5742222222222224e-05, | |
| "loss": 0.4417, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.28142723813626247, | |
| "grad_norm": 1.8984375, | |
| "learning_rate": 1.5653333333333335e-05, | |
| "loss": 0.3823, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.2871706511594515, | |
| "grad_norm": 1.9140625, | |
| "learning_rate": 1.5564444444444445e-05, | |
| "loss": 0.394, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.29291406418264054, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 1.5475555555555556e-05, | |
| "loss": 0.4017, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.2986574772058296, | |
| "grad_norm": 2.4375, | |
| "learning_rate": 1.5386666666666666e-05, | |
| "loss": 0.4121, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.3044008902290186, | |
| "grad_norm": 2.078125, | |
| "learning_rate": 1.5297777777777777e-05, | |
| "loss": 0.3843, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.3101443032522076, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 1.520888888888889e-05, | |
| "loss": 0.4395, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.31588771627539664, | |
| "grad_norm": 2.25, | |
| "learning_rate": 1.5120000000000001e-05, | |
| "loss": 0.3883, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.32163112929858567, | |
| "grad_norm": 1.6171875, | |
| "learning_rate": 1.5031111111111111e-05, | |
| "loss": 0.4089, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3273745423217747, | |
| "grad_norm": 1.734375, | |
| "learning_rate": 1.4942222222222222e-05, | |
| "loss": 0.3895, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.33311795534496375, | |
| "grad_norm": 1.9453125, | |
| "learning_rate": 1.4853333333333336e-05, | |
| "loss": 0.4084, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.3388613683681528, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 1.4764444444444446e-05, | |
| "loss": 0.4083, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.3446047813913418, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 1.4675555555555557e-05, | |
| "loss": 0.4037, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.35034819441453086, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 1.4586666666666667e-05, | |
| "loss": 0.3538, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.3560916074377199, | |
| "grad_norm": 2.015625, | |
| "learning_rate": 1.449777777777778e-05, | |
| "loss": 0.3661, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.3618350204609089, | |
| "grad_norm": 1.984375, | |
| "learning_rate": 1.440888888888889e-05, | |
| "loss": 0.3992, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.3675784334840979, | |
| "grad_norm": 2.328125, | |
| "learning_rate": 1.432e-05, | |
| "loss": 0.3721, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.37332184650728695, | |
| "grad_norm": 2.453125, | |
| "learning_rate": 1.4231111111111114e-05, | |
| "loss": 0.4009, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.379065259530476, | |
| "grad_norm": 2.25, | |
| "learning_rate": 1.4142222222222224e-05, | |
| "loss": 0.401, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.384808672553665, | |
| "grad_norm": 2.46875, | |
| "learning_rate": 1.4053333333333335e-05, | |
| "loss": 0.3783, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.39055208557685406, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 1.3964444444444445e-05, | |
| "loss": 0.4021, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.3962954986000431, | |
| "grad_norm": 2.15625, | |
| "learning_rate": 1.3875555555555557e-05, | |
| "loss": 0.3713, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.40203891162323213, | |
| "grad_norm": 2.125, | |
| "learning_rate": 1.3786666666666668e-05, | |
| "loss": 0.4036, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.4077823246464211, | |
| "grad_norm": 2.03125, | |
| "learning_rate": 1.3697777777777778e-05, | |
| "loss": 0.368, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.41352573766961015, | |
| "grad_norm": 2.5625, | |
| "learning_rate": 1.3608888888888889e-05, | |
| "loss": 0.3578, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.4192691506927992, | |
| "grad_norm": 2.390625, | |
| "learning_rate": 1.3520000000000003e-05, | |
| "loss": 0.3731, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.4250125637159882, | |
| "grad_norm": 2.5, | |
| "learning_rate": 1.3431111111111113e-05, | |
| "loss": 0.3817, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.43075597673917726, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 1.3342222222222224e-05, | |
| "loss": 0.3717, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.4364993897623663, | |
| "grad_norm": 2.203125, | |
| "learning_rate": 1.3253333333333334e-05, | |
| "loss": 0.3629, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.44224280278555533, | |
| "grad_norm": 2.625, | |
| "learning_rate": 1.3164444444444446e-05, | |
| "loss": 0.3698, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.44798621580874437, | |
| "grad_norm": 2.0, | |
| "learning_rate": 1.3075555555555557e-05, | |
| "loss": 0.3661, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.45372962883193335, | |
| "grad_norm": 2.5, | |
| "learning_rate": 1.2986666666666667e-05, | |
| "loss": 0.342, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.4594730418551224, | |
| "grad_norm": 2.046875, | |
| "learning_rate": 1.2897777777777778e-05, | |
| "loss": 0.3862, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.4652164548783114, | |
| "grad_norm": 2.40625, | |
| "learning_rate": 1.2808888888888891e-05, | |
| "loss": 0.3436, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.47095986790150046, | |
| "grad_norm": 2.3125, | |
| "learning_rate": 1.2720000000000002e-05, | |
| "loss": 0.3374, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.4767032809246895, | |
| "grad_norm": 2.15625, | |
| "learning_rate": 1.2631111111111112e-05, | |
| "loss": 0.3648, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.48244669394787854, | |
| "grad_norm": 2.25, | |
| "learning_rate": 1.2542222222222223e-05, | |
| "loss": 0.327, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.48819010697106757, | |
| "grad_norm": 2.3125, | |
| "learning_rate": 1.2453333333333335e-05, | |
| "loss": 0.3406, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.4939335199942566, | |
| "grad_norm": 2.03125, | |
| "learning_rate": 1.2364444444444445e-05, | |
| "loss": 0.3441, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.4996769330174456, | |
| "grad_norm": 2.609375, | |
| "learning_rate": 1.2275555555555556e-05, | |
| "loss": 0.3695, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5054203460406347, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 1.2186666666666666e-05, | |
| "loss": 0.3386, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5111637590638237, | |
| "grad_norm": 2.328125, | |
| "learning_rate": 1.209777777777778e-05, | |
| "loss": 0.3478, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5169071720870128, | |
| "grad_norm": 2.078125, | |
| "learning_rate": 1.200888888888889e-05, | |
| "loss": 0.3274, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.5226505851102017, | |
| "grad_norm": 2.3125, | |
| "learning_rate": 1.1920000000000001e-05, | |
| "loss": 0.3381, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.5283939981333907, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 1.1831111111111112e-05, | |
| "loss": 0.3438, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5341374111565798, | |
| "grad_norm": 2.390625, | |
| "learning_rate": 1.1742222222222224e-05, | |
| "loss": 0.3655, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.5398808241797688, | |
| "grad_norm": 2.203125, | |
| "learning_rate": 1.1653333333333334e-05, | |
| "loss": 0.3219, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.5456242372029578, | |
| "grad_norm": 2.453125, | |
| "learning_rate": 1.1564444444444445e-05, | |
| "loss": 0.3286, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.5513676502261469, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 1.1475555555555557e-05, | |
| "loss": 0.3547, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5571110632493359, | |
| "grad_norm": 2.578125, | |
| "learning_rate": 1.1386666666666669e-05, | |
| "loss": 0.3325, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5628544762725249, | |
| "grad_norm": 2.375, | |
| "learning_rate": 1.129777777777778e-05, | |
| "loss": 0.3711, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.568597889295714, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 1.120888888888889e-05, | |
| "loss": 0.3124, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.574341302318903, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 1.1120000000000002e-05, | |
| "loss": 0.3289, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.580084715342092, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 1.1031111111111112e-05, | |
| "loss": 0.3218, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.5858281283652811, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 1.0942222222222223e-05, | |
| "loss": 0.3503, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.5915715413884701, | |
| "grad_norm": 2.328125, | |
| "learning_rate": 1.0853333333333333e-05, | |
| "loss": 0.3551, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.5973149544116592, | |
| "grad_norm": 2.453125, | |
| "learning_rate": 1.0764444444444445e-05, | |
| "loss": 0.3194, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.6030583674348482, | |
| "grad_norm": 2.1875, | |
| "learning_rate": 1.0675555555555558e-05, | |
| "loss": 0.3478, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.6088017804580372, | |
| "grad_norm": 2.265625, | |
| "learning_rate": 1.0586666666666668e-05, | |
| "loss": 0.3202, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.6145451934812263, | |
| "grad_norm": 1.9140625, | |
| "learning_rate": 1.0497777777777779e-05, | |
| "loss": 0.3095, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.6202886065044152, | |
| "grad_norm": 2.71875, | |
| "learning_rate": 1.040888888888889e-05, | |
| "loss": 0.3112, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.6260320195276042, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 1.0320000000000001e-05, | |
| "loss": 0.3115, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.6317754325507933, | |
| "grad_norm": 2.296875, | |
| "learning_rate": 1.0231111111111112e-05, | |
| "loss": 0.3296, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.6375188455739823, | |
| "grad_norm": 2.703125, | |
| "learning_rate": 1.0142222222222222e-05, | |
| "loss": 0.279, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.6432622585971713, | |
| "grad_norm": 2.71875, | |
| "learning_rate": 1.0053333333333334e-05, | |
| "loss": 0.3153, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.6490056716203604, | |
| "grad_norm": 1.953125, | |
| "learning_rate": 9.964444444444445e-06, | |
| "loss": 0.3039, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.6547490846435494, | |
| "grad_norm": 2.140625, | |
| "learning_rate": 9.875555555555557e-06, | |
| "loss": 0.3238, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.6604924976667385, | |
| "grad_norm": 2.25, | |
| "learning_rate": 9.786666666666667e-06, | |
| "loss": 0.3143, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.6662359106899275, | |
| "grad_norm": 2.546875, | |
| "learning_rate": 9.697777777777778e-06, | |
| "loss": 0.3145, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.6719793237131165, | |
| "grad_norm": 2.15625, | |
| "learning_rate": 9.60888888888889e-06, | |
| "loss": 0.3034, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.6777227367363056, | |
| "grad_norm": 3.875, | |
| "learning_rate": 9.52e-06, | |
| "loss": 0.2813, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.6834661497594946, | |
| "grad_norm": 2.046875, | |
| "learning_rate": 9.431111111111112e-06, | |
| "loss": 0.3211, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.6892095627826836, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 9.342222222222223e-06, | |
| "loss": 0.2954, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.6949529758058727, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 9.253333333333333e-06, | |
| "loss": 0.2969, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.7006963888290617, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 9.164444444444446e-06, | |
| "loss": 0.2942, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.7064398018522507, | |
| "grad_norm": 2.0, | |
| "learning_rate": 9.075555555555556e-06, | |
| "loss": 0.321, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.7121832148754398, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 8.986666666666666e-06, | |
| "loss": 0.2949, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.7179266278986287, | |
| "grad_norm": 2.609375, | |
| "learning_rate": 8.897777777777779e-06, | |
| "loss": 0.3111, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.7236700409218177, | |
| "grad_norm": 2.234375, | |
| "learning_rate": 8.80888888888889e-06, | |
| "loss": 0.3199, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.7294134539450068, | |
| "grad_norm": 3.171875, | |
| "learning_rate": 8.720000000000001e-06, | |
| "loss": 0.3239, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.7351568669681958, | |
| "grad_norm": 2.640625, | |
| "learning_rate": 8.631111111111112e-06, | |
| "loss": 0.3203, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.7409002799913849, | |
| "grad_norm": 2.3125, | |
| "learning_rate": 8.542222222222222e-06, | |
| "loss": 0.3071, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.7466436930145739, | |
| "grad_norm": 2.328125, | |
| "learning_rate": 8.453333333333334e-06, | |
| "loss": 0.3296, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.7523871060377629, | |
| "grad_norm": 2.296875, | |
| "learning_rate": 8.364444444444445e-06, | |
| "loss": 0.2994, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.758130519060952, | |
| "grad_norm": 2.390625, | |
| "learning_rate": 8.275555555555557e-06, | |
| "loss": 0.2661, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.763873932084141, | |
| "grad_norm": 2.34375, | |
| "learning_rate": 8.186666666666667e-06, | |
| "loss": 0.3194, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.76961734510733, | |
| "grad_norm": 2.28125, | |
| "learning_rate": 8.09777777777778e-06, | |
| "loss": 0.283, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.7753607581305191, | |
| "grad_norm": 2.734375, | |
| "learning_rate": 8.00888888888889e-06, | |
| "loss": 0.3358, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.7811041711537081, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 7.92e-06, | |
| "loss": 0.3041, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.7868475841768972, | |
| "grad_norm": 2.671875, | |
| "learning_rate": 7.831111111111111e-06, | |
| "loss": 0.2701, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.7925909972000862, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 7.742222222222223e-06, | |
| "loss": 0.2783, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.7983344102232752, | |
| "grad_norm": 2.296875, | |
| "learning_rate": 7.653333333333333e-06, | |
| "loss": 0.2989, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.8040778232464643, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 7.564444444444446e-06, | |
| "loss": 0.2711, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.8098212362696532, | |
| "grad_norm": 2.21875, | |
| "learning_rate": 7.475555555555556e-06, | |
| "loss": 0.29, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.8155646492928422, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 7.386666666666667e-06, | |
| "loss": 0.3208, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.8213080623160313, | |
| "grad_norm": 2.21875, | |
| "learning_rate": 7.297777777777778e-06, | |
| "loss": 0.2838, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.8270514753392203, | |
| "grad_norm": 2.828125, | |
| "learning_rate": 7.20888888888889e-06, | |
| "loss": 0.3104, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.8327948883624093, | |
| "grad_norm": 2.3125, | |
| "learning_rate": 7.1200000000000004e-06, | |
| "loss": 0.2805, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.8385383013855984, | |
| "grad_norm": 1.71875, | |
| "learning_rate": 7.031111111111112e-06, | |
| "loss": 0.2542, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.8442817144087874, | |
| "grad_norm": 2.1875, | |
| "learning_rate": 6.942222222222222e-06, | |
| "loss": 0.2698, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.8500251274319764, | |
| "grad_norm": 2.671875, | |
| "learning_rate": 6.853333333333334e-06, | |
| "loss": 0.2938, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.8557685404551655, | |
| "grad_norm": 2.484375, | |
| "learning_rate": 6.764444444444445e-06, | |
| "loss": 0.2657, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.8615119534783545, | |
| "grad_norm": 2.125, | |
| "learning_rate": 6.675555555555556e-06, | |
| "loss": 0.3194, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.8672553665015436, | |
| "grad_norm": 3.0, | |
| "learning_rate": 6.5866666666666666e-06, | |
| "loss": 0.2863, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.8729987795247326, | |
| "grad_norm": 2.640625, | |
| "learning_rate": 6.497777777777779e-06, | |
| "loss": 0.2636, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.8787421925479216, | |
| "grad_norm": 2.1875, | |
| "learning_rate": 6.408888888888889e-06, | |
| "loss": 0.2537, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.8844856055711107, | |
| "grad_norm": 2.265625, | |
| "learning_rate": 6.3200000000000005e-06, | |
| "loss": 0.2915, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.8902290185942997, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 6.231111111111111e-06, | |
| "loss": 0.2814, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.8959724316174887, | |
| "grad_norm": 2.421875, | |
| "learning_rate": 6.142222222222223e-06, | |
| "loss": 0.2508, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.9017158446406778, | |
| "grad_norm": 1.8984375, | |
| "learning_rate": 6.0533333333333335e-06, | |
| "loss": 0.2499, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.9074592576638667, | |
| "grad_norm": 3.515625, | |
| "learning_rate": 5.964444444444445e-06, | |
| "loss": 0.2779, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.9132026706870557, | |
| "grad_norm": 2.09375, | |
| "learning_rate": 5.875555555555556e-06, | |
| "loss": 0.2703, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.9189460837102448, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 5.7866666666666674e-06, | |
| "loss": 0.2851, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.9246894967334338, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 5.697777777777779e-06, | |
| "loss": 0.2771, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.9304329097566229, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 5.608888888888889e-06, | |
| "loss": 0.2545, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.9361763227798119, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 5.5200000000000005e-06, | |
| "loss": 0.2534, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.9419197358030009, | |
| "grad_norm": 2.65625, | |
| "learning_rate": 5.431111111111112e-06, | |
| "loss": 0.2851, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.94766314882619, | |
| "grad_norm": 2.109375, | |
| "learning_rate": 5.342222222222223e-06, | |
| "loss": 0.3072, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.953406561849379, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 5.2533333333333336e-06, | |
| "loss": 0.2709, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.959149974872568, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 5.164444444444445e-06, | |
| "loss": 0.2876, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.9648933878957571, | |
| "grad_norm": 2.015625, | |
| "learning_rate": 5.075555555555556e-06, | |
| "loss": 0.2824, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.9706368009189461, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 4.986666666666667e-06, | |
| "loss": 0.2867, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.9763802139421351, | |
| "grad_norm": 2.421875, | |
| "learning_rate": 4.897777777777778e-06, | |
| "loss": 0.2711, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.9821236269653242, | |
| "grad_norm": 2.375, | |
| "learning_rate": 4.808888888888889e-06, | |
| "loss": 0.3008, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.9878670399885132, | |
| "grad_norm": 1.9296875, | |
| "learning_rate": 4.7200000000000005e-06, | |
| "loss": 0.2649, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.9936104530117023, | |
| "grad_norm": 1.9296875, | |
| "learning_rate": 4.631111111111111e-06, | |
| "loss": 0.2625, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.9993538660348912, | |
| "grad_norm": 2.359375, | |
| "learning_rate": 4.542222222222223e-06, | |
| "loss": 0.2995, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.0045947304185512, | |
| "grad_norm": 2.078125, | |
| "learning_rate": 4.453333333333334e-06, | |
| "loss": 0.2862, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.0103381434417402, | |
| "grad_norm": 1.953125, | |
| "learning_rate": 4.364444444444445e-06, | |
| "loss": 0.24, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.0160815564649293, | |
| "grad_norm": 2.28125, | |
| "learning_rate": 4.275555555555556e-06, | |
| "loss": 0.2394, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.0218249694881183, | |
| "grad_norm": 1.796875, | |
| "learning_rate": 4.1866666666666675e-06, | |
| "loss": 0.2177, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.0275683825113073, | |
| "grad_norm": 2.28125, | |
| "learning_rate": 4.097777777777778e-06, | |
| "loss": 0.2514, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.0333117955344964, | |
| "grad_norm": 1.890625, | |
| "learning_rate": 4.008888888888889e-06, | |
| "loss": 0.2158, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.0390552085576854, | |
| "grad_norm": 2.015625, | |
| "learning_rate": 3.920000000000001e-06, | |
| "loss": 0.2377, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.0447986215808744, | |
| "grad_norm": 2.5, | |
| "learning_rate": 3.831111111111112e-06, | |
| "loss": 0.2193, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.0505420346040635, | |
| "grad_norm": 2.40625, | |
| "learning_rate": 3.7422222222222228e-06, | |
| "loss": 0.2354, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.0562854476272525, | |
| "grad_norm": 1.8984375, | |
| "learning_rate": 3.6533333333333336e-06, | |
| "loss": 0.2315, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.0620288606504416, | |
| "grad_norm": 2.484375, | |
| "learning_rate": 3.564444444444445e-06, | |
| "loss": 0.1987, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.0677722736736306, | |
| "grad_norm": 2.5, | |
| "learning_rate": 3.475555555555556e-06, | |
| "loss": 0.2276, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.0735156866968196, | |
| "grad_norm": 2.90625, | |
| "learning_rate": 3.386666666666667e-06, | |
| "loss": 0.225, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.0792590997200087, | |
| "grad_norm": 1.8046875, | |
| "learning_rate": 3.297777777777778e-06, | |
| "loss": 0.1992, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.0850025127431977, | |
| "grad_norm": 2.265625, | |
| "learning_rate": 3.2088888888888893e-06, | |
| "loss": 0.2068, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.0907459257663867, | |
| "grad_norm": 2.296875, | |
| "learning_rate": 3.12e-06, | |
| "loss": 0.225, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.0964893387895758, | |
| "grad_norm": 2.703125, | |
| "learning_rate": 3.0311111111111115e-06, | |
| "loss": 0.2173, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.1022327518127648, | |
| "grad_norm": 2.625, | |
| "learning_rate": 2.9422222222222224e-06, | |
| "loss": 0.2348, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.1079761648359538, | |
| "grad_norm": 1.796875, | |
| "learning_rate": 2.8533333333333337e-06, | |
| "loss": 0.2427, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.1137195778591429, | |
| "grad_norm": 2.0625, | |
| "learning_rate": 2.7644444444444446e-06, | |
| "loss": 0.2386, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.119462990882332, | |
| "grad_norm": 2.140625, | |
| "learning_rate": 2.675555555555556e-06, | |
| "loss": 0.2433, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.125206403905521, | |
| "grad_norm": 2.25, | |
| "learning_rate": 2.5866666666666667e-06, | |
| "loss": 0.2448, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.1309498169287098, | |
| "grad_norm": 2.140625, | |
| "learning_rate": 2.497777777777778e-06, | |
| "loss": 0.2304, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.136693229951899, | |
| "grad_norm": 2.09375, | |
| "learning_rate": 2.4088888888888894e-06, | |
| "loss": 0.2304, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.1424366429750878, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 2.3200000000000002e-06, | |
| "loss": 0.236, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.1481800559982769, | |
| "grad_norm": 2.125, | |
| "learning_rate": 2.2311111111111115e-06, | |
| "loss": 0.2367, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.153923469021466, | |
| "grad_norm": 2.484375, | |
| "learning_rate": 2.1422222222222224e-06, | |
| "loss": 0.2178, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.159666882044655, | |
| "grad_norm": 2.53125, | |
| "learning_rate": 2.0533333333333337e-06, | |
| "loss": 0.2333, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.165410295067844, | |
| "grad_norm": 3.09375, | |
| "learning_rate": 1.9644444444444446e-06, | |
| "loss": 0.2586, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.171153708091033, | |
| "grad_norm": 2.953125, | |
| "learning_rate": 1.8755555555555557e-06, | |
| "loss": 0.2483, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.176897121114222, | |
| "grad_norm": 2.25, | |
| "learning_rate": 1.7866666666666668e-06, | |
| "loss": 0.2505, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.182640534137411, | |
| "grad_norm": 1.96875, | |
| "learning_rate": 1.6977777777777779e-06, | |
| "loss": 0.208, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.1883839471606001, | |
| "grad_norm": 2.390625, | |
| "learning_rate": 1.608888888888889e-06, | |
| "loss": 0.2279, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.1941273601837892, | |
| "grad_norm": 2.171875, | |
| "learning_rate": 1.52e-06, | |
| "loss": 0.2002, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.1998707732069782, | |
| "grad_norm": 2.28125, | |
| "learning_rate": 1.4311111111111111e-06, | |
| "loss": 0.2204, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.2056141862301673, | |
| "grad_norm": 2.453125, | |
| "learning_rate": 1.3422222222222222e-06, | |
| "loss": 0.2339, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.2113575992533563, | |
| "grad_norm": 1.875, | |
| "learning_rate": 1.2533333333333333e-06, | |
| "loss": 0.2362, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.2171010122765453, | |
| "grad_norm": 2.484375, | |
| "learning_rate": 1.1644444444444446e-06, | |
| "loss": 0.2324, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.2228444252997344, | |
| "grad_norm": 2.640625, | |
| "learning_rate": 1.0755555555555557e-06, | |
| "loss": 0.2404, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.2285878383229234, | |
| "grad_norm": 2.265625, | |
| "learning_rate": 9.866666666666668e-07, | |
| "loss": 0.202, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.2343312513461124, | |
| "grad_norm": 2.421875, | |
| "learning_rate": 8.977777777777778e-07, | |
| "loss": 0.2306, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.2400746643693015, | |
| "grad_norm": 2.46875, | |
| "learning_rate": 8.088888888888889e-07, | |
| "loss": 0.2094, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.2458180773924905, | |
| "grad_norm": 2.09375, | |
| "learning_rate": 7.2e-07, | |
| "loss": 0.2177, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.2515614904156795, | |
| "grad_norm": 2.515625, | |
| "learning_rate": 6.311111111111112e-07, | |
| "loss": 0.2508, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.2573049034388686, | |
| "grad_norm": 2.15625, | |
| "learning_rate": 5.422222222222223e-07, | |
| "loss": 0.2231, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.2630483164620576, | |
| "grad_norm": 1.7109375, | |
| "learning_rate": 4.533333333333334e-07, | |
| "loss": 0.2178, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.2687917294852467, | |
| "grad_norm": 2.5625, | |
| "learning_rate": 3.644444444444445e-07, | |
| "loss": 0.194, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.2745351425084357, | |
| "grad_norm": 2.125, | |
| "learning_rate": 2.7555555555555555e-07, | |
| "loss": 0.2302, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.2802785555316247, | |
| "grad_norm": 2.328125, | |
| "learning_rate": 1.866666666666667e-07, | |
| "loss": 0.2347, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.2860219685548138, | |
| "grad_norm": 2.1875, | |
| "learning_rate": 9.777777777777779e-08, | |
| "loss": 0.2061, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.2917653815780028, | |
| "grad_norm": 2.828125, | |
| "learning_rate": 8.88888888888889e-09, | |
| "loss": 0.2463, | |
| "step": 2250 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2250, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0743805199693906e+18, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |