{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.951768488745981, "eval_steps": 500, "global_step": 385, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012861736334405145, "grad_norm": 577.1012573242188, "learning_rate": 8.333333333333335e-09, "loss": 13.9619, "num_input_tokens_seen": 15136, "step": 1 }, { "epoch": 0.02572347266881029, "grad_norm": 552.5532836914062, "learning_rate": 1.666666666666667e-08, "loss": 13.7129, "num_input_tokens_seen": 30208, "step": 2 }, { "epoch": 0.03858520900321544, "grad_norm": 562.699462890625, "learning_rate": 2.5000000000000002e-08, "loss": 13.8474, "num_input_tokens_seen": 45376, "step": 3 }, { "epoch": 0.05144694533762058, "grad_norm": 585.796630859375, "learning_rate": 3.333333333333334e-08, "loss": 13.8844, "num_input_tokens_seen": 59968, "step": 4 }, { "epoch": 0.06430868167202572, "grad_norm": 592.56689453125, "learning_rate": 4.166666666666667e-08, "loss": 14.138, "num_input_tokens_seen": 75136, "step": 5 }, { "epoch": 0.07717041800643087, "grad_norm": 569.5825805664062, "learning_rate": 5.0000000000000004e-08, "loss": 13.9077, "num_input_tokens_seen": 90816, "step": 6 }, { "epoch": 0.09003215434083602, "grad_norm": 578.7579956054688, "learning_rate": 5.833333333333334e-08, "loss": 13.8396, "num_input_tokens_seen": 105600, "step": 7 }, { "epoch": 0.10289389067524116, "grad_norm": 577.330322265625, "learning_rate": 6.666666666666668e-08, "loss": 13.9828, "num_input_tokens_seen": 119872, "step": 8 }, { "epoch": 0.1157556270096463, "grad_norm": 575.2134399414062, "learning_rate": 7.500000000000001e-08, "loss": 14.0361, "num_input_tokens_seen": 135104, "step": 9 }, { "epoch": 0.12861736334405144, "grad_norm": 579.0157470703125, "learning_rate": 8.333333333333334e-08, "loss": 13.9392, "num_input_tokens_seen": 150016, "step": 10 }, { "epoch": 0.1414790996784566, "grad_norm": 585.5687255859375, "learning_rate": 9.166666666666668e-08, "loss": 14.0256, "num_input_tokens_seen": 165216, "step": 11 }, { "epoch": 0.15434083601286175, "grad_norm": 566.5042114257812, "learning_rate": 1.0000000000000001e-07, "loss": 13.6693, "num_input_tokens_seen": 179296, "step": 12 }, { "epoch": 0.16720257234726688, "grad_norm": 574.4198608398438, "learning_rate": 1.0833333333333335e-07, "loss": 13.9031, "num_input_tokens_seen": 193312, "step": 13 }, { "epoch": 0.18006430868167203, "grad_norm": 588.1041259765625, "learning_rate": 1.1666666666666668e-07, "loss": 13.8575, "num_input_tokens_seen": 208768, "step": 14 }, { "epoch": 0.19292604501607716, "grad_norm": 573.4750366210938, "learning_rate": 1.2500000000000002e-07, "loss": 13.8366, "num_input_tokens_seen": 223456, "step": 15 }, { "epoch": 0.2057877813504823, "grad_norm": 572.26220703125, "learning_rate": 1.3333333333333336e-07, "loss": 13.8705, "num_input_tokens_seen": 238656, "step": 16 }, { "epoch": 0.21864951768488747, "grad_norm": 566.3518676757812, "learning_rate": 1.4166666666666668e-07, "loss": 13.1816, "num_input_tokens_seen": 253312, "step": 17 }, { "epoch": 0.2315112540192926, "grad_norm": 560.5255126953125, "learning_rate": 1.5000000000000002e-07, "loss": 13.2292, "num_input_tokens_seen": 268352, "step": 18 }, { "epoch": 0.24437299035369775, "grad_norm": 583.0404663085938, "learning_rate": 1.5833333333333336e-07, "loss": 13.4366, "num_input_tokens_seen": 282560, "step": 19 }, { "epoch": 0.2572347266881029, "grad_norm": 555.8986206054688, "learning_rate": 1.6666666666666668e-07, "loss": 12.9904, "num_input_tokens_seen": 297120, "step": 20 }, { "epoch": 0.27009646302250806, "grad_norm": 557.931396484375, "learning_rate": 1.7500000000000002e-07, "loss": 12.878, "num_input_tokens_seen": 312416, "step": 21 }, { "epoch": 0.2829581993569132, "grad_norm": 562.4014282226562, "learning_rate": 1.8333333333333336e-07, "loss": 12.7794, "num_input_tokens_seen": 327968, "step": 22 }, { "epoch": 0.2958199356913183, "grad_norm": 548.7069702148438, "learning_rate": 1.9166666666666668e-07, "loss": 11.3144, "num_input_tokens_seen": 342880, "step": 23 }, { "epoch": 0.3086816720257235, "grad_norm": 511.2892761230469, "learning_rate": 2.0000000000000002e-07, "loss": 10.8531, "num_input_tokens_seen": 358848, "step": 24 }, { "epoch": 0.3215434083601286, "grad_norm": 496.71868896484375, "learning_rate": 2.0833333333333333e-07, "loss": 10.7149, "num_input_tokens_seen": 373792, "step": 25 }, { "epoch": 0.33440514469453375, "grad_norm": 504.7909851074219, "learning_rate": 2.166666666666667e-07, "loss": 10.5802, "num_input_tokens_seen": 389248, "step": 26 }, { "epoch": 0.34726688102893893, "grad_norm": 495.0346984863281, "learning_rate": 2.2500000000000002e-07, "loss": 10.3671, "num_input_tokens_seen": 404320, "step": 27 }, { "epoch": 0.36012861736334406, "grad_norm": 482.26318359375, "learning_rate": 2.3333333333333336e-07, "loss": 10.1751, "num_input_tokens_seen": 419296, "step": 28 }, { "epoch": 0.3729903536977492, "grad_norm": 480.48193359375, "learning_rate": 2.416666666666667e-07, "loss": 9.7707, "num_input_tokens_seen": 433632, "step": 29 }, { "epoch": 0.3858520900321543, "grad_norm": 467.7600402832031, "learning_rate": 2.5000000000000004e-07, "loss": 9.6489, "num_input_tokens_seen": 448736, "step": 30 }, { "epoch": 0.3987138263665595, "grad_norm": 468.9175720214844, "learning_rate": 2.5833333333333333e-07, "loss": 8.6805, "num_input_tokens_seen": 463840, "step": 31 }, { "epoch": 0.4115755627009646, "grad_norm": 394.77691650390625, "learning_rate": 2.666666666666667e-07, "loss": 5.9207, "num_input_tokens_seen": 478816, "step": 32 }, { "epoch": 0.42443729903536975, "grad_norm": 440.14697265625, "learning_rate": 2.75e-07, "loss": 5.7661, "num_input_tokens_seen": 494464, "step": 33 }, { "epoch": 0.43729903536977494, "grad_norm": 436.6060791015625, "learning_rate": 2.8333333333333336e-07, "loss": 5.6168, "num_input_tokens_seen": 510080, "step": 34 }, { "epoch": 0.45016077170418006, "grad_norm": 388.576416015625, "learning_rate": 2.916666666666667e-07, "loss": 5.3367, "num_input_tokens_seen": 525216, "step": 35 }, { "epoch": 0.4630225080385852, "grad_norm": 327.9267272949219, "learning_rate": 3.0000000000000004e-07, "loss": 4.9751, "num_input_tokens_seen": 539648, "step": 36 }, { "epoch": 0.4758842443729904, "grad_norm": 322.6181335449219, "learning_rate": 3.083333333333334e-07, "loss": 4.7041, "num_input_tokens_seen": 554080, "step": 37 }, { "epoch": 0.4887459807073955, "grad_norm": 313.6160888671875, "learning_rate": 3.166666666666667e-07, "loss": 4.4631, "num_input_tokens_seen": 568960, "step": 38 }, { "epoch": 0.5016077170418006, "grad_norm": 299.7319641113281, "learning_rate": 3.25e-07, "loss": 4.1912, "num_input_tokens_seen": 583904, "step": 39 }, { "epoch": 0.5144694533762058, "grad_norm": 308.4530944824219, "learning_rate": 3.3333333333333335e-07, "loss": 3.9146, "num_input_tokens_seen": 598848, "step": 40 }, { "epoch": 0.5273311897106109, "grad_norm": 315.37396240234375, "learning_rate": 3.416666666666667e-07, "loss": 3.053, "num_input_tokens_seen": 614208, "step": 41 }, { "epoch": 0.5401929260450161, "grad_norm": 263.33258056640625, "learning_rate": 3.5000000000000004e-07, "loss": 1.5544, "num_input_tokens_seen": 629280, "step": 42 }, { "epoch": 0.5530546623794212, "grad_norm": 177.95458984375, "learning_rate": 3.583333333333334e-07, "loss": 1.0549, "num_input_tokens_seen": 643968, "step": 43 }, { "epoch": 0.5659163987138264, "grad_norm": 142.09307861328125, "learning_rate": 3.666666666666667e-07, "loss": 0.711, "num_input_tokens_seen": 659040, "step": 44 }, { "epoch": 0.5787781350482315, "grad_norm": 93.70928192138672, "learning_rate": 3.75e-07, "loss": 0.5127, "num_input_tokens_seen": 674464, "step": 45 }, { "epoch": 0.5916398713826366, "grad_norm": 65.4582290649414, "learning_rate": 3.8333333333333335e-07, "loss": 0.4143, "num_input_tokens_seen": 689216, "step": 46 }, { "epoch": 0.6045016077170418, "grad_norm": 51.321414947509766, "learning_rate": 3.9166666666666675e-07, "loss": 0.4014, "num_input_tokens_seen": 704448, "step": 47 }, { "epoch": 0.617363344051447, "grad_norm": 38.20650100708008, "learning_rate": 4.0000000000000003e-07, "loss": 0.384, "num_input_tokens_seen": 719520, "step": 48 }, { "epoch": 0.6302250803858521, "grad_norm": 70.43089294433594, "learning_rate": 4.083333333333334e-07, "loss": 0.3017, "num_input_tokens_seen": 735520, "step": 49 }, { "epoch": 0.6430868167202572, "grad_norm": 47.982505798339844, "learning_rate": 4.1666666666666667e-07, "loss": 0.3031, "num_input_tokens_seen": 750976, "step": 50 }, { "epoch": 0.6559485530546624, "grad_norm": 27.063344955444336, "learning_rate": 4.2500000000000006e-07, "loss": 0.2818, "num_input_tokens_seen": 765728, "step": 51 }, { "epoch": 0.6688102893890675, "grad_norm": 60.75875473022461, "learning_rate": 4.333333333333334e-07, "loss": 0.2881, "num_input_tokens_seen": 780608, "step": 52 }, { "epoch": 0.6816720257234726, "grad_norm": 56.81061935424805, "learning_rate": 4.416666666666667e-07, "loss": 0.2943, "num_input_tokens_seen": 796192, "step": 53 }, { "epoch": 0.6945337620578779, "grad_norm": 31.529571533203125, "learning_rate": 4.5000000000000003e-07, "loss": 0.2781, "num_input_tokens_seen": 810496, "step": 54 }, { "epoch": 0.707395498392283, "grad_norm": 33.71590805053711, "learning_rate": 4.583333333333333e-07, "loss": 0.2724, "num_input_tokens_seen": 826048, "step": 55 }, { "epoch": 0.7202572347266881, "grad_norm": 26.352554321289062, "learning_rate": 4.666666666666667e-07, "loss": 0.252, "num_input_tokens_seen": 840512, "step": 56 }, { "epoch": 0.7331189710610932, "grad_norm": 15.606438636779785, "learning_rate": 4.7500000000000006e-07, "loss": 0.2482, "num_input_tokens_seen": 855808, "step": 57 }, { "epoch": 0.7459807073954984, "grad_norm": 31.445789337158203, "learning_rate": 4.833333333333334e-07, "loss": 0.2212, "num_input_tokens_seen": 871520, "step": 58 }, { "epoch": 0.7588424437299035, "grad_norm": 9.967268943786621, "learning_rate": 4.916666666666667e-07, "loss": 0.2408, "num_input_tokens_seen": 886144, "step": 59 }, { "epoch": 0.7717041800643086, "grad_norm": 23.841659545898438, "learning_rate": 5.000000000000001e-07, "loss": 0.2133, "num_input_tokens_seen": 901408, "step": 60 }, { "epoch": 0.7845659163987139, "grad_norm": 16.51668357849121, "learning_rate": 5.083333333333334e-07, "loss": 0.2319, "num_input_tokens_seen": 916672, "step": 61 }, { "epoch": 0.797427652733119, "grad_norm": 51.05732345581055, "learning_rate": 5.166666666666667e-07, "loss": 0.2703, "num_input_tokens_seen": 931168, "step": 62 }, { "epoch": 0.8102893890675241, "grad_norm": 61.92572784423828, "learning_rate": 5.250000000000001e-07, "loss": 0.2647, "num_input_tokens_seen": 946816, "step": 63 }, { "epoch": 0.8231511254019293, "grad_norm": 10.283011436462402, "learning_rate": 5.333333333333335e-07, "loss": 0.2288, "num_input_tokens_seen": 961248, "step": 64 }, { "epoch": 0.8360128617363344, "grad_norm": 45.712772369384766, "learning_rate": 5.416666666666667e-07, "loss": 0.2426, "num_input_tokens_seen": 976736, "step": 65 }, { "epoch": 0.8488745980707395, "grad_norm": 22.493711471557617, "learning_rate": 5.5e-07, "loss": 0.1936, "num_input_tokens_seen": 992192, "step": 66 }, { "epoch": 0.8617363344051447, "grad_norm": 10.771052360534668, "learning_rate": 5.583333333333333e-07, "loss": 0.192, "num_input_tokens_seen": 1007808, "step": 67 }, { "epoch": 0.8745980707395499, "grad_norm": 11.903575897216797, "learning_rate": 5.666666666666667e-07, "loss": 0.1956, "num_input_tokens_seen": 1022176, "step": 68 }, { "epoch": 0.887459807073955, "grad_norm": 23.1299991607666, "learning_rate": 5.750000000000001e-07, "loss": 0.1862, "num_input_tokens_seen": 1037280, "step": 69 }, { "epoch": 0.9003215434083601, "grad_norm": 11.018325805664062, "learning_rate": 5.833333333333334e-07, "loss": 0.195, "num_input_tokens_seen": 1051392, "step": 70 }, { "epoch": 0.9131832797427653, "grad_norm": 27.96196174621582, "learning_rate": 5.916666666666667e-07, "loss": 0.1971, "num_input_tokens_seen": 1066432, "step": 71 }, { "epoch": 0.9260450160771704, "grad_norm": 14.441658020019531, "learning_rate": 6.000000000000001e-07, "loss": 0.1683, "num_input_tokens_seen": 1081696, "step": 72 }, { "epoch": 0.9389067524115756, "grad_norm": 17.725786209106445, "learning_rate": 6.083333333333334e-07, "loss": 0.2003, "num_input_tokens_seen": 1096672, "step": 73 }, { "epoch": 0.9517684887459807, "grad_norm": 26.286319732666016, "learning_rate": 6.166666666666668e-07, "loss": 0.1543, "num_input_tokens_seen": 1111488, "step": 74 }, { "epoch": 0.9646302250803859, "grad_norm": 38.698097229003906, "learning_rate": 6.25e-07, "loss": 0.1982, "num_input_tokens_seen": 1126368, "step": 75 }, { "epoch": 0.977491961414791, "grad_norm": 8.287293434143066, "learning_rate": 6.333333333333334e-07, "loss": 0.1545, "num_input_tokens_seen": 1141536, "step": 76 }, { "epoch": 0.9903536977491961, "grad_norm": 44.253807067871094, "learning_rate": 6.416666666666667e-07, "loss": 0.1573, "num_input_tokens_seen": 1156320, "step": 77 }, { "epoch": 1.0032154340836013, "grad_norm": 57.11432647705078, "learning_rate": 6.5e-07, "loss": 0.1788, "num_input_tokens_seen": 1171904, "step": 78 }, { "epoch": 1.0160771704180065, "grad_norm": 21.049449920654297, "learning_rate": 6.583333333333333e-07, "loss": 0.187, "num_input_tokens_seen": 1187232, "step": 79 }, { "epoch": 1.0289389067524115, "grad_norm": 67.61766052246094, "learning_rate": 6.666666666666667e-07, "loss": 0.2009, "num_input_tokens_seen": 1201632, "step": 80 }, { "epoch": 1.0418006430868167, "grad_norm": 91.35673522949219, "learning_rate": 6.750000000000001e-07, "loss": 0.2655, "num_input_tokens_seen": 1216832, "step": 81 }, { "epoch": 1.0546623794212218, "grad_norm": 54.41016387939453, "learning_rate": 6.833333333333334e-07, "loss": 0.2041, "num_input_tokens_seen": 1232128, "step": 82 }, { "epoch": 1.067524115755627, "grad_norm": 10.082817077636719, "learning_rate": 6.916666666666668e-07, "loss": 0.1675, "num_input_tokens_seen": 1246944, "step": 83 }, { "epoch": 1.0803858520900322, "grad_norm": 43.82609558105469, "learning_rate": 7.000000000000001e-07, "loss": 0.2097, "num_input_tokens_seen": 1262624, "step": 84 }, { "epoch": 1.0932475884244373, "grad_norm": 51.51130676269531, "learning_rate": 7.083333333333334e-07, "loss": 0.2171, "num_input_tokens_seen": 1278080, "step": 85 }, { "epoch": 1.1061093247588425, "grad_norm": 35.92145538330078, "learning_rate": 7.166666666666668e-07, "loss": 0.1702, "num_input_tokens_seen": 1293280, "step": 86 }, { "epoch": 1.1189710610932475, "grad_norm": 8.067242622375488, "learning_rate": 7.25e-07, "loss": 0.1255, "num_input_tokens_seen": 1307456, "step": 87 }, { "epoch": 1.1318327974276527, "grad_norm": 28.8727970123291, "learning_rate": 7.333333333333334e-07, "loss": 0.1826, "num_input_tokens_seen": 1323168, "step": 88 }, { "epoch": 1.144694533762058, "grad_norm": 45.813865661621094, "learning_rate": 7.416666666666668e-07, "loss": 0.2039, "num_input_tokens_seen": 1337984, "step": 89 }, { "epoch": 1.157556270096463, "grad_norm": 53.024105072021484, "learning_rate": 7.5e-07, "loss": 0.2337, "num_input_tokens_seen": 1353568, "step": 90 }, { "epoch": 1.1704180064308682, "grad_norm": 17.642776489257812, "learning_rate": 7.583333333333334e-07, "loss": 0.1466, "num_input_tokens_seen": 1368800, "step": 91 }, { "epoch": 1.1832797427652733, "grad_norm": 17.741811752319336, "learning_rate": 7.666666666666667e-07, "loss": 0.1295, "num_input_tokens_seen": 1384192, "step": 92 }, { "epoch": 1.1961414790996785, "grad_norm": 24.114030838012695, "learning_rate": 7.750000000000001e-07, "loss": 0.1525, "num_input_tokens_seen": 1398784, "step": 93 }, { "epoch": 1.2090032154340835, "grad_norm": 27.668102264404297, "learning_rate": 7.833333333333335e-07, "loss": 0.1735, "num_input_tokens_seen": 1414304, "step": 94 }, { "epoch": 1.2218649517684887, "grad_norm": 7.892695426940918, "learning_rate": 7.916666666666667e-07, "loss": 0.1484, "num_input_tokens_seen": 1430880, "step": 95 }, { "epoch": 1.234726688102894, "grad_norm": 12.985011100769043, "learning_rate": 8.000000000000001e-07, "loss": 0.1517, "num_input_tokens_seen": 1445760, "step": 96 }, { "epoch": 1.247588424437299, "grad_norm": 20.373502731323242, "learning_rate": 8.083333333333334e-07, "loss": 0.1429, "num_input_tokens_seen": 1460576, "step": 97 }, { "epoch": 1.2604501607717042, "grad_norm": 8.579370498657227, "learning_rate": 8.166666666666668e-07, "loss": 0.151, "num_input_tokens_seen": 1476032, "step": 98 }, { "epoch": 1.2733118971061093, "grad_norm": 6.366987705230713, "learning_rate": 8.250000000000001e-07, "loss": 0.1217, "num_input_tokens_seen": 1491008, "step": 99 }, { "epoch": 1.2861736334405145, "grad_norm": 19.89562225341797, "learning_rate": 8.333333333333333e-07, "loss": 0.1366, "num_input_tokens_seen": 1505120, "step": 100 }, { "epoch": 1.2990353697749195, "grad_norm": 10.997241020202637, "learning_rate": 8.416666666666667e-07, "loss": 0.1534, "num_input_tokens_seen": 1520736, "step": 101 }, { "epoch": 1.3118971061093248, "grad_norm": 17.20572280883789, "learning_rate": 8.500000000000001e-07, "loss": 0.141, "num_input_tokens_seen": 1535424, "step": 102 }, { "epoch": 1.32475884244373, "grad_norm": 10.88858413696289, "learning_rate": 8.583333333333334e-07, "loss": 0.1238, "num_input_tokens_seen": 1549856, "step": 103 }, { "epoch": 1.337620578778135, "grad_norm": 6.872950077056885, "learning_rate": 8.666666666666668e-07, "loss": 0.1241, "num_input_tokens_seen": 1564512, "step": 104 }, { "epoch": 1.3504823151125402, "grad_norm": 6.931344509124756, "learning_rate": 8.75e-07, "loss": 0.1414, "num_input_tokens_seen": 1578976, "step": 105 }, { "epoch": 1.3633440514469453, "grad_norm": 12.237205505371094, "learning_rate": 8.833333333333334e-07, "loss": 0.1296, "num_input_tokens_seen": 1593952, "step": 106 }, { "epoch": 1.3762057877813505, "grad_norm": 13.147500991821289, "learning_rate": 8.916666666666668e-07, "loss": 0.1232, "num_input_tokens_seen": 1608704, "step": 107 }, { "epoch": 1.3890675241157555, "grad_norm": 23.525880813598633, "learning_rate": 9.000000000000001e-07, "loss": 0.1625, "num_input_tokens_seen": 1623936, "step": 108 }, { "epoch": 1.4019292604501608, "grad_norm": 25.576210021972656, "learning_rate": 9.083333333333335e-07, "loss": 0.1509, "num_input_tokens_seen": 1639264, "step": 109 }, { "epoch": 1.414790996784566, "grad_norm": 18.12908935546875, "learning_rate": 9.166666666666666e-07, "loss": 0.1416, "num_input_tokens_seen": 1654528, "step": 110 }, { "epoch": 1.427652733118971, "grad_norm": 14.664992332458496, "learning_rate": 9.25e-07, "loss": 0.1481, "num_input_tokens_seen": 1669824, "step": 111 }, { "epoch": 1.4405144694533762, "grad_norm": 17.804119110107422, "learning_rate": 9.333333333333334e-07, "loss": 0.1303, "num_input_tokens_seen": 1684800, "step": 112 }, { "epoch": 1.4533762057877815, "grad_norm": 10.886981964111328, "learning_rate": 9.416666666666667e-07, "loss": 0.116, "num_input_tokens_seen": 1699712, "step": 113 }, { "epoch": 1.4662379421221865, "grad_norm": 14.883463859558105, "learning_rate": 9.500000000000001e-07, "loss": 0.0981, "num_input_tokens_seen": 1714208, "step": 114 }, { "epoch": 1.4790996784565915, "grad_norm": 9.800952911376953, "learning_rate": 9.583333333333334e-07, "loss": 0.1174, "num_input_tokens_seen": 1728672, "step": 115 }, { "epoch": 1.4919614147909968, "grad_norm": 8.804801940917969, "learning_rate": 9.666666666666668e-07, "loss": 0.1458, "num_input_tokens_seen": 1743008, "step": 116 }, { "epoch": 1.504823151125402, "grad_norm": 9.836427688598633, "learning_rate": 9.750000000000002e-07, "loss": 0.0952, "num_input_tokens_seen": 1758016, "step": 117 }, { "epoch": 1.517684887459807, "grad_norm": 7.373986721038818, "learning_rate": 9.833333333333334e-07, "loss": 0.1233, "num_input_tokens_seen": 1772512, "step": 118 }, { "epoch": 1.5305466237942122, "grad_norm": 16.394031524658203, "learning_rate": 9.916666666666668e-07, "loss": 0.127, "num_input_tokens_seen": 1787808, "step": 119 }, { "epoch": 1.5434083601286175, "grad_norm": 8.336946487426758, "learning_rate": 1.0000000000000002e-06, "loss": 0.1121, "num_input_tokens_seen": 1803104, "step": 120 }, { "epoch": 1.5562700964630225, "grad_norm": 28.298938751220703, "learning_rate": 1.0083333333333333e-06, "loss": 0.1432, "num_input_tokens_seen": 1817632, "step": 121 }, { "epoch": 1.5691318327974275, "grad_norm": 25.474973678588867, "learning_rate": 1.0166666666666667e-06, "loss": 0.1446, "num_input_tokens_seen": 1832576, "step": 122 }, { "epoch": 1.5819935691318328, "grad_norm": 12.33470630645752, "learning_rate": 1.025e-06, "loss": 0.1056, "num_input_tokens_seen": 1847872, "step": 123 }, { "epoch": 1.594855305466238, "grad_norm": 16.17420768737793, "learning_rate": 1.0333333333333333e-06, "loss": 0.1193, "num_input_tokens_seen": 1863008, "step": 124 }, { "epoch": 1.607717041800643, "grad_norm": 23.109798431396484, "learning_rate": 1.0416666666666667e-06, "loss": 0.1409, "num_input_tokens_seen": 1878688, "step": 125 }, { "epoch": 1.6205787781350482, "grad_norm": 15.091007232666016, "learning_rate": 1.0500000000000001e-06, "loss": 0.1154, "num_input_tokens_seen": 1894144, "step": 126 }, { "epoch": 1.6334405144694535, "grad_norm": 7.409396171569824, "learning_rate": 1.0583333333333335e-06, "loss": 0.1046, "num_input_tokens_seen": 1909504, "step": 127 }, { "epoch": 1.6463022508038585, "grad_norm": 12.553366661071777, "learning_rate": 1.066666666666667e-06, "loss": 0.09, "num_input_tokens_seen": 1924896, "step": 128 }, { "epoch": 1.6591639871382635, "grad_norm": 10.23271369934082, "learning_rate": 1.075e-06, "loss": 0.0858, "num_input_tokens_seen": 1939232, "step": 129 }, { "epoch": 1.6720257234726688, "grad_norm": 6.091500282287598, "learning_rate": 1.0833333333333335e-06, "loss": 0.0782, "num_input_tokens_seen": 1954720, "step": 130 }, { "epoch": 1.684887459807074, "grad_norm": 12.670690536499023, "learning_rate": 1.0916666666666667e-06, "loss": 0.1429, "num_input_tokens_seen": 1969632, "step": 131 }, { "epoch": 1.697749196141479, "grad_norm": 12.42375659942627, "learning_rate": 1.1e-06, "loss": 0.1121, "num_input_tokens_seen": 1984544, "step": 132 }, { "epoch": 1.7106109324758842, "grad_norm": 4.910019874572754, "learning_rate": 1.1083333333333335e-06, "loss": 0.0458, "num_input_tokens_seen": 1999584, "step": 133 }, { "epoch": 1.7234726688102895, "grad_norm": 8.262114524841309, "learning_rate": 1.1166666666666666e-06, "loss": 0.1217, "num_input_tokens_seen": 2014048, "step": 134 }, { "epoch": 1.7363344051446945, "grad_norm": 21.927522659301758, "learning_rate": 1.125e-06, "loss": 0.1253, "num_input_tokens_seen": 2029312, "step": 135 }, { "epoch": 1.7491961414790995, "grad_norm": 8.816388130187988, "learning_rate": 1.1333333333333334e-06, "loss": 0.077, "num_input_tokens_seen": 2044192, "step": 136 }, { "epoch": 1.762057877813505, "grad_norm": 6.877133846282959, "learning_rate": 1.1416666666666668e-06, "loss": 0.0719, "num_input_tokens_seen": 2059552, "step": 137 }, { "epoch": 1.77491961414791, "grad_norm": 5.565485000610352, "learning_rate": 1.1500000000000002e-06, "loss": 0.0916, "num_input_tokens_seen": 2074336, "step": 138 }, { "epoch": 1.787781350482315, "grad_norm": 11.287223815917969, "learning_rate": 1.1583333333333334e-06, "loss": 0.0812, "num_input_tokens_seen": 2089280, "step": 139 }, { "epoch": 1.8006430868167203, "grad_norm": 7.971100807189941, "learning_rate": 1.1666666666666668e-06, "loss": 0.1176, "num_input_tokens_seen": 2104256, "step": 140 }, { "epoch": 1.8135048231511255, "grad_norm": 7.78786563873291, "learning_rate": 1.175e-06, "loss": 0.0631, "num_input_tokens_seen": 2119232, "step": 141 }, { "epoch": 1.8263665594855305, "grad_norm": 8.365852355957031, "learning_rate": 1.1833333333333334e-06, "loss": 0.1137, "num_input_tokens_seen": 2133824, "step": 142 }, { "epoch": 1.8392282958199357, "grad_norm": 16.2335262298584, "learning_rate": 1.1916666666666668e-06, "loss": 0.0958, "num_input_tokens_seen": 2148480, "step": 143 }, { "epoch": 1.852090032154341, "grad_norm": 8.455872535705566, "learning_rate": 1.2000000000000002e-06, "loss": 0.1343, "num_input_tokens_seen": 2163328, "step": 144 }, { "epoch": 1.864951768488746, "grad_norm": 13.241278648376465, "learning_rate": 1.2083333333333333e-06, "loss": 0.1101, "num_input_tokens_seen": 2177568, "step": 145 }, { "epoch": 1.877813504823151, "grad_norm": 13.46755313873291, "learning_rate": 1.2166666666666667e-06, "loss": 0.0914, "num_input_tokens_seen": 2192384, "step": 146 }, { "epoch": 1.8906752411575563, "grad_norm": 7.21525764465332, "learning_rate": 1.2250000000000001e-06, "loss": 0.1114, "num_input_tokens_seen": 2207584, "step": 147 }, { "epoch": 1.9035369774919615, "grad_norm": 5.000830173492432, "learning_rate": 1.2333333333333335e-06, "loss": 0.083, "num_input_tokens_seen": 2222208, "step": 148 }, { "epoch": 1.9163987138263665, "grad_norm": 8.995044708251953, "learning_rate": 1.2416666666666667e-06, "loss": 0.1095, "num_input_tokens_seen": 2237760, "step": 149 }, { "epoch": 1.9292604501607717, "grad_norm": 7.872910976409912, "learning_rate": 1.25e-06, "loss": 0.0662, "num_input_tokens_seen": 2252768, "step": 150 }, { "epoch": 1.942122186495177, "grad_norm": 11.594476699829102, "learning_rate": 1.2583333333333333e-06, "loss": 0.0979, "num_input_tokens_seen": 2268736, "step": 151 }, { "epoch": 1.954983922829582, "grad_norm": 14.245850563049316, "learning_rate": 1.2666666666666669e-06, "loss": 0.0847, "num_input_tokens_seen": 2283840, "step": 152 }, { "epoch": 1.967845659163987, "grad_norm": 12.000141143798828, "learning_rate": 1.275e-06, "loss": 0.0949, "num_input_tokens_seen": 2297696, "step": 153 }, { "epoch": 1.9807073954983923, "grad_norm": 16.763721466064453, "learning_rate": 1.2833333333333335e-06, "loss": 0.1206, "num_input_tokens_seen": 2311936, "step": 154 }, { "epoch": 1.9935691318327975, "grad_norm": 22.926712036132812, "learning_rate": 1.2916666666666669e-06, "loss": 0.139, "num_input_tokens_seen": 2327392, "step": 155 }, { "epoch": 2.0064308681672025, "grad_norm": 7.715733528137207, "learning_rate": 1.3e-06, "loss": 0.07, "num_input_tokens_seen": 2341952, "step": 156 }, { "epoch": 2.0192926045016075, "grad_norm": 5.047181129455566, "learning_rate": 1.3083333333333334e-06, "loss": 0.0562, "num_input_tokens_seen": 2356768, "step": 157 }, { "epoch": 2.032154340836013, "grad_norm": 10.79956340789795, "learning_rate": 1.3166666666666666e-06, "loss": 0.0456, "num_input_tokens_seen": 2371584, "step": 158 }, { "epoch": 2.045016077170418, "grad_norm": 13.864229202270508, "learning_rate": 1.3250000000000002e-06, "loss": 0.0582, "num_input_tokens_seen": 2386880, "step": 159 }, { "epoch": 2.057877813504823, "grad_norm": 7.294124603271484, "learning_rate": 1.3333333333333334e-06, "loss": 0.0452, "num_input_tokens_seen": 2401248, "step": 160 }, { "epoch": 2.0707395498392285, "grad_norm": 5.838871479034424, "learning_rate": 1.3416666666666666e-06, "loss": 0.0553, "num_input_tokens_seen": 2415968, "step": 161 }, { "epoch": 2.0836012861736335, "grad_norm": 20.286243438720703, "learning_rate": 1.3500000000000002e-06, "loss": 0.1108, "num_input_tokens_seen": 2430368, "step": 162 }, { "epoch": 2.0964630225080385, "grad_norm": 12.720942497253418, "learning_rate": 1.3583333333333334e-06, "loss": 0.0791, "num_input_tokens_seen": 2445056, "step": 163 }, { "epoch": 2.1093247588424435, "grad_norm": 9.107832908630371, "learning_rate": 1.3666666666666668e-06, "loss": 0.0637, "num_input_tokens_seen": 2459744, "step": 164 }, { "epoch": 2.122186495176849, "grad_norm": 6.473385334014893, "learning_rate": 1.3750000000000002e-06, "loss": 0.0404, "num_input_tokens_seen": 2474400, "step": 165 }, { "epoch": 2.135048231511254, "grad_norm": 6.510437488555908, "learning_rate": 1.3833333333333336e-06, "loss": 0.0372, "num_input_tokens_seen": 2489280, "step": 166 }, { "epoch": 2.147909967845659, "grad_norm": 8.90833854675293, "learning_rate": 1.3916666666666668e-06, "loss": 0.045, "num_input_tokens_seen": 2503776, "step": 167 }, { "epoch": 2.1607717041800645, "grad_norm": 11.84021282196045, "learning_rate": 1.4000000000000001e-06, "loss": 0.0964, "num_input_tokens_seen": 2518752, "step": 168 }, { "epoch": 2.1736334405144695, "grad_norm": 5.904385566711426, "learning_rate": 1.4083333333333335e-06, "loss": 0.0543, "num_input_tokens_seen": 2533568, "step": 169 }, { "epoch": 2.1864951768488745, "grad_norm": 6.935682773590088, "learning_rate": 1.4166666666666667e-06, "loss": 0.071, "num_input_tokens_seen": 2548640, "step": 170 }, { "epoch": 2.19935691318328, "grad_norm": 3.85262131690979, "learning_rate": 1.425e-06, "loss": 0.0285, "num_input_tokens_seen": 2563232, "step": 171 }, { "epoch": 2.212218649517685, "grad_norm": 5.231224060058594, "learning_rate": 1.4333333333333335e-06, "loss": 0.0399, "num_input_tokens_seen": 2578336, "step": 172 }, { "epoch": 2.22508038585209, "grad_norm": 6.751733779907227, "learning_rate": 1.4416666666666667e-06, "loss": 0.067, "num_input_tokens_seen": 2594336, "step": 173 }, { "epoch": 2.237942122186495, "grad_norm": 4.619236469268799, "learning_rate": 1.45e-06, "loss": 0.0436, "num_input_tokens_seen": 2609408, "step": 174 }, { "epoch": 2.2508038585209005, "grad_norm": 4.799732685089111, "learning_rate": 1.4583333333333335e-06, "loss": 0.0522, "num_input_tokens_seen": 2624288, "step": 175 }, { "epoch": 2.2636655948553055, "grad_norm": 7.050292015075684, "learning_rate": 1.4666666666666669e-06, "loss": 0.0521, "num_input_tokens_seen": 2639296, "step": 176 }, { "epoch": 2.2765273311897105, "grad_norm": 6.788993835449219, "learning_rate": 1.475e-06, "loss": 0.0446, "num_input_tokens_seen": 2653856, "step": 177 }, { "epoch": 2.289389067524116, "grad_norm": 7.368555545806885, "learning_rate": 1.4833333333333337e-06, "loss": 0.0378, "num_input_tokens_seen": 2669216, "step": 178 }, { "epoch": 2.302250803858521, "grad_norm": 6.674912452697754, "learning_rate": 1.4916666666666669e-06, "loss": 0.0387, "num_input_tokens_seen": 2683264, "step": 179 }, { "epoch": 2.315112540192926, "grad_norm": 4.754708766937256, "learning_rate": 1.5e-06, "loss": 0.036, "num_input_tokens_seen": 2698112, "step": 180 }, { "epoch": 2.327974276527331, "grad_norm": 11.09841251373291, "learning_rate": 1.5083333333333336e-06, "loss": 0.0765, "num_input_tokens_seen": 2712800, "step": 181 }, { "epoch": 2.3408360128617365, "grad_norm": 13.097764015197754, "learning_rate": 1.5166666666666668e-06, "loss": 0.0884, "num_input_tokens_seen": 2728032, "step": 182 }, { "epoch": 2.3536977491961415, "grad_norm": 14.366617202758789, "learning_rate": 1.525e-06, "loss": 0.0801, "num_input_tokens_seen": 2742432, "step": 183 }, { "epoch": 2.3665594855305465, "grad_norm": 6.947451591491699, "learning_rate": 1.5333333333333334e-06, "loss": 0.0276, "num_input_tokens_seen": 2758176, "step": 184 }, { "epoch": 2.379421221864952, "grad_norm": 11.02580738067627, "learning_rate": 1.5416666666666668e-06, "loss": 0.0778, "num_input_tokens_seen": 2772768, "step": 185 }, { "epoch": 2.392282958199357, "grad_norm": 15.404431343078613, "learning_rate": 1.5500000000000002e-06, "loss": 0.0726, "num_input_tokens_seen": 2787968, "step": 186 }, { "epoch": 2.405144694533762, "grad_norm": 19.688905715942383, "learning_rate": 1.5583333333333334e-06, "loss": 0.1381, "num_input_tokens_seen": 2802560, "step": 187 }, { "epoch": 2.418006430868167, "grad_norm": 5.732620716094971, "learning_rate": 1.566666666666667e-06, "loss": 0.0408, "num_input_tokens_seen": 2817856, "step": 188 }, { "epoch": 2.4308681672025725, "grad_norm": 9.736536026000977, "learning_rate": 1.5750000000000002e-06, "loss": 0.1066, "num_input_tokens_seen": 2833408, "step": 189 }, { "epoch": 2.4437299035369775, "grad_norm": 9.312838554382324, "learning_rate": 1.5833333333333333e-06, "loss": 0.0686, "num_input_tokens_seen": 2848832, "step": 190 }, { "epoch": 2.4565916398713825, "grad_norm": 7.711926460266113, "learning_rate": 1.591666666666667e-06, "loss": 0.0428, "num_input_tokens_seen": 2863968, "step": 191 }, { "epoch": 2.469453376205788, "grad_norm": 6.855398654937744, "learning_rate": 1.6000000000000001e-06, "loss": 0.0387, "num_input_tokens_seen": 2878944, "step": 192 }, { "epoch": 2.482315112540193, "grad_norm": 3.374630928039551, "learning_rate": 1.6083333333333333e-06, "loss": 0.0489, "num_input_tokens_seen": 2894112, "step": 193 }, { "epoch": 2.495176848874598, "grad_norm": 7.452945709228516, "learning_rate": 1.6166666666666667e-06, "loss": 0.0621, "num_input_tokens_seen": 2908928, "step": 194 }, { "epoch": 2.508038585209003, "grad_norm": 12.478065490722656, "learning_rate": 1.6250000000000001e-06, "loss": 0.0651, "num_input_tokens_seen": 2923680, "step": 195 }, { "epoch": 2.5209003215434085, "grad_norm": 4.181528091430664, "learning_rate": 1.6333333333333335e-06, "loss": 0.0398, "num_input_tokens_seen": 2939136, "step": 196 }, { "epoch": 2.5337620578778135, "grad_norm": 5.443283557891846, "learning_rate": 1.6416666666666667e-06, "loss": 0.0369, "num_input_tokens_seen": 2953760, "step": 197 }, { "epoch": 2.5466237942122185, "grad_norm": 7.586729526519775, "learning_rate": 1.6500000000000003e-06, "loss": 0.0582, "num_input_tokens_seen": 2968640, "step": 198 }, { "epoch": 2.559485530546624, "grad_norm": 6.357609272003174, "learning_rate": 1.6583333333333335e-06, "loss": 0.0479, "num_input_tokens_seen": 2983456, "step": 199 }, { "epoch": 2.572347266881029, "grad_norm": 8.761418342590332, "learning_rate": 1.6666666666666667e-06, "loss": 0.0561, "num_input_tokens_seen": 2998144, "step": 200 }, { "epoch": 2.585209003215434, "grad_norm": 7.058523654937744, "learning_rate": 1.6750000000000003e-06, "loss": 0.0497, "num_input_tokens_seen": 3013248, "step": 201 }, { "epoch": 2.598070739549839, "grad_norm": 8.091582298278809, "learning_rate": 1.6833333333333335e-06, "loss": 0.063, "num_input_tokens_seen": 3028320, "step": 202 }, { "epoch": 2.6109324758842445, "grad_norm": 9.102255821228027, "learning_rate": 1.6916666666666666e-06, "loss": 0.054, "num_input_tokens_seen": 3043232, "step": 203 }, { "epoch": 2.6237942122186495, "grad_norm": 9.521942138671875, "learning_rate": 1.7000000000000002e-06, "loss": 0.0729, "num_input_tokens_seen": 3058848, "step": 204 }, { "epoch": 2.6366559485530545, "grad_norm": 7.026569843292236, "learning_rate": 1.7083333333333334e-06, "loss": 0.0685, "num_input_tokens_seen": 3074080, "step": 205 }, { "epoch": 2.64951768488746, "grad_norm": 13.492867469787598, "learning_rate": 1.7166666666666668e-06, "loss": 0.0823, "num_input_tokens_seen": 3089152, "step": 206 }, { "epoch": 2.662379421221865, "grad_norm": 6.880157947540283, "learning_rate": 1.725e-06, "loss": 0.0322, "num_input_tokens_seen": 3103616, "step": 207 }, { "epoch": 2.67524115755627, "grad_norm": 8.232446670532227, "learning_rate": 1.7333333333333336e-06, "loss": 0.0821, "num_input_tokens_seen": 3119200, "step": 208 }, { "epoch": 2.688102893890675, "grad_norm": 7.365771293640137, "learning_rate": 1.7416666666666668e-06, "loss": 0.0561, "num_input_tokens_seen": 3134400, "step": 209 }, { "epoch": 2.7009646302250805, "grad_norm": 9.879140853881836, "learning_rate": 1.75e-06, "loss": 0.0468, "num_input_tokens_seen": 3149248, "step": 210 }, { "epoch": 2.7138263665594855, "grad_norm": 8.636083602905273, "learning_rate": 1.7583333333333336e-06, "loss": 0.0593, "num_input_tokens_seen": 3164800, "step": 211 }, { "epoch": 2.7266881028938905, "grad_norm": 5.015254497528076, "learning_rate": 1.7666666666666668e-06, "loss": 0.0403, "num_input_tokens_seen": 3179520, "step": 212 }, { "epoch": 2.739549839228296, "grad_norm": 7.730808258056641, "learning_rate": 1.7750000000000002e-06, "loss": 0.0459, "num_input_tokens_seen": 3194016, "step": 213 }, { "epoch": 2.752411575562701, "grad_norm": 6.433826923370361, "learning_rate": 1.7833333333333336e-06, "loss": 0.0509, "num_input_tokens_seen": 3208608, "step": 214 }, { "epoch": 2.765273311897106, "grad_norm": 7.32877779006958, "learning_rate": 1.7916666666666667e-06, "loss": 0.0873, "num_input_tokens_seen": 3224096, "step": 215 }, { "epoch": 2.778135048231511, "grad_norm": 7.283654689788818, "learning_rate": 1.8000000000000001e-06, "loss": 0.0618, "num_input_tokens_seen": 3239136, "step": 216 }, { "epoch": 2.7909967845659165, "grad_norm": 11.044650077819824, "learning_rate": 1.8083333333333335e-06, "loss": 0.1025, "num_input_tokens_seen": 3253184, "step": 217 }, { "epoch": 2.8038585209003215, "grad_norm": 7.287780284881592, "learning_rate": 1.816666666666667e-06, "loss": 0.0669, "num_input_tokens_seen": 3268000, "step": 218 }, { "epoch": 2.816720257234727, "grad_norm": 7.542795658111572, "learning_rate": 1.825e-06, "loss": 0.0447, "num_input_tokens_seen": 3282528, "step": 219 }, { "epoch": 2.829581993569132, "grad_norm": 9.812615394592285, "learning_rate": 1.8333333333333333e-06, "loss": 0.0635, "num_input_tokens_seen": 3297824, "step": 220 }, { "epoch": 2.842443729903537, "grad_norm": 5.7050580978393555, "learning_rate": 1.8416666666666669e-06, "loss": 0.0614, "num_input_tokens_seen": 3312480, "step": 221 }, { "epoch": 2.855305466237942, "grad_norm": 13.298068046569824, "learning_rate": 1.85e-06, "loss": 0.0646, "num_input_tokens_seen": 3327392, "step": 222 }, { "epoch": 2.868167202572347, "grad_norm": 13.815607070922852, "learning_rate": 1.8583333333333335e-06, "loss": 0.0893, "num_input_tokens_seen": 3342944, "step": 223 }, { "epoch": 2.8810289389067525, "grad_norm": 6.954331398010254, "learning_rate": 1.8666666666666669e-06, "loss": 0.0653, "num_input_tokens_seen": 3358112, "step": 224 }, { "epoch": 2.8938906752411575, "grad_norm": 5.133541584014893, "learning_rate": 1.8750000000000003e-06, "loss": 0.0402, "num_input_tokens_seen": 3373472, "step": 225 }, { "epoch": 2.906752411575563, "grad_norm": 7.754403114318848, "learning_rate": 1.8833333333333334e-06, "loss": 0.0407, "num_input_tokens_seen": 3389088, "step": 226 }, { "epoch": 2.919614147909968, "grad_norm": 9.570775985717773, "learning_rate": 1.8916666666666668e-06, "loss": 0.0949, "num_input_tokens_seen": 3403616, "step": 227 }, { "epoch": 2.932475884244373, "grad_norm": 9.21279525756836, "learning_rate": 1.9000000000000002e-06, "loss": 0.0789, "num_input_tokens_seen": 3418944, "step": 228 }, { "epoch": 2.945337620578778, "grad_norm": 4.255781650543213, "learning_rate": 1.9083333333333334e-06, "loss": 0.0438, "num_input_tokens_seen": 3434048, "step": 229 }, { "epoch": 2.958199356913183, "grad_norm": 10.596036911010742, "learning_rate": 1.916666666666667e-06, "loss": 0.0905, "num_input_tokens_seen": 3449952, "step": 230 }, { "epoch": 2.9710610932475885, "grad_norm": 9.656909942626953, "learning_rate": 1.925e-06, "loss": 0.0495, "num_input_tokens_seen": 3464736, "step": 231 }, { "epoch": 2.9839228295819935, "grad_norm": 9.526041984558105, "learning_rate": 1.9333333333333336e-06, "loss": 0.0642, "num_input_tokens_seen": 3480416, "step": 232 }, { "epoch": 2.996784565916399, "grad_norm": 5.788015365600586, "learning_rate": 1.9416666666666666e-06, "loss": 0.0607, "num_input_tokens_seen": 3495264, "step": 233 }, { "epoch": 3.009646302250804, "grad_norm": 6.213415145874023, "learning_rate": 1.9500000000000004e-06, "loss": 0.0363, "num_input_tokens_seen": 3510080, "step": 234 }, { "epoch": 3.022508038585209, "grad_norm": 7.382534980773926, "learning_rate": 1.9583333333333334e-06, "loss": 0.0492, "num_input_tokens_seen": 3524736, "step": 235 }, { "epoch": 3.035369774919614, "grad_norm": 9.54261302947998, "learning_rate": 1.9666666666666668e-06, "loss": 0.0373, "num_input_tokens_seen": 3539968, "step": 236 }, { "epoch": 3.0482315112540195, "grad_norm": 3.8869571685791016, "learning_rate": 1.975e-06, "loss": 0.0203, "num_input_tokens_seen": 3554464, "step": 237 }, { "epoch": 3.0610932475884245, "grad_norm": 2.183758497238159, "learning_rate": 1.9833333333333335e-06, "loss": 0.0175, "num_input_tokens_seen": 3569664, "step": 238 }, { "epoch": 3.0739549839228295, "grad_norm": 3.851379871368408, "learning_rate": 1.991666666666667e-06, "loss": 0.0497, "num_input_tokens_seen": 3585248, "step": 239 }, { "epoch": 3.0868167202572345, "grad_norm": 6.622572898864746, "learning_rate": 2.0000000000000003e-06, "loss": 0.0361, "num_input_tokens_seen": 3600160, "step": 240 }, { "epoch": 3.09967845659164, "grad_norm": 3.4328978061676025, "learning_rate": 2.0083333333333337e-06, "loss": 0.0193, "num_input_tokens_seen": 3615232, "step": 241 }, { "epoch": 3.112540192926045, "grad_norm": 8.150415420532227, "learning_rate": 2.0166666666666667e-06, "loss": 0.0142, "num_input_tokens_seen": 3629312, "step": 242 }, { "epoch": 3.12540192926045, "grad_norm": 5.3073835372924805, "learning_rate": 2.025e-06, "loss": 0.0415, "num_input_tokens_seen": 3644064, "step": 243 }, { "epoch": 3.1382636655948555, "grad_norm": 4.33547830581665, "learning_rate": 2.0333333333333335e-06, "loss": 0.0178, "num_input_tokens_seen": 3659328, "step": 244 }, { "epoch": 3.1511254019292605, "grad_norm": 2.770535469055176, "learning_rate": 2.041666666666667e-06, "loss": 0.0166, "num_input_tokens_seen": 3674720, "step": 245 }, { "epoch": 3.1639871382636655, "grad_norm": 8.939215660095215, "learning_rate": 2.05e-06, "loss": 0.0424, "num_input_tokens_seen": 3689984, "step": 246 }, { "epoch": 3.176848874598071, "grad_norm": 4.35013484954834, "learning_rate": 2.0583333333333337e-06, "loss": 0.0464, "num_input_tokens_seen": 3705216, "step": 247 }, { "epoch": 3.189710610932476, "grad_norm": 4.8758368492126465, "learning_rate": 2.0666666666666666e-06, "loss": 0.0235, "num_input_tokens_seen": 3719712, "step": 248 }, { "epoch": 3.202572347266881, "grad_norm": 2.769258499145508, "learning_rate": 2.075e-06, "loss": 0.0128, "num_input_tokens_seen": 3734272, "step": 249 }, { "epoch": 3.215434083601286, "grad_norm": 4.699342727661133, "learning_rate": 2.0833333333333334e-06, "loss": 0.0319, "num_input_tokens_seen": 3750240, "step": 250 }, { "epoch": 3.2282958199356915, "grad_norm": 3.377748966217041, "learning_rate": 2.091666666666667e-06, "loss": 0.0196, "num_input_tokens_seen": 3764736, "step": 251 }, { "epoch": 3.2411575562700965, "grad_norm": 2.9469077587127686, "learning_rate": 2.1000000000000002e-06, "loss": 0.0326, "num_input_tokens_seen": 3779744, "step": 252 }, { "epoch": 3.2540192926045015, "grad_norm": 3.4729647636413574, "learning_rate": 2.1083333333333336e-06, "loss": 0.017, "num_input_tokens_seen": 3794208, "step": 253 }, { "epoch": 3.266881028938907, "grad_norm": 3.1130521297454834, "learning_rate": 2.116666666666667e-06, "loss": 0.0372, "num_input_tokens_seen": 3809376, "step": 254 }, { "epoch": 3.279742765273312, "grad_norm": 3.0094053745269775, "learning_rate": 2.125e-06, "loss": 0.0165, "num_input_tokens_seen": 3824384, "step": 255 }, { "epoch": 3.292604501607717, "grad_norm": 2.0758721828460693, "learning_rate": 2.133333333333334e-06, "loss": 0.0142, "num_input_tokens_seen": 3840000, "step": 256 }, { "epoch": 3.305466237942122, "grad_norm": 2.0254931449890137, "learning_rate": 2.1416666666666668e-06, "loss": 0.0117, "num_input_tokens_seen": 3854240, "step": 257 }, { "epoch": 3.3183279742765275, "grad_norm": 5.01600980758667, "learning_rate": 2.15e-06, "loss": 0.0264, "num_input_tokens_seen": 3869792, "step": 258 }, { "epoch": 3.3311897106109325, "grad_norm": 7.374160289764404, "learning_rate": 2.1583333333333336e-06, "loss": 0.034, "num_input_tokens_seen": 3885152, "step": 259 }, { "epoch": 3.3440514469453375, "grad_norm": 7.036656856536865, "learning_rate": 2.166666666666667e-06, "loss": 0.031, "num_input_tokens_seen": 3900032, "step": 260 }, { "epoch": 3.356913183279743, "grad_norm": 4.439852237701416, "learning_rate": 2.1750000000000004e-06, "loss": 0.018, "num_input_tokens_seen": 3914432, "step": 261 }, { "epoch": 3.369774919614148, "grad_norm": 8.757259368896484, "learning_rate": 2.1833333333333333e-06, "loss": 0.0461, "num_input_tokens_seen": 3929600, "step": 262 }, { "epoch": 3.382636655948553, "grad_norm": 4.993197917938232, "learning_rate": 2.191666666666667e-06, "loss": 0.0372, "num_input_tokens_seen": 3944032, "step": 263 }, { "epoch": 3.395498392282958, "grad_norm": 5.627339839935303, "learning_rate": 2.2e-06, "loss": 0.0471, "num_input_tokens_seen": 3958464, "step": 264 }, { "epoch": 3.4083601286173635, "grad_norm": 6.644600868225098, "learning_rate": 2.2083333333333335e-06, "loss": 0.037, "num_input_tokens_seen": 3973024, "step": 265 }, { "epoch": 3.4212218649517685, "grad_norm": 4.708102703094482, "learning_rate": 2.216666666666667e-06, "loss": 0.0245, "num_input_tokens_seen": 3987520, "step": 266 }, { "epoch": 3.4340836012861735, "grad_norm": 3.4480528831481934, "learning_rate": 2.2250000000000003e-06, "loss": 0.0233, "num_input_tokens_seen": 4002656, "step": 267 }, { "epoch": 3.446945337620579, "grad_norm": 5.0224175453186035, "learning_rate": 2.2333333333333333e-06, "loss": 0.0256, "num_input_tokens_seen": 4017824, "step": 268 }, { "epoch": 3.459807073954984, "grad_norm": 9.12070369720459, "learning_rate": 2.2416666666666667e-06, "loss": 0.0345, "num_input_tokens_seen": 4032480, "step": 269 }, { "epoch": 3.472668810289389, "grad_norm": 5.336618900299072, "learning_rate": 2.25e-06, "loss": 0.0656, "num_input_tokens_seen": 4047648, "step": 270 }, { "epoch": 3.485530546623794, "grad_norm": 4.003684997558594, "learning_rate": 2.2583333333333335e-06, "loss": 0.0247, "num_input_tokens_seen": 4062688, "step": 271 }, { "epoch": 3.4983922829581995, "grad_norm": 5.631625175476074, "learning_rate": 2.266666666666667e-06, "loss": 0.051, "num_input_tokens_seen": 4077856, "step": 272 }, { "epoch": 3.5112540192926045, "grad_norm": 4.748563766479492, "learning_rate": 2.2750000000000002e-06, "loss": 0.0417, "num_input_tokens_seen": 4092864, "step": 273 }, { "epoch": 3.5241157556270095, "grad_norm": 3.368995189666748, "learning_rate": 2.2833333333333336e-06, "loss": 0.0176, "num_input_tokens_seen": 4107584, "step": 274 }, { "epoch": 3.536977491961415, "grad_norm": 3.463752508163452, "learning_rate": 2.2916666666666666e-06, "loss": 0.0155, "num_input_tokens_seen": 4122080, "step": 275 }, { "epoch": 3.54983922829582, "grad_norm": 4.7195234298706055, "learning_rate": 2.3000000000000004e-06, "loss": 0.0135, "num_input_tokens_seen": 4137600, "step": 276 }, { "epoch": 3.562700964630225, "grad_norm": 6.3303680419921875, "learning_rate": 2.3083333333333334e-06, "loss": 0.0222, "num_input_tokens_seen": 4152960, "step": 277 }, { "epoch": 3.57556270096463, "grad_norm": 3.4051268100738525, "learning_rate": 2.316666666666667e-06, "loss": 0.0289, "num_input_tokens_seen": 4168640, "step": 278 }, { "epoch": 3.5884244372990355, "grad_norm": 3.971222162246704, "learning_rate": 2.325e-06, "loss": 0.0314, "num_input_tokens_seen": 4183616, "step": 279 }, { "epoch": 3.6012861736334405, "grad_norm": 5.509104251861572, "learning_rate": 2.3333333333333336e-06, "loss": 0.0582, "num_input_tokens_seen": 4197536, "step": 280 }, { "epoch": 3.6141479099678455, "grad_norm": 4.121878623962402, "learning_rate": 2.341666666666667e-06, "loss": 0.0339, "num_input_tokens_seen": 4213152, "step": 281 }, { "epoch": 3.627009646302251, "grad_norm": 6.977634429931641, "learning_rate": 2.35e-06, "loss": 0.0656, "num_input_tokens_seen": 4228320, "step": 282 }, { "epoch": 3.639871382636656, "grad_norm": 3.6625123023986816, "learning_rate": 2.3583333333333338e-06, "loss": 0.0355, "num_input_tokens_seen": 4243680, "step": 283 }, { "epoch": 3.652733118971061, "grad_norm": 4.691339492797852, "learning_rate": 2.3666666666666667e-06, "loss": 0.037, "num_input_tokens_seen": 4258464, "step": 284 }, { "epoch": 3.665594855305466, "grad_norm": 3.1770055294036865, "learning_rate": 2.375e-06, "loss": 0.0206, "num_input_tokens_seen": 4273728, "step": 285 }, { "epoch": 3.6784565916398715, "grad_norm": 2.424788475036621, "learning_rate": 2.3833333333333335e-06, "loss": 0.0286, "num_input_tokens_seen": 4288416, "step": 286 }, { "epoch": 3.6913183279742765, "grad_norm": 5.731730937957764, "learning_rate": 2.391666666666667e-06, "loss": 0.0316, "num_input_tokens_seen": 4303712, "step": 287 }, { "epoch": 3.7041800643086815, "grad_norm": 4.939082622528076, "learning_rate": 2.4000000000000003e-06, "loss": 0.0263, "num_input_tokens_seen": 4319104, "step": 288 }, { "epoch": 3.717041800643087, "grad_norm": 2.817589521408081, "learning_rate": 2.4083333333333337e-06, "loss": 0.0265, "num_input_tokens_seen": 4334912, "step": 289 }, { "epoch": 3.729903536977492, "grad_norm": 2.6983306407928467, "learning_rate": 2.4166666666666667e-06, "loss": 0.0251, "num_input_tokens_seen": 4349280, "step": 290 }, { "epoch": 3.742765273311897, "grad_norm": 5.260397434234619, "learning_rate": 2.425e-06, "loss": 0.0293, "num_input_tokens_seen": 4363872, "step": 291 }, { "epoch": 3.755627009646302, "grad_norm": 4.8456034660339355, "learning_rate": 2.4333333333333335e-06, "loss": 0.0278, "num_input_tokens_seen": 4378752, "step": 292 }, { "epoch": 3.7684887459807075, "grad_norm": 4.560141563415527, "learning_rate": 2.441666666666667e-06, "loss": 0.035, "num_input_tokens_seen": 4393792, "step": 293 }, { "epoch": 3.7813504823151125, "grad_norm": 6.035860538482666, "learning_rate": 2.4500000000000003e-06, "loss": 0.0455, "num_input_tokens_seen": 4408192, "step": 294 }, { "epoch": 3.7942122186495175, "grad_norm": 6.634167194366455, "learning_rate": 2.4583333333333332e-06, "loss": 0.0429, "num_input_tokens_seen": 4422432, "step": 295 }, { "epoch": 3.807073954983923, "grad_norm": 9.78260326385498, "learning_rate": 2.466666666666667e-06, "loss": 0.056, "num_input_tokens_seen": 4437728, "step": 296 }, { "epoch": 3.819935691318328, "grad_norm": 6.473577499389648, "learning_rate": 2.475e-06, "loss": 0.0325, "num_input_tokens_seen": 4452928, "step": 297 }, { "epoch": 3.832797427652733, "grad_norm": 8.688652992248535, "learning_rate": 2.4833333333333334e-06, "loss": 0.0401, "num_input_tokens_seen": 4468096, "step": 298 }, { "epoch": 3.845659163987138, "grad_norm": 4.41790771484375, "learning_rate": 2.491666666666667e-06, "loss": 0.0182, "num_input_tokens_seen": 4482976, "step": 299 }, { "epoch": 3.8585209003215435, "grad_norm": 3.7182106971740723, "learning_rate": 2.5e-06, "loss": 0.0344, "num_input_tokens_seen": 4497952, "step": 300 }, { "epoch": 3.8713826366559485, "grad_norm": 4.0538225173950195, "learning_rate": 2.5083333333333336e-06, "loss": 0.0111, "num_input_tokens_seen": 4512896, "step": 301 }, { "epoch": 3.884244372990354, "grad_norm": 3.3492507934570312, "learning_rate": 2.5166666666666666e-06, "loss": 0.0088, "num_input_tokens_seen": 4528448, "step": 302 }, { "epoch": 3.897106109324759, "grad_norm": 3.518641710281372, "learning_rate": 2.5250000000000004e-06, "loss": 0.013, "num_input_tokens_seen": 4543712, "step": 303 }, { "epoch": 3.909967845659164, "grad_norm": 5.015272617340088, "learning_rate": 2.5333333333333338e-06, "loss": 0.0433, "num_input_tokens_seen": 4559008, "step": 304 }, { "epoch": 3.922829581993569, "grad_norm": 3.2924585342407227, "learning_rate": 2.5416666666666668e-06, "loss": 0.0214, "num_input_tokens_seen": 4573696, "step": 305 }, { "epoch": 3.935691318327974, "grad_norm": 8.486410140991211, "learning_rate": 2.55e-06, "loss": 0.0416, "num_input_tokens_seen": 4588384, "step": 306 }, { "epoch": 3.9485530546623795, "grad_norm": 6.4795002937316895, "learning_rate": 2.558333333333334e-06, "loss": 0.0321, "num_input_tokens_seen": 4602944, "step": 307 }, { "epoch": 3.9614147909967845, "grad_norm": 6.211554050445557, "learning_rate": 2.566666666666667e-06, "loss": 0.0526, "num_input_tokens_seen": 4618336, "step": 308 }, { "epoch": 3.97427652733119, "grad_norm": 5.602302074432373, "learning_rate": 2.5750000000000003e-06, "loss": 0.025, "num_input_tokens_seen": 4633472, "step": 309 }, { "epoch": 3.987138263665595, "grad_norm": 9.136455535888672, "learning_rate": 2.5833333333333337e-06, "loss": 0.0787, "num_input_tokens_seen": 4649472, "step": 310 }, { "epoch": 4.0, "grad_norm": 8.398818969726562, "learning_rate": 2.5916666666666667e-06, "loss": 0.0559, "num_input_tokens_seen": 4664384, "step": 311 }, { "epoch": 4.012861736334405, "grad_norm": 2.286210536956787, "learning_rate": 2.6e-06, "loss": 0.0259, "num_input_tokens_seen": 4679200, "step": 312 }, { "epoch": 4.02572347266881, "grad_norm": 6.477627754211426, "learning_rate": 2.608333333333333e-06, "loss": 0.037, "num_input_tokens_seen": 4693696, "step": 313 }, { "epoch": 4.038585209003215, "grad_norm": 4.953974723815918, "learning_rate": 2.616666666666667e-06, "loss": 0.0293, "num_input_tokens_seen": 4708608, "step": 314 }, { "epoch": 4.051446945337621, "grad_norm": 2.7035064697265625, "learning_rate": 2.6250000000000003e-06, "loss": 0.0153, "num_input_tokens_seen": 4722944, "step": 315 }, { "epoch": 4.064308681672026, "grad_norm": 3.8599040508270264, "learning_rate": 2.6333333333333332e-06, "loss": 0.0234, "num_input_tokens_seen": 4739296, "step": 316 }, { "epoch": 4.077170418006431, "grad_norm": 5.4720072746276855, "learning_rate": 2.6416666666666666e-06, "loss": 0.0212, "num_input_tokens_seen": 4754048, "step": 317 }, { "epoch": 4.090032154340836, "grad_norm": 3.8060128688812256, "learning_rate": 2.6500000000000005e-06, "loss": 0.0187, "num_input_tokens_seen": 4769376, "step": 318 }, { "epoch": 4.102893890675241, "grad_norm": 5.436901092529297, "learning_rate": 2.6583333333333334e-06, "loss": 0.0245, "num_input_tokens_seen": 4784640, "step": 319 }, { "epoch": 4.115755627009646, "grad_norm": 1.5260390043258667, "learning_rate": 2.666666666666667e-06, "loss": 0.0032, "num_input_tokens_seen": 4799392, "step": 320 }, { "epoch": 4.128617363344051, "grad_norm": 11.145827293395996, "learning_rate": 2.6750000000000002e-06, "loss": 0.0367, "num_input_tokens_seen": 4813696, "step": 321 }, { "epoch": 4.141479099678457, "grad_norm": 5.737987518310547, "learning_rate": 2.683333333333333e-06, "loss": 0.038, "num_input_tokens_seen": 4828928, "step": 322 }, { "epoch": 4.154340836012862, "grad_norm": 12.955948829650879, "learning_rate": 2.691666666666667e-06, "loss": 0.0194, "num_input_tokens_seen": 4843232, "step": 323 }, { "epoch": 4.167202572347267, "grad_norm": 2.663821220397949, "learning_rate": 2.7000000000000004e-06, "loss": 0.0262, "num_input_tokens_seen": 4858592, "step": 324 }, { "epoch": 4.180064308681672, "grad_norm": 7.183210372924805, "learning_rate": 2.7083333333333334e-06, "loss": 0.0151, "num_input_tokens_seen": 4872928, "step": 325 }, { "epoch": 4.192926045016077, "grad_norm": 5.024378776550293, "learning_rate": 2.7166666666666668e-06, "loss": 0.0267, "num_input_tokens_seen": 4887968, "step": 326 }, { "epoch": 4.205787781350482, "grad_norm": 1.0667839050292969, "learning_rate": 2.7250000000000006e-06, "loss": 0.0041, "num_input_tokens_seen": 4902752, "step": 327 }, { "epoch": 4.218649517684887, "grad_norm": 6.243584632873535, "learning_rate": 2.7333333333333336e-06, "loss": 0.0365, "num_input_tokens_seen": 4917280, "step": 328 }, { "epoch": 4.231511254019293, "grad_norm": 3.255237579345703, "learning_rate": 2.741666666666667e-06, "loss": 0.0151, "num_input_tokens_seen": 4933088, "step": 329 }, { "epoch": 4.244372990353698, "grad_norm": 1.9820326566696167, "learning_rate": 2.7500000000000004e-06, "loss": 0.006, "num_input_tokens_seen": 4948128, "step": 330 }, { "epoch": 4.257234726688103, "grad_norm": 4.553264141082764, "learning_rate": 2.7583333333333333e-06, "loss": 0.0247, "num_input_tokens_seen": 4962752, "step": 331 }, { "epoch": 4.270096463022508, "grad_norm": 3.139723062515259, "learning_rate": 2.766666666666667e-06, "loss": 0.0147, "num_input_tokens_seen": 4978272, "step": 332 }, { "epoch": 4.282958199356913, "grad_norm": 9.011144638061523, "learning_rate": 2.7750000000000005e-06, "loss": 0.0413, "num_input_tokens_seen": 4992384, "step": 333 }, { "epoch": 4.295819935691318, "grad_norm": 11.36248779296875, "learning_rate": 2.7833333333333335e-06, "loss": 0.0318, "num_input_tokens_seen": 5007136, "step": 334 }, { "epoch": 4.308681672025724, "grad_norm": 7.9140305519104, "learning_rate": 2.791666666666667e-06, "loss": 0.0513, "num_input_tokens_seen": 5021600, "step": 335 }, { "epoch": 4.321543408360129, "grad_norm": 4.12211799621582, "learning_rate": 2.8000000000000003e-06, "loss": 0.0165, "num_input_tokens_seen": 5037216, "step": 336 }, { "epoch": 4.334405144694534, "grad_norm": 4.724823951721191, "learning_rate": 2.8083333333333333e-06, "loss": 0.0131, "num_input_tokens_seen": 5052320, "step": 337 }, { "epoch": 4.347266881028939, "grad_norm": 6.9196085929870605, "learning_rate": 2.816666666666667e-06, "loss": 0.0246, "num_input_tokens_seen": 5067552, "step": 338 }, { "epoch": 4.360128617363344, "grad_norm": 5.785276412963867, "learning_rate": 2.825e-06, "loss": 0.0465, "num_input_tokens_seen": 5082240, "step": 339 }, { "epoch": 4.372990353697749, "grad_norm": 1.6201478242874146, "learning_rate": 2.8333333333333335e-06, "loss": 0.0084, "num_input_tokens_seen": 5097408, "step": 340 }, { "epoch": 4.385852090032154, "grad_norm": 5.511898994445801, "learning_rate": 2.841666666666667e-06, "loss": 0.0105, "num_input_tokens_seen": 5111680, "step": 341 }, { "epoch": 4.39871382636656, "grad_norm": 8.025418281555176, "learning_rate": 2.85e-06, "loss": 0.0388, "num_input_tokens_seen": 5127264, "step": 342 }, { "epoch": 4.411575562700965, "grad_norm": 12.538291931152344, "learning_rate": 2.8583333333333336e-06, "loss": 0.0486, "num_input_tokens_seen": 5142464, "step": 343 }, { "epoch": 4.42443729903537, "grad_norm": 10.100396156311035, "learning_rate": 2.866666666666667e-06, "loss": 0.0354, "num_input_tokens_seen": 5157376, "step": 344 }, { "epoch": 4.437299035369775, "grad_norm": 4.218758583068848, "learning_rate": 2.875e-06, "loss": 0.0464, "num_input_tokens_seen": 5171648, "step": 345 }, { "epoch": 4.45016077170418, "grad_norm": 5.751765251159668, "learning_rate": 2.8833333333333334e-06, "loss": 0.0312, "num_input_tokens_seen": 5186752, "step": 346 }, { "epoch": 4.463022508038585, "grad_norm": 6.100085258483887, "learning_rate": 2.8916666666666672e-06, "loss": 0.0352, "num_input_tokens_seen": 5202176, "step": 347 }, { "epoch": 4.47588424437299, "grad_norm": 6.4870991706848145, "learning_rate": 2.9e-06, "loss": 0.0116, "num_input_tokens_seen": 5216896, "step": 348 }, { "epoch": 4.488745980707396, "grad_norm": 1.8258070945739746, "learning_rate": 2.9083333333333336e-06, "loss": 0.0118, "num_input_tokens_seen": 5232448, "step": 349 }, { "epoch": 4.501607717041801, "grad_norm": 2.6930766105651855, "learning_rate": 2.916666666666667e-06, "loss": 0.0138, "num_input_tokens_seen": 5247296, "step": 350 }, { "epoch": 4.514469453376206, "grad_norm": 5.420289993286133, "learning_rate": 2.925e-06, "loss": 0.0344, "num_input_tokens_seen": 5262304, "step": 351 }, { "epoch": 4.527331189710611, "grad_norm": 4.682693958282471, "learning_rate": 2.9333333333333338e-06, "loss": 0.0231, "num_input_tokens_seen": 5277664, "step": 352 }, { "epoch": 4.540192926045016, "grad_norm": 4.194357872009277, "learning_rate": 2.941666666666667e-06, "loss": 0.0267, "num_input_tokens_seen": 5292576, "step": 353 }, { "epoch": 4.553054662379421, "grad_norm": 1.0653674602508545, "learning_rate": 2.95e-06, "loss": 0.0109, "num_input_tokens_seen": 5306304, "step": 354 }, { "epoch": 4.565916398713826, "grad_norm": 1.8720684051513672, "learning_rate": 2.9583333333333335e-06, "loss": 0.0444, "num_input_tokens_seen": 5320768, "step": 355 }, { "epoch": 4.578778135048232, "grad_norm": 4.005693435668945, "learning_rate": 2.9666666666666673e-06, "loss": 0.0142, "num_input_tokens_seen": 5336128, "step": 356 }, { "epoch": 4.591639871382637, "grad_norm": 5.997867107391357, "learning_rate": 2.9750000000000003e-06, "loss": 0.0361, "num_input_tokens_seen": 5351168, "step": 357 }, { "epoch": 4.604501607717042, "grad_norm": 3.8871259689331055, "learning_rate": 2.9833333333333337e-06, "loss": 0.0197, "num_input_tokens_seen": 5366112, "step": 358 }, { "epoch": 4.617363344051447, "grad_norm": 4.269320011138916, "learning_rate": 2.991666666666667e-06, "loss": 0.0373, "num_input_tokens_seen": 5381440, "step": 359 }, { "epoch": 4.630225080385852, "grad_norm": 3.6487338542938232, "learning_rate": 3e-06, "loss": 0.0469, "num_input_tokens_seen": 5397152, "step": 360 }, { "epoch": 4.643086816720257, "grad_norm": 3.9794583320617676, "learning_rate": 3.0083333333333335e-06, "loss": 0.0283, "num_input_tokens_seen": 5412128, "step": 361 }, { "epoch": 4.655948553054662, "grad_norm": 5.071774482727051, "learning_rate": 3.0166666666666673e-06, "loss": 0.0271, "num_input_tokens_seen": 5426816, "step": 362 }, { "epoch": 4.668810289389068, "grad_norm": 3.651871681213379, "learning_rate": 3.0250000000000003e-06, "loss": 0.0303, "num_input_tokens_seen": 5441760, "step": 363 }, { "epoch": 4.681672025723473, "grad_norm": 3.1399481296539307, "learning_rate": 3.0333333333333337e-06, "loss": 0.0207, "num_input_tokens_seen": 5457504, "step": 364 }, { "epoch": 4.694533762057878, "grad_norm": 3.089629888534546, "learning_rate": 3.0416666666666666e-06, "loss": 0.0157, "num_input_tokens_seen": 5473088, "step": 365 }, { "epoch": 4.707395498392283, "grad_norm": 1.4184813499450684, "learning_rate": 3.05e-06, "loss": 0.0128, "num_input_tokens_seen": 5487712, "step": 366 }, { "epoch": 4.720257234726688, "grad_norm": 5.21925687789917, "learning_rate": 3.058333333333334e-06, "loss": 0.0293, "num_input_tokens_seen": 5503680, "step": 367 }, { "epoch": 4.733118971061093, "grad_norm": 3.7801499366760254, "learning_rate": 3.066666666666667e-06, "loss": 0.0211, "num_input_tokens_seen": 5518464, "step": 368 }, { "epoch": 4.745980707395498, "grad_norm": 9.715234756469727, "learning_rate": 3.075e-06, "loss": 0.0929, "num_input_tokens_seen": 5533504, "step": 369 }, { "epoch": 4.758842443729904, "grad_norm": 6.241518974304199, "learning_rate": 3.0833333333333336e-06, "loss": 0.0195, "num_input_tokens_seen": 5548832, "step": 370 }, { "epoch": 4.771704180064309, "grad_norm": 2.6658706665039062, "learning_rate": 3.0916666666666666e-06, "loss": 0.0114, "num_input_tokens_seen": 5564256, "step": 371 }, { "epoch": 4.784565916398714, "grad_norm": 7.38095235824585, "learning_rate": 3.1000000000000004e-06, "loss": 0.0207, "num_input_tokens_seen": 5580064, "step": 372 }, { "epoch": 4.797427652733119, "grad_norm": 5.605347156524658, "learning_rate": 3.1083333333333338e-06, "loss": 0.0144, "num_input_tokens_seen": 5595008, "step": 373 }, { "epoch": 4.810289389067524, "grad_norm": 4.410486221313477, "learning_rate": 3.1166666666666668e-06, "loss": 0.0197, "num_input_tokens_seen": 5610304, "step": 374 }, { "epoch": 4.823151125401929, "grad_norm": 5.259794235229492, "learning_rate": 3.125e-06, "loss": 0.0669, "num_input_tokens_seen": 5626464, "step": 375 }, { "epoch": 4.836012861736334, "grad_norm": 5.21687650680542, "learning_rate": 3.133333333333334e-06, "loss": 0.0156, "num_input_tokens_seen": 5641216, "step": 376 }, { "epoch": 4.84887459807074, "grad_norm": 3.3515796661376953, "learning_rate": 3.141666666666667e-06, "loss": 0.018, "num_input_tokens_seen": 5656832, "step": 377 }, { "epoch": 4.861736334405145, "grad_norm": 7.488304138183594, "learning_rate": 3.1500000000000003e-06, "loss": 0.064, "num_input_tokens_seen": 5671744, "step": 378 }, { "epoch": 4.87459807073955, "grad_norm": 2.625629186630249, "learning_rate": 3.1583333333333337e-06, "loss": 0.0263, "num_input_tokens_seen": 5686496, "step": 379 }, { "epoch": 4.887459807073955, "grad_norm": 3.379544496536255, "learning_rate": 3.1666666666666667e-06, "loss": 0.0256, "num_input_tokens_seen": 5701216, "step": 380 }, { "epoch": 4.90032154340836, "grad_norm": 2.0447638034820557, "learning_rate": 3.175e-06, "loss": 0.0269, "num_input_tokens_seen": 5715904, "step": 381 }, { "epoch": 4.913183279742765, "grad_norm": 3.5290844440460205, "learning_rate": 3.183333333333334e-06, "loss": 0.0281, "num_input_tokens_seen": 5730976, "step": 382 }, { "epoch": 4.92604501607717, "grad_norm": 2.183229923248291, "learning_rate": 3.191666666666667e-06, "loss": 0.0199, "num_input_tokens_seen": 5745824, "step": 383 }, { "epoch": 4.938906752411576, "grad_norm": 1.9661577939987183, "learning_rate": 3.2000000000000003e-06, "loss": 0.0277, "num_input_tokens_seen": 5760160, "step": 384 }, { "epoch": 4.951768488745981, "grad_norm": 5.09304666519165, "learning_rate": 3.2083333333333337e-06, "loss": 0.0158, "num_input_tokens_seen": 5776160, "step": 385 }, { "epoch": 4.951768488745981, "num_input_tokens_seen": 5776160, "step": 385, "total_flos": 2.6009787341943603e+17, "train_loss": 1.2260490821160424, "train_runtime": 5290.2175, "train_samples_per_second": 18.789, "train_steps_per_second": 0.073 } ], "logging_steps": 1, "max_steps": 385, "num_input_tokens_seen": 5776160, "num_train_epochs": 5, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.6009787341943603e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }