{ "best_global_step": 400, "best_metric": 0.17246778309345245, "best_model_checkpoint": "saves/lora/gemma-3-1b-it/train_copa_1745950323/checkpoint-400", "epoch": 444.44444444444446, "eval_steps": 200, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05555555555555555, "grad_norm": 52.05956268310547, "learning_rate": 4.999999876629946e-05, "loss": 6.3509, "num_input_tokens_seen": 1376, "step": 5 }, { "epoch": 0.1111111111111111, "grad_norm": 8.793147087097168, "learning_rate": 4.999999375439123e-05, "loss": 2.1914, "num_input_tokens_seen": 2800, "step": 10 }, { "epoch": 0.16666666666666666, "grad_norm": 14.541181564331055, "learning_rate": 4.9999984887169785e-05, "loss": 0.32, "num_input_tokens_seen": 4240, "step": 15 }, { "epoch": 0.2222222222222222, "grad_norm": 27.98386001586914, "learning_rate": 4.9999972164636506e-05, "loss": 0.3556, "num_input_tokens_seen": 5664, "step": 20 }, { "epoch": 0.2777777777777778, "grad_norm": 6.053948879241943, "learning_rate": 4.999995558679334e-05, "loss": 0.2596, "num_input_tokens_seen": 7072, "step": 25 }, { "epoch": 0.3333333333333333, "grad_norm": 1.9764091968536377, "learning_rate": 4.999993515364287e-05, "loss": 0.1986, "num_input_tokens_seen": 8480, "step": 30 }, { "epoch": 0.3888888888888889, "grad_norm": 4.875693321228027, "learning_rate": 4.999991086518822e-05, "loss": 0.4441, "num_input_tokens_seen": 9840, "step": 35 }, { "epoch": 0.4444444444444444, "grad_norm": 1.1535682678222656, "learning_rate": 4.999988272143315e-05, "loss": 0.3309, "num_input_tokens_seen": 11200, "step": 40 }, { "epoch": 0.5, "grad_norm": 0.5426120758056641, "learning_rate": 4.999985072238199e-05, "loss": 0.172, "num_input_tokens_seen": 12640, "step": 45 }, { "epoch": 0.5555555555555556, "grad_norm": 2.3886866569519043, "learning_rate": 4.999981486803969e-05, "loss": 0.2694, "num_input_tokens_seen": 14032, "step": 50 }, { "epoch": 0.6111111111111112, "grad_norm": 2.607588768005371, "learning_rate": 4.999977515841176e-05, "loss": 0.1698, "num_input_tokens_seen": 15472, "step": 55 }, { "epoch": 0.6666666666666666, "grad_norm": 1.3638684749603271, "learning_rate": 4.9999731593504344e-05, "loss": 0.1864, "num_input_tokens_seen": 16896, "step": 60 }, { "epoch": 0.7222222222222222, "grad_norm": 2.051412582397461, "learning_rate": 4.999968417332415e-05, "loss": 0.1717, "num_input_tokens_seen": 18336, "step": 65 }, { "epoch": 0.7777777777777778, "grad_norm": 1.8135327100753784, "learning_rate": 4.999963289787848e-05, "loss": 0.1992, "num_input_tokens_seen": 19760, "step": 70 }, { "epoch": 0.8333333333333334, "grad_norm": 0.8167670369148254, "learning_rate": 4.999957776717526e-05, "loss": 0.1836, "num_input_tokens_seen": 21136, "step": 75 }, { "epoch": 0.8888888888888888, "grad_norm": 0.6489667296409607, "learning_rate": 4.9999518781222984e-05, "loss": 0.1925, "num_input_tokens_seen": 22544, "step": 80 }, { "epoch": 0.9444444444444444, "grad_norm": 1.4682481288909912, "learning_rate": 4.9999455940030746e-05, "loss": 0.1705, "num_input_tokens_seen": 23920, "step": 85 }, { "epoch": 1.0, "grad_norm": 0.3783438801765442, "learning_rate": 4.999938924360824e-05, "loss": 0.177, "num_input_tokens_seen": 25312, "step": 90 }, { "epoch": 1.0555555555555556, "grad_norm": 0.31409984827041626, "learning_rate": 4.999931869196575e-05, "loss": 0.1814, "num_input_tokens_seen": 26720, "step": 95 }, { "epoch": 1.1111111111111112, "grad_norm": 1.3968892097473145, "learning_rate": 4.999924428511416e-05, "loss": 0.191, "num_input_tokens_seen": 28096, "step": 100 }, { "epoch": 1.1666666666666667, "grad_norm": 0.21890664100646973, "learning_rate": 4.999916602306494e-05, "loss": 0.174, "num_input_tokens_seen": 29488, "step": 105 }, { "epoch": 1.2222222222222223, "grad_norm": 0.70839524269104, "learning_rate": 4.999908390583016e-05, "loss": 0.1748, "num_input_tokens_seen": 30880, "step": 110 }, { "epoch": 1.2777777777777777, "grad_norm": 0.8897643089294434, "learning_rate": 4.999899793342247e-05, "loss": 0.1835, "num_input_tokens_seen": 32256, "step": 115 }, { "epoch": 1.3333333333333333, "grad_norm": 0.2501010298728943, "learning_rate": 4.999890810585516e-05, "loss": 0.1731, "num_input_tokens_seen": 33616, "step": 120 }, { "epoch": 1.3888888888888888, "grad_norm": 0.8290737867355347, "learning_rate": 4.999881442314206e-05, "loss": 0.1687, "num_input_tokens_seen": 35040, "step": 125 }, { "epoch": 1.4444444444444444, "grad_norm": 1.3983407020568848, "learning_rate": 4.9998716885297617e-05, "loss": 0.1833, "num_input_tokens_seen": 36480, "step": 130 }, { "epoch": 1.5, "grad_norm": 0.15893539786338806, "learning_rate": 4.999861549233688e-05, "loss": 0.1851, "num_input_tokens_seen": 37888, "step": 135 }, { "epoch": 1.5555555555555556, "grad_norm": 0.2590113878250122, "learning_rate": 4.999851024427548e-05, "loss": 0.184, "num_input_tokens_seen": 39296, "step": 140 }, { "epoch": 1.6111111111111112, "grad_norm": 0.8140949010848999, "learning_rate": 4.999840114112965e-05, "loss": 0.1938, "num_input_tokens_seen": 40704, "step": 145 }, { "epoch": 1.6666666666666665, "grad_norm": 0.48039984703063965, "learning_rate": 4.999828818291621e-05, "loss": 0.164, "num_input_tokens_seen": 42112, "step": 150 }, { "epoch": 1.7222222222222223, "grad_norm": 1.1865350008010864, "learning_rate": 4.999817136965259e-05, "loss": 0.1801, "num_input_tokens_seen": 43520, "step": 155 }, { "epoch": 1.7777777777777777, "grad_norm": 0.48067542910575867, "learning_rate": 4.9998050701356794e-05, "loss": 0.1703, "num_input_tokens_seen": 44912, "step": 160 }, { "epoch": 1.8333333333333335, "grad_norm": 0.4895879924297333, "learning_rate": 4.999792617804744e-05, "loss": 0.1788, "num_input_tokens_seen": 46272, "step": 165 }, { "epoch": 1.8888888888888888, "grad_norm": 0.1481882929801941, "learning_rate": 4.9997797799743724e-05, "loss": 0.1666, "num_input_tokens_seen": 47664, "step": 170 }, { "epoch": 1.9444444444444444, "grad_norm": 0.15430331230163574, "learning_rate": 4.999766556646545e-05, "loss": 0.177, "num_input_tokens_seen": 49072, "step": 175 }, { "epoch": 2.0, "grad_norm": 0.44250163435935974, "learning_rate": 4.9997529478232996e-05, "loss": 0.1778, "num_input_tokens_seen": 50464, "step": 180 }, { "epoch": 2.0555555555555554, "grad_norm": 0.6995267271995544, "learning_rate": 4.9997389535067365e-05, "loss": 0.1736, "num_input_tokens_seen": 51904, "step": 185 }, { "epoch": 2.111111111111111, "grad_norm": 0.49313464760780334, "learning_rate": 4.999724573699012e-05, "loss": 0.1731, "num_input_tokens_seen": 53280, "step": 190 }, { "epoch": 2.1666666666666665, "grad_norm": 0.579094409942627, "learning_rate": 4.9997098084023457e-05, "loss": 0.1707, "num_input_tokens_seen": 54720, "step": 195 }, { "epoch": 2.2222222222222223, "grad_norm": 0.1603352576494217, "learning_rate": 4.999694657619013e-05, "loss": 0.1727, "num_input_tokens_seen": 56176, "step": 200 }, { "epoch": 2.2222222222222223, "eval_loss": 0.1755761057138443, "eval_runtime": 1.3942, "eval_samples_per_second": 28.69, "eval_steps_per_second": 14.345, "num_input_tokens_seen": 56176, "step": 200 }, { "epoch": 2.2777777777777777, "grad_norm": 0.46050429344177246, "learning_rate": 4.999679121351352e-05, "loss": 0.1796, "num_input_tokens_seen": 57584, "step": 205 }, { "epoch": 2.3333333333333335, "grad_norm": 1.1872974634170532, "learning_rate": 4.9996631996017565e-05, "loss": 0.1803, "num_input_tokens_seen": 58944, "step": 210 }, { "epoch": 2.388888888888889, "grad_norm": 0.10493234544992447, "learning_rate": 4.9996468923726835e-05, "loss": 0.1739, "num_input_tokens_seen": 60352, "step": 215 }, { "epoch": 2.4444444444444446, "grad_norm": 0.10411107540130615, "learning_rate": 4.999630199666647e-05, "loss": 0.1777, "num_input_tokens_seen": 61728, "step": 220 }, { "epoch": 2.5, "grad_norm": 0.46297702193260193, "learning_rate": 4.999613121486222e-05, "loss": 0.1668, "num_input_tokens_seen": 63136, "step": 225 }, { "epoch": 2.5555555555555554, "grad_norm": 0.15567085146903992, "learning_rate": 4.999595657834041e-05, "loss": 0.1742, "num_input_tokens_seen": 64528, "step": 230 }, { "epoch": 2.611111111111111, "grad_norm": 0.3540121614933014, "learning_rate": 4.999577808712798e-05, "loss": 0.1622, "num_input_tokens_seen": 65936, "step": 235 }, { "epoch": 2.6666666666666665, "grad_norm": 1.2461355924606323, "learning_rate": 4.999559574125244e-05, "loss": 0.1895, "num_input_tokens_seen": 67312, "step": 240 }, { "epoch": 2.7222222222222223, "grad_norm": 0.1442859023809433, "learning_rate": 4.9995409540741934e-05, "loss": 0.1847, "num_input_tokens_seen": 68768, "step": 245 }, { "epoch": 2.7777777777777777, "grad_norm": 0.49313804507255554, "learning_rate": 4.999521948562516e-05, "loss": 0.1694, "num_input_tokens_seen": 70144, "step": 250 }, { "epoch": 2.8333333333333335, "grad_norm": 0.5397193431854248, "learning_rate": 4.999502557593143e-05, "loss": 0.171, "num_input_tokens_seen": 71568, "step": 255 }, { "epoch": 2.888888888888889, "grad_norm": 0.16595672070980072, "learning_rate": 4.999482781169066e-05, "loss": 0.1713, "num_input_tokens_seen": 72976, "step": 260 }, { "epoch": 2.9444444444444446, "grad_norm": 0.7038723230361938, "learning_rate": 4.9994626192933324e-05, "loss": 0.163, "num_input_tokens_seen": 74368, "step": 265 }, { "epoch": 3.0, "grad_norm": 0.6314331293106079, "learning_rate": 4.999442071969054e-05, "loss": 0.2079, "num_input_tokens_seen": 75760, "step": 270 }, { "epoch": 3.0555555555555554, "grad_norm": 0.3439156115055084, "learning_rate": 4.999421139199397e-05, "loss": 0.1605, "num_input_tokens_seen": 77168, "step": 275 }, { "epoch": 3.111111111111111, "grad_norm": 0.5454378724098206, "learning_rate": 4.999399820987592e-05, "loss": 0.1707, "num_input_tokens_seen": 78624, "step": 280 }, { "epoch": 3.1666666666666665, "grad_norm": 0.352318674325943, "learning_rate": 4.999378117336924e-05, "loss": 0.1714, "num_input_tokens_seen": 80016, "step": 285 }, { "epoch": 3.2222222222222223, "grad_norm": 0.8695530295372009, "learning_rate": 4.9993560282507415e-05, "loss": 0.1803, "num_input_tokens_seen": 81424, "step": 290 }, { "epoch": 3.2777777777777777, "grad_norm": 0.1707405000925064, "learning_rate": 4.9993335537324495e-05, "loss": 0.1709, "num_input_tokens_seen": 82784, "step": 295 }, { "epoch": 3.3333333333333335, "grad_norm": 0.33875563740730286, "learning_rate": 4.999310693785516e-05, "loss": 0.1735, "num_input_tokens_seen": 84208, "step": 300 }, { "epoch": 3.388888888888889, "grad_norm": 1.208124041557312, "learning_rate": 4.9992874484134653e-05, "loss": 0.1683, "num_input_tokens_seen": 85632, "step": 305 }, { "epoch": 3.4444444444444446, "grad_norm": 0.793390154838562, "learning_rate": 4.999263817619882e-05, "loss": 0.1799, "num_input_tokens_seen": 87040, "step": 310 }, { "epoch": 3.5, "grad_norm": 0.29905351996421814, "learning_rate": 4.9992398014084105e-05, "loss": 0.1675, "num_input_tokens_seen": 88432, "step": 315 }, { "epoch": 3.5555555555555554, "grad_norm": 0.2570115029811859, "learning_rate": 4.999215399782754e-05, "loss": 0.167, "num_input_tokens_seen": 89824, "step": 320 }, { "epoch": 3.611111111111111, "grad_norm": 0.5243192911148071, "learning_rate": 4.999190612746675e-05, "loss": 0.1873, "num_input_tokens_seen": 91216, "step": 325 }, { "epoch": 3.6666666666666665, "grad_norm": 0.571905255317688, "learning_rate": 4.999165440303998e-05, "loss": 0.175, "num_input_tokens_seen": 92640, "step": 330 }, { "epoch": 3.7222222222222223, "grad_norm": 1.2190546989440918, "learning_rate": 4.999139882458603e-05, "loss": 0.1802, "num_input_tokens_seen": 94032, "step": 335 }, { "epoch": 3.7777777777777777, "grad_norm": 0.14073139429092407, "learning_rate": 4.9991139392144314e-05, "loss": 0.1834, "num_input_tokens_seen": 95424, "step": 340 }, { "epoch": 3.8333333333333335, "grad_norm": 0.8736257553100586, "learning_rate": 4.999087610575485e-05, "loss": 0.1738, "num_input_tokens_seen": 96816, "step": 345 }, { "epoch": 3.888888888888889, "grad_norm": 0.4504700005054474, "learning_rate": 4.999060896545824e-05, "loss": 0.167, "num_input_tokens_seen": 98192, "step": 350 }, { "epoch": 3.9444444444444446, "grad_norm": 1.014581561088562, "learning_rate": 4.999033797129568e-05, "loss": 0.1743, "num_input_tokens_seen": 99584, "step": 355 }, { "epoch": 4.0, "grad_norm": 0.27219530940055847, "learning_rate": 4.999006312330894e-05, "loss": 0.1827, "num_input_tokens_seen": 100960, "step": 360 }, { "epoch": 4.055555555555555, "grad_norm": 0.43597641587257385, "learning_rate": 4.998978442154043e-05, "loss": 0.1756, "num_input_tokens_seen": 102400, "step": 365 }, { "epoch": 4.111111111111111, "grad_norm": 0.4349745810031891, "learning_rate": 4.9989501866033125e-05, "loss": 0.1699, "num_input_tokens_seen": 103792, "step": 370 }, { "epoch": 4.166666666666667, "grad_norm": 0.11765392124652863, "learning_rate": 4.998921545683059e-05, "loss": 0.1731, "num_input_tokens_seen": 105136, "step": 375 }, { "epoch": 4.222222222222222, "grad_norm": 0.24583174288272858, "learning_rate": 4.9988925193976996e-05, "loss": 0.1731, "num_input_tokens_seen": 106528, "step": 380 }, { "epoch": 4.277777777777778, "grad_norm": 0.4225457012653351, "learning_rate": 4.998863107751711e-05, "loss": 0.1745, "num_input_tokens_seen": 107936, "step": 385 }, { "epoch": 4.333333333333333, "grad_norm": 0.4553951919078827, "learning_rate": 4.998833310749629e-05, "loss": 0.1764, "num_input_tokens_seen": 109296, "step": 390 }, { "epoch": 4.388888888888889, "grad_norm": 0.42602837085723877, "learning_rate": 4.998803128396047e-05, "loss": 0.1779, "num_input_tokens_seen": 110688, "step": 395 }, { "epoch": 4.444444444444445, "grad_norm": 0.36668962240219116, "learning_rate": 4.9987725606956215e-05, "loss": 0.1686, "num_input_tokens_seen": 112064, "step": 400 }, { "epoch": 4.444444444444445, "eval_loss": 0.17246778309345245, "eval_runtime": 1.3922, "eval_samples_per_second": 28.732, "eval_steps_per_second": 14.366, "num_input_tokens_seen": 112064, "step": 400 }, { "epoch": 4.5, "grad_norm": 0.3358176648616791, "learning_rate": 4.998741607653066e-05, "loss": 0.1658, "num_input_tokens_seen": 113472, "step": 405 }, { "epoch": 4.555555555555555, "grad_norm": 0.8441155552864075, "learning_rate": 4.9987102692731523e-05, "loss": 0.1831, "num_input_tokens_seen": 114880, "step": 410 }, { "epoch": 4.611111111111111, "grad_norm": 0.5721865892410278, "learning_rate": 4.9986785455607157e-05, "loss": 0.1634, "num_input_tokens_seen": 116224, "step": 415 }, { "epoch": 4.666666666666667, "grad_norm": 0.4790472090244293, "learning_rate": 4.9986464365206456e-05, "loss": 0.1779, "num_input_tokens_seen": 117680, "step": 420 }, { "epoch": 4.722222222222222, "grad_norm": 0.4134179353713989, "learning_rate": 4.9986139421578956e-05, "loss": 0.1751, "num_input_tokens_seen": 119072, "step": 425 }, { "epoch": 4.777777777777778, "grad_norm": 0.434089332818985, "learning_rate": 4.998581062477477e-05, "loss": 0.17, "num_input_tokens_seen": 120496, "step": 430 }, { "epoch": 4.833333333333333, "grad_norm": 0.7444345951080322, "learning_rate": 4.998547797484458e-05, "loss": 0.1669, "num_input_tokens_seen": 121936, "step": 435 }, { "epoch": 4.888888888888889, "grad_norm": 0.35180190205574036, "learning_rate": 4.9985141471839706e-05, "loss": 0.1899, "num_input_tokens_seen": 123344, "step": 440 }, { "epoch": 4.944444444444445, "grad_norm": 0.6865087151527405, "learning_rate": 4.998480111581203e-05, "loss": 0.1743, "num_input_tokens_seen": 124736, "step": 445 }, { "epoch": 5.0, "grad_norm": 0.26684269309043884, "learning_rate": 4.998445690681405e-05, "loss": 0.1827, "num_input_tokens_seen": 126144, "step": 450 }, { "epoch": 5.055555555555555, "grad_norm": 1.191989541053772, "learning_rate": 4.9984108844898834e-05, "loss": 0.1639, "num_input_tokens_seen": 127520, "step": 455 }, { "epoch": 5.111111111111111, "grad_norm": 2.8081705570220947, "learning_rate": 4.9983756930120076e-05, "loss": 0.1739, "num_input_tokens_seen": 128944, "step": 460 }, { "epoch": 5.166666666666667, "grad_norm": 0.775079607963562, "learning_rate": 4.9983401162532025e-05, "loss": 0.2019, "num_input_tokens_seen": 130336, "step": 465 }, { "epoch": 5.222222222222222, "grad_norm": 2.300171136856079, "learning_rate": 4.998304154218955e-05, "loss": 0.1835, "num_input_tokens_seen": 131696, "step": 470 }, { "epoch": 5.277777777777778, "grad_norm": 0.27711567282676697, "learning_rate": 4.998267806914812e-05, "loss": 0.1618, "num_input_tokens_seen": 133088, "step": 475 }, { "epoch": 5.333333333333333, "grad_norm": 1.4059696197509766, "learning_rate": 4.998231074346378e-05, "loss": 0.1601, "num_input_tokens_seen": 134464, "step": 480 }, { "epoch": 5.388888888888889, "grad_norm": 0.9249340295791626, "learning_rate": 4.998193956519317e-05, "loss": 0.174, "num_input_tokens_seen": 135856, "step": 485 }, { "epoch": 5.444444444444445, "grad_norm": 2.2096493244171143, "learning_rate": 4.9981564534393545e-05, "loss": 0.1903, "num_input_tokens_seen": 137248, "step": 490 }, { "epoch": 5.5, "grad_norm": 1.7270090579986572, "learning_rate": 4.998118565112272e-05, "loss": 0.1223, "num_input_tokens_seen": 138592, "step": 495 }, { "epoch": 5.555555555555555, "grad_norm": 1.3556010723114014, "learning_rate": 4.998080291543914e-05, "loss": 0.1796, "num_input_tokens_seen": 140016, "step": 500 }, { "epoch": 5.611111111111111, "grad_norm": 1.665784478187561, "learning_rate": 4.9980416327401826e-05, "loss": 0.1288, "num_input_tokens_seen": 141424, "step": 505 }, { "epoch": 5.666666666666667, "grad_norm": 7.016280651092529, "learning_rate": 4.998002588707038e-05, "loss": 0.149, "num_input_tokens_seen": 142864, "step": 510 }, { "epoch": 5.722222222222222, "grad_norm": 2.036130666732788, "learning_rate": 4.997963159450503e-05, "loss": 0.178, "num_input_tokens_seen": 144320, "step": 515 }, { "epoch": 5.777777777777778, "grad_norm": 4.048322677612305, "learning_rate": 4.9979233449766575e-05, "loss": 0.2106, "num_input_tokens_seen": 145712, "step": 520 }, { "epoch": 5.833333333333333, "grad_norm": 2.3726274967193604, "learning_rate": 4.997883145291641e-05, "loss": 0.1233, "num_input_tokens_seen": 147088, "step": 525 }, { "epoch": 5.888888888888889, "grad_norm": 2.6477255821228027, "learning_rate": 4.9978425604016536e-05, "loss": 0.1746, "num_input_tokens_seen": 148464, "step": 530 }, { "epoch": 5.944444444444445, "grad_norm": 1.9229193925857544, "learning_rate": 4.9978015903129536e-05, "loss": 0.1718, "num_input_tokens_seen": 149840, "step": 535 }, { "epoch": 6.0, "grad_norm": 1.0382695198059082, "learning_rate": 4.997760235031859e-05, "loss": 0.1479, "num_input_tokens_seen": 151248, "step": 540 }, { "epoch": 6.055555555555555, "grad_norm": 1.678343653678894, "learning_rate": 4.9977184945647473e-05, "loss": 0.1653, "num_input_tokens_seen": 152624, "step": 545 }, { "epoch": 6.111111111111111, "grad_norm": 1.2994381189346313, "learning_rate": 4.997676368918055e-05, "loss": 0.11, "num_input_tokens_seen": 154032, "step": 550 }, { "epoch": 6.166666666666667, "grad_norm": 6.310273170471191, "learning_rate": 4.9976338580982794e-05, "loss": 0.0592, "num_input_tokens_seen": 155456, "step": 555 }, { "epoch": 6.222222222222222, "grad_norm": 0.33658450841903687, "learning_rate": 4.9975909621119755e-05, "loss": 0.1911, "num_input_tokens_seen": 156864, "step": 560 }, { "epoch": 6.277777777777778, "grad_norm": 24.20230484008789, "learning_rate": 4.997547680965758e-05, "loss": 0.2415, "num_input_tokens_seen": 158304, "step": 565 }, { "epoch": 6.333333333333333, "grad_norm": 0.08835373818874359, "learning_rate": 4.997504014666302e-05, "loss": 0.1619, "num_input_tokens_seen": 159712, "step": 570 }, { "epoch": 6.388888888888889, "grad_norm": 16.907752990722656, "learning_rate": 4.997459963220342e-05, "loss": 0.1208, "num_input_tokens_seen": 161088, "step": 575 }, { "epoch": 6.444444444444445, "grad_norm": 3.8359901905059814, "learning_rate": 4.997415526634671e-05, "loss": 0.1849, "num_input_tokens_seen": 162464, "step": 580 }, { "epoch": 6.5, "grad_norm": 12.966411590576172, "learning_rate": 4.99737070491614e-05, "loss": 0.4248, "num_input_tokens_seen": 163872, "step": 585 }, { "epoch": 6.555555555555555, "grad_norm": 0.8648723363876343, "learning_rate": 4.997325498071663e-05, "loss": 0.082, "num_input_tokens_seen": 165280, "step": 590 }, { "epoch": 6.611111111111111, "grad_norm": 4.166049957275391, "learning_rate": 4.997279906108211e-05, "loss": 0.1458, "num_input_tokens_seen": 166704, "step": 595 }, { "epoch": 6.666666666666667, "grad_norm": 1.024248480796814, "learning_rate": 4.9972339290328155e-05, "loss": 0.061, "num_input_tokens_seen": 168112, "step": 600 }, { "epoch": 6.666666666666667, "eval_loss": 0.23587098717689514, "eval_runtime": 1.388, "eval_samples_per_second": 28.818, "eval_steps_per_second": 14.409, "num_input_tokens_seen": 168112, "step": 600 }, { "epoch": 6.722222222222222, "grad_norm": 11.788779258728027, "learning_rate": 4.9971875668525646e-05, "loss": 0.168, "num_input_tokens_seen": 169504, "step": 605 }, { "epoch": 6.777777777777778, "grad_norm": 6.65277624130249, "learning_rate": 4.997140819574609e-05, "loss": 0.1616, "num_input_tokens_seen": 170880, "step": 610 }, { "epoch": 6.833333333333333, "grad_norm": 4.700632572174072, "learning_rate": 4.997093687206159e-05, "loss": 0.1375, "num_input_tokens_seen": 172256, "step": 615 }, { "epoch": 6.888888888888889, "grad_norm": 5.170658588409424, "learning_rate": 4.997046169754482e-05, "loss": 0.1259, "num_input_tokens_seen": 173664, "step": 620 }, { "epoch": 6.944444444444445, "grad_norm": 2.457627058029175, "learning_rate": 4.996998267226905e-05, "loss": 0.0552, "num_input_tokens_seen": 175024, "step": 625 }, { "epoch": 7.0, "grad_norm": 6.668975353240967, "learning_rate": 4.996949979630817e-05, "loss": 0.0653, "num_input_tokens_seen": 176400, "step": 630 }, { "epoch": 7.055555555555555, "grad_norm": 8.039339065551758, "learning_rate": 4.996901306973663e-05, "loss": 0.0481, "num_input_tokens_seen": 177792, "step": 635 }, { "epoch": 7.111111111111111, "grad_norm": 0.10910012573003769, "learning_rate": 4.996852249262949e-05, "loss": 0.0666, "num_input_tokens_seen": 179168, "step": 640 }, { "epoch": 7.166666666666667, "grad_norm": 0.26580506563186646, "learning_rate": 4.996802806506241e-05, "loss": 0.0685, "num_input_tokens_seen": 180544, "step": 645 }, { "epoch": 7.222222222222222, "grad_norm": 2.9731197357177734, "learning_rate": 4.996752978711164e-05, "loss": 0.1369, "num_input_tokens_seen": 181984, "step": 650 }, { "epoch": 7.277777777777778, "grad_norm": 0.15011097490787506, "learning_rate": 4.996702765885401e-05, "loss": 0.0302, "num_input_tokens_seen": 183392, "step": 655 }, { "epoch": 7.333333333333333, "grad_norm": 2.5298264026641846, "learning_rate": 4.9966521680366964e-05, "loss": 0.1871, "num_input_tokens_seen": 184784, "step": 660 }, { "epoch": 7.388888888888889, "grad_norm": 0.34174296259880066, "learning_rate": 4.9966011851728524e-05, "loss": 0.0344, "num_input_tokens_seen": 186192, "step": 665 }, { "epoch": 7.444444444444445, "grad_norm": 28.602052688598633, "learning_rate": 4.996549817301731e-05, "loss": 0.0447, "num_input_tokens_seen": 187568, "step": 670 }, { "epoch": 7.5, "grad_norm": 2.744209051132202, "learning_rate": 4.9964980644312544e-05, "loss": 0.1074, "num_input_tokens_seen": 188928, "step": 675 }, { "epoch": 7.555555555555555, "grad_norm": 6.414815902709961, "learning_rate": 4.996445926569403e-05, "loss": 0.0247, "num_input_tokens_seen": 190336, "step": 680 }, { "epoch": 7.611111111111111, "grad_norm": 9.614093780517578, "learning_rate": 4.996393403724218e-05, "loss": 0.0817, "num_input_tokens_seen": 191712, "step": 685 }, { "epoch": 7.666666666666667, "grad_norm": 0.36083126068115234, "learning_rate": 4.9963404959037985e-05, "loss": 0.0057, "num_input_tokens_seen": 193056, "step": 690 }, { "epoch": 7.722222222222222, "grad_norm": 0.11737316101789474, "learning_rate": 4.996287203116303e-05, "loss": 0.0048, "num_input_tokens_seen": 194496, "step": 695 }, { "epoch": 7.777777777777778, "grad_norm": 4.259082317352295, "learning_rate": 4.996233525369951e-05, "loss": 0.3374, "num_input_tokens_seen": 195936, "step": 700 }, { "epoch": 7.833333333333333, "grad_norm": 2.2666428089141846, "learning_rate": 4.99617946267302e-05, "loss": 0.1729, "num_input_tokens_seen": 197360, "step": 705 }, { "epoch": 7.888888888888889, "grad_norm": 17.92521095275879, "learning_rate": 4.996125015033846e-05, "loss": 0.082, "num_input_tokens_seen": 198784, "step": 710 }, { "epoch": 7.944444444444445, "grad_norm": 1.2762236595153809, "learning_rate": 4.996070182460827e-05, "loss": 0.0884, "num_input_tokens_seen": 200192, "step": 715 }, { "epoch": 8.0, "grad_norm": 0.5090296864509583, "learning_rate": 4.996014964962418e-05, "loss": 0.1006, "num_input_tokens_seen": 201616, "step": 720 }, { "epoch": 8.055555555555555, "grad_norm": 4.027255058288574, "learning_rate": 4.9959593625471344e-05, "loss": 0.0342, "num_input_tokens_seen": 203024, "step": 725 }, { "epoch": 8.11111111111111, "grad_norm": 4.048319339752197, "learning_rate": 4.995903375223552e-05, "loss": 0.0344, "num_input_tokens_seen": 204400, "step": 730 }, { "epoch": 8.166666666666666, "grad_norm": 0.24375838041305542, "learning_rate": 4.995847003000302e-05, "loss": 0.0084, "num_input_tokens_seen": 205824, "step": 735 }, { "epoch": 8.222222222222221, "grad_norm": 0.135248064994812, "learning_rate": 4.9957902458860804e-05, "loss": 0.0082, "num_input_tokens_seen": 207264, "step": 740 }, { "epoch": 8.277777777777779, "grad_norm": 0.10986128449440002, "learning_rate": 4.995733103889639e-05, "loss": 0.0031, "num_input_tokens_seen": 208656, "step": 745 }, { "epoch": 8.333333333333334, "grad_norm": 40.262962341308594, "learning_rate": 4.99567557701979e-05, "loss": 0.05, "num_input_tokens_seen": 210064, "step": 750 }, { "epoch": 8.38888888888889, "grad_norm": 0.04103904962539673, "learning_rate": 4.995617665285403e-05, "loss": 0.0343, "num_input_tokens_seen": 211440, "step": 755 }, { "epoch": 8.444444444444445, "grad_norm": 4.6744866371154785, "learning_rate": 4.99555936869541e-05, "loss": 0.1719, "num_input_tokens_seen": 212848, "step": 760 }, { "epoch": 8.5, "grad_norm": 0.01972949132323265, "learning_rate": 4.995500687258803e-05, "loss": 0.0005, "num_input_tokens_seen": 214272, "step": 765 }, { "epoch": 8.555555555555555, "grad_norm": 0.019499527290463448, "learning_rate": 4.995441620984628e-05, "loss": 0.0635, "num_input_tokens_seen": 215696, "step": 770 }, { "epoch": 8.61111111111111, "grad_norm": 0.8852298855781555, "learning_rate": 4.995382169881996e-05, "loss": 0.0737, "num_input_tokens_seen": 217088, "step": 775 }, { "epoch": 8.666666666666666, "grad_norm": 0.2136494666337967, "learning_rate": 4.9953223339600755e-05, "loss": 0.0806, "num_input_tokens_seen": 218496, "step": 780 }, { "epoch": 8.722222222222221, "grad_norm": 0.1153506338596344, "learning_rate": 4.995262113228091e-05, "loss": 0.1072, "num_input_tokens_seen": 219840, "step": 785 }, { "epoch": 8.777777777777779, "grad_norm": 0.14334902167320251, "learning_rate": 4.995201507695332e-05, "loss": 0.0022, "num_input_tokens_seen": 221264, "step": 790 }, { "epoch": 8.833333333333334, "grad_norm": 0.037030261009931564, "learning_rate": 4.995140517371144e-05, "loss": 0.0559, "num_input_tokens_seen": 222656, "step": 795 }, { "epoch": 8.88888888888889, "grad_norm": 0.06263444572687149, "learning_rate": 4.995079142264932e-05, "loss": 0.0765, "num_input_tokens_seen": 224048, "step": 800 }, { "epoch": 8.88888888888889, "eval_loss": 0.35959765315055847, "eval_runtime": 1.3861, "eval_samples_per_second": 28.857, "eval_steps_per_second": 14.429, "num_input_tokens_seen": 224048, "step": 800 }, { "epoch": 8.944444444444445, "grad_norm": 3.9700815677642822, "learning_rate": 4.995017382386162e-05, "loss": 0.0503, "num_input_tokens_seen": 225440, "step": 805 }, { "epoch": 9.0, "grad_norm": 0.02025281824171543, "learning_rate": 4.994955237744356e-05, "loss": 0.0785, "num_input_tokens_seen": 226816, "step": 810 }, { "epoch": 9.055555555555555, "grad_norm": 0.07512472569942474, "learning_rate": 4.994892708349101e-05, "loss": 0.0394, "num_input_tokens_seen": 228176, "step": 815 }, { "epoch": 9.11111111111111, "grad_norm": 0.13887622952461243, "learning_rate": 4.994829794210035e-05, "loss": 0.0114, "num_input_tokens_seen": 229584, "step": 820 }, { "epoch": 9.166666666666666, "grad_norm": 3.414530038833618, "learning_rate": 4.994766495336864e-05, "loss": 0.0411, "num_input_tokens_seen": 230944, "step": 825 }, { "epoch": 9.222222222222221, "grad_norm": 0.02765979804098606, "learning_rate": 4.994702811739348e-05, "loss": 0.0179, "num_input_tokens_seen": 232368, "step": 830 }, { "epoch": 9.277777777777779, "grad_norm": 0.07871554046869278, "learning_rate": 4.994638743427308e-05, "loss": 0.0393, "num_input_tokens_seen": 233792, "step": 835 }, { "epoch": 9.333333333333334, "grad_norm": 0.04480154812335968, "learning_rate": 4.994574290410624e-05, "loss": 0.001, "num_input_tokens_seen": 235216, "step": 840 }, { "epoch": 9.38888888888889, "grad_norm": 0.2051994353532791, "learning_rate": 4.9945094526992364e-05, "loss": 0.0015, "num_input_tokens_seen": 236576, "step": 845 }, { "epoch": 9.444444444444445, "grad_norm": 0.03577961400151253, "learning_rate": 4.994444230303142e-05, "loss": 0.0008, "num_input_tokens_seen": 237952, "step": 850 }, { "epoch": 9.5, "grad_norm": 0.043894439935684204, "learning_rate": 4.994378623232402e-05, "loss": 0.0005, "num_input_tokens_seen": 239360, "step": 855 }, { "epoch": 9.555555555555555, "grad_norm": 0.05347692221403122, "learning_rate": 4.99431263149713e-05, "loss": 0.0761, "num_input_tokens_seen": 240784, "step": 860 }, { "epoch": 9.61111111111111, "grad_norm": 0.05136746168136597, "learning_rate": 4.9942462551075056e-05, "loss": 0.0005, "num_input_tokens_seen": 242208, "step": 865 }, { "epoch": 9.666666666666666, "grad_norm": 0.10801062732934952, "learning_rate": 4.994179494073764e-05, "loss": 0.0072, "num_input_tokens_seen": 243584, "step": 870 }, { "epoch": 9.722222222222221, "grad_norm": 0.02358662523329258, "learning_rate": 4.9941123484062e-05, "loss": 0.0519, "num_input_tokens_seen": 244960, "step": 875 }, { "epoch": 9.777777777777779, "grad_norm": 0.16574795544147491, "learning_rate": 4.99404481811517e-05, "loss": 0.0006, "num_input_tokens_seen": 246384, "step": 880 }, { "epoch": 9.833333333333334, "grad_norm": 0.19994547963142395, "learning_rate": 4.9939769032110864e-05, "loss": 0.0007, "num_input_tokens_seen": 247840, "step": 885 }, { "epoch": 9.88888888888889, "grad_norm": 17.124847412109375, "learning_rate": 4.993908603704423e-05, "loss": 0.1091, "num_input_tokens_seen": 249216, "step": 890 }, { "epoch": 9.944444444444445, "grad_norm": 0.04390018805861473, "learning_rate": 4.9938399196057126e-05, "loss": 0.0007, "num_input_tokens_seen": 250592, "step": 895 }, { "epoch": 10.0, "grad_norm": 0.03742208704352379, "learning_rate": 4.993770850925547e-05, "loss": 0.047, "num_input_tokens_seen": 251968, "step": 900 }, { "epoch": 10.055555555555555, "grad_norm": 0.030842816457152367, "learning_rate": 4.993701397674577e-05, "loss": 0.0007, "num_input_tokens_seen": 253392, "step": 905 }, { "epoch": 10.11111111111111, "grad_norm": 0.019408447667956352, "learning_rate": 4.993631559863515e-05, "loss": 0.0007, "num_input_tokens_seen": 254752, "step": 910 }, { "epoch": 10.166666666666666, "grad_norm": 0.016620345413684845, "learning_rate": 4.9935613375031283e-05, "loss": 0.0003, "num_input_tokens_seen": 256176, "step": 915 }, { "epoch": 10.222222222222221, "grad_norm": 0.12508945167064667, "learning_rate": 4.993490730604248e-05, "loss": 0.0007, "num_input_tokens_seen": 257584, "step": 920 }, { "epoch": 10.277777777777779, "grad_norm": 0.020689988508820534, "learning_rate": 4.993419739177761e-05, "loss": 0.0003, "num_input_tokens_seen": 259008, "step": 925 }, { "epoch": 10.333333333333334, "grad_norm": 0.011085642501711845, "learning_rate": 4.9933483632346164e-05, "loss": 0.0003, "num_input_tokens_seen": 260416, "step": 930 }, { "epoch": 10.38888888888889, "grad_norm": 0.04516610503196716, "learning_rate": 4.993276602785821e-05, "loss": 0.0325, "num_input_tokens_seen": 261872, "step": 935 }, { "epoch": 10.444444444444445, "grad_norm": 0.002245417097583413, "learning_rate": 4.993204457842441e-05, "loss": 0.0004, "num_input_tokens_seen": 263232, "step": 940 }, { "epoch": 10.5, "grad_norm": 0.01526965107768774, "learning_rate": 4.993131928415602e-05, "loss": 0.0003, "num_input_tokens_seen": 264656, "step": 945 }, { "epoch": 10.555555555555555, "grad_norm": 0.03349163755774498, "learning_rate": 4.993059014516489e-05, "loss": 0.0002, "num_input_tokens_seen": 266016, "step": 950 }, { "epoch": 10.61111111111111, "grad_norm": 0.0074921706691384315, "learning_rate": 4.9929857161563464e-05, "loss": 0.0001, "num_input_tokens_seen": 267408, "step": 955 }, { "epoch": 10.666666666666666, "grad_norm": 0.0030619909521192312, "learning_rate": 4.992912033346477e-05, "loss": 0.0002, "num_input_tokens_seen": 268800, "step": 960 }, { "epoch": 10.722222222222221, "grad_norm": 9.71147632598877, "learning_rate": 4.992837966098245e-05, "loss": 0.0443, "num_input_tokens_seen": 270208, "step": 965 }, { "epoch": 10.777777777777779, "grad_norm": 0.024364158511161804, "learning_rate": 4.992763514423071e-05, "loss": 0.0001, "num_input_tokens_seen": 271552, "step": 970 }, { "epoch": 10.833333333333334, "grad_norm": 11.987972259521484, "learning_rate": 4.992688678332437e-05, "loss": 0.069, "num_input_tokens_seen": 272928, "step": 975 }, { "epoch": 10.88888888888889, "grad_norm": 0.003983642905950546, "learning_rate": 4.992613457837884e-05, "loss": 0.0551, "num_input_tokens_seen": 274352, "step": 980 }, { "epoch": 10.944444444444445, "grad_norm": 0.015770645812153816, "learning_rate": 4.992537852951011e-05, "loss": 0.0001, "num_input_tokens_seen": 275760, "step": 985 }, { "epoch": 11.0, "grad_norm": 0.12733624875545502, "learning_rate": 4.9924618636834785e-05, "loss": 0.0005, "num_input_tokens_seen": 277152, "step": 990 }, { "epoch": 11.055555555555555, "grad_norm": 0.002660186728462577, "learning_rate": 4.9923854900470046e-05, "loss": 0.0002, "num_input_tokens_seen": 278512, "step": 995 }, { "epoch": 11.11111111111111, "grad_norm": 0.010107087902724743, "learning_rate": 4.992308732053367e-05, "loss": 0.0003, "num_input_tokens_seen": 279904, "step": 1000 }, { "epoch": 11.11111111111111, "eval_loss": 0.4669176936149597, "eval_runtime": 1.3948, "eval_samples_per_second": 28.679, "eval_steps_per_second": 14.339, "num_input_tokens_seen": 279904, "step": 1000 }, { "epoch": 11.166666666666666, "grad_norm": 0.012315575033426285, "learning_rate": 4.992231589714402e-05, "loss": 0.0001, "num_input_tokens_seen": 281360, "step": 1005 }, { "epoch": 11.222222222222221, "grad_norm": 24.5928955078125, "learning_rate": 4.992154063042007e-05, "loss": 0.0066, "num_input_tokens_seen": 282752, "step": 1010 }, { "epoch": 11.277777777777779, "grad_norm": 0.010293233208358288, "learning_rate": 4.992076152048136e-05, "loss": 0.0002, "num_input_tokens_seen": 284160, "step": 1015 }, { "epoch": 11.333333333333334, "grad_norm": 0.0015340570826083422, "learning_rate": 4.991997856744807e-05, "loss": 0.0001, "num_input_tokens_seen": 285536, "step": 1020 }, { "epoch": 11.38888888888889, "grad_norm": 0.0015279524959623814, "learning_rate": 4.9919191771440905e-05, "loss": 0.0, "num_input_tokens_seen": 286896, "step": 1025 }, { "epoch": 11.444444444444445, "grad_norm": 0.037485234439373016, "learning_rate": 4.991840113258122e-05, "loss": 0.0001, "num_input_tokens_seen": 288256, "step": 1030 }, { "epoch": 11.5, "grad_norm": 0.0017047245055437088, "learning_rate": 4.9917606650990933e-05, "loss": 0.0001, "num_input_tokens_seen": 289680, "step": 1035 }, { "epoch": 11.555555555555555, "grad_norm": 0.0025834767147898674, "learning_rate": 4.9916808326792566e-05, "loss": 0.0, "num_input_tokens_seen": 291040, "step": 1040 }, { "epoch": 11.61111111111111, "grad_norm": 0.022313855588436127, "learning_rate": 4.9916006160109235e-05, "loss": 0.0001, "num_input_tokens_seen": 292400, "step": 1045 }, { "epoch": 11.666666666666666, "grad_norm": 0.008967495523393154, "learning_rate": 4.991520015106464e-05, "loss": 0.0001, "num_input_tokens_seen": 293824, "step": 1050 }, { "epoch": 11.722222222222221, "grad_norm": 0.003285276470705867, "learning_rate": 4.991439029978308e-05, "loss": 0.0003, "num_input_tokens_seen": 295232, "step": 1055 }, { "epoch": 11.777777777777779, "grad_norm": 0.001463075983338058, "learning_rate": 4.9913576606389434e-05, "loss": 0.0001, "num_input_tokens_seen": 296640, "step": 1060 }, { "epoch": 11.833333333333334, "grad_norm": 0.0031799189746379852, "learning_rate": 4.991275907100919e-05, "loss": 0.0001, "num_input_tokens_seen": 298048, "step": 1065 }, { "epoch": 11.88888888888889, "grad_norm": 0.0010242857970297337, "learning_rate": 4.9911937693768434e-05, "loss": 0.0001, "num_input_tokens_seen": 299456, "step": 1070 }, { "epoch": 11.944444444444445, "grad_norm": 0.001280217315070331, "learning_rate": 4.991111247479382e-05, "loss": 0.0006, "num_input_tokens_seen": 300848, "step": 1075 }, { "epoch": 12.0, "grad_norm": 0.00015693908790126443, "learning_rate": 4.9910283414212605e-05, "loss": 0.0, "num_input_tokens_seen": 302224, "step": 1080 }, { "epoch": 12.055555555555555, "grad_norm": 0.0009421953582204878, "learning_rate": 4.990945051215265e-05, "loss": 0.0002, "num_input_tokens_seen": 303648, "step": 1085 }, { "epoch": 12.11111111111111, "grad_norm": 0.005614342167973518, "learning_rate": 4.99086137687424e-05, "loss": 0.0006, "num_input_tokens_seen": 305024, "step": 1090 }, { "epoch": 12.166666666666666, "grad_norm": 0.0014418581267818809, "learning_rate": 4.9907773184110874e-05, "loss": 0.0002, "num_input_tokens_seen": 306448, "step": 1095 }, { "epoch": 12.222222222222221, "grad_norm": 0.0023614426609128714, "learning_rate": 4.9906928758387715e-05, "loss": 0.0, "num_input_tokens_seen": 307856, "step": 1100 }, { "epoch": 12.277777777777779, "grad_norm": 0.0021121089812368155, "learning_rate": 4.9906080491703146e-05, "loss": 0.0, "num_input_tokens_seen": 309280, "step": 1105 }, { "epoch": 12.333333333333334, "grad_norm": 0.0012867911718785763, "learning_rate": 4.990522838418797e-05, "loss": 0.0, "num_input_tokens_seen": 310704, "step": 1110 }, { "epoch": 12.38888888888889, "grad_norm": 0.0014517122181132436, "learning_rate": 4.9904372435973604e-05, "loss": 0.0, "num_input_tokens_seen": 312080, "step": 1115 }, { "epoch": 12.444444444444445, "grad_norm": 0.004178262781351805, "learning_rate": 4.990351264719203e-05, "loss": 0.0, "num_input_tokens_seen": 313520, "step": 1120 }, { "epoch": 12.5, "grad_norm": 0.002247866475954652, "learning_rate": 4.990264901797586e-05, "loss": 0.028, "num_input_tokens_seen": 314912, "step": 1125 }, { "epoch": 12.555555555555555, "grad_norm": 0.0010443027131259441, "learning_rate": 4.990178154845826e-05, "loss": 0.0, "num_input_tokens_seen": 316336, "step": 1130 }, { "epoch": 12.61111111111111, "grad_norm": 0.007503070868551731, "learning_rate": 4.9900910238773014e-05, "loss": 0.0, "num_input_tokens_seen": 317744, "step": 1135 }, { "epoch": 12.666666666666666, "grad_norm": 0.0016888609388843179, "learning_rate": 4.990003508905448e-05, "loss": 0.0, "num_input_tokens_seen": 319136, "step": 1140 }, { "epoch": 12.722222222222221, "grad_norm": 0.0007850857218727469, "learning_rate": 4.989915609943763e-05, "loss": 0.0, "num_input_tokens_seen": 320528, "step": 1145 }, { "epoch": 12.777777777777779, "grad_norm": 0.00016866129590198398, "learning_rate": 4.9898273270058e-05, "loss": 0.0, "num_input_tokens_seen": 321936, "step": 1150 }, { "epoch": 12.833333333333334, "grad_norm": 0.004441921133548021, "learning_rate": 4.989738660105174e-05, "loss": 0.0, "num_input_tokens_seen": 323344, "step": 1155 }, { "epoch": 12.88888888888889, "grad_norm": 0.00401266198605299, "learning_rate": 4.989649609255559e-05, "loss": 0.0, "num_input_tokens_seen": 324784, "step": 1160 }, { "epoch": 12.944444444444445, "grad_norm": 0.0008980792481452227, "learning_rate": 4.989560174470687e-05, "loss": 0.0, "num_input_tokens_seen": 326208, "step": 1165 }, { "epoch": 13.0, "grad_norm": 0.0013239652616903186, "learning_rate": 4.989470355764351e-05, "loss": 0.0, "num_input_tokens_seen": 327600, "step": 1170 }, { "epoch": 13.055555555555555, "grad_norm": 0.0003628781414590776, "learning_rate": 4.9893801531504e-05, "loss": 0.0, "num_input_tokens_seen": 329024, "step": 1175 }, { "epoch": 13.11111111111111, "grad_norm": 0.0010009472025558352, "learning_rate": 4.9892895666427475e-05, "loss": 0.0, "num_input_tokens_seen": 330400, "step": 1180 }, { "epoch": 13.166666666666666, "grad_norm": 0.03768958896398544, "learning_rate": 4.9891985962553606e-05, "loss": 0.0001, "num_input_tokens_seen": 331808, "step": 1185 }, { "epoch": 13.222222222222221, "grad_norm": 0.009641497395932674, "learning_rate": 4.989107242002269e-05, "loss": 0.0, "num_input_tokens_seen": 333200, "step": 1190 }, { "epoch": 13.277777777777779, "grad_norm": 0.0006912300013937056, "learning_rate": 4.989015503897561e-05, "loss": 0.0, "num_input_tokens_seen": 334640, "step": 1195 }, { "epoch": 13.333333333333334, "grad_norm": 0.00037567151593975723, "learning_rate": 4.988923381955383e-05, "loss": 0.0, "num_input_tokens_seen": 336032, "step": 1200 }, { "epoch": 13.333333333333334, "eval_loss": 0.69465172290802, "eval_runtime": 1.3847, "eval_samples_per_second": 28.888, "eval_steps_per_second": 14.444, "num_input_tokens_seen": 336032, "step": 1200 }, { "epoch": 13.38888888888889, "grad_norm": 0.056749068200588226, "learning_rate": 4.988830876189942e-05, "loss": 0.0001, "num_input_tokens_seen": 337440, "step": 1205 }, { "epoch": 13.444444444444445, "grad_norm": 0.0033829989843070507, "learning_rate": 4.988737986615503e-05, "loss": 0.0, "num_input_tokens_seen": 338832, "step": 1210 }, { "epoch": 13.5, "grad_norm": 0.0011844121618196368, "learning_rate": 4.988644713246391e-05, "loss": 0.0, "num_input_tokens_seen": 340224, "step": 1215 }, { "epoch": 13.555555555555555, "grad_norm": 0.008158550597727299, "learning_rate": 4.988551056096991e-05, "loss": 0.0, "num_input_tokens_seen": 341616, "step": 1220 }, { "epoch": 13.61111111111111, "grad_norm": 0.0007474473677575588, "learning_rate": 4.988457015181743e-05, "loss": 0.0, "num_input_tokens_seen": 342992, "step": 1225 }, { "epoch": 13.666666666666666, "grad_norm": 0.003371564671397209, "learning_rate": 4.988362590515153e-05, "loss": 0.0, "num_input_tokens_seen": 344384, "step": 1230 }, { "epoch": 13.722222222222221, "grad_norm": 0.009013683535158634, "learning_rate": 4.9882677821117805e-05, "loss": 0.0, "num_input_tokens_seen": 345824, "step": 1235 }, { "epoch": 13.777777777777779, "grad_norm": 0.0007282248116098344, "learning_rate": 4.988172589986246e-05, "loss": 0.0, "num_input_tokens_seen": 347248, "step": 1240 }, { "epoch": 13.833333333333334, "grad_norm": 0.0018894876120612025, "learning_rate": 4.9880770141532304e-05, "loss": 0.0, "num_input_tokens_seen": 348672, "step": 1245 }, { "epoch": 13.88888888888889, "grad_norm": 0.00047606683801859617, "learning_rate": 4.987981054627472e-05, "loss": 0.0, "num_input_tokens_seen": 350032, "step": 1250 }, { "epoch": 13.944444444444445, "grad_norm": 0.00033190863905474544, "learning_rate": 4.987884711423769e-05, "loss": 0.0, "num_input_tokens_seen": 351392, "step": 1255 }, { "epoch": 14.0, "grad_norm": 0.0013995933113619685, "learning_rate": 4.9877879845569784e-05, "loss": 0.0, "num_input_tokens_seen": 352784, "step": 1260 }, { "epoch": 14.055555555555555, "grad_norm": 0.003328617662191391, "learning_rate": 4.9876908740420175e-05, "loss": 0.0, "num_input_tokens_seen": 354176, "step": 1265 }, { "epoch": 14.11111111111111, "grad_norm": 0.0004669480840675533, "learning_rate": 4.987593379893861e-05, "loss": 0.0, "num_input_tokens_seen": 355568, "step": 1270 }, { "epoch": 14.166666666666666, "grad_norm": 0.0002796306798700243, "learning_rate": 4.987495502127545e-05, "loss": 0.0, "num_input_tokens_seen": 356960, "step": 1275 }, { "epoch": 14.222222222222221, "grad_norm": 0.0004675764066632837, "learning_rate": 4.987397240758162e-05, "loss": 0.0, "num_input_tokens_seen": 358368, "step": 1280 }, { "epoch": 14.277777777777779, "grad_norm": 0.0017654354451224208, "learning_rate": 4.9872985958008664e-05, "loss": 0.0, "num_input_tokens_seen": 359760, "step": 1285 }, { "epoch": 14.333333333333334, "grad_norm": 0.0001061224756995216, "learning_rate": 4.987199567270871e-05, "loss": 0.0, "num_input_tokens_seen": 361200, "step": 1290 }, { "epoch": 14.38888888888889, "grad_norm": 0.002756010740995407, "learning_rate": 4.9871001551834444e-05, "loss": 0.0, "num_input_tokens_seen": 362656, "step": 1295 }, { "epoch": 14.444444444444445, "grad_norm": 0.002624095417559147, "learning_rate": 4.98700035955392e-05, "loss": 0.0, "num_input_tokens_seen": 364032, "step": 1300 }, { "epoch": 14.5, "grad_norm": 0.00017280258180107921, "learning_rate": 4.986900180397686e-05, "loss": 0.0, "num_input_tokens_seen": 365408, "step": 1305 }, { "epoch": 14.555555555555555, "grad_norm": 0.0013871699338778853, "learning_rate": 4.9867996177301926e-05, "loss": 0.0, "num_input_tokens_seen": 366784, "step": 1310 }, { "epoch": 14.61111111111111, "grad_norm": 0.00021458178525790572, "learning_rate": 4.9866986715669464e-05, "loss": 0.0, "num_input_tokens_seen": 368176, "step": 1315 }, { "epoch": 14.666666666666666, "grad_norm": 0.0002293186407769099, "learning_rate": 4.9865973419235155e-05, "loss": 0.0, "num_input_tokens_seen": 369552, "step": 1320 }, { "epoch": 14.722222222222221, "grad_norm": 0.0004672967188525945, "learning_rate": 4.986495628815526e-05, "loss": 0.0, "num_input_tokens_seen": 370944, "step": 1325 }, { "epoch": 14.777777777777779, "grad_norm": 0.002825044095516205, "learning_rate": 4.986393532258663e-05, "loss": 0.0, "num_input_tokens_seen": 372368, "step": 1330 }, { "epoch": 14.833333333333334, "grad_norm": 0.00021190402912907302, "learning_rate": 4.986291052268671e-05, "loss": 0.0, "num_input_tokens_seen": 373760, "step": 1335 }, { "epoch": 14.88888888888889, "grad_norm": 0.0002641167666297406, "learning_rate": 4.986188188861355e-05, "loss": 0.0, "num_input_tokens_seen": 375120, "step": 1340 }, { "epoch": 14.944444444444445, "grad_norm": 0.0003483938053250313, "learning_rate": 4.9860849420525766e-05, "loss": 0.0, "num_input_tokens_seen": 376528, "step": 1345 }, { "epoch": 15.0, "grad_norm": 0.00023569990298710763, "learning_rate": 4.9859813118582575e-05, "loss": 0.0, "num_input_tokens_seen": 377952, "step": 1350 }, { "epoch": 15.055555555555555, "grad_norm": 0.00035506789572536945, "learning_rate": 4.98587729829438e-05, "loss": 0.0, "num_input_tokens_seen": 379328, "step": 1355 }, { "epoch": 15.11111111111111, "grad_norm": 0.004679740406572819, "learning_rate": 4.985772901376983e-05, "loss": 0.0, "num_input_tokens_seen": 380720, "step": 1360 }, { "epoch": 15.166666666666666, "grad_norm": 0.00022841374448034912, "learning_rate": 4.9856681211221666e-05, "loss": 0.0, "num_input_tokens_seen": 382096, "step": 1365 }, { "epoch": 15.222222222222221, "grad_norm": 0.002719972748309374, "learning_rate": 4.985562957546089e-05, "loss": 0.0, "num_input_tokens_seen": 383488, "step": 1370 }, { "epoch": 15.277777777777779, "grad_norm": 0.0012940933229401708, "learning_rate": 4.9854574106649686e-05, "loss": 0.0, "num_input_tokens_seen": 384848, "step": 1375 }, { "epoch": 15.333333333333334, "grad_norm": 0.000778093293774873, "learning_rate": 4.985351480495081e-05, "loss": 0.0, "num_input_tokens_seen": 386256, "step": 1380 }, { "epoch": 15.38888888888889, "grad_norm": 0.0006262014503590763, "learning_rate": 4.985245167052762e-05, "loss": 0.0, "num_input_tokens_seen": 387696, "step": 1385 }, { "epoch": 15.444444444444445, "grad_norm": 0.003923195879906416, "learning_rate": 4.9851384703544066e-05, "loss": 0.0, "num_input_tokens_seen": 389072, "step": 1390 }, { "epoch": 15.5, "grad_norm": 0.0003193231823388487, "learning_rate": 4.985031390416469e-05, "loss": 0.0, "num_input_tokens_seen": 390464, "step": 1395 }, { "epoch": 15.555555555555555, "grad_norm": 0.00017896643839776516, "learning_rate": 4.984923927255461e-05, "loss": 0.0, "num_input_tokens_seen": 391904, "step": 1400 }, { "epoch": 15.555555555555555, "eval_loss": 0.7522537112236023, "eval_runtime": 1.3878, "eval_samples_per_second": 28.822, "eval_steps_per_second": 14.411, "num_input_tokens_seen": 391904, "step": 1400 }, { "epoch": 15.61111111111111, "grad_norm": 0.002309726318344474, "learning_rate": 4.984816080887958e-05, "loss": 0.0, "num_input_tokens_seen": 393344, "step": 1405 }, { "epoch": 15.666666666666666, "grad_norm": 0.001190578448586166, "learning_rate": 4.9847078513305875e-05, "loss": 0.0, "num_input_tokens_seen": 394784, "step": 1410 }, { "epoch": 15.722222222222221, "grad_norm": 0.0001942941889865324, "learning_rate": 4.984599238600043e-05, "loss": 0.0, "num_input_tokens_seen": 396176, "step": 1415 }, { "epoch": 15.777777777777779, "grad_norm": 0.00010351234959671274, "learning_rate": 4.9844902427130716e-05, "loss": 0.0, "num_input_tokens_seen": 397552, "step": 1420 }, { "epoch": 15.833333333333334, "grad_norm": 8.065612200880423e-05, "learning_rate": 4.984380863686482e-05, "loss": 0.0, "num_input_tokens_seen": 398960, "step": 1425 }, { "epoch": 15.88888888888889, "grad_norm": 0.0029633785597980022, "learning_rate": 4.984271101537143e-05, "loss": 0.0, "num_input_tokens_seen": 400384, "step": 1430 }, { "epoch": 15.944444444444445, "grad_norm": 0.00027499679708853364, "learning_rate": 4.9841609562819816e-05, "loss": 0.0, "num_input_tokens_seen": 401776, "step": 1435 }, { "epoch": 16.0, "grad_norm": 0.0002889605239033699, "learning_rate": 4.984050427937983e-05, "loss": 0.0, "num_input_tokens_seen": 403168, "step": 1440 }, { "epoch": 16.055555555555557, "grad_norm": 0.0004454855225048959, "learning_rate": 4.983939516522191e-05, "loss": 0.0, "num_input_tokens_seen": 404592, "step": 1445 }, { "epoch": 16.11111111111111, "grad_norm": 0.00023809142294339836, "learning_rate": 4.983828222051711e-05, "loss": 0.0, "num_input_tokens_seen": 405984, "step": 1450 }, { "epoch": 16.166666666666668, "grad_norm": 0.004155315924435854, "learning_rate": 4.983716544543705e-05, "loss": 0.0, "num_input_tokens_seen": 407392, "step": 1455 }, { "epoch": 16.22222222222222, "grad_norm": 9.548653179081157e-05, "learning_rate": 4.983604484015395e-05, "loss": 0.0, "num_input_tokens_seen": 408768, "step": 1460 }, { "epoch": 16.27777777777778, "grad_norm": 0.0005125017487443984, "learning_rate": 4.983492040484064e-05, "loss": 0.0, "num_input_tokens_seen": 410144, "step": 1465 }, { "epoch": 16.333333333333332, "grad_norm": 0.0014373287558555603, "learning_rate": 4.98337921396705e-05, "loss": 0.0, "num_input_tokens_seen": 411568, "step": 1470 }, { "epoch": 16.38888888888889, "grad_norm": 0.0024377782829105854, "learning_rate": 4.983266004481753e-05, "loss": 0.0, "num_input_tokens_seen": 412976, "step": 1475 }, { "epoch": 16.444444444444443, "grad_norm": 0.0004369731468614191, "learning_rate": 4.9831524120456316e-05, "loss": 0.0, "num_input_tokens_seen": 414368, "step": 1480 }, { "epoch": 16.5, "grad_norm": 0.00019569652795325965, "learning_rate": 4.9830384366762026e-05, "loss": 0.0, "num_input_tokens_seen": 415760, "step": 1485 }, { "epoch": 16.555555555555557, "grad_norm": 0.00020500282698776573, "learning_rate": 4.9829240783910436e-05, "loss": 0.0, "num_input_tokens_seen": 417184, "step": 1490 }, { "epoch": 16.61111111111111, "grad_norm": 0.00014867508434690535, "learning_rate": 4.982809337207789e-05, "loss": 0.0, "num_input_tokens_seen": 418544, "step": 1495 }, { "epoch": 16.666666666666668, "grad_norm": 0.00033428138704039156, "learning_rate": 4.9826942131441337e-05, "loss": 0.0, "num_input_tokens_seen": 419952, "step": 1500 }, { "epoch": 16.72222222222222, "grad_norm": 0.00037858367431908846, "learning_rate": 4.9825787062178315e-05, "loss": 0.0, "num_input_tokens_seen": 421392, "step": 1505 }, { "epoch": 16.77777777777778, "grad_norm": 0.0010514426976442337, "learning_rate": 4.9824628164466945e-05, "loss": 0.0, "num_input_tokens_seen": 422784, "step": 1510 }, { "epoch": 16.833333333333332, "grad_norm": 0.00018736151105258614, "learning_rate": 4.982346543848595e-05, "loss": 0.0, "num_input_tokens_seen": 424192, "step": 1515 }, { "epoch": 16.88888888888889, "grad_norm": 0.002021395368501544, "learning_rate": 4.9822298884414626e-05, "loss": 0.0, "num_input_tokens_seen": 425568, "step": 1520 }, { "epoch": 16.944444444444443, "grad_norm": 0.00012745274580083787, "learning_rate": 4.982112850243288e-05, "loss": 0.0, "num_input_tokens_seen": 426960, "step": 1525 }, { "epoch": 17.0, "grad_norm": 0.0022161155939102173, "learning_rate": 4.98199542927212e-05, "loss": 0.0, "num_input_tokens_seen": 428384, "step": 1530 }, { "epoch": 17.055555555555557, "grad_norm": 0.00037282888661138713, "learning_rate": 4.981877625546066e-05, "loss": 0.0, "num_input_tokens_seen": 429776, "step": 1535 }, { "epoch": 17.11111111111111, "grad_norm": 0.0007710931822657585, "learning_rate": 4.981759439083293e-05, "loss": 0.0, "num_input_tokens_seen": 431248, "step": 1540 }, { "epoch": 17.166666666666668, "grad_norm": 0.002957237185910344, "learning_rate": 4.981640869902027e-05, "loss": 0.0, "num_input_tokens_seen": 432656, "step": 1545 }, { "epoch": 17.22222222222222, "grad_norm": 0.00019065476953983307, "learning_rate": 4.9815219180205517e-05, "loss": 0.0, "num_input_tokens_seen": 434048, "step": 1550 }, { "epoch": 17.27777777777778, "grad_norm": 0.00011246955546084791, "learning_rate": 4.9814025834572126e-05, "loss": 0.0, "num_input_tokens_seen": 435488, "step": 1555 }, { "epoch": 17.333333333333332, "grad_norm": 0.0007554307812824845, "learning_rate": 4.981282866230411e-05, "loss": 0.0, "num_input_tokens_seen": 436880, "step": 1560 }, { "epoch": 17.38888888888889, "grad_norm": 0.00019298451661597937, "learning_rate": 4.981162766358611e-05, "loss": 0.0, "num_input_tokens_seen": 438256, "step": 1565 }, { "epoch": 17.444444444444443, "grad_norm": 8.104767766781151e-05, "learning_rate": 4.9810422838603316e-05, "loss": 0.0, "num_input_tokens_seen": 439664, "step": 1570 }, { "epoch": 17.5, "grad_norm": 8.615315891802311e-05, "learning_rate": 4.9809214187541533e-05, "loss": 0.0, "num_input_tokens_seen": 441088, "step": 1575 }, { "epoch": 17.555555555555557, "grad_norm": 0.0006851017824374139, "learning_rate": 4.980800171058715e-05, "loss": 0.0, "num_input_tokens_seen": 442496, "step": 1580 }, { "epoch": 17.61111111111111, "grad_norm": 0.00023122392303775996, "learning_rate": 4.980678540792715e-05, "loss": 0.0, "num_input_tokens_seen": 443872, "step": 1585 }, { "epoch": 17.666666666666668, "grad_norm": 9.19198791962117e-05, "learning_rate": 4.980556527974909e-05, "loss": 0.0, "num_input_tokens_seen": 445312, "step": 1590 }, { "epoch": 17.72222222222222, "grad_norm": 0.0004957696073688567, "learning_rate": 4.980434132624114e-05, "loss": 0.0, "num_input_tokens_seen": 446736, "step": 1595 }, { "epoch": 17.77777777777778, "grad_norm": 0.0012996296864002943, "learning_rate": 4.980311354759205e-05, "loss": 0.0, "num_input_tokens_seen": 448112, "step": 1600 }, { "epoch": 17.77777777777778, "eval_loss": 0.7925209999084473, "eval_runtime": 1.3912, "eval_samples_per_second": 28.753, "eval_steps_per_second": 14.377, "num_input_tokens_seen": 448112, "step": 1600 }, { "epoch": 17.833333333333332, "grad_norm": 0.0001805934007279575, "learning_rate": 4.980188194399116e-05, "loss": 0.0, "num_input_tokens_seen": 449504, "step": 1605 }, { "epoch": 17.88888888888889, "grad_norm": 5.6420751207042485e-05, "learning_rate": 4.9800646515628384e-05, "loss": 0.0, "num_input_tokens_seen": 450896, "step": 1610 }, { "epoch": 17.944444444444443, "grad_norm": 7.06026767147705e-05, "learning_rate": 4.979940726269426e-05, "loss": 0.0, "num_input_tokens_seen": 452256, "step": 1615 }, { "epoch": 18.0, "grad_norm": 0.0005946660530753434, "learning_rate": 4.979816418537988e-05, "loss": 0.0, "num_input_tokens_seen": 453648, "step": 1620 }, { "epoch": 18.055555555555557, "grad_norm": 8.645749039715156e-05, "learning_rate": 4.979691728387696e-05, "loss": 0.0, "num_input_tokens_seen": 455024, "step": 1625 }, { "epoch": 18.11111111111111, "grad_norm": 5.6363485782640055e-05, "learning_rate": 4.979566655837776e-05, "loss": 0.0, "num_input_tokens_seen": 456384, "step": 1630 }, { "epoch": 18.166666666666668, "grad_norm": 0.0008357074111700058, "learning_rate": 4.9794412009075184e-05, "loss": 0.0, "num_input_tokens_seen": 457824, "step": 1635 }, { "epoch": 18.22222222222222, "grad_norm": 0.0003079876769334078, "learning_rate": 4.979315363616269e-05, "loss": 0.0, "num_input_tokens_seen": 459136, "step": 1640 }, { "epoch": 18.27777777777778, "grad_norm": 0.00011375745089026168, "learning_rate": 4.979189143983434e-05, "loss": 0.0, "num_input_tokens_seen": 460544, "step": 1645 }, { "epoch": 18.333333333333332, "grad_norm": 0.00033707849797792733, "learning_rate": 4.979062542028478e-05, "loss": 0.0, "num_input_tokens_seen": 461968, "step": 1650 }, { "epoch": 18.38888888888889, "grad_norm": 0.001867177663370967, "learning_rate": 4.978935557770923e-05, "loss": 0.0, "num_input_tokens_seen": 463376, "step": 1655 }, { "epoch": 18.444444444444443, "grad_norm": 0.001031277934089303, "learning_rate": 4.978808191230353e-05, "loss": 0.0, "num_input_tokens_seen": 464768, "step": 1660 }, { "epoch": 18.5, "grad_norm": 0.00015013136726338416, "learning_rate": 4.9786804424264085e-05, "loss": 0.0, "num_input_tokens_seen": 466144, "step": 1665 }, { "epoch": 18.555555555555557, "grad_norm": 0.00016382666944991797, "learning_rate": 4.978552311378792e-05, "loss": 0.0, "num_input_tokens_seen": 467504, "step": 1670 }, { "epoch": 18.61111111111111, "grad_norm": 0.0010015044827014208, "learning_rate": 4.978423798107261e-05, "loss": 0.0, "num_input_tokens_seen": 468912, "step": 1675 }, { "epoch": 18.666666666666668, "grad_norm": 0.00010270297207171097, "learning_rate": 4.978294902631635e-05, "loss": 0.0, "num_input_tokens_seen": 470304, "step": 1680 }, { "epoch": 18.72222222222222, "grad_norm": 9.552429401082918e-05, "learning_rate": 4.9781656249717914e-05, "loss": 0.0, "num_input_tokens_seen": 471712, "step": 1685 }, { "epoch": 18.77777777777778, "grad_norm": 0.00011139902198920026, "learning_rate": 4.9780359651476645e-05, "loss": 0.0, "num_input_tokens_seen": 473072, "step": 1690 }, { "epoch": 18.833333333333332, "grad_norm": 5.7503617426846176e-05, "learning_rate": 4.977905923179251e-05, "loss": 0.0, "num_input_tokens_seen": 474512, "step": 1695 }, { "epoch": 18.88888888888889, "grad_norm": 5.095069718663581e-05, "learning_rate": 4.977775499086606e-05, "loss": 0.0, "num_input_tokens_seen": 475936, "step": 1700 }, { "epoch": 18.944444444444443, "grad_norm": 8.869833254721016e-05, "learning_rate": 4.97764469288984e-05, "loss": 0.0, "num_input_tokens_seen": 477392, "step": 1705 }, { "epoch": 19.0, "grad_norm": 0.00010512317385291681, "learning_rate": 4.977513504609127e-05, "loss": 0.0, "num_input_tokens_seen": 478768, "step": 1710 }, { "epoch": 19.055555555555557, "grad_norm": 0.0008224218618124723, "learning_rate": 4.9773819342646965e-05, "loss": 0.0, "num_input_tokens_seen": 480176, "step": 1715 }, { "epoch": 19.11111111111111, "grad_norm": 0.0001066598342731595, "learning_rate": 4.97724998187684e-05, "loss": 0.0, "num_input_tokens_seen": 481552, "step": 1720 }, { "epoch": 19.166666666666668, "grad_norm": 0.00012286397395655513, "learning_rate": 4.9771176474659045e-05, "loss": 0.0, "num_input_tokens_seen": 482928, "step": 1725 }, { "epoch": 19.22222222222222, "grad_norm": 7.251569331856444e-05, "learning_rate": 4.976984931052299e-05, "loss": 0.0, "num_input_tokens_seen": 484320, "step": 1730 }, { "epoch": 19.27777777777778, "grad_norm": 7.488499977625906e-05, "learning_rate": 4.976851832656489e-05, "loss": 0.0, "num_input_tokens_seen": 485712, "step": 1735 }, { "epoch": 19.333333333333332, "grad_norm": 0.0012924324255436659, "learning_rate": 4.9767183522990004e-05, "loss": 0.0, "num_input_tokens_seen": 487168, "step": 1740 }, { "epoch": 19.38888888888889, "grad_norm": 4.154927228228189e-05, "learning_rate": 4.9765844900004176e-05, "loss": 0.0, "num_input_tokens_seen": 488528, "step": 1745 }, { "epoch": 19.444444444444443, "grad_norm": 0.0004510214494075626, "learning_rate": 4.9764502457813834e-05, "loss": 0.0, "num_input_tokens_seen": 489920, "step": 1750 }, { "epoch": 19.5, "grad_norm": 0.00044796906877309084, "learning_rate": 4.9763156196626005e-05, "loss": 0.0, "num_input_tokens_seen": 491328, "step": 1755 }, { "epoch": 19.555555555555557, "grad_norm": 0.00013572006719186902, "learning_rate": 4.97618061166483e-05, "loss": 0.0, "num_input_tokens_seen": 492704, "step": 1760 }, { "epoch": 19.61111111111111, "grad_norm": 0.00012221597717143595, "learning_rate": 4.9760452218088915e-05, "loss": 0.0, "num_input_tokens_seen": 494096, "step": 1765 }, { "epoch": 19.666666666666668, "grad_norm": 0.00019606169371400028, "learning_rate": 4.975909450115663e-05, "loss": 0.0, "num_input_tokens_seen": 495504, "step": 1770 }, { "epoch": 19.72222222222222, "grad_norm": 3.378487235750072e-05, "learning_rate": 4.975773296606084e-05, "loss": 0.0, "num_input_tokens_seen": 496944, "step": 1775 }, { "epoch": 19.77777777777778, "grad_norm": 0.00010729785572038963, "learning_rate": 4.97563676130115e-05, "loss": 0.0, "num_input_tokens_seen": 498336, "step": 1780 }, { "epoch": 19.833333333333332, "grad_norm": 8.452758629573509e-05, "learning_rate": 4.9754998442219166e-05, "loss": 0.0, "num_input_tokens_seen": 499776, "step": 1785 }, { "epoch": 19.88888888888889, "grad_norm": 0.00031103866058401763, "learning_rate": 4.9753625453894984e-05, "loss": 0.0, "num_input_tokens_seen": 501120, "step": 1790 }, { "epoch": 19.944444444444443, "grad_norm": 0.0003434326499700546, "learning_rate": 4.975224864825068e-05, "loss": 0.0, "num_input_tokens_seen": 502480, "step": 1795 }, { "epoch": 20.0, "grad_norm": 0.00016531688743270934, "learning_rate": 4.9750868025498576e-05, "loss": 0.0, "num_input_tokens_seen": 503920, "step": 1800 }, { "epoch": 20.0, "eval_loss": 0.8222076296806335, "eval_runtime": 1.3996, "eval_samples_per_second": 28.58, "eval_steps_per_second": 14.29, "num_input_tokens_seen": 503920, "step": 1800 }, { "epoch": 20.055555555555557, "grad_norm": 5.4542750149266794e-05, "learning_rate": 4.974948358585158e-05, "loss": 0.0, "num_input_tokens_seen": 505296, "step": 1805 }, { "epoch": 20.11111111111111, "grad_norm": 0.0002679562021512538, "learning_rate": 4.9748095329523205e-05, "loss": 0.0, "num_input_tokens_seen": 506688, "step": 1810 }, { "epoch": 20.166666666666668, "grad_norm": 9.193173900712281e-05, "learning_rate": 4.974670325672752e-05, "loss": 0.0, "num_input_tokens_seen": 508112, "step": 1815 }, { "epoch": 20.22222222222222, "grad_norm": 8.52837401907891e-05, "learning_rate": 4.974530736767921e-05, "loss": 0.0, "num_input_tokens_seen": 509488, "step": 1820 }, { "epoch": 20.27777777777778, "grad_norm": 0.000394130329368636, "learning_rate": 4.9743907662593524e-05, "loss": 0.0, "num_input_tokens_seen": 510928, "step": 1825 }, { "epoch": 20.333333333333332, "grad_norm": 0.0004559578374028206, "learning_rate": 4.974250414168633e-05, "loss": 0.0, "num_input_tokens_seen": 512352, "step": 1830 }, { "epoch": 20.38888888888889, "grad_norm": 0.00017352827126160264, "learning_rate": 4.974109680517407e-05, "loss": 0.0, "num_input_tokens_seen": 513792, "step": 1835 }, { "epoch": 20.444444444444443, "grad_norm": 0.00011737380555132404, "learning_rate": 4.973968565327376e-05, "loss": 0.0, "num_input_tokens_seen": 515232, "step": 1840 }, { "epoch": 20.5, "grad_norm": 0.00011081498087150976, "learning_rate": 4.973827068620303e-05, "loss": 0.0, "num_input_tokens_seen": 516608, "step": 1845 }, { "epoch": 20.555555555555557, "grad_norm": 0.0008738667238503695, "learning_rate": 4.973685190418008e-05, "loss": 0.0, "num_input_tokens_seen": 517984, "step": 1850 }, { "epoch": 20.61111111111111, "grad_norm": 0.00011784715024987236, "learning_rate": 4.97354293074237e-05, "loss": 0.0, "num_input_tokens_seen": 519392, "step": 1855 }, { "epoch": 20.666666666666668, "grad_norm": 0.0006040948210284114, "learning_rate": 4.9734002896153276e-05, "loss": 0.0, "num_input_tokens_seen": 520768, "step": 1860 }, { "epoch": 20.72222222222222, "grad_norm": 0.0003986984957009554, "learning_rate": 4.973257267058877e-05, "loss": 0.0, "num_input_tokens_seen": 522192, "step": 1865 }, { "epoch": 20.77777777777778, "grad_norm": 4.9194848543265834e-05, "learning_rate": 4.973113863095076e-05, "loss": 0.0, "num_input_tokens_seen": 523616, "step": 1870 }, { "epoch": 20.833333333333332, "grad_norm": 0.00011139233538415283, "learning_rate": 4.9729700777460384e-05, "loss": 0.0, "num_input_tokens_seen": 525008, "step": 1875 }, { "epoch": 20.88888888888889, "grad_norm": 0.0007251373026520014, "learning_rate": 4.972825911033937e-05, "loss": 0.0, "num_input_tokens_seen": 526384, "step": 1880 }, { "epoch": 20.944444444444443, "grad_norm": 5.4022148106014356e-05, "learning_rate": 4.9726813629810056e-05, "loss": 0.0, "num_input_tokens_seen": 527808, "step": 1885 }, { "epoch": 21.0, "grad_norm": 0.00020635471446439624, "learning_rate": 4.9725364336095326e-05, "loss": 0.0, "num_input_tokens_seen": 529216, "step": 1890 }, { "epoch": 21.055555555555557, "grad_norm": 0.0002467419544700533, "learning_rate": 4.972391122941871e-05, "loss": 0.0, "num_input_tokens_seen": 530624, "step": 1895 }, { "epoch": 21.11111111111111, "grad_norm": 0.00010597435175441206, "learning_rate": 4.972245431000428e-05, "loss": 0.0, "num_input_tokens_seen": 532032, "step": 1900 }, { "epoch": 21.166666666666668, "grad_norm": 2.769455204543192e-05, "learning_rate": 4.972099357807671e-05, "loss": 0.0, "num_input_tokens_seen": 533408, "step": 1905 }, { "epoch": 21.22222222222222, "grad_norm": 4.01767210860271e-05, "learning_rate": 4.971952903386127e-05, "loss": 0.0, "num_input_tokens_seen": 534816, "step": 1910 }, { "epoch": 21.27777777777778, "grad_norm": 1.733384488034062e-05, "learning_rate": 4.971806067758381e-05, "loss": 0.0, "num_input_tokens_seen": 536208, "step": 1915 }, { "epoch": 21.333333333333332, "grad_norm": 7.532102608820423e-05, "learning_rate": 4.971658850947076e-05, "loss": 0.0, "num_input_tokens_seen": 537632, "step": 1920 }, { "epoch": 21.38888888888889, "grad_norm": 0.00032594683580100536, "learning_rate": 4.9715112529749165e-05, "loss": 0.0, "num_input_tokens_seen": 539040, "step": 1925 }, { "epoch": 21.444444444444443, "grad_norm": 3.1064751965459436e-05, "learning_rate": 4.9713632738646624e-05, "loss": 0.0, "num_input_tokens_seen": 540464, "step": 1930 }, { "epoch": 21.5, "grad_norm": 0.0005710162804462016, "learning_rate": 4.971214913639134e-05, "loss": 0.0, "num_input_tokens_seen": 541872, "step": 1935 }, { "epoch": 21.555555555555557, "grad_norm": 9.109647362492979e-05, "learning_rate": 4.9710661723212104e-05, "loss": 0.0, "num_input_tokens_seen": 543264, "step": 1940 }, { "epoch": 21.61111111111111, "grad_norm": 4.968086795997806e-05, "learning_rate": 4.9709170499338295e-05, "loss": 0.0, "num_input_tokens_seen": 544688, "step": 1945 }, { "epoch": 21.666666666666668, "grad_norm": 8.631202217657119e-05, "learning_rate": 4.9707675464999895e-05, "loss": 0.0, "num_input_tokens_seen": 546096, "step": 1950 }, { "epoch": 21.72222222222222, "grad_norm": 0.00015864372835494578, "learning_rate": 4.970617662042743e-05, "loss": 0.0, "num_input_tokens_seen": 547472, "step": 1955 }, { "epoch": 21.77777777777778, "grad_norm": 7.792789983795956e-05, "learning_rate": 4.970467396585206e-05, "loss": 0.0, "num_input_tokens_seen": 548912, "step": 1960 }, { "epoch": 21.833333333333332, "grad_norm": 3.568080501281656e-05, "learning_rate": 4.97031675015055e-05, "loss": 0.0, "num_input_tokens_seen": 550304, "step": 1965 }, { "epoch": 21.88888888888889, "grad_norm": 0.0001224165316671133, "learning_rate": 4.9701657227620075e-05, "loss": 0.0, "num_input_tokens_seen": 551664, "step": 1970 }, { "epoch": 21.944444444444443, "grad_norm": 0.0002568525087554008, "learning_rate": 4.9700143144428685e-05, "loss": 0.0, "num_input_tokens_seen": 553040, "step": 1975 }, { "epoch": 22.0, "grad_norm": 0.00011700252798618749, "learning_rate": 4.969862525216482e-05, "loss": 0.0, "num_input_tokens_seen": 554400, "step": 1980 }, { "epoch": 22.055555555555557, "grad_norm": 6.400218262569979e-05, "learning_rate": 4.9697103551062556e-05, "loss": 0.0, "num_input_tokens_seen": 555856, "step": 1985 }, { "epoch": 22.11111111111111, "grad_norm": 2.570304241089616e-05, "learning_rate": 4.9695578041356565e-05, "loss": 0.0, "num_input_tokens_seen": 557232, "step": 1990 }, { "epoch": 22.166666666666668, "grad_norm": 0.0013917731121182442, "learning_rate": 4.969404872328209e-05, "loss": 0.0, "num_input_tokens_seen": 558640, "step": 1995 }, { "epoch": 22.22222222222222, "grad_norm": 0.0002819855872076005, "learning_rate": 4.969251559707498e-05, "loss": 0.0, "num_input_tokens_seen": 560016, "step": 2000 }, { "epoch": 22.22222222222222, "eval_loss": 0.8502686619758606, "eval_runtime": 1.379, "eval_samples_per_second": 29.007, "eval_steps_per_second": 14.504, "num_input_tokens_seen": 560016, "step": 2000 }, { "epoch": 22.27777777777778, "grad_norm": 5.157985651749186e-05, "learning_rate": 4.9690978662971674e-05, "loss": 0.0, "num_input_tokens_seen": 561408, "step": 2005 }, { "epoch": 22.333333333333332, "grad_norm": 6.462297460529953e-05, "learning_rate": 4.968943792120916e-05, "loss": 0.0, "num_input_tokens_seen": 562784, "step": 2010 }, { "epoch": 22.38888888888889, "grad_norm": 7.692455255892128e-05, "learning_rate": 4.9687893372025046e-05, "loss": 0.0, "num_input_tokens_seen": 564224, "step": 2015 }, { "epoch": 22.444444444444443, "grad_norm": 5.622641765512526e-05, "learning_rate": 4.9686345015657535e-05, "loss": 0.0, "num_input_tokens_seen": 565616, "step": 2020 }, { "epoch": 22.5, "grad_norm": 0.00017372416914440691, "learning_rate": 4.968479285234538e-05, "loss": 0.0, "num_input_tokens_seen": 566992, "step": 2025 }, { "epoch": 22.555555555555557, "grad_norm": 0.0006057779537513852, "learning_rate": 4.9683236882327974e-05, "loss": 0.0, "num_input_tokens_seen": 568384, "step": 2030 }, { "epoch": 22.61111111111111, "grad_norm": 0.0001934084139065817, "learning_rate": 4.968167710584526e-05, "loss": 0.0, "num_input_tokens_seen": 569792, "step": 2035 }, { "epoch": 22.666666666666668, "grad_norm": 0.00037938213790766895, "learning_rate": 4.968011352313775e-05, "loss": 0.0, "num_input_tokens_seen": 571216, "step": 2040 }, { "epoch": 22.72222222222222, "grad_norm": 0.00018987689691130072, "learning_rate": 4.967854613444659e-05, "loss": 0.0, "num_input_tokens_seen": 572640, "step": 2045 }, { "epoch": 22.77777777777778, "grad_norm": 0.0005457923398353159, "learning_rate": 4.967697494001349e-05, "loss": 0.0, "num_input_tokens_seen": 573968, "step": 2050 }, { "epoch": 22.833333333333332, "grad_norm": 0.0003721605462487787, "learning_rate": 4.9675399940080736e-05, "loss": 0.0, "num_input_tokens_seen": 575360, "step": 2055 }, { "epoch": 22.88888888888889, "grad_norm": 3.8544341805391014e-05, "learning_rate": 4.9673821134891226e-05, "loss": 0.0, "num_input_tokens_seen": 576752, "step": 2060 }, { "epoch": 22.944444444444443, "grad_norm": 6.849401688668877e-05, "learning_rate": 4.967223852468842e-05, "loss": 0.0, "num_input_tokens_seen": 578192, "step": 2065 }, { "epoch": 23.0, "grad_norm": 0.00042144101462326944, "learning_rate": 4.967065210971639e-05, "loss": 0.0, "num_input_tokens_seen": 579584, "step": 2070 }, { "epoch": 23.055555555555557, "grad_norm": 0.0001993042096728459, "learning_rate": 4.966906189021977e-05, "loss": 0.0, "num_input_tokens_seen": 580992, "step": 2075 }, { "epoch": 23.11111111111111, "grad_norm": 0.0005612937384285033, "learning_rate": 4.966746786644379e-05, "loss": 0.0, "num_input_tokens_seen": 582368, "step": 2080 }, { "epoch": 23.166666666666668, "grad_norm": 0.00010695023956941441, "learning_rate": 4.966587003863429e-05, "loss": 0.0, "num_input_tokens_seen": 583760, "step": 2085 }, { "epoch": 23.22222222222222, "grad_norm": 3.2119438401423395e-05, "learning_rate": 4.966426840703765e-05, "loss": 0.0, "num_input_tokens_seen": 585184, "step": 2090 }, { "epoch": 23.27777777777778, "grad_norm": 0.00019546409021131694, "learning_rate": 4.9662662971900875e-05, "loss": 0.0, "num_input_tokens_seen": 586544, "step": 2095 }, { "epoch": 23.333333333333332, "grad_norm": 2.637158650031779e-05, "learning_rate": 4.9661053733471534e-05, "loss": 0.0, "num_input_tokens_seen": 587952, "step": 2100 }, { "epoch": 23.38888888888889, "grad_norm": 2.5202383767464198e-05, "learning_rate": 4.965944069199781e-05, "loss": 0.0, "num_input_tokens_seen": 589376, "step": 2105 }, { "epoch": 23.444444444444443, "grad_norm": 0.0002927422756329179, "learning_rate": 4.965782384772842e-05, "loss": 0.0, "num_input_tokens_seen": 590832, "step": 2110 }, { "epoch": 23.5, "grad_norm": 0.00012294189946260303, "learning_rate": 4.9656203200912734e-05, "loss": 0.0, "num_input_tokens_seen": 592208, "step": 2115 }, { "epoch": 23.555555555555557, "grad_norm": 4.318052015150897e-05, "learning_rate": 4.965457875180067e-05, "loss": 0.0, "num_input_tokens_seen": 593584, "step": 2120 }, { "epoch": 23.61111111111111, "grad_norm": 5.57230414415244e-05, "learning_rate": 4.9652950500642724e-05, "loss": 0.0, "num_input_tokens_seen": 594928, "step": 2125 }, { "epoch": 23.666666666666668, "grad_norm": 6.197045149747282e-05, "learning_rate": 4.965131844769001e-05, "loss": 0.0, "num_input_tokens_seen": 596304, "step": 2130 }, { "epoch": 23.72222222222222, "grad_norm": 0.00016503101505804807, "learning_rate": 4.96496825931942e-05, "loss": 0.0, "num_input_tokens_seen": 597712, "step": 2135 }, { "epoch": 23.77777777777778, "grad_norm": 0.0001153392149717547, "learning_rate": 4.9648042937407566e-05, "loss": 0.0, "num_input_tokens_seen": 599120, "step": 2140 }, { "epoch": 23.833333333333332, "grad_norm": 9.553593554301187e-05, "learning_rate": 4.964639948058297e-05, "loss": 0.0, "num_input_tokens_seen": 600528, "step": 2145 }, { "epoch": 23.88888888888889, "grad_norm": 5.8084711781702936e-05, "learning_rate": 4.9644752222973846e-05, "loss": 0.0, "num_input_tokens_seen": 601920, "step": 2150 }, { "epoch": 23.944444444444443, "grad_norm": 1.7461303286836483e-05, "learning_rate": 4.964310116483422e-05, "loss": 0.0, "num_input_tokens_seen": 603296, "step": 2155 }, { "epoch": 24.0, "grad_norm": 3.380742055014707e-05, "learning_rate": 4.964144630641872e-05, "loss": 0.0, "num_input_tokens_seen": 604720, "step": 2160 }, { "epoch": 24.055555555555557, "grad_norm": 8.028733282117173e-05, "learning_rate": 4.9639787647982525e-05, "loss": 0.0, "num_input_tokens_seen": 606112, "step": 2165 }, { "epoch": 24.11111111111111, "grad_norm": 0.00048083445290103555, "learning_rate": 4.963812518978143e-05, "loss": 0.0, "num_input_tokens_seen": 607488, "step": 2170 }, { "epoch": 24.166666666666668, "grad_norm": 1.4123955224931706e-05, "learning_rate": 4.963645893207182e-05, "loss": 0.0, "num_input_tokens_seen": 608912, "step": 2175 }, { "epoch": 24.22222222222222, "grad_norm": 2.1610374460578896e-05, "learning_rate": 4.963478887511063e-05, "loss": 0.0, "num_input_tokens_seen": 610320, "step": 2180 }, { "epoch": 24.27777777777778, "grad_norm": 4.276915205991827e-05, "learning_rate": 4.963311501915542e-05, "loss": 0.0, "num_input_tokens_seen": 611760, "step": 2185 }, { "epoch": 24.333333333333332, "grad_norm": 0.0001376907021040097, "learning_rate": 4.963143736446432e-05, "loss": 0.0, "num_input_tokens_seen": 613120, "step": 2190 }, { "epoch": 24.38888888888889, "grad_norm": 3.372023275005631e-05, "learning_rate": 4.962975591129603e-05, "loss": 0.0, "num_input_tokens_seen": 614544, "step": 2195 }, { "epoch": 24.444444444444443, "grad_norm": 8.958809485193342e-05, "learning_rate": 4.962807065990986e-05, "loss": 0.0, "num_input_tokens_seen": 615952, "step": 2200 }, { "epoch": 24.444444444444443, "eval_loss": 0.8652166128158569, "eval_runtime": 1.3824, "eval_samples_per_second": 28.936, "eval_steps_per_second": 14.468, "num_input_tokens_seen": 615952, "step": 2200 }, { "epoch": 24.5, "grad_norm": 0.00040576246101409197, "learning_rate": 4.9626381610565714e-05, "loss": 0.0, "num_input_tokens_seen": 617360, "step": 2205 }, { "epoch": 24.555555555555557, "grad_norm": 6.024059257470071e-05, "learning_rate": 4.9624688763524043e-05, "loss": 0.0, "num_input_tokens_seen": 618736, "step": 2210 }, { "epoch": 24.61111111111111, "grad_norm": 6.059467341401614e-05, "learning_rate": 4.962299211904591e-05, "loss": 0.0, "num_input_tokens_seen": 620128, "step": 2215 }, { "epoch": 24.666666666666668, "grad_norm": 0.000886657100636512, "learning_rate": 4.962129167739296e-05, "loss": 0.0, "num_input_tokens_seen": 621488, "step": 2220 }, { "epoch": 24.72222222222222, "grad_norm": 1.7720296455081552e-05, "learning_rate": 4.961958743882742e-05, "loss": 0.0, "num_input_tokens_seen": 622880, "step": 2225 }, { "epoch": 24.77777777777778, "grad_norm": 5.51643934159074e-05, "learning_rate": 4.961787940361211e-05, "loss": 0.0, "num_input_tokens_seen": 624304, "step": 2230 }, { "epoch": 24.833333333333332, "grad_norm": 1.9458564565866254e-05, "learning_rate": 4.961616757201043e-05, "loss": 0.0, "num_input_tokens_seen": 625696, "step": 2235 }, { "epoch": 24.88888888888889, "grad_norm": 9.721668175188825e-05, "learning_rate": 4.961445194428637e-05, "loss": 0.0, "num_input_tokens_seen": 627120, "step": 2240 }, { "epoch": 24.944444444444443, "grad_norm": 5.16153741045855e-05, "learning_rate": 4.9612732520704486e-05, "loss": 0.0, "num_input_tokens_seen": 628528, "step": 2245 }, { "epoch": 25.0, "grad_norm": 0.00016544442041777074, "learning_rate": 4.961100930152994e-05, "loss": 0.0, "num_input_tokens_seen": 630000, "step": 2250 }, { "epoch": 25.055555555555557, "grad_norm": 5.578760465141386e-05, "learning_rate": 4.960928228702849e-05, "loss": 0.0, "num_input_tokens_seen": 631408, "step": 2255 }, { "epoch": 25.11111111111111, "grad_norm": 4.176267248112708e-05, "learning_rate": 4.960755147746645e-05, "loss": 0.0, "num_input_tokens_seen": 632800, "step": 2260 }, { "epoch": 25.166666666666668, "grad_norm": 0.00012132609117543325, "learning_rate": 4.9605816873110736e-05, "loss": 0.0, "num_input_tokens_seen": 634240, "step": 2265 }, { "epoch": 25.22222222222222, "grad_norm": 9.320865501649678e-05, "learning_rate": 4.960407847422883e-05, "loss": 0.0, "num_input_tokens_seen": 635632, "step": 2270 }, { "epoch": 25.27777777777778, "grad_norm": 4.781084135174751e-05, "learning_rate": 4.960233628108885e-05, "loss": 0.0, "num_input_tokens_seen": 637008, "step": 2275 }, { "epoch": 25.333333333333332, "grad_norm": 0.00022698378597851843, "learning_rate": 4.960059029395942e-05, "loss": 0.0, "num_input_tokens_seen": 638384, "step": 2280 }, { "epoch": 25.38888888888889, "grad_norm": 0.0010108143324032426, "learning_rate": 4.959884051310983e-05, "loss": 0.0, "num_input_tokens_seen": 639808, "step": 2285 }, { "epoch": 25.444444444444443, "grad_norm": 0.0005294315051287413, "learning_rate": 4.959708693880991e-05, "loss": 0.0, "num_input_tokens_seen": 641232, "step": 2290 }, { "epoch": 25.5, "grad_norm": 0.00022696018277201802, "learning_rate": 4.9595329571330074e-05, "loss": 0.0, "num_input_tokens_seen": 642608, "step": 2295 }, { "epoch": 25.555555555555557, "grad_norm": 2.8993717933190055e-05, "learning_rate": 4.9593568410941326e-05, "loss": 0.0, "num_input_tokens_seen": 644032, "step": 2300 }, { "epoch": 25.61111111111111, "grad_norm": 6.00336461502593e-05, "learning_rate": 4.959180345791528e-05, "loss": 0.0, "num_input_tokens_seen": 645440, "step": 2305 }, { "epoch": 25.666666666666668, "grad_norm": 8.727703971089795e-05, "learning_rate": 4.9590034712524086e-05, "loss": 0.0, "num_input_tokens_seen": 646816, "step": 2310 }, { "epoch": 25.72222222222222, "grad_norm": 7.356452260864899e-05, "learning_rate": 4.958826217504053e-05, "loss": 0.0, "num_input_tokens_seen": 648240, "step": 2315 }, { "epoch": 25.77777777777778, "grad_norm": 3.331639163661748e-05, "learning_rate": 4.958648584573795e-05, "loss": 0.0, "num_input_tokens_seen": 649632, "step": 2320 }, { "epoch": 25.833333333333332, "grad_norm": 2.4345878046005964e-05, "learning_rate": 4.958470572489028e-05, "loss": 0.0, "num_input_tokens_seen": 651040, "step": 2325 }, { "epoch": 25.88888888888889, "grad_norm": 3.715162893058732e-05, "learning_rate": 4.958292181277203e-05, "loss": 0.0, "num_input_tokens_seen": 652432, "step": 2330 }, { "epoch": 25.944444444444443, "grad_norm": 4.074284152011387e-05, "learning_rate": 4.958113410965832e-05, "loss": 0.0, "num_input_tokens_seen": 653808, "step": 2335 }, { "epoch": 26.0, "grad_norm": 1.784297091944609e-05, "learning_rate": 4.957934261582481e-05, "loss": 0.0, "num_input_tokens_seen": 655200, "step": 2340 }, { "epoch": 26.055555555555557, "grad_norm": 5.018157753511332e-05, "learning_rate": 4.95775473315478e-05, "loss": 0.0, "num_input_tokens_seen": 656624, "step": 2345 }, { "epoch": 26.11111111111111, "grad_norm": 0.00038048424175940454, "learning_rate": 4.9575748257104124e-05, "loss": 0.0, "num_input_tokens_seen": 658048, "step": 2350 }, { "epoch": 26.166666666666668, "grad_norm": 0.00014599399582948536, "learning_rate": 4.9573945392771224e-05, "loss": 0.0, "num_input_tokens_seen": 659456, "step": 2355 }, { "epoch": 26.22222222222222, "grad_norm": 4.231333878124133e-05, "learning_rate": 4.9572138738827134e-05, "loss": 0.0, "num_input_tokens_seen": 660864, "step": 2360 }, { "epoch": 26.27777777777778, "grad_norm": 0.00026025797706097364, "learning_rate": 4.957032829555046e-05, "loss": 0.0, "num_input_tokens_seen": 662224, "step": 2365 }, { "epoch": 26.333333333333332, "grad_norm": 0.00017484156705904752, "learning_rate": 4.956851406322039e-05, "loss": 0.0, "num_input_tokens_seen": 663680, "step": 2370 }, { "epoch": 26.38888888888889, "grad_norm": 0.0002955698291771114, "learning_rate": 4.9566696042116704e-05, "loss": 0.0, "num_input_tokens_seen": 665104, "step": 2375 }, { "epoch": 26.444444444444443, "grad_norm": 4.9519156164024025e-05, "learning_rate": 4.9564874232519766e-05, "loss": 0.0, "num_input_tokens_seen": 666496, "step": 2380 }, { "epoch": 26.5, "grad_norm": 4.3825119064422324e-05, "learning_rate": 4.9563048634710516e-05, "loss": 0.0, "num_input_tokens_seen": 667904, "step": 2385 }, { "epoch": 26.555555555555557, "grad_norm": 4.135424751439132e-05, "learning_rate": 4.956121924897049e-05, "loss": 0.0, "num_input_tokens_seen": 669280, "step": 2390 }, { "epoch": 26.61111111111111, "grad_norm": 8.658198930788785e-05, "learning_rate": 4.955938607558181e-05, "loss": 0.0, "num_input_tokens_seen": 670720, "step": 2395 }, { "epoch": 26.666666666666668, "grad_norm": 0.0004204284050501883, "learning_rate": 4.955754911482715e-05, "loss": 0.0, "num_input_tokens_seen": 672080, "step": 2400 }, { "epoch": 26.666666666666668, "eval_loss": 0.8761671185493469, "eval_runtime": 1.402, "eval_samples_per_second": 28.531, "eval_steps_per_second": 14.266, "num_input_tokens_seen": 672080, "step": 2400 }, { "epoch": 26.72222222222222, "grad_norm": 4.222455754643306e-05, "learning_rate": 4.9555708366989804e-05, "loss": 0.0, "num_input_tokens_seen": 673440, "step": 2405 }, { "epoch": 26.77777777777778, "grad_norm": 4.134628034080379e-05, "learning_rate": 4.9553863832353655e-05, "loss": 0.0, "num_input_tokens_seen": 674816, "step": 2410 }, { "epoch": 26.833333333333332, "grad_norm": 2.9197892217780463e-05, "learning_rate": 4.955201551120313e-05, "loss": 0.0, "num_input_tokens_seen": 676176, "step": 2415 }, { "epoch": 26.88888888888889, "grad_norm": 2.395533192611765e-05, "learning_rate": 4.955016340382328e-05, "loss": 0.0, "num_input_tokens_seen": 677552, "step": 2420 }, { "epoch": 26.944444444444443, "grad_norm": 0.0008047012379392982, "learning_rate": 4.954830751049972e-05, "loss": 0.0, "num_input_tokens_seen": 678960, "step": 2425 }, { "epoch": 27.0, "grad_norm": 0.00016575933841522783, "learning_rate": 4.954644783151864e-05, "loss": 0.0, "num_input_tokens_seen": 680432, "step": 2430 }, { "epoch": 27.055555555555557, "grad_norm": 3.195412500645034e-05, "learning_rate": 4.954458436716684e-05, "loss": 0.0, "num_input_tokens_seen": 681808, "step": 2435 }, { "epoch": 27.11111111111111, "grad_norm": 3.013573405041825e-05, "learning_rate": 4.954271711773168e-05, "loss": 0.0, "num_input_tokens_seen": 683216, "step": 2440 }, { "epoch": 27.166666666666668, "grad_norm": 0.00010391337127657607, "learning_rate": 4.9540846083501115e-05, "loss": 0.0, "num_input_tokens_seen": 684608, "step": 2445 }, { "epoch": 27.22222222222222, "grad_norm": 5.287347448756918e-05, "learning_rate": 4.953897126476369e-05, "loss": 0.0, "num_input_tokens_seen": 686000, "step": 2450 }, { "epoch": 27.27777777777778, "grad_norm": 4.66477322333958e-05, "learning_rate": 4.9537092661808514e-05, "loss": 0.0, "num_input_tokens_seen": 687408, "step": 2455 }, { "epoch": 27.333333333333332, "grad_norm": 1.393284310324816e-05, "learning_rate": 4.9535210274925306e-05, "loss": 0.0, "num_input_tokens_seen": 688800, "step": 2460 }, { "epoch": 27.38888888888889, "grad_norm": 1.5105432794371154e-05, "learning_rate": 4.953332410440435e-05, "loss": 0.0, "num_input_tokens_seen": 690208, "step": 2465 }, { "epoch": 27.444444444444443, "grad_norm": 4.7234036173904315e-05, "learning_rate": 4.9531434150536496e-05, "loss": 0.0, "num_input_tokens_seen": 691568, "step": 2470 }, { "epoch": 27.5, "grad_norm": 3.758081584237516e-05, "learning_rate": 4.952954041361322e-05, "loss": 0.0, "num_input_tokens_seen": 692976, "step": 2475 }, { "epoch": 27.555555555555557, "grad_norm": 5.246984801487997e-05, "learning_rate": 4.952764289392655e-05, "loss": 0.0, "num_input_tokens_seen": 694336, "step": 2480 }, { "epoch": 27.61111111111111, "grad_norm": 8.123254519887269e-05, "learning_rate": 4.952574159176912e-05, "loss": 0.0, "num_input_tokens_seen": 695744, "step": 2485 }, { "epoch": 27.666666666666668, "grad_norm": 0.00011323363287374377, "learning_rate": 4.952383650743413e-05, "loss": 0.0, "num_input_tokens_seen": 697152, "step": 2490 }, { "epoch": 27.72222222222222, "grad_norm": 2.9316546715563163e-05, "learning_rate": 4.952192764121536e-05, "loss": 0.0, "num_input_tokens_seen": 698592, "step": 2495 }, { "epoch": 27.77777777777778, "grad_norm": 4.79363625345286e-05, "learning_rate": 4.9520014993407185e-05, "loss": 0.0, "num_input_tokens_seen": 700000, "step": 2500 }, { "epoch": 27.833333333333332, "grad_norm": 2.844012669811491e-05, "learning_rate": 4.951809856430456e-05, "loss": 0.0, "num_input_tokens_seen": 701424, "step": 2505 }, { "epoch": 27.88888888888889, "grad_norm": 0.0006006441544741392, "learning_rate": 4.951617835420303e-05, "loss": 0.0, "num_input_tokens_seen": 702816, "step": 2510 }, { "epoch": 27.944444444444443, "grad_norm": 0.00010569362348178402, "learning_rate": 4.951425436339869e-05, "loss": 0.0, "num_input_tokens_seen": 704208, "step": 2515 }, { "epoch": 28.0, "grad_norm": 0.00034518507891334593, "learning_rate": 4.9512326592188274e-05, "loss": 0.0, "num_input_tokens_seen": 705616, "step": 2520 }, { "epoch": 28.055555555555557, "grad_norm": 0.0001576580834807828, "learning_rate": 4.9510395040869054e-05, "loss": 0.0, "num_input_tokens_seen": 706992, "step": 2525 }, { "epoch": 28.11111111111111, "grad_norm": 0.000102562531537842, "learning_rate": 4.9508459709738905e-05, "loss": 0.0, "num_input_tokens_seen": 708448, "step": 2530 }, { "epoch": 28.166666666666668, "grad_norm": 7.52553969505243e-05, "learning_rate": 4.950652059909627e-05, "loss": 0.0, "num_input_tokens_seen": 709840, "step": 2535 }, { "epoch": 28.22222222222222, "grad_norm": 0.00015264125249814242, "learning_rate": 4.95045777092402e-05, "loss": 0.0, "num_input_tokens_seen": 711216, "step": 2540 }, { "epoch": 28.27777777777778, "grad_norm": 3.5410332202445716e-05, "learning_rate": 4.950263104047031e-05, "loss": 0.0, "num_input_tokens_seen": 712624, "step": 2545 }, { "epoch": 28.333333333333332, "grad_norm": 8.286369120469317e-05, "learning_rate": 4.9500680593086775e-05, "loss": 0.0, "num_input_tokens_seen": 714048, "step": 2550 }, { "epoch": 28.38888888888889, "grad_norm": 6.282031972659752e-05, "learning_rate": 4.94987263673904e-05, "loss": 0.0, "num_input_tokens_seen": 715424, "step": 2555 }, { "epoch": 28.444444444444443, "grad_norm": 7.42892807465978e-05, "learning_rate": 4.949676836368256e-05, "loss": 0.0, "num_input_tokens_seen": 716800, "step": 2560 }, { "epoch": 28.5, "grad_norm": 7.949929567985237e-05, "learning_rate": 4.949480658226518e-05, "loss": 0.0, "num_input_tokens_seen": 718208, "step": 2565 }, { "epoch": 28.555555555555557, "grad_norm": 3.334287612233311e-05, "learning_rate": 4.949284102344082e-05, "loss": 0.0, "num_input_tokens_seen": 719600, "step": 2570 }, { "epoch": 28.61111111111111, "grad_norm": 3.4819044230971485e-05, "learning_rate": 4.9490871687512565e-05, "loss": 0.0, "num_input_tokens_seen": 720976, "step": 2575 }, { "epoch": 28.666666666666668, "grad_norm": 0.0007882231730036438, "learning_rate": 4.948889857478413e-05, "loss": 0.0, "num_input_tokens_seen": 722368, "step": 2580 }, { "epoch": 28.72222222222222, "grad_norm": 2.5774119421839714e-05, "learning_rate": 4.948692168555978e-05, "loss": 0.0, "num_input_tokens_seen": 723808, "step": 2585 }, { "epoch": 28.77777777777778, "grad_norm": 0.0003026557678822428, "learning_rate": 4.94849410201444e-05, "loss": 0.0, "num_input_tokens_seen": 725216, "step": 2590 }, { "epoch": 28.833333333333332, "grad_norm": 2.6463409085408784e-05, "learning_rate": 4.948295657884341e-05, "loss": 0.0, "num_input_tokens_seen": 726624, "step": 2595 }, { "epoch": 28.88888888888889, "grad_norm": 0.00028928450774401426, "learning_rate": 4.9480968361962835e-05, "loss": 0.0, "num_input_tokens_seen": 728000, "step": 2600 }, { "epoch": 28.88888888888889, "eval_loss": 0.8936017155647278, "eval_runtime": 1.387, "eval_samples_per_second": 28.838, "eval_steps_per_second": 14.419, "num_input_tokens_seen": 728000, "step": 2600 }, { "epoch": 28.944444444444443, "grad_norm": 3.402322909096256e-05, "learning_rate": 4.9478976369809305e-05, "loss": 0.0, "num_input_tokens_seen": 729424, "step": 2605 }, { "epoch": 29.0, "grad_norm": 7.093116437317804e-05, "learning_rate": 4.947698060268999e-05, "loss": 0.0, "num_input_tokens_seen": 730816, "step": 2610 }, { "epoch": 29.055555555555557, "grad_norm": 0.0001817973970901221, "learning_rate": 4.9474981060912665e-05, "loss": 0.0, "num_input_tokens_seen": 732208, "step": 2615 }, { "epoch": 29.11111111111111, "grad_norm": 6.541317270603031e-05, "learning_rate": 4.94729777447857e-05, "loss": 0.0, "num_input_tokens_seen": 733584, "step": 2620 }, { "epoch": 29.166666666666668, "grad_norm": 3.0849925678921863e-05, "learning_rate": 4.947097065461801e-05, "loss": 0.0, "num_input_tokens_seen": 734992, "step": 2625 }, { "epoch": 29.22222222222222, "grad_norm": 0.00028002358158119023, "learning_rate": 4.9468959790719125e-05, "loss": 0.0, "num_input_tokens_seen": 736400, "step": 2630 }, { "epoch": 29.27777777777778, "grad_norm": 0.0003106495423708111, "learning_rate": 4.9466945153399146e-05, "loss": 0.0, "num_input_tokens_seen": 737776, "step": 2635 }, { "epoch": 29.333333333333332, "grad_norm": 3.446101618465036e-05, "learning_rate": 4.9464926742968755e-05, "loss": 0.0, "num_input_tokens_seen": 739152, "step": 2640 }, { "epoch": 29.38888888888889, "grad_norm": 8.06869356893003e-05, "learning_rate": 4.946290455973921e-05, "loss": 0.0, "num_input_tokens_seen": 740576, "step": 2645 }, { "epoch": 29.444444444444443, "grad_norm": 6.693419709336013e-05, "learning_rate": 4.9460878604022365e-05, "loss": 0.0, "num_input_tokens_seen": 741968, "step": 2650 }, { "epoch": 29.5, "grad_norm": 2.922880412370432e-05, "learning_rate": 4.945884887613065e-05, "loss": 0.0, "num_input_tokens_seen": 743328, "step": 2655 }, { "epoch": 29.555555555555557, "grad_norm": 4.15995491493959e-05, "learning_rate": 4.9456815376377055e-05, "loss": 0.0, "num_input_tokens_seen": 744704, "step": 2660 }, { "epoch": 29.61111111111111, "grad_norm": 6.279857188928872e-05, "learning_rate": 4.9454778105075195e-05, "loss": 0.0, "num_input_tokens_seen": 746112, "step": 2665 }, { "epoch": 29.666666666666668, "grad_norm": 2.1573479898506775e-05, "learning_rate": 4.945273706253924e-05, "loss": 0.0, "num_input_tokens_seen": 747536, "step": 2670 }, { "epoch": 29.72222222222222, "grad_norm": 3.726744762388989e-05, "learning_rate": 4.9450692249083925e-05, "loss": 0.0, "num_input_tokens_seen": 748944, "step": 2675 }, { "epoch": 29.77777777777778, "grad_norm": 9.75472794380039e-05, "learning_rate": 4.9448643665024605e-05, "loss": 0.0, "num_input_tokens_seen": 750368, "step": 2680 }, { "epoch": 29.833333333333332, "grad_norm": 9.64464052231051e-05, "learning_rate": 4.944659131067719e-05, "loss": 0.0, "num_input_tokens_seen": 751776, "step": 2685 }, { "epoch": 29.88888888888889, "grad_norm": 2.6942305339616723e-05, "learning_rate": 4.944453518635818e-05, "loss": 0.0, "num_input_tokens_seen": 753184, "step": 2690 }, { "epoch": 29.944444444444443, "grad_norm": 3.1172407034318894e-05, "learning_rate": 4.944247529238465e-05, "loss": 0.0, "num_input_tokens_seen": 754544, "step": 2695 }, { "epoch": 30.0, "grad_norm": 5.2656927437055856e-05, "learning_rate": 4.944041162907427e-05, "loss": 0.0, "num_input_tokens_seen": 755952, "step": 2700 }, { "epoch": 30.055555555555557, "grad_norm": 0.0002037132508121431, "learning_rate": 4.943834419674529e-05, "loss": 0.0, "num_input_tokens_seen": 757408, "step": 2705 }, { "epoch": 30.11111111111111, "grad_norm": 1.8543238184065558e-05, "learning_rate": 4.9436272995716506e-05, "loss": 0.0, "num_input_tokens_seen": 758816, "step": 2710 }, { "epoch": 30.166666666666668, "grad_norm": 0.000358207878889516, "learning_rate": 4.943419802630735e-05, "loss": 0.0, "num_input_tokens_seen": 760224, "step": 2715 }, { "epoch": 30.22222222222222, "grad_norm": 1.0885726624110248e-05, "learning_rate": 4.94321192888378e-05, "loss": 0.0, "num_input_tokens_seen": 761616, "step": 2720 }, { "epoch": 30.27777777777778, "grad_norm": 2.482199306541588e-05, "learning_rate": 4.943003678362842e-05, "loss": 0.0, "num_input_tokens_seen": 762992, "step": 2725 }, { "epoch": 30.333333333333332, "grad_norm": 8.475797949358821e-05, "learning_rate": 4.942795051100036e-05, "loss": 0.0, "num_input_tokens_seen": 764384, "step": 2730 }, { "epoch": 30.38888888888889, "grad_norm": 2.3109540052246302e-05, "learning_rate": 4.942586047127536e-05, "loss": 0.0, "num_input_tokens_seen": 765760, "step": 2735 }, { "epoch": 30.444444444444443, "grad_norm": 6.067793947295286e-05, "learning_rate": 4.942376666477571e-05, "loss": 0.0, "num_input_tokens_seen": 767168, "step": 2740 }, { "epoch": 30.5, "grad_norm": 2.8392863896442577e-05, "learning_rate": 4.9421669091824304e-05, "loss": 0.0, "num_input_tokens_seen": 768560, "step": 2745 }, { "epoch": 30.555555555555557, "grad_norm": 4.93154875584878e-05, "learning_rate": 4.9419567752744634e-05, "loss": 0.0, "num_input_tokens_seen": 769936, "step": 2750 }, { "epoch": 30.61111111111111, "grad_norm": 2.2820549929747358e-05, "learning_rate": 4.941746264786074e-05, "loss": 0.0, "num_input_tokens_seen": 771360, "step": 2755 }, { "epoch": 30.666666666666668, "grad_norm": 2.242299888166599e-05, "learning_rate": 4.9415353777497254e-05, "loss": 0.0, "num_input_tokens_seen": 772816, "step": 2760 }, { "epoch": 30.72222222222222, "grad_norm": 1.780792445060797e-05, "learning_rate": 4.9413241141979394e-05, "loss": 0.0, "num_input_tokens_seen": 774208, "step": 2765 }, { "epoch": 30.77777777777778, "grad_norm": 2.4210488845710643e-05, "learning_rate": 4.9411124741632956e-05, "loss": 0.0, "num_input_tokens_seen": 775584, "step": 2770 }, { "epoch": 30.833333333333332, "grad_norm": 9.19766680453904e-05, "learning_rate": 4.940900457678431e-05, "loss": 0.0, "num_input_tokens_seen": 776944, "step": 2775 }, { "epoch": 30.88888888888889, "grad_norm": 4.294555037631653e-05, "learning_rate": 4.9406880647760425e-05, "loss": 0.0, "num_input_tokens_seen": 778352, "step": 2780 }, { "epoch": 30.944444444444443, "grad_norm": 4.3104475480504334e-05, "learning_rate": 4.9404752954888824e-05, "loss": 0.0, "num_input_tokens_seen": 779744, "step": 2785 }, { "epoch": 31.0, "grad_norm": 5.0455244490876794e-05, "learning_rate": 4.940262149849762e-05, "loss": 0.0, "num_input_tokens_seen": 781088, "step": 2790 }, { "epoch": 31.055555555555557, "grad_norm": 0.00013652181951329112, "learning_rate": 4.9400486278915526e-05, "loss": 0.0, "num_input_tokens_seen": 782480, "step": 2795 }, { "epoch": 31.11111111111111, "grad_norm": 0.00010665217268979177, "learning_rate": 4.939834729647181e-05, "loss": 0.0, "num_input_tokens_seen": 783872, "step": 2800 }, { "epoch": 31.11111111111111, "eval_loss": 0.9060349464416504, "eval_runtime": 1.3956, "eval_samples_per_second": 28.661, "eval_steps_per_second": 14.331, "num_input_tokens_seen": 783872, "step": 2800 }, { "epoch": 31.166666666666668, "grad_norm": 2.3685730411671102e-05, "learning_rate": 4.9396204551496326e-05, "loss": 0.0, "num_input_tokens_seen": 785264, "step": 2805 }, { "epoch": 31.22222222222222, "grad_norm": 9.22456274565775e-06, "learning_rate": 4.939405804431952e-05, "loss": 0.0, "num_input_tokens_seen": 786704, "step": 2810 }, { "epoch": 31.27777777777778, "grad_norm": 6.926157220732421e-05, "learning_rate": 4.9391907775272414e-05, "loss": 0.0, "num_input_tokens_seen": 788064, "step": 2815 }, { "epoch": 31.333333333333332, "grad_norm": 9.623382538848091e-06, "learning_rate": 4.9389753744686604e-05, "loss": 0.0, "num_input_tokens_seen": 789472, "step": 2820 }, { "epoch": 31.38888888888889, "grad_norm": 1.0079033927468117e-05, "learning_rate": 4.938759595289426e-05, "loss": 0.0, "num_input_tokens_seen": 790848, "step": 2825 }, { "epoch": 31.444444444444443, "grad_norm": 0.00019652933406177908, "learning_rate": 4.938543440022815e-05, "loss": 0.0, "num_input_tokens_seen": 792192, "step": 2830 }, { "epoch": 31.5, "grad_norm": 0.0009511198149994016, "learning_rate": 4.938326908702161e-05, "loss": 0.0, "num_input_tokens_seen": 793648, "step": 2835 }, { "epoch": 31.555555555555557, "grad_norm": 1.3460256923281122e-05, "learning_rate": 4.9381100013608554e-05, "loss": 0.0, "num_input_tokens_seen": 795056, "step": 2840 }, { "epoch": 31.61111111111111, "grad_norm": 0.0003775043587666005, "learning_rate": 4.9378927180323485e-05, "loss": 0.0, "num_input_tokens_seen": 796480, "step": 2845 }, { "epoch": 31.666666666666668, "grad_norm": 9.660058276494965e-05, "learning_rate": 4.937675058750148e-05, "loss": 0.0, "num_input_tokens_seen": 797888, "step": 2850 }, { "epoch": 31.72222222222222, "grad_norm": 3.429213393246755e-05, "learning_rate": 4.937457023547819e-05, "loss": 0.0, "num_input_tokens_seen": 799328, "step": 2855 }, { "epoch": 31.77777777777778, "grad_norm": 4.993983748136088e-05, "learning_rate": 4.9372386124589876e-05, "loss": 0.0, "num_input_tokens_seen": 800752, "step": 2860 }, { "epoch": 31.833333333333332, "grad_norm": 0.0001524067047284916, "learning_rate": 4.937019825517333e-05, "loss": 0.0, "num_input_tokens_seen": 802144, "step": 2865 }, { "epoch": 31.88888888888889, "grad_norm": 5.687443263013847e-05, "learning_rate": 4.9368006627565954e-05, "loss": 0.0, "num_input_tokens_seen": 803504, "step": 2870 }, { "epoch": 31.944444444444443, "grad_norm": 5.158103886060417e-05, "learning_rate": 4.936581124210573e-05, "loss": 0.0, "num_input_tokens_seen": 804944, "step": 2875 }, { "epoch": 32.0, "grad_norm": 2.3892072931630537e-05, "learning_rate": 4.9363612099131216e-05, "loss": 0.0, "num_input_tokens_seen": 806352, "step": 2880 }, { "epoch": 32.05555555555556, "grad_norm": 3.009045212820638e-05, "learning_rate": 4.936140919898155e-05, "loss": 0.0, "num_input_tokens_seen": 807760, "step": 2885 }, { "epoch": 32.111111111111114, "grad_norm": 9.607342690287624e-06, "learning_rate": 4.9359202541996426e-05, "loss": 0.0, "num_input_tokens_seen": 809184, "step": 2890 }, { "epoch": 32.166666666666664, "grad_norm": 1.1664268640743103e-05, "learning_rate": 4.935699212851616e-05, "loss": 0.0, "num_input_tokens_seen": 810576, "step": 2895 }, { "epoch": 32.22222222222222, "grad_norm": 6.515478889923543e-05, "learning_rate": 4.935477795888162e-05, "loss": 0.0, "num_input_tokens_seen": 812000, "step": 2900 }, { "epoch": 32.27777777777778, "grad_norm": 5.4413267207564786e-05, "learning_rate": 4.935256003343426e-05, "loss": 0.0, "num_input_tokens_seen": 813392, "step": 2905 }, { "epoch": 32.333333333333336, "grad_norm": 5.4605519835604355e-05, "learning_rate": 4.93503383525161e-05, "loss": 0.0, "num_input_tokens_seen": 814800, "step": 2910 }, { "epoch": 32.388888888888886, "grad_norm": 1.7502508853795007e-05, "learning_rate": 4.934811291646977e-05, "loss": 0.0, "num_input_tokens_seen": 816176, "step": 2915 }, { "epoch": 32.44444444444444, "grad_norm": 6.470891094068065e-05, "learning_rate": 4.934588372563845e-05, "loss": 0.0, "num_input_tokens_seen": 817584, "step": 2920 }, { "epoch": 32.5, "grad_norm": 1.341591269010678e-05, "learning_rate": 4.93436507803659e-05, "loss": 0.0, "num_input_tokens_seen": 818944, "step": 2925 }, { "epoch": 32.55555555555556, "grad_norm": 1.4883313269820064e-05, "learning_rate": 4.934141408099649e-05, "loss": 0.0, "num_input_tokens_seen": 820336, "step": 2930 }, { "epoch": 32.611111111111114, "grad_norm": 2.4310547814820893e-05, "learning_rate": 4.9339173627875135e-05, "loss": 0.0, "num_input_tokens_seen": 821728, "step": 2935 }, { "epoch": 32.666666666666664, "grad_norm": 0.00023722834885120392, "learning_rate": 4.9336929421347335e-05, "loss": 0.0, "num_input_tokens_seen": 823136, "step": 2940 }, { "epoch": 32.72222222222222, "grad_norm": 7.623303827131167e-05, "learning_rate": 4.933468146175918e-05, "loss": 0.0, "num_input_tokens_seen": 824528, "step": 2945 }, { "epoch": 32.77777777777778, "grad_norm": 1.0246523743262514e-05, "learning_rate": 4.933242974945734e-05, "loss": 0.0, "num_input_tokens_seen": 825984, "step": 2950 }, { "epoch": 32.833333333333336, "grad_norm": 5.197724385652691e-05, "learning_rate": 4.933017428478906e-05, "loss": 0.0, "num_input_tokens_seen": 827360, "step": 2955 }, { "epoch": 32.888888888888886, "grad_norm": 2.030177893175278e-05, "learning_rate": 4.932791506810214e-05, "loss": 0.0, "num_input_tokens_seen": 828752, "step": 2960 }, { "epoch": 32.94444444444444, "grad_norm": 4.3926593207288533e-05, "learning_rate": 4.932565209974499e-05, "loss": 0.0, "num_input_tokens_seen": 830160, "step": 2965 }, { "epoch": 33.0, "grad_norm": 0.00016196367505472153, "learning_rate": 4.93233853800666e-05, "loss": 0.0, "num_input_tokens_seen": 831520, "step": 2970 }, { "epoch": 33.05555555555556, "grad_norm": 9.748858428793028e-05, "learning_rate": 4.932111490941651e-05, "loss": 0.0, "num_input_tokens_seen": 832880, "step": 2975 }, { "epoch": 33.111111111111114, "grad_norm": 3.983426358900033e-05, "learning_rate": 4.9318840688144876e-05, "loss": 0.0, "num_input_tokens_seen": 834240, "step": 2980 }, { "epoch": 33.166666666666664, "grad_norm": 2.121118814102374e-05, "learning_rate": 4.9316562716602387e-05, "loss": 0.0, "num_input_tokens_seen": 835616, "step": 2985 }, { "epoch": 33.22222222222222, "grad_norm": 1.3889893125451636e-05, "learning_rate": 4.9314280995140346e-05, "loss": 0.0, "num_input_tokens_seen": 837024, "step": 2990 }, { "epoch": 33.27777777777778, "grad_norm": 3.726709837792441e-05, "learning_rate": 4.931199552411063e-05, "loss": 0.0, "num_input_tokens_seen": 838416, "step": 2995 }, { "epoch": 33.333333333333336, "grad_norm": 0.00011408672435209155, "learning_rate": 4.930970630386568e-05, "loss": 0.0, "num_input_tokens_seen": 839808, "step": 3000 }, { "epoch": 33.333333333333336, "eval_loss": 0.9416044354438782, "eval_runtime": 1.3918, "eval_samples_per_second": 28.74, "eval_steps_per_second": 14.37, "num_input_tokens_seen": 839808, "step": 3000 }, { "epoch": 33.388888888888886, "grad_norm": 1.908085323520936e-05, "learning_rate": 4.9307413334758524e-05, "loss": 0.0, "num_input_tokens_seen": 841200, "step": 3005 }, { "epoch": 33.44444444444444, "grad_norm": 1.1180766705365386e-05, "learning_rate": 4.930511661714276e-05, "loss": 0.0, "num_input_tokens_seen": 842560, "step": 3010 }, { "epoch": 33.5, "grad_norm": 0.00024904843303374946, "learning_rate": 4.9302816151372576e-05, "loss": 0.0, "num_input_tokens_seen": 843984, "step": 3015 }, { "epoch": 33.55555555555556, "grad_norm": 0.0002434856432955712, "learning_rate": 4.930051193780274e-05, "loss": 0.0, "num_input_tokens_seen": 845392, "step": 3020 }, { "epoch": 33.611111111111114, "grad_norm": 4.3283842387609184e-05, "learning_rate": 4.929820397678858e-05, "loss": 0.0, "num_input_tokens_seen": 846832, "step": 3025 }, { "epoch": 33.666666666666664, "grad_norm": 0.00029469613218680024, "learning_rate": 4.9295892268686015e-05, "loss": 0.0, "num_input_tokens_seen": 848224, "step": 3030 }, { "epoch": 33.72222222222222, "grad_norm": 1.1522362001414876e-05, "learning_rate": 4.9293576813851536e-05, "loss": 0.0, "num_input_tokens_seen": 849616, "step": 3035 }, { "epoch": 33.77777777777778, "grad_norm": 0.000310637493385002, "learning_rate": 4.929125761264223e-05, "loss": 0.0, "num_input_tokens_seen": 851040, "step": 3040 }, { "epoch": 33.833333333333336, "grad_norm": 3.406820178497583e-05, "learning_rate": 4.928893466541573e-05, "loss": 0.0, "num_input_tokens_seen": 852480, "step": 3045 }, { "epoch": 33.888888888888886, "grad_norm": 5.335138121154159e-05, "learning_rate": 4.928660797253027e-05, "loss": 0.0, "num_input_tokens_seen": 853872, "step": 3050 }, { "epoch": 33.94444444444444, "grad_norm": 1.1625776096479967e-05, "learning_rate": 4.928427753434467e-05, "loss": 0.0, "num_input_tokens_seen": 855280, "step": 3055 }, { "epoch": 34.0, "grad_norm": 8.320362212543841e-06, "learning_rate": 4.9281943351218286e-05, "loss": 0.0, "num_input_tokens_seen": 856688, "step": 3060 }, { "epoch": 34.05555555555556, "grad_norm": 2.3802611394785345e-05, "learning_rate": 4.9279605423511095e-05, "loss": 0.0, "num_input_tokens_seen": 858128, "step": 3065 }, { "epoch": 34.111111111111114, "grad_norm": 4.533951505436562e-05, "learning_rate": 4.927726375158363e-05, "loss": 0.0, "num_input_tokens_seen": 859520, "step": 3070 }, { "epoch": 34.166666666666664, "grad_norm": 1.3318187484401278e-05, "learning_rate": 4.9274918335797004e-05, "loss": 0.0, "num_input_tokens_seen": 860928, "step": 3075 }, { "epoch": 34.22222222222222, "grad_norm": 1.0146065505978186e-05, "learning_rate": 4.927256917651292e-05, "loss": 0.0, "num_input_tokens_seen": 862320, "step": 3080 }, { "epoch": 34.27777777777778, "grad_norm": 5.815005715703592e-05, "learning_rate": 4.927021627409364e-05, "loss": 0.0, "num_input_tokens_seen": 863712, "step": 3085 }, { "epoch": 34.333333333333336, "grad_norm": 1.3346343621378765e-05, "learning_rate": 4.9267859628902005e-05, "loss": 0.0, "num_input_tokens_seen": 865104, "step": 3090 }, { "epoch": 34.388888888888886, "grad_norm": 0.0001033988592098467, "learning_rate": 4.9265499241301454e-05, "loss": 0.0, "num_input_tokens_seen": 866464, "step": 3095 }, { "epoch": 34.44444444444444, "grad_norm": 3.886171543854289e-05, "learning_rate": 4.926313511165598e-05, "loss": 0.0, "num_input_tokens_seen": 867856, "step": 3100 }, { "epoch": 34.5, "grad_norm": 2.0323483113315888e-05, "learning_rate": 4.926076724033016e-05, "loss": 0.0, "num_input_tokens_seen": 869216, "step": 3105 }, { "epoch": 34.55555555555556, "grad_norm": 0.00020424305694177747, "learning_rate": 4.9258395627689146e-05, "loss": 0.0, "num_input_tokens_seen": 870656, "step": 3110 }, { "epoch": 34.611111111111114, "grad_norm": 0.00010091394506162032, "learning_rate": 4.925602027409868e-05, "loss": 0.0, "num_input_tokens_seen": 872032, "step": 3115 }, { "epoch": 34.666666666666664, "grad_norm": 2.558763480919879e-05, "learning_rate": 4.925364117992507e-05, "loss": 0.0, "num_input_tokens_seen": 873440, "step": 3120 }, { "epoch": 34.72222222222222, "grad_norm": 2.929257789219264e-05, "learning_rate": 4.92512583455352e-05, "loss": 0.0, "num_input_tokens_seen": 874864, "step": 3125 }, { "epoch": 34.77777777777778, "grad_norm": 9.40219615586102e-06, "learning_rate": 4.9248871771296536e-05, "loss": 0.0, "num_input_tokens_seen": 876288, "step": 3130 }, { "epoch": 34.833333333333336, "grad_norm": 3.172839933540672e-05, "learning_rate": 4.924648145757711e-05, "loss": 0.0, "num_input_tokens_seen": 877744, "step": 3135 }, { "epoch": 34.888888888888886, "grad_norm": 1.693709782557562e-05, "learning_rate": 4.924408740474554e-05, "loss": 0.0, "num_input_tokens_seen": 879152, "step": 3140 }, { "epoch": 34.94444444444444, "grad_norm": 2.5411101887584664e-05, "learning_rate": 4.924168961317103e-05, "loss": 0.0, "num_input_tokens_seen": 880544, "step": 3145 }, { "epoch": 35.0, "grad_norm": 0.0002899378596339375, "learning_rate": 4.9239288083223334e-05, "loss": 0.0, "num_input_tokens_seen": 881984, "step": 3150 }, { "epoch": 35.05555555555556, "grad_norm": 7.428654498653486e-05, "learning_rate": 4.9236882815272803e-05, "loss": 0.0, "num_input_tokens_seen": 883392, "step": 3155 }, { "epoch": 35.111111111111114, "grad_norm": 9.57120028033387e-06, "learning_rate": 4.9234473809690365e-05, "loss": 0.0, "num_input_tokens_seen": 884848, "step": 3160 }, { "epoch": 35.166666666666664, "grad_norm": 8.225474448408931e-05, "learning_rate": 4.923206106684752e-05, "loss": 0.0, "num_input_tokens_seen": 886272, "step": 3165 }, { "epoch": 35.22222222222222, "grad_norm": 4.584119233186357e-05, "learning_rate": 4.922964458711634e-05, "loss": 0.0, "num_input_tokens_seen": 887680, "step": 3170 }, { "epoch": 35.27777777777778, "grad_norm": 3.251077578170225e-05, "learning_rate": 4.9227224370869474e-05, "loss": 0.0, "num_input_tokens_seen": 889056, "step": 3175 }, { "epoch": 35.333333333333336, "grad_norm": 0.00010729269706644118, "learning_rate": 4.9224800418480155e-05, "loss": 0.0, "num_input_tokens_seen": 890400, "step": 3180 }, { "epoch": 35.388888888888886, "grad_norm": 8.844755939207971e-06, "learning_rate": 4.9222372730322176e-05, "loss": 0.0, "num_input_tokens_seen": 891792, "step": 3185 }, { "epoch": 35.44444444444444, "grad_norm": 2.6135563530260697e-05, "learning_rate": 4.921994130676993e-05, "loss": 0.0, "num_input_tokens_seen": 893248, "step": 3190 }, { "epoch": 35.5, "grad_norm": 7.975802873261273e-05, "learning_rate": 4.9217506148198366e-05, "loss": 0.0, "num_input_tokens_seen": 894672, "step": 3195 }, { "epoch": 35.55555555555556, "grad_norm": 5.819603393319994e-05, "learning_rate": 4.921506725498302e-05, "loss": 0.0, "num_input_tokens_seen": 896064, "step": 3200 }, { "epoch": 35.55555555555556, "eval_loss": 0.9401968717575073, "eval_runtime": 1.391, "eval_samples_per_second": 28.755, "eval_steps_per_second": 14.378, "num_input_tokens_seen": 896064, "step": 3200 }, { "epoch": 35.611111111111114, "grad_norm": 2.130534267053008e-05, "learning_rate": 4.9212624627499994e-05, "loss": 0.0, "num_input_tokens_seen": 897456, "step": 3205 }, { "epoch": 35.666666666666664, "grad_norm": 4.316502236179076e-05, "learning_rate": 4.921017826612597e-05, "loss": 0.0, "num_input_tokens_seen": 898832, "step": 3210 }, { "epoch": 35.72222222222222, "grad_norm": 8.060666914389003e-06, "learning_rate": 4.9207728171238223e-05, "loss": 0.0, "num_input_tokens_seen": 900192, "step": 3215 }, { "epoch": 35.77777777777778, "grad_norm": 8.750689630687702e-06, "learning_rate": 4.920527434321458e-05, "loss": 0.0, "num_input_tokens_seen": 901536, "step": 3220 }, { "epoch": 35.833333333333336, "grad_norm": 3.576601739041507e-05, "learning_rate": 4.920281678243345e-05, "loss": 0.0, "num_input_tokens_seen": 902960, "step": 3225 }, { "epoch": 35.888888888888886, "grad_norm": 0.00012040990986861289, "learning_rate": 4.920035548927381e-05, "loss": 0.0, "num_input_tokens_seen": 904384, "step": 3230 }, { "epoch": 35.94444444444444, "grad_norm": 2.627782305353321e-05, "learning_rate": 4.919789046411525e-05, "loss": 0.0, "num_input_tokens_seen": 905792, "step": 3235 }, { "epoch": 36.0, "grad_norm": 3.787496461882256e-05, "learning_rate": 4.919542170733787e-05, "loss": 0.0, "num_input_tokens_seen": 907168, "step": 3240 }, { "epoch": 36.05555555555556, "grad_norm": 6.78621290717274e-05, "learning_rate": 4.919294921932242e-05, "loss": 0.0, "num_input_tokens_seen": 908544, "step": 3245 }, { "epoch": 36.111111111111114, "grad_norm": 5.1724855438806117e-05, "learning_rate": 4.919047300045016e-05, "loss": 0.0, "num_input_tokens_seen": 909952, "step": 3250 }, { "epoch": 36.166666666666664, "grad_norm": 0.00037893353146500885, "learning_rate": 4.918799305110299e-05, "loss": 0.0, "num_input_tokens_seen": 911328, "step": 3255 }, { "epoch": 36.22222222222222, "grad_norm": 5.377413617679849e-05, "learning_rate": 4.918550937166331e-05, "loss": 0.0, "num_input_tokens_seen": 912704, "step": 3260 }, { "epoch": 36.27777777777778, "grad_norm": 5.551828508032486e-06, "learning_rate": 4.918302196251415e-05, "loss": 0.0, "num_input_tokens_seen": 914128, "step": 3265 }, { "epoch": 36.333333333333336, "grad_norm": 1.99692185560707e-05, "learning_rate": 4.91805308240391e-05, "loss": 0.0, "num_input_tokens_seen": 915536, "step": 3270 }, { "epoch": 36.388888888888886, "grad_norm": 1.3621553080156446e-05, "learning_rate": 4.9178035956622326e-05, "loss": 0.0, "num_input_tokens_seen": 916944, "step": 3275 }, { "epoch": 36.44444444444444, "grad_norm": 2.8700129405478947e-05, "learning_rate": 4.917553736064857e-05, "loss": 0.0, "num_input_tokens_seen": 918368, "step": 3280 }, { "epoch": 36.5, "grad_norm": 7.59587564971298e-05, "learning_rate": 4.917303503650314e-05, "loss": 0.0, "num_input_tokens_seen": 919792, "step": 3285 }, { "epoch": 36.55555555555556, "grad_norm": 2.316149402759038e-05, "learning_rate": 4.917052898457194e-05, "loss": 0.0, "num_input_tokens_seen": 921200, "step": 3290 }, { "epoch": 36.611111111111114, "grad_norm": 8.034970960579813e-05, "learning_rate": 4.916801920524141e-05, "loss": 0.0, "num_input_tokens_seen": 922592, "step": 3295 }, { "epoch": 36.666666666666664, "grad_norm": 2.6258254365529865e-05, "learning_rate": 4.916550569889862e-05, "loss": 0.0, "num_input_tokens_seen": 924000, "step": 3300 }, { "epoch": 36.72222222222222, "grad_norm": 0.00031294231303036213, "learning_rate": 4.916298846593116e-05, "loss": 0.0, "num_input_tokens_seen": 925376, "step": 3305 }, { "epoch": 36.77777777777778, "grad_norm": 1.372493352391757e-05, "learning_rate": 4.916046750672722e-05, "loss": 0.0, "num_input_tokens_seen": 926816, "step": 3310 }, { "epoch": 36.833333333333336, "grad_norm": 1.9434932255535387e-05, "learning_rate": 4.915794282167559e-05, "loss": 0.0, "num_input_tokens_seen": 928176, "step": 3315 }, { "epoch": 36.888888888888886, "grad_norm": 0.00016161463281605393, "learning_rate": 4.915541441116558e-05, "loss": 0.0, "num_input_tokens_seen": 929568, "step": 3320 }, { "epoch": 36.94444444444444, "grad_norm": 1.4892173567204736e-05, "learning_rate": 4.915288227558711e-05, "loss": 0.0, "num_input_tokens_seen": 930992, "step": 3325 }, { "epoch": 37.0, "grad_norm": 1.6816415154607967e-05, "learning_rate": 4.915034641533066e-05, "loss": 0.0, "num_input_tokens_seen": 932384, "step": 3330 }, { "epoch": 37.05555555555556, "grad_norm": 2.7614260034170002e-05, "learning_rate": 4.914780683078731e-05, "loss": 0.0, "num_input_tokens_seen": 933776, "step": 3335 }, { "epoch": 37.111111111111114, "grad_norm": 3.6146462662145495e-05, "learning_rate": 4.9145263522348695e-05, "loss": 0.0, "num_input_tokens_seen": 935184, "step": 3340 }, { "epoch": 37.166666666666664, "grad_norm": 0.00011746589007088915, "learning_rate": 4.9142716490407e-05, "loss": 0.0, "num_input_tokens_seen": 936576, "step": 3345 }, { "epoch": 37.22222222222222, "grad_norm": 1.710352262307424e-05, "learning_rate": 4.914016573535504e-05, "loss": 0.0, "num_input_tokens_seen": 938032, "step": 3350 }, { "epoch": 37.27777777777778, "grad_norm": 5.260747639113106e-05, "learning_rate": 4.9137611257586154e-05, "loss": 0.0, "num_input_tokens_seen": 939456, "step": 3355 }, { "epoch": 37.333333333333336, "grad_norm": 1.3213931197242346e-05, "learning_rate": 4.9135053057494274e-05, "loss": 0.0, "num_input_tokens_seen": 940800, "step": 3360 }, { "epoch": 37.388888888888886, "grad_norm": 1.8148512026527897e-05, "learning_rate": 4.913249113547392e-05, "loss": 0.0, "num_input_tokens_seen": 942160, "step": 3365 }, { "epoch": 37.44444444444444, "grad_norm": 0.00023203101591207087, "learning_rate": 4.912992549192016e-05, "loss": 0.0, "num_input_tokens_seen": 943536, "step": 3370 }, { "epoch": 37.5, "grad_norm": 3.392543658264913e-05, "learning_rate": 4.9127356127228665e-05, "loss": 0.0, "num_input_tokens_seen": 944928, "step": 3375 }, { "epoch": 37.55555555555556, "grad_norm": 2.085809683194384e-05, "learning_rate": 4.912478304179564e-05, "loss": 0.0, "num_input_tokens_seen": 946336, "step": 3380 }, { "epoch": 37.611111111111114, "grad_norm": 0.0006182750803418458, "learning_rate": 4.9122206236017896e-05, "loss": 0.0, "num_input_tokens_seen": 947712, "step": 3385 }, { "epoch": 37.666666666666664, "grad_norm": 3.2331292459275573e-05, "learning_rate": 4.911962571029282e-05, "loss": 0.0, "num_input_tokens_seen": 949104, "step": 3390 }, { "epoch": 37.72222222222222, "grad_norm": 0.0001496963668614626, "learning_rate": 4.9117041465018353e-05, "loss": 0.0, "num_input_tokens_seen": 950496, "step": 3395 }, { "epoch": 37.77777777777778, "grad_norm": 3.206559267709963e-05, "learning_rate": 4.911445350059302e-05, "loss": 0.0, "num_input_tokens_seen": 951888, "step": 3400 }, { "epoch": 37.77777777777778, "eval_loss": 0.9591429829597473, "eval_runtime": 1.3897, "eval_samples_per_second": 28.783, "eval_steps_per_second": 14.392, "num_input_tokens_seen": 951888, "step": 3400 }, { "epoch": 37.833333333333336, "grad_norm": 0.0003007891646120697, "learning_rate": 4.9111861817415905e-05, "loss": 0.0, "num_input_tokens_seen": 953264, "step": 3405 }, { "epoch": 37.888888888888886, "grad_norm": 8.321240602526814e-05, "learning_rate": 4.91092664158867e-05, "loss": 0.0, "num_input_tokens_seen": 954656, "step": 3410 }, { "epoch": 37.94444444444444, "grad_norm": 1.5960840755724348e-05, "learning_rate": 4.910666729640563e-05, "loss": 0.0, "num_input_tokens_seen": 956032, "step": 3415 }, { "epoch": 38.0, "grad_norm": 1.3160165508452337e-05, "learning_rate": 4.910406445937353e-05, "loss": 0.0, "num_input_tokens_seen": 957424, "step": 3420 }, { "epoch": 38.05555555555556, "grad_norm": 5.1129673011018895e-06, "learning_rate": 4.9101457905191774e-05, "loss": 0.0, "num_input_tokens_seen": 958768, "step": 3425 }, { "epoch": 38.111111111111114, "grad_norm": 3.759616447496228e-05, "learning_rate": 4.909884763426233e-05, "loss": 0.0, "num_input_tokens_seen": 960192, "step": 3430 }, { "epoch": 38.166666666666664, "grad_norm": 1.1551617717486806e-05, "learning_rate": 4.9096233646987736e-05, "loss": 0.0, "num_input_tokens_seen": 961600, "step": 3435 }, { "epoch": 38.22222222222222, "grad_norm": 6.181382923386991e-05, "learning_rate": 4.9093615943771104e-05, "loss": 0.0, "num_input_tokens_seen": 963024, "step": 3440 }, { "epoch": 38.27777777777778, "grad_norm": 1.3489377124642488e-05, "learning_rate": 4.909099452501611e-05, "loss": 0.0, "num_input_tokens_seen": 964384, "step": 3445 }, { "epoch": 38.333333333333336, "grad_norm": 2.4430115445284173e-05, "learning_rate": 4.908836939112702e-05, "loss": 0.0, "num_input_tokens_seen": 965760, "step": 3450 }, { "epoch": 38.388888888888886, "grad_norm": 2.8689171813311987e-05, "learning_rate": 4.908574054250865e-05, "loss": 0.0, "num_input_tokens_seen": 967152, "step": 3455 }, { "epoch": 38.44444444444444, "grad_norm": 2.450553438393399e-05, "learning_rate": 4.9083107979566414e-05, "loss": 0.0, "num_input_tokens_seen": 968528, "step": 3460 }, { "epoch": 38.5, "grad_norm": 0.0002462016709614545, "learning_rate": 4.908047170270628e-05, "loss": 0.0, "num_input_tokens_seen": 969968, "step": 3465 }, { "epoch": 38.55555555555556, "grad_norm": 9.369742656417657e-06, "learning_rate": 4.9077831712334784e-05, "loss": 0.0, "num_input_tokens_seen": 971376, "step": 3470 }, { "epoch": 38.611111111111114, "grad_norm": 5.916678674111608e-06, "learning_rate": 4.907518800885907e-05, "loss": 0.0, "num_input_tokens_seen": 972816, "step": 3475 }, { "epoch": 38.666666666666664, "grad_norm": 2.0194833268760704e-05, "learning_rate": 4.907254059268681e-05, "loss": 0.0, "num_input_tokens_seen": 974192, "step": 3480 }, { "epoch": 38.72222222222222, "grad_norm": 1.1268874004599638e-05, "learning_rate": 4.906988946422628e-05, "loss": 0.0, "num_input_tokens_seen": 975600, "step": 3485 }, { "epoch": 38.77777777777778, "grad_norm": 2.9254008040879853e-05, "learning_rate": 4.9067234623886315e-05, "loss": 0.0, "num_input_tokens_seen": 976992, "step": 3490 }, { "epoch": 38.833333333333336, "grad_norm": 2.6735864594229497e-05, "learning_rate": 4.9064576072076316e-05, "loss": 0.0, "num_input_tokens_seen": 978384, "step": 3495 }, { "epoch": 38.888888888888886, "grad_norm": 1.419605814589886e-05, "learning_rate": 4.906191380920628e-05, "loss": 0.0, "num_input_tokens_seen": 979776, "step": 3500 }, { "epoch": 38.94444444444444, "grad_norm": 5.3724270401289687e-05, "learning_rate": 4.905924783568675e-05, "loss": 0.0, "num_input_tokens_seen": 981200, "step": 3505 }, { "epoch": 39.0, "grad_norm": 0.00025918486062437296, "learning_rate": 4.905657815192886e-05, "loss": 0.0, "num_input_tokens_seen": 982624, "step": 3510 }, { "epoch": 39.05555555555556, "grad_norm": 2.530536221456714e-05, "learning_rate": 4.90539047583443e-05, "loss": 0.0, "num_input_tokens_seen": 984048, "step": 3515 }, { "epoch": 39.111111111111114, "grad_norm": 1.9700579287018627e-05, "learning_rate": 4.905122765534534e-05, "loss": 0.0, "num_input_tokens_seen": 985472, "step": 3520 }, { "epoch": 39.166666666666664, "grad_norm": 6.952212970645633e-06, "learning_rate": 4.9048546843344846e-05, "loss": 0.0, "num_input_tokens_seen": 986880, "step": 3525 }, { "epoch": 39.22222222222222, "grad_norm": 2.689552457013633e-05, "learning_rate": 4.9045862322756206e-05, "loss": 0.0, "num_input_tokens_seen": 988320, "step": 3530 }, { "epoch": 39.27777777777778, "grad_norm": 3.745743742911145e-05, "learning_rate": 4.904317409399342e-05, "loss": 0.0, "num_input_tokens_seen": 989680, "step": 3535 }, { "epoch": 39.333333333333336, "grad_norm": 7.325001934077591e-05, "learning_rate": 4.904048215747104e-05, "loss": 0.0, "num_input_tokens_seen": 991136, "step": 3540 }, { "epoch": 39.388888888888886, "grad_norm": 3.593415021896362e-05, "learning_rate": 4.90377865136042e-05, "loss": 0.0, "num_input_tokens_seen": 992528, "step": 3545 }, { "epoch": 39.44444444444444, "grad_norm": 3.8476853660540655e-05, "learning_rate": 4.90350871628086e-05, "loss": 0.0, "num_input_tokens_seen": 993920, "step": 3550 }, { "epoch": 39.5, "grad_norm": 4.645724402507767e-06, "learning_rate": 4.903238410550052e-05, "loss": 0.0, "num_input_tokens_seen": 995312, "step": 3555 }, { "epoch": 39.55555555555556, "grad_norm": 4.975459160050377e-05, "learning_rate": 4.90296773420968e-05, "loss": 0.0, "num_input_tokens_seen": 996656, "step": 3560 }, { "epoch": 39.611111111111114, "grad_norm": 3.4027398214675486e-05, "learning_rate": 4.902696687301486e-05, "loss": 0.0, "num_input_tokens_seen": 998064, "step": 3565 }, { "epoch": 39.666666666666664, "grad_norm": 1.8443422959535383e-05, "learning_rate": 4.902425269867268e-05, "loss": 0.0, "num_input_tokens_seen": 999504, "step": 3570 }, { "epoch": 39.72222222222222, "grad_norm": 1.2364193935354706e-05, "learning_rate": 4.902153481948883e-05, "loss": 0.0, "num_input_tokens_seen": 1000864, "step": 3575 }, { "epoch": 39.77777777777778, "grad_norm": 2.0440806110855192e-05, "learning_rate": 4.901881323588244e-05, "loss": 0.0, "num_input_tokens_seen": 1002272, "step": 3580 }, { "epoch": 39.833333333333336, "grad_norm": 0.00011492156772874296, "learning_rate": 4.90160879482732e-05, "loss": 0.0, "num_input_tokens_seen": 1003648, "step": 3585 }, { "epoch": 39.888888888888886, "grad_norm": 6.47884517093189e-05, "learning_rate": 4.9013358957081405e-05, "loss": 0.0, "num_input_tokens_seen": 1005008, "step": 3590 }, { "epoch": 39.94444444444444, "grad_norm": 0.0001401575282216072, "learning_rate": 4.901062626272789e-05, "loss": 0.0, "num_input_tokens_seen": 1006384, "step": 3595 }, { "epoch": 40.0, "grad_norm": 4.069664282724261e-05, "learning_rate": 4.900788986563406e-05, "loss": 0.0, "num_input_tokens_seen": 1007760, "step": 3600 }, { "epoch": 40.0, "eval_loss": 0.9697567820549011, "eval_runtime": 1.3965, "eval_samples_per_second": 28.643, "eval_steps_per_second": 14.321, "num_input_tokens_seen": 1007760, "step": 3600 }, { "epoch": 40.05555555555556, "grad_norm": 0.00017948588356375694, "learning_rate": 4.9005149766221915e-05, "loss": 0.0, "num_input_tokens_seen": 1009168, "step": 3605 }, { "epoch": 40.111111111111114, "grad_norm": 9.73176156549016e-06, "learning_rate": 4.9002405964914e-05, "loss": 0.0, "num_input_tokens_seen": 1010576, "step": 3610 }, { "epoch": 40.166666666666664, "grad_norm": 2.3648330170544796e-05, "learning_rate": 4.899965846213346e-05, "loss": 0.0, "num_input_tokens_seen": 1011952, "step": 3615 }, { "epoch": 40.22222222222222, "grad_norm": 1.6226531442953274e-05, "learning_rate": 4.899690725830399e-05, "loss": 0.0, "num_input_tokens_seen": 1013408, "step": 3620 }, { "epoch": 40.27777777777778, "grad_norm": 2.6295459974789992e-05, "learning_rate": 4.899415235384985e-05, "loss": 0.0, "num_input_tokens_seen": 1014832, "step": 3625 }, { "epoch": 40.333333333333336, "grad_norm": 1.265124501514947e-05, "learning_rate": 4.899139374919589e-05, "loss": 0.0, "num_input_tokens_seen": 1016256, "step": 3630 }, { "epoch": 40.388888888888886, "grad_norm": 7.840478247089777e-06, "learning_rate": 4.898863144476752e-05, "loss": 0.0, "num_input_tokens_seen": 1017632, "step": 3635 }, { "epoch": 40.44444444444444, "grad_norm": 2.870014213840477e-05, "learning_rate": 4.898586544099072e-05, "loss": 0.0, "num_input_tokens_seen": 1019008, "step": 3640 }, { "epoch": 40.5, "grad_norm": 1.1239450941502582e-05, "learning_rate": 4.898309573829204e-05, "loss": 0.0, "num_input_tokens_seen": 1020432, "step": 3645 }, { "epoch": 40.55555555555556, "grad_norm": 4.5835851778974757e-05, "learning_rate": 4.898032233709862e-05, "loss": 0.0, "num_input_tokens_seen": 1021776, "step": 3650 }, { "epoch": 40.611111111111114, "grad_norm": 1.2468741260818206e-05, "learning_rate": 4.8977545237838123e-05, "loss": 0.0, "num_input_tokens_seen": 1023120, "step": 3655 }, { "epoch": 40.666666666666664, "grad_norm": 1.33121457110974e-05, "learning_rate": 4.8974764440938836e-05, "loss": 0.0, "num_input_tokens_seen": 1024512, "step": 3660 }, { "epoch": 40.72222222222222, "grad_norm": 2.5537954570609145e-05, "learning_rate": 4.897197994682959e-05, "loss": 0.0, "num_input_tokens_seen": 1025920, "step": 3665 }, { "epoch": 40.77777777777778, "grad_norm": 9.528394730295986e-05, "learning_rate": 4.8969191755939786e-05, "loss": 0.0, "num_input_tokens_seen": 1027360, "step": 3670 }, { "epoch": 40.833333333333336, "grad_norm": 6.390779162757099e-05, "learning_rate": 4.8966399868699396e-05, "loss": 0.0, "num_input_tokens_seen": 1028752, "step": 3675 }, { "epoch": 40.888888888888886, "grad_norm": 7.770858792355284e-05, "learning_rate": 4.8963604285538965e-05, "loss": 0.0, "num_input_tokens_seen": 1030128, "step": 3680 }, { "epoch": 40.94444444444444, "grad_norm": 3.593760266085155e-05, "learning_rate": 4.8960805006889604e-05, "loss": 0.0, "num_input_tokens_seen": 1031584, "step": 3685 }, { "epoch": 41.0, "grad_norm": 6.509187642222969e-06, "learning_rate": 4.8958002033183004e-05, "loss": 0.0, "num_input_tokens_seen": 1032960, "step": 3690 }, { "epoch": 41.05555555555556, "grad_norm": 2.1018007828388363e-05, "learning_rate": 4.8955195364851414e-05, "loss": 0.0, "num_input_tokens_seen": 1034416, "step": 3695 }, { "epoch": 41.111111111111114, "grad_norm": 1.2992968549951911e-05, "learning_rate": 4.895238500232766e-05, "loss": 0.0, "num_input_tokens_seen": 1035824, "step": 3700 }, { "epoch": 41.166666666666664, "grad_norm": 0.00013018546451348811, "learning_rate": 4.8949570946045143e-05, "loss": 0.0, "num_input_tokens_seen": 1037184, "step": 3705 }, { "epoch": 41.22222222222222, "grad_norm": 1.6633000996080227e-05, "learning_rate": 4.89467531964378e-05, "loss": 0.0, "num_input_tokens_seen": 1038544, "step": 3710 }, { "epoch": 41.27777777777778, "grad_norm": 8.391301889787428e-06, "learning_rate": 4.894393175394019e-05, "loss": 0.0, "num_input_tokens_seen": 1039920, "step": 3715 }, { "epoch": 41.333333333333336, "grad_norm": 1.2231388609507121e-05, "learning_rate": 4.8941106618987406e-05, "loss": 0.0, "num_input_tokens_seen": 1041344, "step": 3720 }, { "epoch": 41.388888888888886, "grad_norm": 5.762180535384687e-06, "learning_rate": 4.893827779201512e-05, "loss": 0.0, "num_input_tokens_seen": 1042704, "step": 3725 }, { "epoch": 41.44444444444444, "grad_norm": 8.850935955706518e-06, "learning_rate": 4.893544527345957e-05, "loss": 0.0, "num_input_tokens_seen": 1044080, "step": 3730 }, { "epoch": 41.5, "grad_norm": 7.128150627977448e-06, "learning_rate": 4.8932609063757563e-05, "loss": 0.0, "num_input_tokens_seen": 1045520, "step": 3735 }, { "epoch": 41.55555555555556, "grad_norm": 1.9519184206728823e-05, "learning_rate": 4.8929769163346484e-05, "loss": 0.0, "num_input_tokens_seen": 1046944, "step": 3740 }, { "epoch": 41.611111111111114, "grad_norm": 1.8841119526769035e-05, "learning_rate": 4.892692557266429e-05, "loss": 0.0, "num_input_tokens_seen": 1048336, "step": 3745 }, { "epoch": 41.666666666666664, "grad_norm": 0.00022129544231574982, "learning_rate": 4.8924078292149464e-05, "loss": 0.0, "num_input_tokens_seen": 1049760, "step": 3750 }, { "epoch": 41.72222222222222, "grad_norm": 1.1079952855652664e-05, "learning_rate": 4.892122732224114e-05, "loss": 0.0, "num_input_tokens_seen": 1051088, "step": 3755 }, { "epoch": 41.77777777777778, "grad_norm": 1.5442576113855466e-05, "learning_rate": 4.8918372663378944e-05, "loss": 0.0, "num_input_tokens_seen": 1052448, "step": 3760 }, { "epoch": 41.833333333333336, "grad_norm": 1.6472920833621174e-05, "learning_rate": 4.89155143160031e-05, "loss": 0.0, "num_input_tokens_seen": 1053840, "step": 3765 }, { "epoch": 41.888888888888886, "grad_norm": 0.00019632003386504948, "learning_rate": 4.891265228055441e-05, "loss": 0.0, "num_input_tokens_seen": 1055232, "step": 3770 }, { "epoch": 41.94444444444444, "grad_norm": 2.4787997972453013e-05, "learning_rate": 4.890978655747424e-05, "loss": 0.0, "num_input_tokens_seen": 1056672, "step": 3775 }, { "epoch": 42.0, "grad_norm": 7.210585408756742e-06, "learning_rate": 4.89069171472045e-05, "loss": 0.0, "num_input_tokens_seen": 1058080, "step": 3780 }, { "epoch": 42.05555555555556, "grad_norm": 6.389891950675519e-06, "learning_rate": 4.890404405018772e-05, "loss": 0.0, "num_input_tokens_seen": 1059440, "step": 3785 }, { "epoch": 42.111111111111114, "grad_norm": 0.0001015025918604806, "learning_rate": 4.8901167266866934e-05, "loss": 0.0, "num_input_tokens_seen": 1060864, "step": 3790 }, { "epoch": 42.166666666666664, "grad_norm": 6.218269118107855e-05, "learning_rate": 4.88982867976858e-05, "loss": 0.0, "num_input_tokens_seen": 1062224, "step": 3795 }, { "epoch": 42.22222222222222, "grad_norm": 3.8713998947059736e-05, "learning_rate": 4.889540264308852e-05, "loss": 0.0, "num_input_tokens_seen": 1063648, "step": 3800 }, { "epoch": 42.22222222222222, "eval_loss": 0.9947481155395508, "eval_runtime": 1.383, "eval_samples_per_second": 28.923, "eval_steps_per_second": 14.462, "num_input_tokens_seen": 1063648, "step": 3800 }, { "epoch": 42.27777777777778, "grad_norm": 5.406211130321026e-06, "learning_rate": 4.889251480351986e-05, "loss": 0.0, "num_input_tokens_seen": 1065024, "step": 3805 }, { "epoch": 42.333333333333336, "grad_norm": 1.9978408090537414e-05, "learning_rate": 4.888962327942517e-05, "loss": 0.0, "num_input_tokens_seen": 1066432, "step": 3810 }, { "epoch": 42.388888888888886, "grad_norm": 3.844921229756437e-05, "learning_rate": 4.8886728071250356e-05, "loss": 0.0, "num_input_tokens_seen": 1067872, "step": 3815 }, { "epoch": 42.44444444444444, "grad_norm": 9.561254955769982e-06, "learning_rate": 4.8883829179441884e-05, "loss": 0.0, "num_input_tokens_seen": 1069264, "step": 3820 }, { "epoch": 42.5, "grad_norm": 8.046648872550577e-06, "learning_rate": 4.888092660444682e-05, "loss": 0.0, "num_input_tokens_seen": 1070688, "step": 3825 }, { "epoch": 42.55555555555556, "grad_norm": 7.544962500105612e-06, "learning_rate": 4.887802034671276e-05, "loss": 0.0, "num_input_tokens_seen": 1072096, "step": 3830 }, { "epoch": 42.611111111111114, "grad_norm": 1.1562066902115475e-05, "learning_rate": 4.88751104066879e-05, "loss": 0.0, "num_input_tokens_seen": 1073456, "step": 3835 }, { "epoch": 42.666666666666664, "grad_norm": 6.909148942213506e-05, "learning_rate": 4.887219678482098e-05, "loss": 0.0, "num_input_tokens_seen": 1074864, "step": 3840 }, { "epoch": 42.72222222222222, "grad_norm": 6.64989638607949e-05, "learning_rate": 4.8869279481561316e-05, "loss": 0.0, "num_input_tokens_seen": 1076256, "step": 3845 }, { "epoch": 42.77777777777778, "grad_norm": 7.444792117894394e-06, "learning_rate": 4.88663584973588e-05, "loss": 0.0, "num_input_tokens_seen": 1077664, "step": 3850 }, { "epoch": 42.833333333333336, "grad_norm": 1.6886795492609963e-05, "learning_rate": 4.8863433832663874e-05, "loss": 0.0, "num_input_tokens_seen": 1079072, "step": 3855 }, { "epoch": 42.888888888888886, "grad_norm": 1.3411765394266695e-05, "learning_rate": 4.886050548792757e-05, "loss": 0.0, "num_input_tokens_seen": 1080480, "step": 3860 }, { "epoch": 42.94444444444444, "grad_norm": 9.531348950986285e-06, "learning_rate": 4.8857573463601465e-05, "loss": 0.0, "num_input_tokens_seen": 1081888, "step": 3865 }, { "epoch": 43.0, "grad_norm": 5.162848174222745e-05, "learning_rate": 4.885463776013772e-05, "loss": 0.0, "num_input_tokens_seen": 1083264, "step": 3870 }, { "epoch": 43.05555555555556, "grad_norm": 3.413949889363721e-05, "learning_rate": 4.8851698377989056e-05, "loss": 0.0, "num_input_tokens_seen": 1084688, "step": 3875 }, { "epoch": 43.111111111111114, "grad_norm": 5.398789198807208e-06, "learning_rate": 4.884875531760876e-05, "loss": 0.0, "num_input_tokens_seen": 1086096, "step": 3880 }, { "epoch": 43.166666666666664, "grad_norm": 2.9425393222481944e-05, "learning_rate": 4.88458085794507e-05, "loss": 0.0, "num_input_tokens_seen": 1087456, "step": 3885 }, { "epoch": 43.22222222222222, "grad_norm": 0.00020399101777002215, "learning_rate": 4.884285816396929e-05, "loss": 0.0, "num_input_tokens_seen": 1088864, "step": 3890 }, { "epoch": 43.27777777777778, "grad_norm": 1.4099992768024094e-05, "learning_rate": 4.8839904071619526e-05, "loss": 0.0, "num_input_tokens_seen": 1090256, "step": 3895 }, { "epoch": 43.333333333333336, "grad_norm": 5.765522291767411e-05, "learning_rate": 4.8836946302856955e-05, "loss": 0.0, "num_input_tokens_seen": 1091696, "step": 3900 }, { "epoch": 43.388888888888886, "grad_norm": 7.263702627824387e-06, "learning_rate": 4.8833984858137715e-05, "loss": 0.0, "num_input_tokens_seen": 1093040, "step": 3905 }, { "epoch": 43.44444444444444, "grad_norm": 9.437569133297075e-06, "learning_rate": 4.8831019737918494e-05, "loss": 0.0, "num_input_tokens_seen": 1094464, "step": 3910 }, { "epoch": 43.5, "grad_norm": 2.0979421606170945e-05, "learning_rate": 4.882805094265655e-05, "loss": 0.0, "num_input_tokens_seen": 1095856, "step": 3915 }, { "epoch": 43.55555555555556, "grad_norm": 6.480584488599561e-06, "learning_rate": 4.8825078472809706e-05, "loss": 0.0, "num_input_tokens_seen": 1097264, "step": 3920 }, { "epoch": 43.611111111111114, "grad_norm": 0.00011651279055513442, "learning_rate": 4.882210232883635e-05, "loss": 0.0, "num_input_tokens_seen": 1098656, "step": 3925 }, { "epoch": 43.666666666666664, "grad_norm": 7.59135127736954e-06, "learning_rate": 4.881912251119546e-05, "loss": 0.0, "num_input_tokens_seen": 1100096, "step": 3930 }, { "epoch": 43.72222222222222, "grad_norm": 5.404611147241667e-05, "learning_rate": 4.881613902034654e-05, "loss": 0.0, "num_input_tokens_seen": 1101568, "step": 3935 }, { "epoch": 43.77777777777778, "grad_norm": 4.7655539674451575e-06, "learning_rate": 4.88131518567497e-05, "loss": 0.0, "num_input_tokens_seen": 1102976, "step": 3940 }, { "epoch": 43.833333333333336, "grad_norm": 1.8016726244241e-05, "learning_rate": 4.881016102086558e-05, "loss": 0.0, "num_input_tokens_seen": 1104352, "step": 3945 }, { "epoch": 43.888888888888886, "grad_norm": 1.249903561983956e-05, "learning_rate": 4.8807166513155425e-05, "loss": 0.0, "num_input_tokens_seen": 1105776, "step": 3950 }, { "epoch": 43.94444444444444, "grad_norm": 1.356204847979825e-05, "learning_rate": 4.8804168334081004e-05, "loss": 0.0, "num_input_tokens_seen": 1107168, "step": 3955 }, { "epoch": 44.0, "grad_norm": 5.009686356061138e-05, "learning_rate": 4.880116648410468e-05, "loss": 0.0, "num_input_tokens_seen": 1108592, "step": 3960 }, { "epoch": 44.05555555555556, "grad_norm": 4.791670107806567e-06, "learning_rate": 4.879816096368939e-05, "loss": 0.0, "num_input_tokens_seen": 1110000, "step": 3965 }, { "epoch": 44.111111111111114, "grad_norm": 5.071765190223232e-05, "learning_rate": 4.879515177329861e-05, "loss": 0.0, "num_input_tokens_seen": 1111376, "step": 3970 }, { "epoch": 44.166666666666664, "grad_norm": 8.892788173398003e-05, "learning_rate": 4.8792138913396394e-05, "loss": 0.0, "num_input_tokens_seen": 1112800, "step": 3975 }, { "epoch": 44.22222222222222, "grad_norm": 1.3407037840806879e-05, "learning_rate": 4.8789122384447374e-05, "loss": 0.0, "num_input_tokens_seen": 1114224, "step": 3980 }, { "epoch": 44.27777777777778, "grad_norm": 0.00013091121218167245, "learning_rate": 4.878610218691673e-05, "loss": 0.0, "num_input_tokens_seen": 1115600, "step": 3985 }, { "epoch": 44.333333333333336, "grad_norm": 1.8648353943717666e-05, "learning_rate": 4.87830783212702e-05, "loss": 0.0, "num_input_tokens_seen": 1116976, "step": 3990 }, { "epoch": 44.388888888888886, "grad_norm": 7.4130171014985535e-06, "learning_rate": 4.878005078797413e-05, "loss": 0.0, "num_input_tokens_seen": 1118336, "step": 3995 }, { "epoch": 44.44444444444444, "grad_norm": 1.5243986126733944e-05, "learning_rate": 4.877701958749539e-05, "loss": 0.0, "num_input_tokens_seen": 1119744, "step": 4000 }, { "epoch": 44.44444444444444, "eval_loss": 0.9982012510299683, "eval_runtime": 1.383, "eval_samples_per_second": 28.924, "eval_steps_per_second": 14.462, "num_input_tokens_seen": 1119744, "step": 4000 }, { "epoch": 44.5, "grad_norm": 1.874399822554551e-05, "learning_rate": 4.877398472030142e-05, "loss": 0.0, "num_input_tokens_seen": 1121088, "step": 4005 }, { "epoch": 44.55555555555556, "grad_norm": 3.5786717489827424e-05, "learning_rate": 4.877094618686024e-05, "loss": 0.0, "num_input_tokens_seen": 1122512, "step": 4010 }, { "epoch": 44.611111111111114, "grad_norm": 1.4240418749977835e-05, "learning_rate": 4.876790398764045e-05, "loss": 0.0, "num_input_tokens_seen": 1123936, "step": 4015 }, { "epoch": 44.666666666666664, "grad_norm": 4.5883374696131796e-05, "learning_rate": 4.8764858123111167e-05, "loss": 0.0, "num_input_tokens_seen": 1125312, "step": 4020 }, { "epoch": 44.72222222222222, "grad_norm": 2.3487353246309794e-05, "learning_rate": 4.876180859374212e-05, "loss": 0.0, "num_input_tokens_seen": 1126736, "step": 4025 }, { "epoch": 44.77777777777778, "grad_norm": 7.321646990021691e-06, "learning_rate": 4.875875540000357e-05, "loss": 0.0, "num_input_tokens_seen": 1128176, "step": 4030 }, { "epoch": 44.833333333333336, "grad_norm": 5.146178955328651e-05, "learning_rate": 4.8755698542366376e-05, "loss": 0.0, "num_input_tokens_seen": 1129568, "step": 4035 }, { "epoch": 44.888888888888886, "grad_norm": 0.00011678854934871197, "learning_rate": 4.875263802130193e-05, "loss": 0.0, "num_input_tokens_seen": 1131008, "step": 4040 }, { "epoch": 44.94444444444444, "grad_norm": 3.472471871646121e-05, "learning_rate": 4.8749573837282207e-05, "loss": 0.0, "num_input_tokens_seen": 1132368, "step": 4045 }, { "epoch": 45.0, "grad_norm": 1.1766745956265368e-05, "learning_rate": 4.874650599077974e-05, "loss": 0.0, "num_input_tokens_seen": 1133744, "step": 4050 }, { "epoch": 45.05555555555556, "grad_norm": 7.056467438815162e-06, "learning_rate": 4.874343448226764e-05, "loss": 0.0, "num_input_tokens_seen": 1135152, "step": 4055 }, { "epoch": 45.111111111111114, "grad_norm": 6.713475613651099e-06, "learning_rate": 4.874035931221955e-05, "loss": 0.0, "num_input_tokens_seen": 1136560, "step": 4060 }, { "epoch": 45.166666666666664, "grad_norm": 3.895638201356633e-06, "learning_rate": 4.8737280481109724e-05, "loss": 0.0, "num_input_tokens_seen": 1137952, "step": 4065 }, { "epoch": 45.22222222222222, "grad_norm": 1.2308641998970415e-05, "learning_rate": 4.873419798941294e-05, "loss": 0.0, "num_input_tokens_seen": 1139328, "step": 4070 }, { "epoch": 45.27777777777778, "grad_norm": 3.7688814700231887e-06, "learning_rate": 4.873111183760458e-05, "loss": 0.0, "num_input_tokens_seen": 1140752, "step": 4075 }, { "epoch": 45.333333333333336, "grad_norm": 1.4086392184253782e-05, "learning_rate": 4.8728022026160537e-05, "loss": 0.0, "num_input_tokens_seen": 1142096, "step": 4080 }, { "epoch": 45.388888888888886, "grad_norm": 9.797677194001153e-06, "learning_rate": 4.872492855555732e-05, "loss": 0.0, "num_input_tokens_seen": 1143440, "step": 4085 }, { "epoch": 45.44444444444444, "grad_norm": 8.574225830670912e-06, "learning_rate": 4.8721831426271956e-05, "loss": 0.0, "num_input_tokens_seen": 1144832, "step": 4090 }, { "epoch": 45.5, "grad_norm": 0.00013070815475657582, "learning_rate": 4.87187306387821e-05, "loss": 0.0, "num_input_tokens_seen": 1146240, "step": 4095 }, { "epoch": 45.55555555555556, "grad_norm": 2.3450935259461403e-05, "learning_rate": 4.87156261935659e-05, "loss": 0.0, "num_input_tokens_seen": 1147648, "step": 4100 }, { "epoch": 45.611111111111114, "grad_norm": 2.24289542529732e-05, "learning_rate": 4.871251809110211e-05, "loss": 0.0, "num_input_tokens_seen": 1149040, "step": 4105 }, { "epoch": 45.666666666666664, "grad_norm": 4.609889583662152e-05, "learning_rate": 4.8709406331870044e-05, "loss": 0.0, "num_input_tokens_seen": 1150448, "step": 4110 }, { "epoch": 45.72222222222222, "grad_norm": 1.4200148143572733e-05, "learning_rate": 4.8706290916349574e-05, "loss": 0.0, "num_input_tokens_seen": 1151856, "step": 4115 }, { "epoch": 45.77777777777778, "grad_norm": 0.00012287896242924035, "learning_rate": 4.8703171845021134e-05, "loss": 0.0, "num_input_tokens_seen": 1153264, "step": 4120 }, { "epoch": 45.833333333333336, "grad_norm": 4.3149444536538795e-05, "learning_rate": 4.870004911836572e-05, "loss": 0.0, "num_input_tokens_seen": 1154656, "step": 4125 }, { "epoch": 45.888888888888886, "grad_norm": 8.970830822363496e-05, "learning_rate": 4.869692273686489e-05, "loss": 0.0, "num_input_tokens_seen": 1156096, "step": 4130 }, { "epoch": 45.94444444444444, "grad_norm": 5.434768172563054e-05, "learning_rate": 4.869379270100079e-05, "loss": 0.0, "num_input_tokens_seen": 1157488, "step": 4135 }, { "epoch": 46.0, "grad_norm": 3.137778185191564e-05, "learning_rate": 4.86906590112561e-05, "loss": 0.0, "num_input_tokens_seen": 1158864, "step": 4140 }, { "epoch": 46.05555555555556, "grad_norm": 2.848404437827412e-05, "learning_rate": 4.8687521668114064e-05, "loss": 0.0, "num_input_tokens_seen": 1160256, "step": 4145 }, { "epoch": 46.111111111111114, "grad_norm": 6.165747436170932e-06, "learning_rate": 4.868438067205853e-05, "loss": 0.0, "num_input_tokens_seen": 1161616, "step": 4150 }, { "epoch": 46.166666666666664, "grad_norm": 2.6230318326270208e-05, "learning_rate": 4.8681236023573844e-05, "loss": 0.0, "num_input_tokens_seen": 1163040, "step": 4155 }, { "epoch": 46.22222222222222, "grad_norm": 1.604597673576791e-05, "learning_rate": 4.867808772314497e-05, "loss": 0.0, "num_input_tokens_seen": 1164448, "step": 4160 }, { "epoch": 46.27777777777778, "grad_norm": 1.2258011338417418e-05, "learning_rate": 4.867493577125741e-05, "loss": 0.0, "num_input_tokens_seen": 1165888, "step": 4165 }, { "epoch": 46.333333333333336, "grad_norm": 4.656665623770095e-06, "learning_rate": 4.867178016839725e-05, "loss": 0.0, "num_input_tokens_seen": 1167296, "step": 4170 }, { "epoch": 46.388888888888886, "grad_norm": 1.2248271559656132e-05, "learning_rate": 4.8668620915051094e-05, "loss": 0.0, "num_input_tokens_seen": 1168688, "step": 4175 }, { "epoch": 46.44444444444444, "grad_norm": 7.028240815998288e-06, "learning_rate": 4.866545801170616e-05, "loss": 0.0, "num_input_tokens_seen": 1170064, "step": 4180 }, { "epoch": 46.5, "grad_norm": 2.6089512175531127e-05, "learning_rate": 4.86622914588502e-05, "loss": 0.0, "num_input_tokens_seen": 1171472, "step": 4185 }, { "epoch": 46.55555555555556, "grad_norm": 2.067297464236617e-05, "learning_rate": 4.865912125697154e-05, "loss": 0.0, "num_input_tokens_seen": 1172880, "step": 4190 }, { "epoch": 46.611111111111114, "grad_norm": 1.0689784176065587e-05, "learning_rate": 4.865594740655907e-05, "loss": 0.0, "num_input_tokens_seen": 1174288, "step": 4195 }, { "epoch": 46.666666666666664, "grad_norm": 5.631502517644549e-06, "learning_rate": 4.865276990810222e-05, "loss": 0.0, "num_input_tokens_seen": 1175680, "step": 4200 }, { "epoch": 46.666666666666664, "eval_loss": 1.01814866065979, "eval_runtime": 1.386, "eval_samples_per_second": 28.859, "eval_steps_per_second": 14.43, "num_input_tokens_seen": 1175680, "step": 4200 }, { "epoch": 46.72222222222222, "grad_norm": 1.8662527509150095e-05, "learning_rate": 4.8649588762091016e-05, "loss": 0.0, "num_input_tokens_seen": 1177056, "step": 4205 }, { "epoch": 46.77777777777778, "grad_norm": 1.3988634236739017e-05, "learning_rate": 4.8646403969016016e-05, "loss": 0.0, "num_input_tokens_seen": 1178464, "step": 4210 }, { "epoch": 46.833333333333336, "grad_norm": 1.1871605238411576e-05, "learning_rate": 4.864321552936838e-05, "loss": 0.0, "num_input_tokens_seen": 1179872, "step": 4215 }, { "epoch": 46.888888888888886, "grad_norm": 8.428411092609167e-06, "learning_rate": 4.864002344363978e-05, "loss": 0.0, "num_input_tokens_seen": 1181280, "step": 4220 }, { "epoch": 46.94444444444444, "grad_norm": 2.946536369563546e-05, "learning_rate": 4.863682771232248e-05, "loss": 0.0, "num_input_tokens_seen": 1182704, "step": 4225 }, { "epoch": 47.0, "grad_norm": 1.0210087566520087e-05, "learning_rate": 4.8633628335909324e-05, "loss": 0.0, "num_input_tokens_seen": 1184112, "step": 4230 }, { "epoch": 47.05555555555556, "grad_norm": 2.4110391677822918e-05, "learning_rate": 4.8630425314893676e-05, "loss": 0.0, "num_input_tokens_seen": 1185488, "step": 4235 }, { "epoch": 47.111111111111114, "grad_norm": 1.0949479474220425e-05, "learning_rate": 4.862721864976948e-05, "loss": 0.0, "num_input_tokens_seen": 1186848, "step": 4240 }, { "epoch": 47.166666666666664, "grad_norm": 3.408659904380329e-05, "learning_rate": 4.862400834103125e-05, "loss": 0.0, "num_input_tokens_seen": 1188240, "step": 4245 }, { "epoch": 47.22222222222222, "grad_norm": 9.483890607953072e-05, "learning_rate": 4.862079438917406e-05, "loss": 0.0, "num_input_tokens_seen": 1189680, "step": 4250 }, { "epoch": 47.27777777777778, "grad_norm": 1.6270570995402522e-05, "learning_rate": 4.8617576794693536e-05, "loss": 0.0, "num_input_tokens_seen": 1191072, "step": 4255 }, { "epoch": 47.333333333333336, "grad_norm": 2.609620787552558e-05, "learning_rate": 4.8614355558085875e-05, "loss": 0.0, "num_input_tokens_seen": 1192480, "step": 4260 }, { "epoch": 47.388888888888886, "grad_norm": 1.016026726574637e-05, "learning_rate": 4.861113067984783e-05, "loss": 0.0, "num_input_tokens_seen": 1193872, "step": 4265 }, { "epoch": 47.44444444444444, "grad_norm": 1.201215127366595e-05, "learning_rate": 4.860790216047671e-05, "loss": 0.0, "num_input_tokens_seen": 1195296, "step": 4270 }, { "epoch": 47.5, "grad_norm": 2.4572287657065317e-05, "learning_rate": 4.860467000047041e-05, "loss": 0.0, "num_input_tokens_seen": 1196672, "step": 4275 }, { "epoch": 47.55555555555556, "grad_norm": 3.7259076634654775e-05, "learning_rate": 4.860143420032737e-05, "loss": 0.0, "num_input_tokens_seen": 1198064, "step": 4280 }, { "epoch": 47.611111111111114, "grad_norm": 0.0001294278190471232, "learning_rate": 4.859819476054657e-05, "loss": 0.0, "num_input_tokens_seen": 1199456, "step": 4285 }, { "epoch": 47.666666666666664, "grad_norm": 6.276782642089529e-06, "learning_rate": 4.859495168162758e-05, "loss": 0.0, "num_input_tokens_seen": 1200864, "step": 4290 }, { "epoch": 47.72222222222222, "grad_norm": 3.94511480408255e-05, "learning_rate": 4.859170496407054e-05, "loss": 0.0, "num_input_tokens_seen": 1202224, "step": 4295 }, { "epoch": 47.77777777777778, "grad_norm": 1.068559504346922e-05, "learning_rate": 4.8588454608376114e-05, "loss": 0.0, "num_input_tokens_seen": 1203632, "step": 4300 }, { "epoch": 47.833333333333336, "grad_norm": 1.5473258827114478e-05, "learning_rate": 4.8585200615045555e-05, "loss": 0.0, "num_input_tokens_seen": 1205040, "step": 4305 }, { "epoch": 47.888888888888886, "grad_norm": 3.907824429916218e-05, "learning_rate": 4.8581942984580674e-05, "loss": 0.0, "num_input_tokens_seen": 1206480, "step": 4310 }, { "epoch": 47.94444444444444, "grad_norm": 2.2973574232310057e-05, "learning_rate": 4.857868171748384e-05, "loss": 0.0, "num_input_tokens_seen": 1207952, "step": 4315 }, { "epoch": 48.0, "grad_norm": 4.4163180064060725e-06, "learning_rate": 4.8575416814257976e-05, "loss": 0.0, "num_input_tokens_seen": 1209328, "step": 4320 }, { "epoch": 48.05555555555556, "grad_norm": 2.008791489060968e-05, "learning_rate": 4.857214827540657e-05, "loss": 0.0, "num_input_tokens_seen": 1210768, "step": 4325 }, { "epoch": 48.111111111111114, "grad_norm": 4.193325366941281e-05, "learning_rate": 4.856887610143367e-05, "loss": 0.0, "num_input_tokens_seen": 1212096, "step": 4330 }, { "epoch": 48.166666666666664, "grad_norm": 7.373726020887261e-06, "learning_rate": 4.8565600292843896e-05, "loss": 0.0, "num_input_tokens_seen": 1213504, "step": 4335 }, { "epoch": 48.22222222222222, "grad_norm": 1.1630280823737849e-05, "learning_rate": 4.856232085014241e-05, "loss": 0.0, "num_input_tokens_seen": 1214944, "step": 4340 }, { "epoch": 48.27777777777778, "grad_norm": 2.6659115974325687e-05, "learning_rate": 4.855903777383495e-05, "loss": 0.0, "num_input_tokens_seen": 1216336, "step": 4345 }, { "epoch": 48.333333333333336, "grad_norm": 2.5015129722305574e-05, "learning_rate": 4.85557510644278e-05, "loss": 0.0, "num_input_tokens_seen": 1217760, "step": 4350 }, { "epoch": 48.388888888888886, "grad_norm": 9.005412721307948e-06, "learning_rate": 4.855246072242782e-05, "loss": 0.0, "num_input_tokens_seen": 1219136, "step": 4355 }, { "epoch": 48.44444444444444, "grad_norm": 4.078761412529275e-05, "learning_rate": 4.8549166748342414e-05, "loss": 0.0, "num_input_tokens_seen": 1220512, "step": 4360 }, { "epoch": 48.5, "grad_norm": 4.942756731907139e-06, "learning_rate": 4.8545869142679556e-05, "loss": 0.0, "num_input_tokens_seen": 1221888, "step": 4365 }, { "epoch": 48.55555555555556, "grad_norm": 6.818886959081283e-06, "learning_rate": 4.8542567905947776e-05, "loss": 0.0, "num_input_tokens_seen": 1223264, "step": 4370 }, { "epoch": 48.611111111111114, "grad_norm": 4.178365543339169e-06, "learning_rate": 4.853926303865618e-05, "loss": 0.0, "num_input_tokens_seen": 1224640, "step": 4375 }, { "epoch": 48.666666666666664, "grad_norm": 1.3555958503275178e-05, "learning_rate": 4.853595454131441e-05, "loss": 0.0, "num_input_tokens_seen": 1226016, "step": 4380 }, { "epoch": 48.72222222222222, "grad_norm": 1.3979371942696162e-05, "learning_rate": 4.8532642414432674e-05, "loss": 0.0, "num_input_tokens_seen": 1227424, "step": 4385 }, { "epoch": 48.77777777777778, "grad_norm": 1.6446938388980925e-05, "learning_rate": 4.8529326658521754e-05, "loss": 0.0, "num_input_tokens_seen": 1228864, "step": 4390 }, { "epoch": 48.833333333333336, "grad_norm": 1.8650312995305285e-05, "learning_rate": 4.8526007274092965e-05, "loss": 0.0, "num_input_tokens_seen": 1230272, "step": 4395 }, { "epoch": 48.888888888888886, "grad_norm": 6.0454494814621285e-05, "learning_rate": 4.852268426165822e-05, "loss": 0.0, "num_input_tokens_seen": 1231696, "step": 4400 }, { "epoch": 48.888888888888886, "eval_loss": 1.007348656654358, "eval_runtime": 1.3852, "eval_samples_per_second": 28.876, "eval_steps_per_second": 14.438, "num_input_tokens_seen": 1231696, "step": 4400 }, { "epoch": 48.94444444444444, "grad_norm": 5.197961854719324e-06, "learning_rate": 4.851935762172995e-05, "loss": 0.0, "num_input_tokens_seen": 1233136, "step": 4405 }, { "epoch": 49.0, "grad_norm": 0.00018619511683937162, "learning_rate": 4.8516027354821175e-05, "loss": 0.0, "num_input_tokens_seen": 1234528, "step": 4410 }, { "epoch": 49.05555555555556, "grad_norm": 0.0001178443999378942, "learning_rate": 4.851269346144546e-05, "loss": 0.0, "num_input_tokens_seen": 1235904, "step": 4415 }, { "epoch": 49.111111111111114, "grad_norm": 2.5855946660158224e-05, "learning_rate": 4.850935594211693e-05, "loss": 0.0, "num_input_tokens_seen": 1237312, "step": 4420 }, { "epoch": 49.166666666666664, "grad_norm": 1.014825284073595e-05, "learning_rate": 4.850601479735029e-05, "loss": 0.0, "num_input_tokens_seen": 1238736, "step": 4425 }, { "epoch": 49.22222222222222, "grad_norm": 1.0987683708663099e-05, "learning_rate": 4.850267002766076e-05, "loss": 0.0, "num_input_tokens_seen": 1240128, "step": 4430 }, { "epoch": 49.27777777777778, "grad_norm": 1.1468091543065384e-05, "learning_rate": 4.849932163356417e-05, "loss": 0.0, "num_input_tokens_seen": 1241568, "step": 4435 }, { "epoch": 49.333333333333336, "grad_norm": 1.716876977297943e-05, "learning_rate": 4.8495969615576864e-05, "loss": 0.0, "num_input_tokens_seen": 1242976, "step": 4440 }, { "epoch": 49.388888888888886, "grad_norm": 5.18307797392481e-06, "learning_rate": 4.849261397421577e-05, "loss": 0.0, "num_input_tokens_seen": 1244384, "step": 4445 }, { "epoch": 49.44444444444444, "grad_norm": 5.7223082876589615e-06, "learning_rate": 4.848925470999839e-05, "loss": 0.0, "num_input_tokens_seen": 1245776, "step": 4450 }, { "epoch": 49.5, "grad_norm": 1.8268212443217635e-05, "learning_rate": 4.848589182344273e-05, "loss": 0.0, "num_input_tokens_seen": 1247200, "step": 4455 }, { "epoch": 49.55555555555556, "grad_norm": 1.8349521269556135e-05, "learning_rate": 4.848252531506742e-05, "loss": 0.0, "num_input_tokens_seen": 1248656, "step": 4460 }, { "epoch": 49.611111111111114, "grad_norm": 6.435978866647929e-05, "learning_rate": 4.847915518539161e-05, "loss": 0.0, "num_input_tokens_seen": 1250048, "step": 4465 }, { "epoch": 49.666666666666664, "grad_norm": 1.3771969861409161e-05, "learning_rate": 4.847578143493501e-05, "loss": 0.0, "num_input_tokens_seen": 1251424, "step": 4470 }, { "epoch": 49.72222222222222, "grad_norm": 4.960507794748992e-06, "learning_rate": 4.847240406421789e-05, "loss": 0.0, "num_input_tokens_seen": 1252816, "step": 4475 }, { "epoch": 49.77777777777778, "grad_norm": 3.983898932347074e-05, "learning_rate": 4.84690230737611e-05, "loss": 0.0, "num_input_tokens_seen": 1254160, "step": 4480 }, { "epoch": 49.833333333333336, "grad_norm": 5.761582542618271e-06, "learning_rate": 4.846563846408602e-05, "loss": 0.0, "num_input_tokens_seen": 1255536, "step": 4485 }, { "epoch": 49.888888888888886, "grad_norm": 1.4549897059623618e-05, "learning_rate": 4.84622502357146e-05, "loss": 0.0, "num_input_tokens_seen": 1256960, "step": 4490 }, { "epoch": 49.94444444444444, "grad_norm": 7.966753764776513e-05, "learning_rate": 4.8458858389169345e-05, "loss": 0.0, "num_input_tokens_seen": 1258352, "step": 4495 }, { "epoch": 50.0, "grad_norm": 5.30976103618741e-06, "learning_rate": 4.8455462924973334e-05, "loss": 0.0, "num_input_tokens_seen": 1259744, "step": 4500 }, { "epoch": 50.05555555555556, "grad_norm": 4.382005499792285e-06, "learning_rate": 4.845206384365018e-05, "loss": 0.0, "num_input_tokens_seen": 1261152, "step": 4505 }, { "epoch": 50.111111111111114, "grad_norm": 1.4258395822253078e-05, "learning_rate": 4.844866114572405e-05, "loss": 0.0, "num_input_tokens_seen": 1262528, "step": 4510 }, { "epoch": 50.166666666666664, "grad_norm": 1.4892532817611936e-05, "learning_rate": 4.8445254831719706e-05, "loss": 0.0, "num_input_tokens_seen": 1263936, "step": 4515 }, { "epoch": 50.22222222222222, "grad_norm": 4.7122644900809973e-05, "learning_rate": 4.8441844902162434e-05, "loss": 0.0, "num_input_tokens_seen": 1265312, "step": 4520 }, { "epoch": 50.27777777777778, "grad_norm": 1.2572809282573871e-05, "learning_rate": 4.843843135757809e-05, "loss": 0.0, "num_input_tokens_seen": 1266720, "step": 4525 }, { "epoch": 50.333333333333336, "grad_norm": 4.356831050245091e-05, "learning_rate": 4.843501419849308e-05, "loss": 0.0, "num_input_tokens_seen": 1268144, "step": 4530 }, { "epoch": 50.388888888888886, "grad_norm": 2.5841203751042485e-05, "learning_rate": 4.8431593425434386e-05, "loss": 0.0, "num_input_tokens_seen": 1269568, "step": 4535 }, { "epoch": 50.44444444444444, "grad_norm": 1.9984281607321464e-05, "learning_rate": 4.8428169038929526e-05, "loss": 0.0, "num_input_tokens_seen": 1270896, "step": 4540 }, { "epoch": 50.5, "grad_norm": 1.7111107808887027e-05, "learning_rate": 4.8424741039506575e-05, "loss": 0.0, "num_input_tokens_seen": 1272304, "step": 4545 }, { "epoch": 50.55555555555556, "grad_norm": 6.1506411839218345e-06, "learning_rate": 4.842130942769419e-05, "loss": 0.0, "num_input_tokens_seen": 1273712, "step": 4550 }, { "epoch": 50.611111111111114, "grad_norm": 5.202456577535486e-06, "learning_rate": 4.841787420402156e-05, "loss": 0.0, "num_input_tokens_seen": 1275120, "step": 4555 }, { "epoch": 50.666666666666664, "grad_norm": 1.4827103768766392e-05, "learning_rate": 4.841443536901844e-05, "loss": 0.0, "num_input_tokens_seen": 1276528, "step": 4560 }, { "epoch": 50.72222222222222, "grad_norm": 5.2057662287552375e-06, "learning_rate": 4.841099292321514e-05, "loss": 0.0, "num_input_tokens_seen": 1277952, "step": 4565 }, { "epoch": 50.77777777777778, "grad_norm": 2.8532185751828365e-05, "learning_rate": 4.8407546867142525e-05, "loss": 0.0, "num_input_tokens_seen": 1279312, "step": 4570 }, { "epoch": 50.833333333333336, "grad_norm": 0.00011204696056665853, "learning_rate": 4.840409720133203e-05, "loss": 0.0, "num_input_tokens_seen": 1280672, "step": 4575 }, { "epoch": 50.888888888888886, "grad_norm": 9.508792572887614e-05, "learning_rate": 4.8400643926315634e-05, "loss": 0.0, "num_input_tokens_seen": 1282096, "step": 4580 }, { "epoch": 50.94444444444444, "grad_norm": 1.2500226148404181e-05, "learning_rate": 4.839718704262587e-05, "loss": 0.0, "num_input_tokens_seen": 1283520, "step": 4585 }, { "epoch": 51.0, "grad_norm": 9.364207653561607e-05, "learning_rate": 4.839372655079585e-05, "loss": 0.0, "num_input_tokens_seen": 1284912, "step": 4590 }, { "epoch": 51.05555555555556, "grad_norm": 5.981763024465181e-06, "learning_rate": 4.83902624513592e-05, "loss": 0.0, "num_input_tokens_seen": 1286320, "step": 4595 }, { "epoch": 51.111111111111114, "grad_norm": 4.0835038817022e-06, "learning_rate": 4.838679474485014e-05, "loss": 0.0, "num_input_tokens_seen": 1287744, "step": 4600 }, { "epoch": 51.111111111111114, "eval_loss": 1.0176149606704712, "eval_runtime": 1.3981, "eval_samples_per_second": 28.61, "eval_steps_per_second": 14.305, "num_input_tokens_seen": 1287744, "step": 4600 }, { "epoch": 51.166666666666664, "grad_norm": 3.2674088288331404e-05, "learning_rate": 4.838332343180343e-05, "loss": 0.0, "num_input_tokens_seen": 1289136, "step": 4605 }, { "epoch": 51.22222222222222, "grad_norm": 5.538525783777004e-06, "learning_rate": 4.83798485127544e-05, "loss": 0.0, "num_input_tokens_seen": 1290544, "step": 4610 }, { "epoch": 51.27777777777778, "grad_norm": 1.4559297596861143e-05, "learning_rate": 4.837636998823892e-05, "loss": 0.0, "num_input_tokens_seen": 1291920, "step": 4615 }, { "epoch": 51.333333333333336, "grad_norm": 1.0417134035378695e-05, "learning_rate": 4.8372887858793414e-05, "loss": 0.0, "num_input_tokens_seen": 1293248, "step": 4620 }, { "epoch": 51.388888888888886, "grad_norm": 1.6450969269499183e-05, "learning_rate": 4.836940212495489e-05, "loss": 0.0, "num_input_tokens_seen": 1294688, "step": 4625 }, { "epoch": 51.44444444444444, "grad_norm": 1.1374606401659548e-05, "learning_rate": 4.836591278726087e-05, "loss": 0.0, "num_input_tokens_seen": 1296128, "step": 4630 }, { "epoch": 51.5, "grad_norm": 1.7986380044021644e-05, "learning_rate": 4.836241984624947e-05, "loss": 0.0, "num_input_tokens_seen": 1297616, "step": 4635 }, { "epoch": 51.55555555555556, "grad_norm": 1.038412119669374e-05, "learning_rate": 4.8358923302459336e-05, "loss": 0.0, "num_input_tokens_seen": 1298992, "step": 4640 }, { "epoch": 51.611111111111114, "grad_norm": 1.1789261407102458e-05, "learning_rate": 4.835542315642968e-05, "loss": 0.0, "num_input_tokens_seen": 1300336, "step": 4645 }, { "epoch": 51.666666666666664, "grad_norm": 1.0956297046504915e-05, "learning_rate": 4.8351919408700274e-05, "loss": 0.0, "num_input_tokens_seen": 1301728, "step": 4650 }, { "epoch": 51.72222222222222, "grad_norm": 7.341713171626907e-06, "learning_rate": 4.834841205981144e-05, "loss": 0.0, "num_input_tokens_seen": 1303120, "step": 4655 }, { "epoch": 51.77777777777778, "grad_norm": 5.465956746775191e-06, "learning_rate": 4.8344901110304054e-05, "loss": 0.0, "num_input_tokens_seen": 1304544, "step": 4660 }, { "epoch": 51.833333333333336, "grad_norm": 1.1542704669409432e-05, "learning_rate": 4.8341386560719534e-05, "loss": 0.0, "num_input_tokens_seen": 1305920, "step": 4665 }, { "epoch": 51.888888888888886, "grad_norm": 4.2669922549976036e-05, "learning_rate": 4.833786841159989e-05, "loss": 0.0, "num_input_tokens_seen": 1307312, "step": 4670 }, { "epoch": 51.94444444444444, "grad_norm": 2.609346847748384e-05, "learning_rate": 4.833434666348765e-05, "loss": 0.0, "num_input_tokens_seen": 1308752, "step": 4675 }, { "epoch": 52.0, "grad_norm": 5.481998414325062e-06, "learning_rate": 4.833082131692592e-05, "loss": 0.0, "num_input_tokens_seen": 1310128, "step": 4680 }, { "epoch": 52.05555555555556, "grad_norm": 1.8156071746489033e-05, "learning_rate": 4.832729237245835e-05, "loss": 0.0, "num_input_tokens_seen": 1311488, "step": 4685 }, { "epoch": 52.111111111111114, "grad_norm": 1.1992542567895725e-05, "learning_rate": 4.8323759830629145e-05, "loss": 0.0, "num_input_tokens_seen": 1312896, "step": 4690 }, { "epoch": 52.166666666666664, "grad_norm": 1.5423061995534226e-05, "learning_rate": 4.8320223691983066e-05, "loss": 0.0, "num_input_tokens_seen": 1314288, "step": 4695 }, { "epoch": 52.22222222222222, "grad_norm": 3.198564445483498e-05, "learning_rate": 4.831668395706544e-05, "loss": 0.0, "num_input_tokens_seen": 1315696, "step": 4700 }, { "epoch": 52.27777777777778, "grad_norm": 0.00017776447930373251, "learning_rate": 4.8313140626422125e-05, "loss": 0.0, "num_input_tokens_seen": 1317120, "step": 4705 }, { "epoch": 52.333333333333336, "grad_norm": 2.8429290978237987e-05, "learning_rate": 4.830959370059956e-05, "loss": 0.0, "num_input_tokens_seen": 1318528, "step": 4710 }, { "epoch": 52.388888888888886, "grad_norm": 1.182051892101299e-05, "learning_rate": 4.830604318014472e-05, "loss": 0.0, "num_input_tokens_seen": 1319904, "step": 4715 }, { "epoch": 52.44444444444444, "grad_norm": 4.553983217192581e-06, "learning_rate": 4.830248906560514e-05, "loss": 0.0, "num_input_tokens_seen": 1321312, "step": 4720 }, { "epoch": 52.5, "grad_norm": 5.3466585086425766e-05, "learning_rate": 4.829893135752891e-05, "loss": 0.0, "num_input_tokens_seen": 1322784, "step": 4725 }, { "epoch": 52.55555555555556, "grad_norm": 3.065295095439069e-05, "learning_rate": 4.829537005646466e-05, "loss": 0.0, "num_input_tokens_seen": 1324160, "step": 4730 }, { "epoch": 52.611111111111114, "grad_norm": 1.8736796846496873e-05, "learning_rate": 4.8291805162961615e-05, "loss": 0.0, "num_input_tokens_seen": 1325536, "step": 4735 }, { "epoch": 52.666666666666664, "grad_norm": 4.693441496783635e-06, "learning_rate": 4.82882366775695e-05, "loss": 0.0, "num_input_tokens_seen": 1326880, "step": 4740 }, { "epoch": 52.72222222222222, "grad_norm": 5.649164813803509e-05, "learning_rate": 4.828466460083864e-05, "loss": 0.0, "num_input_tokens_seen": 1328256, "step": 4745 }, { "epoch": 52.77777777777778, "grad_norm": 8.735125447856262e-06, "learning_rate": 4.8281088933319877e-05, "loss": 0.0, "num_input_tokens_seen": 1329648, "step": 4750 }, { "epoch": 52.833333333333336, "grad_norm": 1.4429778275371064e-05, "learning_rate": 4.827750967556464e-05, "loss": 0.0, "num_input_tokens_seen": 1331056, "step": 4755 }, { "epoch": 52.888888888888886, "grad_norm": 3.0111436899460386e-06, "learning_rate": 4.827392682812488e-05, "loss": 0.0, "num_input_tokens_seen": 1332464, "step": 4760 }, { "epoch": 52.94444444444444, "grad_norm": 4.1599696487537585e-06, "learning_rate": 4.827034039155312e-05, "loss": 0.0, "num_input_tokens_seen": 1333856, "step": 4765 }, { "epoch": 53.0, "grad_norm": 1.1050534340029117e-05, "learning_rate": 4.8266750366402445e-05, "loss": 0.0, "num_input_tokens_seen": 1335264, "step": 4770 }, { "epoch": 53.05555555555556, "grad_norm": 8.21605499368161e-06, "learning_rate": 4.8263156753226476e-05, "loss": 0.0, "num_input_tokens_seen": 1336672, "step": 4775 }, { "epoch": 53.111111111111114, "grad_norm": 8.359993444173597e-06, "learning_rate": 4.8259559552579394e-05, "loss": 0.0, "num_input_tokens_seen": 1338048, "step": 4780 }, { "epoch": 53.166666666666664, "grad_norm": 6.17092737229541e-05, "learning_rate": 4.825595876501593e-05, "loss": 0.0, "num_input_tokens_seen": 1339488, "step": 4785 }, { "epoch": 53.22222222222222, "grad_norm": 1.4665434719063342e-05, "learning_rate": 4.825235439109137e-05, "loss": 0.0, "num_input_tokens_seen": 1340912, "step": 4790 }, { "epoch": 53.27777777777778, "grad_norm": 2.4754499463597313e-06, "learning_rate": 4.824874643136156e-05, "loss": 0.0, "num_input_tokens_seen": 1342352, "step": 4795 }, { "epoch": 53.333333333333336, "grad_norm": 7.844178981031291e-06, "learning_rate": 4.824513488638288e-05, "loss": 0.0, "num_input_tokens_seen": 1343760, "step": 4800 }, { "epoch": 53.333333333333336, "eval_loss": 1.0251001119613647, "eval_runtime": 1.3845, "eval_samples_per_second": 28.892, "eval_steps_per_second": 14.446, "num_input_tokens_seen": 1343760, "step": 4800 }, { "epoch": 53.388888888888886, "grad_norm": 3.2669461234036135e-06, "learning_rate": 4.8241519756712293e-05, "loss": 0.0, "num_input_tokens_seen": 1345152, "step": 4805 }, { "epoch": 53.44444444444444, "grad_norm": 5.577478077611886e-06, "learning_rate": 4.8237901042907285e-05, "loss": 0.0, "num_input_tokens_seen": 1346576, "step": 4810 }, { "epoch": 53.5, "grad_norm": 7.81426570028998e-06, "learning_rate": 4.823427874552591e-05, "loss": 0.0, "num_input_tokens_seen": 1348016, "step": 4815 }, { "epoch": 53.55555555555556, "grad_norm": 1.4095667211222462e-05, "learning_rate": 4.823065286512677e-05, "loss": 0.0, "num_input_tokens_seen": 1349408, "step": 4820 }, { "epoch": 53.611111111111114, "grad_norm": 1.1132587133033667e-05, "learning_rate": 4.8227023402269025e-05, "loss": 0.0, "num_input_tokens_seen": 1350816, "step": 4825 }, { "epoch": 53.666666666666664, "grad_norm": 1.277211049455218e-05, "learning_rate": 4.822339035751239e-05, "loss": 0.0, "num_input_tokens_seen": 1352176, "step": 4830 }, { "epoch": 53.72222222222222, "grad_norm": 2.0693749320344068e-05, "learning_rate": 4.8219753731417104e-05, "loss": 0.0, "num_input_tokens_seen": 1353552, "step": 4835 }, { "epoch": 53.77777777777778, "grad_norm": 2.4214172299252823e-05, "learning_rate": 4.821611352454401e-05, "loss": 0.0, "num_input_tokens_seen": 1354928, "step": 4840 }, { "epoch": 53.833333333333336, "grad_norm": 7.487408765882719e-06, "learning_rate": 4.8212469737454444e-05, "loss": 0.0, "num_input_tokens_seen": 1356336, "step": 4845 }, { "epoch": 53.888888888888886, "grad_norm": 1.0178526281379163e-05, "learning_rate": 4.820882237071035e-05, "loss": 0.0, "num_input_tokens_seen": 1357776, "step": 4850 }, { "epoch": 53.94444444444444, "grad_norm": 1.578985757078044e-05, "learning_rate": 4.820517142487417e-05, "loss": 0.0, "num_input_tokens_seen": 1359184, "step": 4855 }, { "epoch": 54.0, "grad_norm": 5.953730124019785e-06, "learning_rate": 4.8201516900508956e-05, "loss": 0.0, "num_input_tokens_seen": 1360592, "step": 4860 }, { "epoch": 54.05555555555556, "grad_norm": 4.021707718493417e-05, "learning_rate": 4.819785879817827e-05, "loss": 0.0, "num_input_tokens_seen": 1361936, "step": 4865 }, { "epoch": 54.111111111111114, "grad_norm": 9.980745744542219e-06, "learning_rate": 4.8194197118446226e-05, "loss": 0.0, "num_input_tokens_seen": 1363360, "step": 4870 }, { "epoch": 54.166666666666664, "grad_norm": 1.1941519005631562e-05, "learning_rate": 4.819053186187752e-05, "loss": 0.0, "num_input_tokens_seen": 1364784, "step": 4875 }, { "epoch": 54.22222222222222, "grad_norm": 1.9484035874484107e-05, "learning_rate": 4.818686302903736e-05, "loss": 0.0, "num_input_tokens_seen": 1366176, "step": 4880 }, { "epoch": 54.27777777777778, "grad_norm": 1.7350692360196263e-05, "learning_rate": 4.818319062049154e-05, "loss": 0.0, "num_input_tokens_seen": 1367568, "step": 4885 }, { "epoch": 54.333333333333336, "grad_norm": 2.969183924506069e-06, "learning_rate": 4.817951463680639e-05, "loss": 0.0, "num_input_tokens_seen": 1368944, "step": 4890 }, { "epoch": 54.388888888888886, "grad_norm": 6.468000719905831e-06, "learning_rate": 4.817583507854879e-05, "loss": 0.0, "num_input_tokens_seen": 1370352, "step": 4895 }, { "epoch": 54.44444444444444, "grad_norm": 5.804417924082372e-06, "learning_rate": 4.817215194628617e-05, "loss": 0.0, "num_input_tokens_seen": 1371760, "step": 4900 }, { "epoch": 54.5, "grad_norm": 5.089932892587967e-05, "learning_rate": 4.816846524058653e-05, "loss": 0.0, "num_input_tokens_seen": 1373168, "step": 4905 }, { "epoch": 54.55555555555556, "grad_norm": 3.218695928808302e-05, "learning_rate": 4.816477496201839e-05, "loss": 0.0, "num_input_tokens_seen": 1374544, "step": 4910 }, { "epoch": 54.611111111111114, "grad_norm": 1.3484113878803328e-05, "learning_rate": 4.8161081111150845e-05, "loss": 0.0, "num_input_tokens_seen": 1375936, "step": 4915 }, { "epoch": 54.666666666666664, "grad_norm": 5.188477189221885e-06, "learning_rate": 4.815738368855354e-05, "loss": 0.0, "num_input_tokens_seen": 1377344, "step": 4920 }, { "epoch": 54.72222222222222, "grad_norm": 1.1819650353572797e-05, "learning_rate": 4.815368269479664e-05, "loss": 0.0, "num_input_tokens_seen": 1378752, "step": 4925 }, { "epoch": 54.77777777777778, "grad_norm": 7.973812898853794e-06, "learning_rate": 4.814997813045092e-05, "loss": 0.0, "num_input_tokens_seen": 1380112, "step": 4930 }, { "epoch": 54.833333333333336, "grad_norm": 7.204032499430468e-06, "learning_rate": 4.814626999608764e-05, "loss": 0.0, "num_input_tokens_seen": 1381536, "step": 4935 }, { "epoch": 54.888888888888886, "grad_norm": 3.4876341032941127e-06, "learning_rate": 4.814255829227865e-05, "loss": 0.0, "num_input_tokens_seen": 1382944, "step": 4940 }, { "epoch": 54.94444444444444, "grad_norm": 6.28023553872481e-06, "learning_rate": 4.813884301959635e-05, "loss": 0.0, "num_input_tokens_seen": 1384384, "step": 4945 }, { "epoch": 55.0, "grad_norm": 9.030189539771527e-05, "learning_rate": 4.813512417861368e-05, "loss": 0.0, "num_input_tokens_seen": 1385760, "step": 4950 }, { "epoch": 55.05555555555556, "grad_norm": 8.253332453023177e-06, "learning_rate": 4.813140176990411e-05, "loss": 0.0, "num_input_tokens_seen": 1387136, "step": 4955 }, { "epoch": 55.111111111111114, "grad_norm": 8.716631782590412e-06, "learning_rate": 4.8127675794041714e-05, "loss": 0.0, "num_input_tokens_seen": 1388560, "step": 4960 }, { "epoch": 55.166666666666664, "grad_norm": 4.8896986299951095e-06, "learning_rate": 4.812394625160107e-05, "loss": 0.0, "num_input_tokens_seen": 1389952, "step": 4965 }, { "epoch": 55.22222222222222, "grad_norm": 7.293286671483656e-06, "learning_rate": 4.812021314315732e-05, "loss": 0.0, "num_input_tokens_seen": 1391344, "step": 4970 }, { "epoch": 55.27777777777778, "grad_norm": 5.303075158735737e-05, "learning_rate": 4.811647646928616e-05, "loss": 0.0, "num_input_tokens_seen": 1392752, "step": 4975 }, { "epoch": 55.333333333333336, "grad_norm": 4.755154805025086e-06, "learning_rate": 4.8112736230563814e-05, "loss": 0.0, "num_input_tokens_seen": 1394112, "step": 4980 }, { "epoch": 55.388888888888886, "grad_norm": 1.018365492200246e-05, "learning_rate": 4.81089924275671e-05, "loss": 0.0, "num_input_tokens_seen": 1395520, "step": 4985 }, { "epoch": 55.44444444444444, "grad_norm": 8.150958456099033e-05, "learning_rate": 4.810524506087335e-05, "loss": 0.0, "num_input_tokens_seen": 1396960, "step": 4990 }, { "epoch": 55.5, "grad_norm": 6.413825303752674e-06, "learning_rate": 4.810149413106044e-05, "loss": 0.0, "num_input_tokens_seen": 1398416, "step": 4995 }, { "epoch": 55.55555555555556, "grad_norm": 8.637874088890385e-06, "learning_rate": 4.809773963870684e-05, "loss": 0.0, "num_input_tokens_seen": 1399856, "step": 5000 }, { "epoch": 55.55555555555556, "eval_loss": 1.043311357498169, "eval_runtime": 1.3862, "eval_samples_per_second": 28.855, "eval_steps_per_second": 14.428, "num_input_tokens_seen": 1399856, "step": 5000 }, { "epoch": 55.611111111111114, "grad_norm": 1.0213449968432542e-05, "learning_rate": 4.809398158439151e-05, "loss": 0.0, "num_input_tokens_seen": 1401264, "step": 5005 }, { "epoch": 55.666666666666664, "grad_norm": 1.0708641639212146e-05, "learning_rate": 4.8090219968694005e-05, "loss": 0.0, "num_input_tokens_seen": 1402656, "step": 5010 }, { "epoch": 55.72222222222222, "grad_norm": 5.052780579717364e-06, "learning_rate": 4.808645479219442e-05, "loss": 0.0, "num_input_tokens_seen": 1404032, "step": 5015 }, { "epoch": 55.77777777777778, "grad_norm": 6.3332995523524005e-06, "learning_rate": 4.8082686055473375e-05, "loss": 0.0, "num_input_tokens_seen": 1405424, "step": 5020 }, { "epoch": 55.833333333333336, "grad_norm": 4.942423402098939e-06, "learning_rate": 4.8078913759112066e-05, "loss": 0.0, "num_input_tokens_seen": 1406800, "step": 5025 }, { "epoch": 55.888888888888886, "grad_norm": 1.1747948519769125e-05, "learning_rate": 4.807513790369223e-05, "loss": 0.0, "num_input_tokens_seen": 1408192, "step": 5030 }, { "epoch": 55.94444444444444, "grad_norm": 4.929333954351023e-06, "learning_rate": 4.8071358489796145e-05, "loss": 0.0, "num_input_tokens_seen": 1409552, "step": 5035 }, { "epoch": 56.0, "grad_norm": 4.609660209098365e-06, "learning_rate": 4.806757551800665e-05, "loss": 0.0, "num_input_tokens_seen": 1410992, "step": 5040 }, { "epoch": 56.05555555555556, "grad_norm": 1.079376397683518e-05, "learning_rate": 4.806378898890713e-05, "loss": 0.0, "num_input_tokens_seen": 1412400, "step": 5045 }, { "epoch": 56.111111111111114, "grad_norm": 5.855455128767062e-06, "learning_rate": 4.80599989030815e-05, "loss": 0.0, "num_input_tokens_seen": 1413792, "step": 5050 }, { "epoch": 56.166666666666664, "grad_norm": 2.444529309286736e-05, "learning_rate": 4.805620526111426e-05, "loss": 0.0, "num_input_tokens_seen": 1415200, "step": 5055 }, { "epoch": 56.22222222222222, "grad_norm": 3.8359557947842404e-06, "learning_rate": 4.805240806359042e-05, "loss": 0.0, "num_input_tokens_seen": 1416608, "step": 5060 }, { "epoch": 56.27777777777778, "grad_norm": 6.09436892773374e-06, "learning_rate": 4.804860731109557e-05, "loss": 0.0, "num_input_tokens_seen": 1418000, "step": 5065 }, { "epoch": 56.333333333333336, "grad_norm": 4.326214821048779e-06, "learning_rate": 4.804480300421581e-05, "loss": 0.0, "num_input_tokens_seen": 1419376, "step": 5070 }, { "epoch": 56.388888888888886, "grad_norm": 6.6548427639645524e-06, "learning_rate": 4.804099514353784e-05, "loss": 0.0, "num_input_tokens_seen": 1420752, "step": 5075 }, { "epoch": 56.44444444444444, "grad_norm": 4.080721737409476e-06, "learning_rate": 4.8037183729648867e-05, "loss": 0.0, "num_input_tokens_seen": 1422112, "step": 5080 }, { "epoch": 56.5, "grad_norm": 1.864629848569166e-05, "learning_rate": 4.803336876313666e-05, "loss": 0.0, "num_input_tokens_seen": 1423536, "step": 5085 }, { "epoch": 56.55555555555556, "grad_norm": 4.546452601061901e-06, "learning_rate": 4.802955024458953e-05, "loss": 0.0, "num_input_tokens_seen": 1424944, "step": 5090 }, { "epoch": 56.611111111111114, "grad_norm": 3.210683189536212e-06, "learning_rate": 4.802572817459634e-05, "loss": 0.0, "num_input_tokens_seen": 1426416, "step": 5095 }, { "epoch": 56.666666666666664, "grad_norm": 5.693486400559777e-06, "learning_rate": 4.802190255374651e-05, "loss": 0.0, "num_input_tokens_seen": 1427856, "step": 5100 }, { "epoch": 56.72222222222222, "grad_norm": 3.850918346870458e-06, "learning_rate": 4.801807338263e-05, "loss": 0.0, "num_input_tokens_seen": 1429168, "step": 5105 }, { "epoch": 56.77777777777778, "grad_norm": 5.2945037168683484e-05, "learning_rate": 4.8014240661837306e-05, "loss": 0.0, "num_input_tokens_seen": 1430576, "step": 5110 }, { "epoch": 56.833333333333336, "grad_norm": 9.245031833415851e-06, "learning_rate": 4.80104043919595e-05, "loss": 0.0, "num_input_tokens_seen": 1431952, "step": 5115 }, { "epoch": 56.888888888888886, "grad_norm": 1.1965943485847674e-05, "learning_rate": 4.800656457358815e-05, "loss": 0.0, "num_input_tokens_seen": 1433344, "step": 5120 }, { "epoch": 56.94444444444444, "grad_norm": 1.3144856893632095e-05, "learning_rate": 4.800272120731544e-05, "loss": 0.0, "num_input_tokens_seen": 1434752, "step": 5125 }, { "epoch": 57.0, "grad_norm": 1.895770651572093e-06, "learning_rate": 4.799887429373404e-05, "loss": 0.0, "num_input_tokens_seen": 1436160, "step": 5130 }, { "epoch": 57.05555555555556, "grad_norm": 7.94053812569473e-06, "learning_rate": 4.79950238334372e-05, "loss": 0.0, "num_input_tokens_seen": 1437600, "step": 5135 }, { "epoch": 57.111111111111114, "grad_norm": 1.6072805010480806e-05, "learning_rate": 4.799116982701872e-05, "loss": 0.0, "num_input_tokens_seen": 1439024, "step": 5140 }, { "epoch": 57.166666666666664, "grad_norm": 6.066579317121068e-06, "learning_rate": 4.7987312275072926e-05, "loss": 0.0, "num_input_tokens_seen": 1440432, "step": 5145 }, { "epoch": 57.22222222222222, "grad_norm": 8.825619261187967e-06, "learning_rate": 4.79834511781947e-05, "loss": 0.0, "num_input_tokens_seen": 1441824, "step": 5150 }, { "epoch": 57.27777777777778, "grad_norm": 4.285803697712254e-06, "learning_rate": 4.797958653697947e-05, "loss": 0.0, "num_input_tokens_seen": 1443200, "step": 5155 }, { "epoch": 57.333333333333336, "grad_norm": 2.8492120691225864e-05, "learning_rate": 4.7975718352023225e-05, "loss": 0.0, "num_input_tokens_seen": 1444592, "step": 5160 }, { "epoch": 57.388888888888886, "grad_norm": 2.1154490241315216e-05, "learning_rate": 4.7971846623922476e-05, "loss": 0.0, "num_input_tokens_seen": 1446016, "step": 5165 }, { "epoch": 57.44444444444444, "grad_norm": 3.2315656426362693e-06, "learning_rate": 4.7967971353274294e-05, "loss": 0.0, "num_input_tokens_seen": 1447408, "step": 5170 }, { "epoch": 57.5, "grad_norm": 1.9088851331616752e-05, "learning_rate": 4.79640925406763e-05, "loss": 0.0, "num_input_tokens_seen": 1448800, "step": 5175 }, { "epoch": 57.55555555555556, "grad_norm": 4.934441221848829e-06, "learning_rate": 4.796021018672664e-05, "loss": 0.0, "num_input_tokens_seen": 1450176, "step": 5180 }, { "epoch": 57.611111111111114, "grad_norm": 3.291177563369274e-06, "learning_rate": 4.795632429202405e-05, "loss": 0.0, "num_input_tokens_seen": 1451568, "step": 5185 }, { "epoch": 57.666666666666664, "grad_norm": 3.363099722264451e-06, "learning_rate": 4.795243485716775e-05, "loss": 0.0, "num_input_tokens_seen": 1452992, "step": 5190 }, { "epoch": 57.72222222222222, "grad_norm": 1.1068165804317687e-05, "learning_rate": 4.794854188275757e-05, "loss": 0.0, "num_input_tokens_seen": 1454416, "step": 5195 }, { "epoch": 57.77777777777778, "grad_norm": 2.3822940420359373e-05, "learning_rate": 4.794464536939384e-05, "loss": 0.0, "num_input_tokens_seen": 1455808, "step": 5200 }, { "epoch": 57.77777777777778, "eval_loss": 1.0433063507080078, "eval_runtime": 1.3836, "eval_samples_per_second": 28.911, "eval_steps_per_second": 14.455, "num_input_tokens_seen": 1455808, "step": 5200 }, { "epoch": 57.833333333333336, "grad_norm": 1.0574087355053052e-05, "learning_rate": 4.794074531767745e-05, "loss": 0.0, "num_input_tokens_seen": 1457200, "step": 5205 }, { "epoch": 57.888888888888886, "grad_norm": 5.570050689129857e-06, "learning_rate": 4.7936841728209834e-05, "loss": 0.0, "num_input_tokens_seen": 1458560, "step": 5210 }, { "epoch": 57.94444444444444, "grad_norm": 2.175614281441085e-05, "learning_rate": 4.7932934601593e-05, "loss": 0.0, "num_input_tokens_seen": 1460000, "step": 5215 }, { "epoch": 58.0, "grad_norm": 9.88949068414513e-06, "learning_rate": 4.792902393842943e-05, "loss": 0.0, "num_input_tokens_seen": 1461440, "step": 5220 }, { "epoch": 58.05555555555556, "grad_norm": 5.498030441231094e-05, "learning_rate": 4.792510973932225e-05, "loss": 0.0, "num_input_tokens_seen": 1462800, "step": 5225 }, { "epoch": 58.111111111111114, "grad_norm": 5.568079814111115e-06, "learning_rate": 4.7921192004875036e-05, "loss": 0.0, "num_input_tokens_seen": 1464192, "step": 5230 }, { "epoch": 58.166666666666664, "grad_norm": 7.03472596796928e-06, "learning_rate": 4.791727073569198e-05, "loss": 0.0, "num_input_tokens_seen": 1465568, "step": 5235 }, { "epoch": 58.22222222222222, "grad_norm": 3.4197608329122886e-05, "learning_rate": 4.7913345932377775e-05, "loss": 0.0, "num_input_tokens_seen": 1466976, "step": 5240 }, { "epoch": 58.27777777777778, "grad_norm": 1.7365637177135795e-05, "learning_rate": 4.790941759553769e-05, "loss": 0.0, "num_input_tokens_seen": 1468368, "step": 5245 }, { "epoch": 58.333333333333336, "grad_norm": 1.0918750376731623e-05, "learning_rate": 4.79054857257775e-05, "loss": 0.0, "num_input_tokens_seen": 1469760, "step": 5250 }, { "epoch": 58.388888888888886, "grad_norm": 3.3933565646293573e-06, "learning_rate": 4.790155032370357e-05, "loss": 0.0, "num_input_tokens_seen": 1471152, "step": 5255 }, { "epoch": 58.44444444444444, "grad_norm": 4.89747571918997e-06, "learning_rate": 4.789761138992278e-05, "loss": 0.0, "num_input_tokens_seen": 1472608, "step": 5260 }, { "epoch": 58.5, "grad_norm": 5.7486186051392e-06, "learning_rate": 4.7893668925042565e-05, "loss": 0.0, "num_input_tokens_seen": 1474016, "step": 5265 }, { "epoch": 58.55555555555556, "grad_norm": 2.6278230507159606e-05, "learning_rate": 4.78897229296709e-05, "loss": 0.0, "num_input_tokens_seen": 1475424, "step": 5270 }, { "epoch": 58.611111111111114, "grad_norm": 1.5024357935544685e-06, "learning_rate": 4.7885773404416315e-05, "loss": 0.0, "num_input_tokens_seen": 1476832, "step": 5275 }, { "epoch": 58.666666666666664, "grad_norm": 8.270330908999313e-06, "learning_rate": 4.788182034988786e-05, "loss": 0.0, "num_input_tokens_seen": 1478240, "step": 5280 }, { "epoch": 58.72222222222222, "grad_norm": 5.387049441196723e-06, "learning_rate": 4.787786376669516e-05, "loss": 0.0, "num_input_tokens_seen": 1479648, "step": 5285 }, { "epoch": 58.77777777777778, "grad_norm": 1.632014573260676e-05, "learning_rate": 4.787390365544837e-05, "loss": 0.0, "num_input_tokens_seen": 1481088, "step": 5290 }, { "epoch": 58.833333333333336, "grad_norm": 9.949586819857359e-05, "learning_rate": 4.786994001675818e-05, "loss": 0.0, "num_input_tokens_seen": 1482512, "step": 5295 }, { "epoch": 58.888888888888886, "grad_norm": 5.415655778051587e-06, "learning_rate": 4.786597285123584e-05, "loss": 0.0, "num_input_tokens_seen": 1483888, "step": 5300 }, { "epoch": 58.94444444444444, "grad_norm": 1.0206645129073877e-05, "learning_rate": 4.7862002159493135e-05, "loss": 0.0, "num_input_tokens_seen": 1485280, "step": 5305 }, { "epoch": 59.0, "grad_norm": 1.821257137635257e-05, "learning_rate": 4.785802794214239e-05, "loss": 0.0, "num_input_tokens_seen": 1486688, "step": 5310 }, { "epoch": 59.05555555555556, "grad_norm": 2.4370197934331372e-05, "learning_rate": 4.7854050199796495e-05, "loss": 0.0, "num_input_tokens_seen": 1488080, "step": 5315 }, { "epoch": 59.111111111111114, "grad_norm": 2.2135320705274353e-06, "learning_rate": 4.7850068933068845e-05, "loss": 0.0, "num_input_tokens_seen": 1489472, "step": 5320 }, { "epoch": 59.166666666666664, "grad_norm": 2.705543920455966e-06, "learning_rate": 4.7846084142573425e-05, "loss": 0.0, "num_input_tokens_seen": 1490880, "step": 5325 }, { "epoch": 59.22222222222222, "grad_norm": 8.46155162435025e-05, "learning_rate": 4.7842095828924725e-05, "loss": 0.0, "num_input_tokens_seen": 1492240, "step": 5330 }, { "epoch": 59.27777777777778, "grad_norm": 4.349918526713736e-06, "learning_rate": 4.783810399273779e-05, "loss": 0.0, "num_input_tokens_seen": 1493664, "step": 5335 }, { "epoch": 59.333333333333336, "grad_norm": 3.3977087241510162e-06, "learning_rate": 4.7834108634628226e-05, "loss": 0.0, "num_input_tokens_seen": 1495072, "step": 5340 }, { "epoch": 59.388888888888886, "grad_norm": 8.023366717679892e-06, "learning_rate": 4.783010975521216e-05, "loss": 0.0, "num_input_tokens_seen": 1496512, "step": 5345 }, { "epoch": 59.44444444444444, "grad_norm": 5.033320121583529e-06, "learning_rate": 4.782610735510626e-05, "loss": 0.0, "num_input_tokens_seen": 1497904, "step": 5350 }, { "epoch": 59.5, "grad_norm": 2.820967665684293e-06, "learning_rate": 4.782210143492776e-05, "loss": 0.0, "num_input_tokens_seen": 1499280, "step": 5355 }, { "epoch": 59.55555555555556, "grad_norm": 2.5864396775432397e-06, "learning_rate": 4.781809199529442e-05, "loss": 0.0, "num_input_tokens_seen": 1500688, "step": 5360 }, { "epoch": 59.611111111111114, "grad_norm": 6.992476755840471e-06, "learning_rate": 4.781407903682454e-05, "loss": 0.0, "num_input_tokens_seen": 1502112, "step": 5365 }, { "epoch": 59.666666666666664, "grad_norm": 4.7915887989802286e-05, "learning_rate": 4.781006256013698e-05, "loss": 0.0, "num_input_tokens_seen": 1503520, "step": 5370 }, { "epoch": 59.72222222222222, "grad_norm": 8.634563528175931e-06, "learning_rate": 4.7806042565851115e-05, "loss": 0.0, "num_input_tokens_seen": 1504928, "step": 5375 }, { "epoch": 59.77777777777778, "grad_norm": 5.478571893036133e-06, "learning_rate": 4.7802019054586895e-05, "loss": 0.0, "num_input_tokens_seen": 1506288, "step": 5380 }, { "epoch": 59.833333333333336, "grad_norm": 3.0779548978898674e-05, "learning_rate": 4.779799202696479e-05, "loss": 0.0, "num_input_tokens_seen": 1507712, "step": 5385 }, { "epoch": 59.888888888888886, "grad_norm": 4.490403625823092e-06, "learning_rate": 4.779396148360581e-05, "loss": 0.0, "num_input_tokens_seen": 1509088, "step": 5390 }, { "epoch": 59.94444444444444, "grad_norm": 0.0001677871769061312, "learning_rate": 4.7789927425131517e-05, "loss": 0.0, "num_input_tokens_seen": 1510464, "step": 5395 }, { "epoch": 60.0, "grad_norm": 5.475509624375263e-06, "learning_rate": 4.778588985216403e-05, "loss": 0.0, "num_input_tokens_seen": 1511856, "step": 5400 }, { "epoch": 60.0, "eval_loss": 1.0579884052276611, "eval_runtime": 1.3783, "eval_samples_per_second": 29.021, "eval_steps_per_second": 14.511, "num_input_tokens_seen": 1511856, "step": 5400 }, { "epoch": 60.05555555555556, "grad_norm": 5.124403742229333e-06, "learning_rate": 4.778184876532598e-05, "loss": 0.0, "num_input_tokens_seen": 1513280, "step": 5405 }, { "epoch": 60.111111111111114, "grad_norm": 7.829033711459488e-05, "learning_rate": 4.7777804165240556e-05, "loss": 0.0, "num_input_tokens_seen": 1514656, "step": 5410 }, { "epoch": 60.166666666666664, "grad_norm": 7.003494374657748e-06, "learning_rate": 4.7773756052531485e-05, "loss": 0.0, "num_input_tokens_seen": 1516064, "step": 5415 }, { "epoch": 60.22222222222222, "grad_norm": 8.172506568371318e-06, "learning_rate": 4.7769704427823035e-05, "loss": 0.0, "num_input_tokens_seen": 1517456, "step": 5420 }, { "epoch": 60.27777777777778, "grad_norm": 3.075215863646008e-05, "learning_rate": 4.776564929174003e-05, "loss": 0.0, "num_input_tokens_seen": 1518832, "step": 5425 }, { "epoch": 60.333333333333336, "grad_norm": 4.0630447983858176e-06, "learning_rate": 4.7761590644907806e-05, "loss": 0.0, "num_input_tokens_seen": 1520256, "step": 5430 }, { "epoch": 60.388888888888886, "grad_norm": 1.7951226254808716e-05, "learning_rate": 4.7757528487952263e-05, "loss": 0.0, "num_input_tokens_seen": 1521632, "step": 5435 }, { "epoch": 60.44444444444444, "grad_norm": 2.108851731463801e-05, "learning_rate": 4.7753462821499836e-05, "loss": 0.0, "num_input_tokens_seen": 1523056, "step": 5440 }, { "epoch": 60.5, "grad_norm": 9.038638381753117e-06, "learning_rate": 4.774939364617751e-05, "loss": 0.0, "num_input_tokens_seen": 1524480, "step": 5445 }, { "epoch": 60.55555555555556, "grad_norm": 4.356722274678759e-05, "learning_rate": 4.7745320962612795e-05, "loss": 0.0, "num_input_tokens_seen": 1525920, "step": 5450 }, { "epoch": 60.611111111111114, "grad_norm": 8.372195225092582e-06, "learning_rate": 4.7741244771433756e-05, "loss": 0.0, "num_input_tokens_seen": 1527344, "step": 5455 }, { "epoch": 60.666666666666664, "grad_norm": 3.8088392102508806e-06, "learning_rate": 4.7737165073268985e-05, "loss": 0.0, "num_input_tokens_seen": 1528704, "step": 5460 }, { "epoch": 60.72222222222222, "grad_norm": 5.508929916686611e-06, "learning_rate": 4.7733081868747626e-05, "loss": 0.0, "num_input_tokens_seen": 1530080, "step": 5465 }, { "epoch": 60.77777777777778, "grad_norm": 1.856148628576193e-05, "learning_rate": 4.772899515849936e-05, "loss": 0.0, "num_input_tokens_seen": 1531440, "step": 5470 }, { "epoch": 60.833333333333336, "grad_norm": 2.7741743906517513e-05, "learning_rate": 4.7724904943154414e-05, "loss": 0.0, "num_input_tokens_seen": 1532832, "step": 5475 }, { "epoch": 60.888888888888886, "grad_norm": 9.132407285505906e-06, "learning_rate": 4.772081122334354e-05, "loss": 0.0, "num_input_tokens_seen": 1534224, "step": 5480 }, { "epoch": 60.94444444444444, "grad_norm": 1.694737875368446e-05, "learning_rate": 4.771671399969806e-05, "loss": 0.0, "num_input_tokens_seen": 1535632, "step": 5485 }, { "epoch": 61.0, "grad_norm": 5.670963673765073e-06, "learning_rate": 4.7712613272849794e-05, "loss": 0.0, "num_input_tokens_seen": 1537008, "step": 5490 }, { "epoch": 61.05555555555556, "grad_norm": 7.967204510350712e-06, "learning_rate": 4.770850904343114e-05, "loss": 0.0, "num_input_tokens_seen": 1538432, "step": 5495 }, { "epoch": 61.111111111111114, "grad_norm": 8.070207513810601e-06, "learning_rate": 4.770440131207502e-05, "loss": 0.0, "num_input_tokens_seen": 1539840, "step": 5500 }, { "epoch": 61.166666666666664, "grad_norm": 4.166251983406255e-06, "learning_rate": 4.7700290079414896e-05, "loss": 0.0, "num_input_tokens_seen": 1541248, "step": 5505 }, { "epoch": 61.22222222222222, "grad_norm": 3.398940862098243e-06, "learning_rate": 4.769617534608477e-05, "loss": 0.0, "num_input_tokens_seen": 1542688, "step": 5510 }, { "epoch": 61.27777777777778, "grad_norm": 9.693119864095934e-06, "learning_rate": 4.7692057112719193e-05, "loss": 0.0, "num_input_tokens_seen": 1544128, "step": 5515 }, { "epoch": 61.333333333333336, "grad_norm": 7.5473394645086955e-06, "learning_rate": 4.7687935379953234e-05, "loss": 0.0, "num_input_tokens_seen": 1545504, "step": 5520 }, { "epoch": 61.388888888888886, "grad_norm": 7.4785239121410996e-06, "learning_rate": 4.7683810148422534e-05, "loss": 0.0, "num_input_tokens_seen": 1546864, "step": 5525 }, { "epoch": 61.44444444444444, "grad_norm": 8.977417564892676e-06, "learning_rate": 4.767968141876324e-05, "loss": 0.0, "num_input_tokens_seen": 1548272, "step": 5530 }, { "epoch": 61.5, "grad_norm": 2.215911763414624e-06, "learning_rate": 4.767554919161207e-05, "loss": 0.0, "num_input_tokens_seen": 1549648, "step": 5535 }, { "epoch": 61.55555555555556, "grad_norm": 1.0291619219060522e-05, "learning_rate": 4.767141346760624e-05, "loss": 0.0, "num_input_tokens_seen": 1551072, "step": 5540 }, { "epoch": 61.611111111111114, "grad_norm": 1.2228042578499299e-05, "learning_rate": 4.766727424738356e-05, "loss": 0.0, "num_input_tokens_seen": 1552464, "step": 5545 }, { "epoch": 61.666666666666664, "grad_norm": 7.30691363060032e-06, "learning_rate": 4.7663131531582325e-05, "loss": 0.0, "num_input_tokens_seen": 1553776, "step": 5550 }, { "epoch": 61.72222222222222, "grad_norm": 2.082020182569977e-05, "learning_rate": 4.765898532084142e-05, "loss": 0.0, "num_input_tokens_seen": 1555216, "step": 5555 }, { "epoch": 61.77777777777778, "grad_norm": 3.0272406092990423e-06, "learning_rate": 4.765483561580022e-05, "loss": 0.0, "num_input_tokens_seen": 1556608, "step": 5560 }, { "epoch": 61.833333333333336, "grad_norm": 4.393470135255484e-06, "learning_rate": 4.7650682417098666e-05, "loss": 0.0, "num_input_tokens_seen": 1558000, "step": 5565 }, { "epoch": 61.888888888888886, "grad_norm": 3.18897764373105e-06, "learning_rate": 4.7646525725377244e-05, "loss": 0.0, "num_input_tokens_seen": 1559408, "step": 5570 }, { "epoch": 61.94444444444444, "grad_norm": 4.075374818057753e-06, "learning_rate": 4.764236554127696e-05, "loss": 0.0, "num_input_tokens_seen": 1560832, "step": 5575 }, { "epoch": 62.0, "grad_norm": 5.447195690067019e-06, "learning_rate": 4.7638201865439356e-05, "loss": 0.0, "num_input_tokens_seen": 1562240, "step": 5580 }, { "epoch": 62.05555555555556, "grad_norm": 1.8074062609230168e-05, "learning_rate": 4.7634034698506545e-05, "loss": 0.0, "num_input_tokens_seen": 1563680, "step": 5585 }, { "epoch": 62.111111111111114, "grad_norm": 7.558448851341382e-05, "learning_rate": 4.762986404112115e-05, "loss": 0.0, "num_input_tokens_seen": 1565056, "step": 5590 }, { "epoch": 62.166666666666664, "grad_norm": 9.583503015164752e-06, "learning_rate": 4.762568989392633e-05, "loss": 0.0, "num_input_tokens_seen": 1566432, "step": 5595 }, { "epoch": 62.22222222222222, "grad_norm": 3.588444997149054e-06, "learning_rate": 4.76215122575658e-05, "loss": 0.0, "num_input_tokens_seen": 1567808, "step": 5600 }, { "epoch": 62.22222222222222, "eval_loss": 1.0587141513824463, "eval_runtime": 1.384, "eval_samples_per_second": 28.902, "eval_steps_per_second": 14.451, "num_input_tokens_seen": 1567808, "step": 5600 }, { "epoch": 62.27777777777778, "grad_norm": 1.203896044899011e-05, "learning_rate": 4.7617331132683795e-05, "loss": 0.0, "num_input_tokens_seen": 1569248, "step": 5605 }, { "epoch": 62.333333333333336, "grad_norm": 2.0558301912387833e-05, "learning_rate": 4.7613146519925105e-05, "loss": 0.0, "num_input_tokens_seen": 1570640, "step": 5610 }, { "epoch": 62.388888888888886, "grad_norm": 1.3469309124047868e-05, "learning_rate": 4.7608958419935045e-05, "loss": 0.0, "num_input_tokens_seen": 1572032, "step": 5615 }, { "epoch": 62.44444444444444, "grad_norm": 1.4633425962529145e-05, "learning_rate": 4.760476683335948e-05, "loss": 0.0, "num_input_tokens_seen": 1573456, "step": 5620 }, { "epoch": 62.5, "grad_norm": 8.202665412682109e-06, "learning_rate": 4.760057176084479e-05, "loss": 0.0, "num_input_tokens_seen": 1574896, "step": 5625 }, { "epoch": 62.55555555555556, "grad_norm": 3.7363040519267088e-06, "learning_rate": 4.759637320303793e-05, "loss": 0.0, "num_input_tokens_seen": 1576320, "step": 5630 }, { "epoch": 62.611111111111114, "grad_norm": 6.056357051420491e-06, "learning_rate": 4.759217116058635e-05, "loss": 0.0, "num_input_tokens_seen": 1577680, "step": 5635 }, { "epoch": 62.666666666666664, "grad_norm": 1.3661649973073509e-05, "learning_rate": 4.758796563413807e-05, "loss": 0.0, "num_input_tokens_seen": 1579040, "step": 5640 }, { "epoch": 62.72222222222222, "grad_norm": 9.456509360461496e-06, "learning_rate": 4.758375662434163e-05, "loss": 0.0, "num_input_tokens_seen": 1580432, "step": 5645 }, { "epoch": 62.77777777777778, "grad_norm": 3.986028787039686e-06, "learning_rate": 4.7579544131846114e-05, "loss": 0.0, "num_input_tokens_seen": 1581872, "step": 5650 }, { "epoch": 62.833333333333336, "grad_norm": 1.785099993867334e-05, "learning_rate": 4.757532815730114e-05, "loss": 0.0, "num_input_tokens_seen": 1583296, "step": 5655 }, { "epoch": 62.888888888888886, "grad_norm": 1.5894647731329314e-05, "learning_rate": 4.7571108701356865e-05, "loss": 0.0, "num_input_tokens_seen": 1584688, "step": 5660 }, { "epoch": 62.94444444444444, "grad_norm": 1.2899966350232717e-05, "learning_rate": 4.756688576466398e-05, "loss": 0.0, "num_input_tokens_seen": 1586096, "step": 5665 }, { "epoch": 63.0, "grad_norm": 6.45081490802113e-06, "learning_rate": 4.756265934787372e-05, "loss": 0.0, "num_input_tokens_seen": 1587488, "step": 5670 }, { "epoch": 63.05555555555556, "grad_norm": 6.674707037745975e-06, "learning_rate": 4.755842945163785e-05, "loss": 0.0, "num_input_tokens_seen": 1588880, "step": 5675 }, { "epoch": 63.111111111111114, "grad_norm": 1.0899292647081893e-05, "learning_rate": 4.755419607660867e-05, "loss": 0.0, "num_input_tokens_seen": 1590272, "step": 5680 }, { "epoch": 63.166666666666664, "grad_norm": 2.9899827040935634e-06, "learning_rate": 4.7549959223439016e-05, "loss": 0.0, "num_input_tokens_seen": 1591696, "step": 5685 }, { "epoch": 63.22222222222222, "grad_norm": 1.736553167575039e-05, "learning_rate": 4.754571889278228e-05, "loss": 0.0, "num_input_tokens_seen": 1593024, "step": 5690 }, { "epoch": 63.27777777777778, "grad_norm": 6.438378477469087e-06, "learning_rate": 4.754147508529235e-05, "loss": 0.0, "num_input_tokens_seen": 1594432, "step": 5695 }, { "epoch": 63.333333333333336, "grad_norm": 4.340948635217501e-06, "learning_rate": 4.75372278016237e-05, "loss": 0.0, "num_input_tokens_seen": 1595808, "step": 5700 }, { "epoch": 63.388888888888886, "grad_norm": 6.806728197261691e-05, "learning_rate": 4.753297704243129e-05, "loss": 0.0, "num_input_tokens_seen": 1597216, "step": 5705 }, { "epoch": 63.44444444444444, "grad_norm": 3.185611785738729e-05, "learning_rate": 4.752872280837066e-05, "loss": 0.0, "num_input_tokens_seen": 1598624, "step": 5710 }, { "epoch": 63.5, "grad_norm": 3.3335572879877873e-06, "learning_rate": 4.752446510009786e-05, "loss": 0.0, "num_input_tokens_seen": 1600000, "step": 5715 }, { "epoch": 63.55555555555556, "grad_norm": 5.328755833033938e-06, "learning_rate": 4.7520203918269476e-05, "loss": 0.0, "num_input_tokens_seen": 1601392, "step": 5720 }, { "epoch": 63.611111111111114, "grad_norm": 3.992436631961027e-06, "learning_rate": 4.751593926354265e-05, "loss": 0.0, "num_input_tokens_seen": 1602768, "step": 5725 }, { "epoch": 63.666666666666664, "grad_norm": 9.592641617928166e-06, "learning_rate": 4.751167113657503e-05, "loss": 0.0, "num_input_tokens_seen": 1604160, "step": 5730 }, { "epoch": 63.72222222222222, "grad_norm": 5.902696102566551e-06, "learning_rate": 4.7507399538024834e-05, "loss": 0.0, "num_input_tokens_seen": 1605584, "step": 5735 }, { "epoch": 63.77777777777778, "grad_norm": 1.1776034625654574e-05, "learning_rate": 4.750312446855077e-05, "loss": 0.0, "num_input_tokens_seen": 1607008, "step": 5740 }, { "epoch": 63.833333333333336, "grad_norm": 1.525581865280401e-05, "learning_rate": 4.749884592881212e-05, "loss": 0.0, "num_input_tokens_seen": 1608400, "step": 5745 }, { "epoch": 63.888888888888886, "grad_norm": 1.1424906915635802e-05, "learning_rate": 4.74945639194687e-05, "loss": 0.0, "num_input_tokens_seen": 1609808, "step": 5750 }, { "epoch": 63.94444444444444, "grad_norm": 6.661660790996393e-06, "learning_rate": 4.749027844118083e-05, "loss": 0.0, "num_input_tokens_seen": 1611232, "step": 5755 }, { "epoch": 64.0, "grad_norm": 1.7263148492929759e-06, "learning_rate": 4.7485989494609395e-05, "loss": 0.0, "num_input_tokens_seen": 1612624, "step": 5760 }, { "epoch": 64.05555555555556, "grad_norm": 3.477070777080371e-06, "learning_rate": 4.748169708041581e-05, "loss": 0.0, "num_input_tokens_seen": 1614048, "step": 5765 }, { "epoch": 64.11111111111111, "grad_norm": 7.586362244182965e-06, "learning_rate": 4.7477401199262004e-05, "loss": 0.0, "num_input_tokens_seen": 1615408, "step": 5770 }, { "epoch": 64.16666666666667, "grad_norm": 3.2498826385563007e-06, "learning_rate": 4.747310185181048e-05, "loss": 0.0, "num_input_tokens_seen": 1616800, "step": 5775 }, { "epoch": 64.22222222222223, "grad_norm": 1.857386450865306e-05, "learning_rate": 4.746879903872422e-05, "loss": 0.0, "num_input_tokens_seen": 1618192, "step": 5780 }, { "epoch": 64.27777777777777, "grad_norm": 8.213328328565694e-06, "learning_rate": 4.746449276066679e-05, "loss": 0.0, "num_input_tokens_seen": 1619568, "step": 5785 }, { "epoch": 64.33333333333333, "grad_norm": 3.7186698591540335e-06, "learning_rate": 4.746018301830227e-05, "loss": 0.0, "num_input_tokens_seen": 1620944, "step": 5790 }, { "epoch": 64.38888888888889, "grad_norm": 4.294221980671864e-06, "learning_rate": 4.7455869812295275e-05, "loss": 0.0, "num_input_tokens_seen": 1622384, "step": 5795 }, { "epoch": 64.44444444444444, "grad_norm": 2.0326067442510976e-06, "learning_rate": 4.7451553143310964e-05, "loss": 0.0, "num_input_tokens_seen": 1623744, "step": 5800 }, { "epoch": 64.44444444444444, "eval_loss": 1.0723211765289307, "eval_runtime": 1.3871, "eval_samples_per_second": 28.837, "eval_steps_per_second": 14.418, "num_input_tokens_seen": 1623744, "step": 5800 }, { "epoch": 64.5, "grad_norm": 2.880659167203703e-06, "learning_rate": 4.744723301201501e-05, "loss": 0.0, "num_input_tokens_seen": 1625152, "step": 5805 }, { "epoch": 64.55555555555556, "grad_norm": 6.7425107772578485e-06, "learning_rate": 4.744290941907364e-05, "loss": 0.0, "num_input_tokens_seen": 1626560, "step": 5810 }, { "epoch": 64.61111111111111, "grad_norm": 1.4656370694865473e-05, "learning_rate": 4.7438582365153594e-05, "loss": 0.0, "num_input_tokens_seen": 1627952, "step": 5815 }, { "epoch": 64.66666666666667, "grad_norm": 1.2520399650384206e-05, "learning_rate": 4.743425185092217e-05, "loss": 0.0, "num_input_tokens_seen": 1629392, "step": 5820 }, { "epoch": 64.72222222222223, "grad_norm": 3.199234197381884e-05, "learning_rate": 4.742991787704719e-05, "loss": 0.0, "num_input_tokens_seen": 1630784, "step": 5825 }, { "epoch": 64.77777777777777, "grad_norm": 1.2285480806895066e-05, "learning_rate": 4.7425580444196994e-05, "loss": 0.0, "num_input_tokens_seen": 1632160, "step": 5830 }, { "epoch": 64.83333333333333, "grad_norm": 3.4463839710952016e-06, "learning_rate": 4.742123955304048e-05, "loss": 0.0, "num_input_tokens_seen": 1633600, "step": 5835 }, { "epoch": 64.88888888888889, "grad_norm": 3.7674778923246777e-06, "learning_rate": 4.741689520424706e-05, "loss": 0.0, "num_input_tokens_seen": 1635008, "step": 5840 }, { "epoch": 64.94444444444444, "grad_norm": 3.3703197459544754e-06, "learning_rate": 4.741254739848669e-05, "loss": 0.0, "num_input_tokens_seen": 1636416, "step": 5845 }, { "epoch": 65.0, "grad_norm": 3.1298018257075455e-06, "learning_rate": 4.740819613642987e-05, "loss": 0.0, "num_input_tokens_seen": 1637824, "step": 5850 }, { "epoch": 65.05555555555556, "grad_norm": 5.802279247291153e-06, "learning_rate": 4.74038414187476e-05, "loss": 0.0, "num_input_tokens_seen": 1639216, "step": 5855 }, { "epoch": 65.11111111111111, "grad_norm": 9.67716368904803e-06, "learning_rate": 4.739948324611144e-05, "loss": 0.0, "num_input_tokens_seen": 1640624, "step": 5860 }, { "epoch": 65.16666666666667, "grad_norm": 5.825457719765836e-06, "learning_rate": 4.7395121619193465e-05, "loss": 0.0, "num_input_tokens_seen": 1642000, "step": 5865 }, { "epoch": 65.22222222222223, "grad_norm": 2.9172813356126426e-06, "learning_rate": 4.7390756538666313e-05, "loss": 0.0, "num_input_tokens_seen": 1643376, "step": 5870 }, { "epoch": 65.27777777777777, "grad_norm": 1.8150869436794892e-05, "learning_rate": 4.738638800520311e-05, "loss": 0.0, "num_input_tokens_seen": 1644832, "step": 5875 }, { "epoch": 65.33333333333333, "grad_norm": 1.1746114068955649e-05, "learning_rate": 4.738201601947757e-05, "loss": 0.0, "num_input_tokens_seen": 1646208, "step": 5880 }, { "epoch": 65.38888888888889, "grad_norm": 7.416795142489718e-06, "learning_rate": 4.7377640582163876e-05, "loss": 0.0, "num_input_tokens_seen": 1647600, "step": 5885 }, { "epoch": 65.44444444444444, "grad_norm": 7.86536475061439e-05, "learning_rate": 4.7373261693936786e-05, "loss": 0.0, "num_input_tokens_seen": 1648976, "step": 5890 }, { "epoch": 65.5, "grad_norm": 5.099703230371233e-06, "learning_rate": 4.7368879355471595e-05, "loss": 0.0, "num_input_tokens_seen": 1650352, "step": 5895 }, { "epoch": 65.55555555555556, "grad_norm": 6.008851869410137e-06, "learning_rate": 4.736449356744409e-05, "loss": 0.0, "num_input_tokens_seen": 1651744, "step": 5900 }, { "epoch": 65.61111111111111, "grad_norm": 1.2567368685267866e-05, "learning_rate": 4.736010433053064e-05, "loss": 0.0, "num_input_tokens_seen": 1653152, "step": 5905 }, { "epoch": 65.66666666666667, "grad_norm": 4.0026807255344465e-06, "learning_rate": 4.73557116454081e-05, "loss": 0.0, "num_input_tokens_seen": 1654576, "step": 5910 }, { "epoch": 65.72222222222223, "grad_norm": 3.5724206099985167e-06, "learning_rate": 4.735131551275389e-05, "loss": 0.0, "num_input_tokens_seen": 1656032, "step": 5915 }, { "epoch": 65.77777777777777, "grad_norm": 2.655283196872915e-06, "learning_rate": 4.734691593324594e-05, "loss": 0.0, "num_input_tokens_seen": 1657408, "step": 5920 }, { "epoch": 65.83333333333333, "grad_norm": 1.5544592315563932e-05, "learning_rate": 4.734251290756272e-05, "loss": 0.0, "num_input_tokens_seen": 1658800, "step": 5925 }, { "epoch": 65.88888888888889, "grad_norm": 7.335404006880708e-06, "learning_rate": 4.7338106436383246e-05, "loss": 0.0, "num_input_tokens_seen": 1660192, "step": 5930 }, { "epoch": 65.94444444444444, "grad_norm": 9.67000778473448e-06, "learning_rate": 4.733369652038703e-05, "loss": 0.0, "num_input_tokens_seen": 1661616, "step": 5935 }, { "epoch": 66.0, "grad_norm": 3.6140515931037953e-06, "learning_rate": 4.7329283160254156e-05, "loss": 0.0, "num_input_tokens_seen": 1663072, "step": 5940 }, { "epoch": 66.05555555555556, "grad_norm": 7.154695049393922e-05, "learning_rate": 4.732486635666521e-05, "loss": 0.0, "num_input_tokens_seen": 1664432, "step": 5945 }, { "epoch": 66.11111111111111, "grad_norm": 2.951196393041755e-06, "learning_rate": 4.732044611030132e-05, "loss": 0.0, "num_input_tokens_seen": 1665808, "step": 5950 }, { "epoch": 66.16666666666667, "grad_norm": 6.160901648399886e-06, "learning_rate": 4.731602242184414e-05, "loss": 0.0, "num_input_tokens_seen": 1667200, "step": 5955 }, { "epoch": 66.22222222222223, "grad_norm": 2.0373114239191636e-05, "learning_rate": 4.7311595291975864e-05, "loss": 0.0, "num_input_tokens_seen": 1668608, "step": 5960 }, { "epoch": 66.27777777777777, "grad_norm": 7.2568859650345985e-06, "learning_rate": 4.7307164721379216e-05, "loss": 0.0, "num_input_tokens_seen": 1669984, "step": 5965 }, { "epoch": 66.33333333333333, "grad_norm": 3.527578201101278e-06, "learning_rate": 4.730273071073743e-05, "loss": 0.0, "num_input_tokens_seen": 1671440, "step": 5970 }, { "epoch": 66.38888888888889, "grad_norm": 1.5136489309952594e-05, "learning_rate": 4.729829326073429e-05, "loss": 0.0, "num_input_tokens_seen": 1672816, "step": 5975 }, { "epoch": 66.44444444444444, "grad_norm": 2.5330306016257964e-05, "learning_rate": 4.7293852372054126e-05, "loss": 0.0, "num_input_tokens_seen": 1674240, "step": 5980 }, { "epoch": 66.5, "grad_norm": 1.1209795047761872e-05, "learning_rate": 4.728940804538176e-05, "loss": 0.0, "num_input_tokens_seen": 1675648, "step": 5985 }, { "epoch": 66.55555555555556, "grad_norm": 5.0745479711622465e-06, "learning_rate": 4.7284960281402556e-05, "loss": 0.0, "num_input_tokens_seen": 1677024, "step": 5990 }, { "epoch": 66.61111111111111, "grad_norm": 3.920170911442256e-06, "learning_rate": 4.728050908080244e-05, "loss": 0.0, "num_input_tokens_seen": 1678464, "step": 5995 }, { "epoch": 66.66666666666667, "grad_norm": 5.062370473751798e-05, "learning_rate": 4.727605444426782e-05, "loss": 0.0, "num_input_tokens_seen": 1679888, "step": 6000 }, { "epoch": 66.66666666666667, "eval_loss": 1.0874003171920776, "eval_runtime": 1.4013, "eval_samples_per_second": 28.545, "eval_steps_per_second": 14.272, "num_input_tokens_seen": 1679888, "step": 6000 }, { "epoch": 66.72222222222223, "grad_norm": 2.7093878998130094e-06, "learning_rate": 4.727159637248567e-05, "loss": 0.0, "num_input_tokens_seen": 1681264, "step": 6005 }, { "epoch": 66.77777777777777, "grad_norm": 1.431433793186443e-05, "learning_rate": 4.7267134866143474e-05, "loss": 0.0, "num_input_tokens_seen": 1682688, "step": 6010 }, { "epoch": 66.83333333333333, "grad_norm": 2.9038553748250706e-06, "learning_rate": 4.726266992592926e-05, "loss": 0.0, "num_input_tokens_seen": 1684064, "step": 6015 }, { "epoch": 66.88888888888889, "grad_norm": 3.435769940551836e-06, "learning_rate": 4.725820155253157e-05, "loss": 0.0, "num_input_tokens_seen": 1685440, "step": 6020 }, { "epoch": 66.94444444444444, "grad_norm": 5.185645932215266e-06, "learning_rate": 4.725372974663948e-05, "loss": 0.0, "num_input_tokens_seen": 1686832, "step": 6025 }, { "epoch": 67.0, "grad_norm": 1.93095138456556e-06, "learning_rate": 4.724925450894262e-05, "loss": 0.0, "num_input_tokens_seen": 1688272, "step": 6030 }, { "epoch": 67.05555555555556, "grad_norm": 3.7196246012172196e-06, "learning_rate": 4.72447758401311e-05, "loss": 0.0, "num_input_tokens_seen": 1689680, "step": 6035 }, { "epoch": 67.11111111111111, "grad_norm": 3.5933944673161022e-06, "learning_rate": 4.7240293740895616e-05, "loss": 0.0, "num_input_tokens_seen": 1691104, "step": 6040 }, { "epoch": 67.16666666666667, "grad_norm": 1.0691766874515451e-05, "learning_rate": 4.723580821192733e-05, "loss": 0.0, "num_input_tokens_seen": 1692512, "step": 6045 }, { "epoch": 67.22222222222223, "grad_norm": 4.5441920519806445e-06, "learning_rate": 4.7231319253917996e-05, "loss": 0.0, "num_input_tokens_seen": 1693936, "step": 6050 }, { "epoch": 67.27777777777777, "grad_norm": 7.37771642889129e-06, "learning_rate": 4.722682686755986e-05, "loss": 0.0, "num_input_tokens_seen": 1695344, "step": 6055 }, { "epoch": 67.33333333333333, "grad_norm": 4.303026344132377e-06, "learning_rate": 4.722233105354569e-05, "loss": 0.0, "num_input_tokens_seen": 1696736, "step": 6060 }, { "epoch": 67.38888888888889, "grad_norm": 5.688041710527614e-06, "learning_rate": 4.7217831812568815e-05, "loss": 0.0, "num_input_tokens_seen": 1698128, "step": 6065 }, { "epoch": 67.44444444444444, "grad_norm": 1.0165968888031784e-05, "learning_rate": 4.721332914532307e-05, "loss": 0.0, "num_input_tokens_seen": 1699504, "step": 6070 }, { "epoch": 67.5, "grad_norm": 8.094531949609518e-06, "learning_rate": 4.720882305250281e-05, "loss": 0.0, "num_input_tokens_seen": 1700896, "step": 6075 }, { "epoch": 67.55555555555556, "grad_norm": 3.567249223124236e-06, "learning_rate": 4.720431353480295e-05, "loss": 0.0, "num_input_tokens_seen": 1702272, "step": 6080 }, { "epoch": 67.61111111111111, "grad_norm": 3.90625564250513e-06, "learning_rate": 4.719980059291891e-05, "loss": 0.0, "num_input_tokens_seen": 1703680, "step": 6085 }, { "epoch": 67.66666666666667, "grad_norm": 4.479530616663396e-06, "learning_rate": 4.7195284227546634e-05, "loss": 0.0, "num_input_tokens_seen": 1705088, "step": 6090 }, { "epoch": 67.72222222222223, "grad_norm": 5.0362927140668035e-05, "learning_rate": 4.7190764439382604e-05, "loss": 0.0, "num_input_tokens_seen": 1706496, "step": 6095 }, { "epoch": 67.77777777777777, "grad_norm": 3.770898729271721e-06, "learning_rate": 4.7186241229123826e-05, "loss": 0.0, "num_input_tokens_seen": 1707904, "step": 6100 }, { "epoch": 67.83333333333333, "grad_norm": 4.088764853804605e-06, "learning_rate": 4.718171459746785e-05, "loss": 0.0, "num_input_tokens_seen": 1709296, "step": 6105 }, { "epoch": 67.88888888888889, "grad_norm": 1.8898712141890428e-06, "learning_rate": 4.717718454511273e-05, "loss": 0.0, "num_input_tokens_seen": 1710672, "step": 6110 }, { "epoch": 67.94444444444444, "grad_norm": 6.829007816122612e-06, "learning_rate": 4.7172651072757056e-05, "loss": 0.0, "num_input_tokens_seen": 1712048, "step": 6115 }, { "epoch": 68.0, "grad_norm": 5.073869033367373e-06, "learning_rate": 4.7168114181099945e-05, "loss": 0.0, "num_input_tokens_seen": 1713488, "step": 6120 }, { "epoch": 68.05555555555556, "grad_norm": 1.448428702133242e-05, "learning_rate": 4.716357387084105e-05, "loss": 0.0, "num_input_tokens_seen": 1714864, "step": 6125 }, { "epoch": 68.11111111111111, "grad_norm": 5.275680905469926e-06, "learning_rate": 4.715903014268054e-05, "loss": 0.0, "num_input_tokens_seen": 1716240, "step": 6130 }, { "epoch": 68.16666666666667, "grad_norm": 5.634458830172662e-06, "learning_rate": 4.715448299731911e-05, "loss": 0.0, "num_input_tokens_seen": 1717648, "step": 6135 }, { "epoch": 68.22222222222223, "grad_norm": 1.8947255739476532e-05, "learning_rate": 4.7149932435457986e-05, "loss": 0.0, "num_input_tokens_seen": 1719072, "step": 6140 }, { "epoch": 68.27777777777777, "grad_norm": 7.847857887099963e-06, "learning_rate": 4.714537845779894e-05, "loss": 0.0, "num_input_tokens_seen": 1720480, "step": 6145 }, { "epoch": 68.33333333333333, "grad_norm": 3.227325214538723e-05, "learning_rate": 4.714082106504423e-05, "loss": 0.0, "num_input_tokens_seen": 1721888, "step": 6150 }, { "epoch": 68.38888888888889, "grad_norm": 2.5499527964711888e-06, "learning_rate": 4.713626025789667e-05, "loss": 0.0, "num_input_tokens_seen": 1723312, "step": 6155 }, { "epoch": 68.44444444444444, "grad_norm": 3.3733522286638618e-06, "learning_rate": 4.7131696037059606e-05, "loss": 0.0, "num_input_tokens_seen": 1724672, "step": 6160 }, { "epoch": 68.5, "grad_norm": 2.962132748507429e-05, "learning_rate": 4.712712840323689e-05, "loss": 0.0, "num_input_tokens_seen": 1726048, "step": 6165 }, { "epoch": 68.55555555555556, "grad_norm": 2.896270052588079e-05, "learning_rate": 4.71225573571329e-05, "loss": 0.0, "num_input_tokens_seen": 1727504, "step": 6170 }, { "epoch": 68.61111111111111, "grad_norm": 4.824913048651069e-06, "learning_rate": 4.711798289945256e-05, "loss": 0.0, "num_input_tokens_seen": 1728928, "step": 6175 }, { "epoch": 68.66666666666667, "grad_norm": 9.351326298201457e-06, "learning_rate": 4.71134050309013e-05, "loss": 0.0, "num_input_tokens_seen": 1730368, "step": 6180 }, { "epoch": 68.72222222222223, "grad_norm": 9.509264600637835e-06, "learning_rate": 4.710882375218509e-05, "loss": 0.0, "num_input_tokens_seen": 1731792, "step": 6185 }, { "epoch": 68.77777777777777, "grad_norm": 5.912971118959831e-06, "learning_rate": 4.7104239064010424e-05, "loss": 0.0, "num_input_tokens_seen": 1733184, "step": 6190 }, { "epoch": 68.83333333333333, "grad_norm": 5.020177468395559e-06, "learning_rate": 4.709965096708432e-05, "loss": 0.0, "num_input_tokens_seen": 1734576, "step": 6195 }, { "epoch": 68.88888888888889, "grad_norm": 1.1334376722516026e-05, "learning_rate": 4.709505946211431e-05, "loss": 0.0, "num_input_tokens_seen": 1735952, "step": 6200 }, { "epoch": 68.88888888888889, "eval_loss": 1.061773657798767, "eval_runtime": 1.3904, "eval_samples_per_second": 28.768, "eval_steps_per_second": 14.384, "num_input_tokens_seen": 1735952, "step": 6200 }, { "epoch": 68.94444444444444, "grad_norm": 6.298865173448576e-06, "learning_rate": 4.709046454980846e-05, "loss": 0.0, "num_input_tokens_seen": 1737344, "step": 6205 }, { "epoch": 69.0, "grad_norm": 1.7175430002680514e-06, "learning_rate": 4.708586623087538e-05, "loss": 0.0, "num_input_tokens_seen": 1738752, "step": 6210 }, { "epoch": 69.05555555555556, "grad_norm": 2.1157235096325167e-05, "learning_rate": 4.708126450602418e-05, "loss": 0.0, "num_input_tokens_seen": 1740176, "step": 6215 }, { "epoch": 69.11111111111111, "grad_norm": 3.2103073408507043e-06, "learning_rate": 4.7076659375964495e-05, "loss": 0.0, "num_input_tokens_seen": 1741536, "step": 6220 }, { "epoch": 69.16666666666667, "grad_norm": 2.7259277430857765e-06, "learning_rate": 4.707205084140651e-05, "loss": 0.0, "num_input_tokens_seen": 1742944, "step": 6225 }, { "epoch": 69.22222222222223, "grad_norm": 5.509361926669953e-06, "learning_rate": 4.7067438903060904e-05, "loss": 0.0, "num_input_tokens_seen": 1744384, "step": 6230 }, { "epoch": 69.27777777777777, "grad_norm": 2.14285478250531e-06, "learning_rate": 4.70628235616389e-05, "loss": 0.0, "num_input_tokens_seen": 1745760, "step": 6235 }, { "epoch": 69.33333333333333, "grad_norm": 1.4887582437950186e-05, "learning_rate": 4.7058204817852256e-05, "loss": 0.0, "num_input_tokens_seen": 1747136, "step": 6240 }, { "epoch": 69.38888888888889, "grad_norm": 3.489564051051275e-06, "learning_rate": 4.705358267241322e-05, "loss": 0.0, "num_input_tokens_seen": 1748544, "step": 6245 }, { "epoch": 69.44444444444444, "grad_norm": 4.222636562190019e-06, "learning_rate": 4.704895712603459e-05, "loss": 0.0, "num_input_tokens_seen": 1749984, "step": 6250 }, { "epoch": 69.5, "grad_norm": 4.16800367020187e-06, "learning_rate": 4.704432817942969e-05, "loss": 0.0, "num_input_tokens_seen": 1751392, "step": 6255 }, { "epoch": 69.55555555555556, "grad_norm": 1.5349427485489286e-05, "learning_rate": 4.703969583331236e-05, "loss": 0.0, "num_input_tokens_seen": 1752784, "step": 6260 }, { "epoch": 69.61111111111111, "grad_norm": 4.2641563595680054e-06, "learning_rate": 4.7035060088396965e-05, "loss": 0.0, "num_input_tokens_seen": 1754160, "step": 6265 }, { "epoch": 69.66666666666667, "grad_norm": 4.070252543897368e-05, "learning_rate": 4.703042094539839e-05, "loss": 0.0, "num_input_tokens_seen": 1755584, "step": 6270 }, { "epoch": 69.72222222222223, "grad_norm": 8.61157604958862e-06, "learning_rate": 4.702577840503206e-05, "loss": 0.0, "num_input_tokens_seen": 1756992, "step": 6275 }, { "epoch": 69.77777777777777, "grad_norm": 2.5572171580279246e-06, "learning_rate": 4.70211324680139e-05, "loss": 0.0, "num_input_tokens_seen": 1758384, "step": 6280 }, { "epoch": 69.83333333333333, "grad_norm": 7.262079179781722e-06, "learning_rate": 4.7016483135060386e-05, "loss": 0.0, "num_input_tokens_seen": 1759792, "step": 6285 }, { "epoch": 69.88888888888889, "grad_norm": 4.479323706618743e-06, "learning_rate": 4.701183040688849e-05, "loss": 0.0, "num_input_tokens_seen": 1761152, "step": 6290 }, { "epoch": 69.94444444444444, "grad_norm": 2.5364122848259285e-05, "learning_rate": 4.700717428421573e-05, "loss": 0.0, "num_input_tokens_seen": 1762608, "step": 6295 }, { "epoch": 70.0, "grad_norm": 1.4215561350283679e-05, "learning_rate": 4.700251476776014e-05, "loss": 0.0, "num_input_tokens_seen": 1764000, "step": 6300 }, { "epoch": 70.05555555555556, "grad_norm": 3.6458134218264604e-06, "learning_rate": 4.699785185824026e-05, "loss": 0.0, "num_input_tokens_seen": 1765376, "step": 6305 }, { "epoch": 70.11111111111111, "grad_norm": 3.04344739561202e-05, "learning_rate": 4.699318555637519e-05, "loss": 0.0, "num_input_tokens_seen": 1766752, "step": 6310 }, { "epoch": 70.16666666666667, "grad_norm": 3.636017254393664e-06, "learning_rate": 4.6988515862884525e-05, "loss": 0.0, "num_input_tokens_seen": 1768160, "step": 6315 }, { "epoch": 70.22222222222223, "grad_norm": 3.773500566239818e-06, "learning_rate": 4.698384277848838e-05, "loss": 0.0, "num_input_tokens_seen": 1769632, "step": 6320 }, { "epoch": 70.27777777777777, "grad_norm": 3.5933448998548556e-06, "learning_rate": 4.6979166303907425e-05, "loss": 0.0, "num_input_tokens_seen": 1771008, "step": 6325 }, { "epoch": 70.33333333333333, "grad_norm": 3.466188672973658e-06, "learning_rate": 4.697448643986281e-05, "loss": 0.0, "num_input_tokens_seen": 1772368, "step": 6330 }, { "epoch": 70.38888888888889, "grad_norm": 2.039588707702933e-06, "learning_rate": 4.696980318707624e-05, "loss": 0.0, "num_input_tokens_seen": 1773744, "step": 6335 }, { "epoch": 70.44444444444444, "grad_norm": 2.032694283116143e-06, "learning_rate": 4.6965116546269924e-05, "loss": 0.0, "num_input_tokens_seen": 1775168, "step": 6340 }, { "epoch": 70.5, "grad_norm": 4.6281516006274614e-06, "learning_rate": 4.6960426518166615e-05, "loss": 0.0, "num_input_tokens_seen": 1776544, "step": 6345 }, { "epoch": 70.55555555555556, "grad_norm": 3.085565913352184e-05, "learning_rate": 4.6955733103489556e-05, "loss": 0.0, "num_input_tokens_seen": 1777952, "step": 6350 }, { "epoch": 70.61111111111111, "grad_norm": 1.7977230527321808e-05, "learning_rate": 4.695103630296255e-05, "loss": 0.0, "num_input_tokens_seen": 1779328, "step": 6355 }, { "epoch": 70.66666666666667, "grad_norm": 4.668662313633831e-06, "learning_rate": 4.694633611730988e-05, "loss": 0.0, "num_input_tokens_seen": 1780736, "step": 6360 }, { "epoch": 70.72222222222223, "grad_norm": 3.5539087548386306e-05, "learning_rate": 4.694163254725639e-05, "loss": 0.0, "num_input_tokens_seen": 1782112, "step": 6365 }, { "epoch": 70.77777777777777, "grad_norm": 2.6031973447970813e-06, "learning_rate": 4.693692559352743e-05, "loss": 0.0, "num_input_tokens_seen": 1783520, "step": 6370 }, { "epoch": 70.83333333333333, "grad_norm": 7.84318308433285e-06, "learning_rate": 4.693221525684886e-05, "loss": 0.0, "num_input_tokens_seen": 1784928, "step": 6375 }, { "epoch": 70.88888888888889, "grad_norm": 3.823069710051641e-05, "learning_rate": 4.6927501537947084e-05, "loss": 0.0, "num_input_tokens_seen": 1786304, "step": 6380 }, { "epoch": 70.94444444444444, "grad_norm": 1.4396588085219264e-05, "learning_rate": 4.692278443754901e-05, "loss": 0.0, "num_input_tokens_seen": 1787712, "step": 6385 }, { "epoch": 71.0, "grad_norm": 3.1244242109096376e-06, "learning_rate": 4.691806395638208e-05, "loss": 0.0, "num_input_tokens_seen": 1789088, "step": 6390 }, { "epoch": 71.05555555555556, "grad_norm": 4.79862274005427e-06, "learning_rate": 4.6913340095174255e-05, "loss": 0.0, "num_input_tokens_seen": 1790496, "step": 6395 }, { "epoch": 71.11111111111111, "grad_norm": 6.255561856960412e-06, "learning_rate": 4.690861285465399e-05, "loss": 0.0, "num_input_tokens_seen": 1791904, "step": 6400 }, { "epoch": 71.11111111111111, "eval_loss": 1.0959649085998535, "eval_runtime": 1.3942, "eval_samples_per_second": 28.691, "eval_steps_per_second": 14.345, "num_input_tokens_seen": 1791904, "step": 6400 }, { "epoch": 71.16666666666667, "grad_norm": 4.444942078407621e-06, "learning_rate": 4.690388223555031e-05, "loss": 0.0, "num_input_tokens_seen": 1793296, "step": 6405 }, { "epoch": 71.22222222222223, "grad_norm": 2.7550697723199846e-06, "learning_rate": 4.689914823859273e-05, "loss": 0.0, "num_input_tokens_seen": 1794688, "step": 6410 }, { "epoch": 71.27777777777777, "grad_norm": 7.29158045942313e-06, "learning_rate": 4.689441086451129e-05, "loss": 0.0, "num_input_tokens_seen": 1796112, "step": 6415 }, { "epoch": 71.33333333333333, "grad_norm": 1.600623591002659e-06, "learning_rate": 4.688967011403655e-05, "loss": 0.0, "num_input_tokens_seen": 1797520, "step": 6420 }, { "epoch": 71.38888888888889, "grad_norm": 2.6137722670682706e-06, "learning_rate": 4.68849259878996e-05, "loss": 0.0, "num_input_tokens_seen": 1798896, "step": 6425 }, { "epoch": 71.44444444444444, "grad_norm": 1.4167115978125366e-06, "learning_rate": 4.6880178486832036e-05, "loss": 0.0, "num_input_tokens_seen": 1800304, "step": 6430 }, { "epoch": 71.5, "grad_norm": 2.4240155198640423e-06, "learning_rate": 4.687542761156598e-05, "loss": 0.0, "num_input_tokens_seen": 1801664, "step": 6435 }, { "epoch": 71.55555555555556, "grad_norm": 1.0104132343258243e-05, "learning_rate": 4.6870673362834096e-05, "loss": 0.0, "num_input_tokens_seen": 1803056, "step": 6440 }, { "epoch": 71.61111111111111, "grad_norm": 4.280268967704615e-06, "learning_rate": 4.6865915741369526e-05, "loss": 0.0, "num_input_tokens_seen": 1804432, "step": 6445 }, { "epoch": 71.66666666666667, "grad_norm": 2.8417896828614175e-05, "learning_rate": 4.686115474790597e-05, "loss": 0.0, "num_input_tokens_seen": 1805856, "step": 6450 }, { "epoch": 71.72222222222223, "grad_norm": 2.532648795749992e-05, "learning_rate": 4.685639038317762e-05, "loss": 0.0, "num_input_tokens_seen": 1807264, "step": 6455 }, { "epoch": 71.77777777777777, "grad_norm": 7.587328127556248e-06, "learning_rate": 4.685162264791921e-05, "loss": 0.0, "num_input_tokens_seen": 1808688, "step": 6460 }, { "epoch": 71.83333333333333, "grad_norm": 1.0395489880465902e-05, "learning_rate": 4.684685154286599e-05, "loss": 0.0, "num_input_tokens_seen": 1810096, "step": 6465 }, { "epoch": 71.88888888888889, "grad_norm": 3.1365398172056302e-06, "learning_rate": 4.684207706875371e-05, "loss": 0.0, "num_input_tokens_seen": 1811488, "step": 6470 }, { "epoch": 71.94444444444444, "grad_norm": 6.647811460425146e-06, "learning_rate": 4.683729922631866e-05, "loss": 0.0, "num_input_tokens_seen": 1812896, "step": 6475 }, { "epoch": 72.0, "grad_norm": 1.2070335287717171e-05, "learning_rate": 4.683251801629765e-05, "loss": 0.0, "num_input_tokens_seen": 1814288, "step": 6480 }, { "epoch": 72.05555555555556, "grad_norm": 1.2868487829109654e-05, "learning_rate": 4.6827733439428e-05, "loss": 0.0, "num_input_tokens_seen": 1815696, "step": 6485 }, { "epoch": 72.11111111111111, "grad_norm": 2.051702495009522e-06, "learning_rate": 4.682294549644754e-05, "loss": 0.0, "num_input_tokens_seen": 1817104, "step": 6490 }, { "epoch": 72.16666666666667, "grad_norm": 1.6768233308539493e-06, "learning_rate": 4.681815418809464e-05, "loss": 0.0, "num_input_tokens_seen": 1818528, "step": 6495 }, { "epoch": 72.22222222222223, "grad_norm": 1.0069246627608663e-06, "learning_rate": 4.681335951510819e-05, "loss": 0.0, "num_input_tokens_seen": 1819920, "step": 6500 }, { "epoch": 72.27777777777777, "grad_norm": 1.8512641872803215e-06, "learning_rate": 4.6808561478227576e-05, "loss": 0.0, "num_input_tokens_seen": 1821312, "step": 6505 }, { "epoch": 72.33333333333333, "grad_norm": 2.1877319795748917e-06, "learning_rate": 4.680376007819271e-05, "loss": 0.0, "num_input_tokens_seen": 1822720, "step": 6510 }, { "epoch": 72.38888888888889, "grad_norm": 3.547651431290433e-05, "learning_rate": 4.679895531574405e-05, "loss": 0.0, "num_input_tokens_seen": 1824096, "step": 6515 }, { "epoch": 72.44444444444444, "grad_norm": 1.461910596844973e-05, "learning_rate": 4.679414719162253e-05, "loss": 0.0, "num_input_tokens_seen": 1825472, "step": 6520 }, { "epoch": 72.5, "grad_norm": 8.98991129361093e-06, "learning_rate": 4.6789335706569635e-05, "loss": 0.0, "num_input_tokens_seen": 1826832, "step": 6525 }, { "epoch": 72.55555555555556, "grad_norm": 1.3479269910021685e-05, "learning_rate": 4.678452086132734e-05, "loss": 0.0, "num_input_tokens_seen": 1828272, "step": 6530 }, { "epoch": 72.61111111111111, "grad_norm": 7.01144153936184e-06, "learning_rate": 4.677970265663818e-05, "loss": 0.0, "num_input_tokens_seen": 1829680, "step": 6535 }, { "epoch": 72.66666666666667, "grad_norm": 2.2248595996643417e-06, "learning_rate": 4.677488109324517e-05, "loss": 0.0, "num_input_tokens_seen": 1831072, "step": 6540 }, { "epoch": 72.72222222222223, "grad_norm": 5.906169462832622e-06, "learning_rate": 4.6770056171891846e-05, "loss": 0.0, "num_input_tokens_seen": 1832480, "step": 6545 }, { "epoch": 72.77777777777777, "grad_norm": 2.0776324163307436e-06, "learning_rate": 4.6765227893322286e-05, "loss": 0.0, "num_input_tokens_seen": 1833872, "step": 6550 }, { "epoch": 72.83333333333333, "grad_norm": 3.5277712413517293e-06, "learning_rate": 4.676039625828107e-05, "loss": 0.0, "num_input_tokens_seen": 1835296, "step": 6555 }, { "epoch": 72.88888888888889, "grad_norm": 9.424409654457122e-06, "learning_rate": 4.675556126751328e-05, "loss": 0.0, "num_input_tokens_seen": 1836704, "step": 6560 }, { "epoch": 72.94444444444444, "grad_norm": 2.259171651530778e-06, "learning_rate": 4.6750722921764556e-05, "loss": 0.0, "num_input_tokens_seen": 1838112, "step": 6565 }, { "epoch": 73.0, "grad_norm": 2.983797912747832e-06, "learning_rate": 4.674588122178102e-05, "loss": 0.0, "num_input_tokens_seen": 1839536, "step": 6570 }, { "epoch": 73.05555555555556, "grad_norm": 1.6561521078983787e-06, "learning_rate": 4.674103616830931e-05, "loss": 0.0, "num_input_tokens_seen": 1840944, "step": 6575 }, { "epoch": 73.11111111111111, "grad_norm": 2.105256726281368e-06, "learning_rate": 4.673618776209663e-05, "loss": 0.0, "num_input_tokens_seen": 1842352, "step": 6580 }, { "epoch": 73.16666666666667, "grad_norm": 8.863483344612177e-06, "learning_rate": 4.673133600389063e-05, "loss": 0.0, "num_input_tokens_seen": 1843776, "step": 6585 }, { "epoch": 73.22222222222223, "grad_norm": 3.5973400827060686e-06, "learning_rate": 4.672648089443953e-05, "loss": 0.0, "num_input_tokens_seen": 1845152, "step": 6590 }, { "epoch": 73.27777777777777, "grad_norm": 1.747944043017924e-05, "learning_rate": 4.672162243449204e-05, "loss": 0.0, "num_input_tokens_seen": 1846512, "step": 6595 }, { "epoch": 73.33333333333333, "grad_norm": 3.3966823593800655e-06, "learning_rate": 4.67167606247974e-05, "loss": 0.0, "num_input_tokens_seen": 1847888, "step": 6600 }, { "epoch": 73.33333333333333, "eval_loss": 1.120741605758667, "eval_runtime": 1.4385, "eval_samples_per_second": 27.806, "eval_steps_per_second": 13.903, "num_input_tokens_seen": 1847888, "step": 6600 }, { "epoch": 73.38888888888889, "grad_norm": 1.0541858500801027e-05, "learning_rate": 4.671189546610536e-05, "loss": 0.0, "num_input_tokens_seen": 1849264, "step": 6605 }, { "epoch": 73.44444444444444, "grad_norm": 7.541529157606419e-06, "learning_rate": 4.67070269591662e-05, "loss": 0.0, "num_input_tokens_seen": 1850656, "step": 6610 }, { "epoch": 73.5, "grad_norm": 2.4518817554053385e-06, "learning_rate": 4.670215510473068e-05, "loss": 0.0, "num_input_tokens_seen": 1852064, "step": 6615 }, { "epoch": 73.55555555555556, "grad_norm": 3.867416580760619e-06, "learning_rate": 4.669727990355013e-05, "loss": 0.0, "num_input_tokens_seen": 1853472, "step": 6620 }, { "epoch": 73.61111111111111, "grad_norm": 2.2371768864104524e-06, "learning_rate": 4.669240135637635e-05, "loss": 0.0, "num_input_tokens_seen": 1854832, "step": 6625 }, { "epoch": 73.66666666666667, "grad_norm": 2.100163101204089e-06, "learning_rate": 4.6687519463961675e-05, "loss": 0.0, "num_input_tokens_seen": 1856224, "step": 6630 }, { "epoch": 73.72222222222223, "grad_norm": 4.333860033511883e-06, "learning_rate": 4.668263422705896e-05, "loss": 0.0, "num_input_tokens_seen": 1857648, "step": 6635 }, { "epoch": 73.77777777777777, "grad_norm": 5.0227795327373315e-06, "learning_rate": 4.667774564642156e-05, "loss": 0.0, "num_input_tokens_seen": 1859072, "step": 6640 }, { "epoch": 73.83333333333333, "grad_norm": 4.165625796304084e-06, "learning_rate": 4.6672853722803365e-05, "loss": 0.0, "num_input_tokens_seen": 1860496, "step": 6645 }, { "epoch": 73.88888888888889, "grad_norm": 3.419127551751444e-06, "learning_rate": 4.666795845695877e-05, "loss": 0.0, "num_input_tokens_seen": 1861920, "step": 6650 }, { "epoch": 73.94444444444444, "grad_norm": 6.047072929504793e-06, "learning_rate": 4.666305984964269e-05, "loss": 0.0, "num_input_tokens_seen": 1863344, "step": 6655 }, { "epoch": 74.0, "grad_norm": 3.2553541586821666e-06, "learning_rate": 4.6658157901610535e-05, "loss": 0.0, "num_input_tokens_seen": 1864752, "step": 6660 }, { "epoch": 74.05555555555556, "grad_norm": 7.399526111839805e-06, "learning_rate": 4.665325261361826e-05, "loss": 0.0, "num_input_tokens_seen": 1866176, "step": 6665 }, { "epoch": 74.11111111111111, "grad_norm": 3.4307993246329715e-06, "learning_rate": 4.664834398642232e-05, "loss": 0.0, "num_input_tokens_seen": 1867568, "step": 6670 }, { "epoch": 74.16666666666667, "grad_norm": 3.798533043664065e-06, "learning_rate": 4.6643432020779686e-05, "loss": 0.0, "num_input_tokens_seen": 1868928, "step": 6675 }, { "epoch": 74.22222222222223, "grad_norm": 2.5898587409756146e-05, "learning_rate": 4.663851671744786e-05, "loss": 0.0, "num_input_tokens_seen": 1870336, "step": 6680 }, { "epoch": 74.27777777777777, "grad_norm": 7.108214958861936e-06, "learning_rate": 4.6633598077184815e-05, "loss": 0.0, "num_input_tokens_seen": 1871744, "step": 6685 }, { "epoch": 74.33333333333333, "grad_norm": 1.5874262317083776e-05, "learning_rate": 4.662867610074908e-05, "loss": 0.0, "num_input_tokens_seen": 1873136, "step": 6690 }, { "epoch": 74.38888888888889, "grad_norm": 5.378125479182927e-06, "learning_rate": 4.6623750788899696e-05, "loss": 0.0, "num_input_tokens_seen": 1874528, "step": 6695 }, { "epoch": 74.44444444444444, "grad_norm": 2.4148864667949965e-06, "learning_rate": 4.6618822142396195e-05, "loss": 0.0, "num_input_tokens_seen": 1875888, "step": 6700 }, { "epoch": 74.5, "grad_norm": 4.4919506763108075e-06, "learning_rate": 4.661389016199864e-05, "loss": 0.0, "num_input_tokens_seen": 1877312, "step": 6705 }, { "epoch": 74.55555555555556, "grad_norm": 4.774757144332398e-06, "learning_rate": 4.660895484846761e-05, "loss": 0.0, "num_input_tokens_seen": 1878736, "step": 6710 }, { "epoch": 74.61111111111111, "grad_norm": 2.885868525481783e-06, "learning_rate": 4.660401620256418e-05, "loss": 0.0, "num_input_tokens_seen": 1880176, "step": 6715 }, { "epoch": 74.66666666666667, "grad_norm": 9.358501529277419e-07, "learning_rate": 4.659907422504997e-05, "loss": 0.0, "num_input_tokens_seen": 1881568, "step": 6720 }, { "epoch": 74.72222222222223, "grad_norm": 1.4170032955007628e-05, "learning_rate": 4.6594128916687074e-05, "loss": 0.0, "num_input_tokens_seen": 1882992, "step": 6725 }, { "epoch": 74.77777777777777, "grad_norm": 3.912676220352296e-06, "learning_rate": 4.658918027823813e-05, "loss": 0.0, "num_input_tokens_seen": 1884352, "step": 6730 }, { "epoch": 74.83333333333333, "grad_norm": 2.085383357552928e-06, "learning_rate": 4.658422831046628e-05, "loss": 0.0, "num_input_tokens_seen": 1885696, "step": 6735 }, { "epoch": 74.88888888888889, "grad_norm": 4.968853772879811e-06, "learning_rate": 4.657927301413518e-05, "loss": 0.0, "num_input_tokens_seen": 1887104, "step": 6740 }, { "epoch": 74.94444444444444, "grad_norm": 7.045335223665461e-06, "learning_rate": 4.657431439000901e-05, "loss": 0.0, "num_input_tokens_seen": 1888512, "step": 6745 }, { "epoch": 75.0, "grad_norm": 6.5611579884716775e-06, "learning_rate": 4.656935243885243e-05, "loss": 0.0, "num_input_tokens_seen": 1889904, "step": 6750 }, { "epoch": 75.05555555555556, "grad_norm": 4.146959327044897e-06, "learning_rate": 4.656438716143066e-05, "loss": 0.0, "num_input_tokens_seen": 1891312, "step": 6755 }, { "epoch": 75.11111111111111, "grad_norm": 3.1622309961676365e-06, "learning_rate": 4.6559418558509384e-05, "loss": 0.0, "num_input_tokens_seen": 1892704, "step": 6760 }, { "epoch": 75.16666666666667, "grad_norm": 5.956680524832336e-06, "learning_rate": 4.6554446630854833e-05, "loss": 0.0, "num_input_tokens_seen": 1894112, "step": 6765 }, { "epoch": 75.22222222222223, "grad_norm": 3.2820255000842735e-05, "learning_rate": 4.654947137923374e-05, "loss": 0.0, "num_input_tokens_seen": 1895552, "step": 6770 }, { "epoch": 75.27777777777777, "grad_norm": 1.0406397450424265e-05, "learning_rate": 4.654449280441335e-05, "loss": 0.0, "num_input_tokens_seen": 1896992, "step": 6775 }, { "epoch": 75.33333333333333, "grad_norm": 2.1124408249306725e-06, "learning_rate": 4.653951090716143e-05, "loss": 0.0, "num_input_tokens_seen": 1898384, "step": 6780 }, { "epoch": 75.38888888888889, "grad_norm": 3.5625614600576228e-06, "learning_rate": 4.653452568824625e-05, "loss": 0.0, "num_input_tokens_seen": 1899776, "step": 6785 }, { "epoch": 75.44444444444444, "grad_norm": 3.908101916749729e-06, "learning_rate": 4.6529537148436585e-05, "loss": 0.0, "num_input_tokens_seen": 1901168, "step": 6790 }, { "epoch": 75.5, "grad_norm": 6.277057309489464e-06, "learning_rate": 4.6524545288501734e-05, "loss": 0.0, "num_input_tokens_seen": 1902592, "step": 6795 }, { "epoch": 75.55555555555556, "grad_norm": 1.557341647639987e-06, "learning_rate": 4.6519550109211506e-05, "loss": 0.0, "num_input_tokens_seen": 1904000, "step": 6800 }, { "epoch": 75.55555555555556, "eval_loss": 1.1112022399902344, "eval_runtime": 1.4019, "eval_samples_per_second": 28.532, "eval_steps_per_second": 14.266, "num_input_tokens_seen": 1904000, "step": 6800 }, { "epoch": 75.61111111111111, "grad_norm": 4.86725548398681e-06, "learning_rate": 4.651455161133622e-05, "loss": 0.0, "num_input_tokens_seen": 1905392, "step": 6805 }, { "epoch": 75.66666666666667, "grad_norm": 1.4184391147864517e-05, "learning_rate": 4.6509549795646704e-05, "loss": 0.0, "num_input_tokens_seen": 1906736, "step": 6810 }, { "epoch": 75.72222222222223, "grad_norm": 3.573838739612256e-06, "learning_rate": 4.6504544662914306e-05, "loss": 0.0, "num_input_tokens_seen": 1908128, "step": 6815 }, { "epoch": 75.77777777777777, "grad_norm": 6.873994607303757e-06, "learning_rate": 4.6499536213910876e-05, "loss": 0.0, "num_input_tokens_seen": 1909552, "step": 6820 }, { "epoch": 75.83333333333333, "grad_norm": 2.757822130661225e-06, "learning_rate": 4.6494524449408786e-05, "loss": 0.0, "num_input_tokens_seen": 1910960, "step": 6825 }, { "epoch": 75.88888888888889, "grad_norm": 6.3395159486390185e-06, "learning_rate": 4.6489509370180903e-05, "loss": 0.0, "num_input_tokens_seen": 1912320, "step": 6830 }, { "epoch": 75.94444444444444, "grad_norm": 9.01245221029967e-06, "learning_rate": 4.648449097700063e-05, "loss": 0.0, "num_input_tokens_seen": 1913744, "step": 6835 }, { "epoch": 76.0, "grad_norm": 3.98884185415227e-06, "learning_rate": 4.647946927064185e-05, "loss": 0.0, "num_input_tokens_seen": 1915072, "step": 6840 }, { "epoch": 76.05555555555556, "grad_norm": 7.5548337008513045e-06, "learning_rate": 4.647444425187898e-05, "loss": 0.0, "num_input_tokens_seen": 1916480, "step": 6845 }, { "epoch": 76.11111111111111, "grad_norm": 2.062452267637127e-06, "learning_rate": 4.646941592148695e-05, "loss": 0.0, "num_input_tokens_seen": 1917888, "step": 6850 }, { "epoch": 76.16666666666667, "grad_norm": 3.5793668757833075e-06, "learning_rate": 4.646438428024117e-05, "loss": 0.0, "num_input_tokens_seen": 1919216, "step": 6855 }, { "epoch": 76.22222222222223, "grad_norm": 9.757103725860361e-06, "learning_rate": 4.64593493289176e-05, "loss": 0.0, "num_input_tokens_seen": 1920624, "step": 6860 }, { "epoch": 76.27777777777777, "grad_norm": 3.977271717303665e-06, "learning_rate": 4.64543110682927e-05, "loss": 0.0, "num_input_tokens_seen": 1922016, "step": 6865 }, { "epoch": 76.33333333333333, "grad_norm": 5.857990345248254e-06, "learning_rate": 4.644926949914341e-05, "loss": 0.0, "num_input_tokens_seen": 1923488, "step": 6870 }, { "epoch": 76.38888888888889, "grad_norm": 2.579110287115327e-06, "learning_rate": 4.644422462224722e-05, "loss": 0.0, "num_input_tokens_seen": 1924880, "step": 6875 }, { "epoch": 76.44444444444444, "grad_norm": 1.5521416571573354e-05, "learning_rate": 4.643917643838211e-05, "loss": 0.0, "num_input_tokens_seen": 1926288, "step": 6880 }, { "epoch": 76.5, "grad_norm": 2.151751459678053e-06, "learning_rate": 4.6434124948326564e-05, "loss": 0.0, "num_input_tokens_seen": 1927616, "step": 6885 }, { "epoch": 76.55555555555556, "grad_norm": 1.7623007124711876e-06, "learning_rate": 4.6429070152859594e-05, "loss": 0.0, "num_input_tokens_seen": 1929056, "step": 6890 }, { "epoch": 76.61111111111111, "grad_norm": 2.1632006337313214e-06, "learning_rate": 4.6424012052760714e-05, "loss": 0.0, "num_input_tokens_seen": 1930400, "step": 6895 }, { "epoch": 76.66666666666667, "grad_norm": 2.6474724563740892e-06, "learning_rate": 4.6418950648809945e-05, "loss": 0.0, "num_input_tokens_seen": 1931840, "step": 6900 }, { "epoch": 76.72222222222223, "grad_norm": 2.3492791569879046e-06, "learning_rate": 4.641388594178782e-05, "loss": 0.0, "num_input_tokens_seen": 1933216, "step": 6905 }, { "epoch": 76.77777777777777, "grad_norm": 8.235198947659228e-06, "learning_rate": 4.640881793247538e-05, "loss": 0.0, "num_input_tokens_seen": 1934640, "step": 6910 }, { "epoch": 76.83333333333333, "grad_norm": 5.9797744142997544e-06, "learning_rate": 4.6403746621654173e-05, "loss": 0.0, "num_input_tokens_seen": 1936000, "step": 6915 }, { "epoch": 76.88888888888889, "grad_norm": 2.2416354113374837e-05, "learning_rate": 4.639867201010626e-05, "loss": 0.0, "num_input_tokens_seen": 1937392, "step": 6920 }, { "epoch": 76.94444444444444, "grad_norm": 3.2800510325614596e-06, "learning_rate": 4.6393594098614204e-05, "loss": 0.0, "num_input_tokens_seen": 1938832, "step": 6925 }, { "epoch": 77.0, "grad_norm": 4.024130248581059e-06, "learning_rate": 4.63885128879611e-05, "loss": 0.0, "num_input_tokens_seen": 1940288, "step": 6930 }, { "epoch": 77.05555555555556, "grad_norm": 6.541678885696456e-06, "learning_rate": 4.638342837893052e-05, "loss": 0.0, "num_input_tokens_seen": 1941728, "step": 6935 }, { "epoch": 77.11111111111111, "grad_norm": 1.0185656719841063e-05, "learning_rate": 4.6378340572306565e-05, "loss": 0.0, "num_input_tokens_seen": 1943152, "step": 6940 }, { "epoch": 77.16666666666667, "grad_norm": 3.832848960882984e-06, "learning_rate": 4.6373249468873833e-05, "loss": 0.0, "num_input_tokens_seen": 1944560, "step": 6945 }, { "epoch": 77.22222222222223, "grad_norm": 2.4873324946383946e-06, "learning_rate": 4.636815506941744e-05, "loss": 0.0, "num_input_tokens_seen": 1945952, "step": 6950 }, { "epoch": 77.27777777777777, "grad_norm": 3.0850808343529934e-06, "learning_rate": 4.6363057374723004e-05, "loss": 0.0, "num_input_tokens_seen": 1947312, "step": 6955 }, { "epoch": 77.33333333333333, "grad_norm": 2.631968163768761e-05, "learning_rate": 4.635795638557666e-05, "loss": 0.0, "num_input_tokens_seen": 1948704, "step": 6960 }, { "epoch": 77.38888888888889, "grad_norm": 6.882508387207054e-06, "learning_rate": 4.635285210276504e-05, "loss": 0.0, "num_input_tokens_seen": 1950128, "step": 6965 }, { "epoch": 77.44444444444444, "grad_norm": 5.427695668913657e-06, "learning_rate": 4.6347744527075295e-05, "loss": 0.0, "num_input_tokens_seen": 1951568, "step": 6970 }, { "epoch": 77.5, "grad_norm": 4.07666584578692e-06, "learning_rate": 4.634263365929506e-05, "loss": 0.0, "num_input_tokens_seen": 1953008, "step": 6975 }, { "epoch": 77.55555555555556, "grad_norm": 3.429060598136857e-06, "learning_rate": 4.6337519500212515e-05, "loss": 0.0, "num_input_tokens_seen": 1954368, "step": 6980 }, { "epoch": 77.61111111111111, "grad_norm": 4.974131115886848e-06, "learning_rate": 4.633240205061632e-05, "loss": 0.0, "num_input_tokens_seen": 1955728, "step": 6985 }, { "epoch": 77.66666666666667, "grad_norm": 8.42720692162402e-06, "learning_rate": 4.632728131129565e-05, "loss": 0.0, "num_input_tokens_seen": 1957104, "step": 6990 }, { "epoch": 77.72222222222223, "grad_norm": 3.937220299121691e-06, "learning_rate": 4.632215728304018e-05, "loss": 0.0, "num_input_tokens_seen": 1958528, "step": 6995 }, { "epoch": 77.77777777777777, "grad_norm": 3.4167012472607894e-06, "learning_rate": 4.63170299666401e-05, "loss": 0.0, "num_input_tokens_seen": 1959920, "step": 7000 }, { "epoch": 77.77777777777777, "eval_loss": 1.123112678527832, "eval_runtime": 1.381, "eval_samples_per_second": 28.964, "eval_steps_per_second": 14.482, "num_input_tokens_seen": 1959920, "step": 7000 }, { "epoch": 77.83333333333333, "grad_norm": 2.9423779324133648e-06, "learning_rate": 4.631189936288612e-05, "loss": 0.0, "num_input_tokens_seen": 1961328, "step": 7005 }, { "epoch": 77.88888888888889, "grad_norm": 2.093798912028433e-06, "learning_rate": 4.630676547256944e-05, "loss": 0.0, "num_input_tokens_seen": 1962720, "step": 7010 }, { "epoch": 77.94444444444444, "grad_norm": 3.503969310258981e-06, "learning_rate": 4.630162829648176e-05, "loss": 0.0, "num_input_tokens_seen": 1964064, "step": 7015 }, { "epoch": 78.0, "grad_norm": 3.3793139664339833e-06, "learning_rate": 4.629648783541531e-05, "loss": 0.0, "num_input_tokens_seen": 1965472, "step": 7020 }, { "epoch": 78.05555555555556, "grad_norm": 1.7695807400741614e-05, "learning_rate": 4.6291344090162804e-05, "loss": 0.0, "num_input_tokens_seen": 1966896, "step": 7025 }, { "epoch": 78.11111111111111, "grad_norm": 1.9055596567341127e-06, "learning_rate": 4.628619706151748e-05, "loss": 0.0, "num_input_tokens_seen": 1968272, "step": 7030 }, { "epoch": 78.16666666666667, "grad_norm": 8.032633559196256e-06, "learning_rate": 4.628104675027306e-05, "loss": 0.0, "num_input_tokens_seen": 1969696, "step": 7035 }, { "epoch": 78.22222222222223, "grad_norm": 8.16738247522153e-06, "learning_rate": 4.6275893157223805e-05, "loss": 0.0, "num_input_tokens_seen": 1971136, "step": 7040 }, { "epoch": 78.27777777777777, "grad_norm": 2.501053131709341e-06, "learning_rate": 4.627073628316445e-05, "loss": 0.0, "num_input_tokens_seen": 1972528, "step": 7045 }, { "epoch": 78.33333333333333, "grad_norm": 1.0307686352462042e-05, "learning_rate": 4.626557612889026e-05, "loss": 0.0, "num_input_tokens_seen": 1973920, "step": 7050 }, { "epoch": 78.38888888888889, "grad_norm": 3.4986780974577414e-06, "learning_rate": 4.626041269519699e-05, "loss": 0.0, "num_input_tokens_seen": 1975296, "step": 7055 }, { "epoch": 78.44444444444444, "grad_norm": 5.737283117923653e-06, "learning_rate": 4.6255245982880905e-05, "loss": 0.0, "num_input_tokens_seen": 1976656, "step": 7060 }, { "epoch": 78.5, "grad_norm": 7.647437996638473e-06, "learning_rate": 4.625007599273879e-05, "loss": 0.0, "num_input_tokens_seen": 1978048, "step": 7065 }, { "epoch": 78.55555555555556, "grad_norm": 3.06545462080976e-06, "learning_rate": 4.6244902725567895e-05, "loss": 0.0, "num_input_tokens_seen": 1979408, "step": 7070 }, { "epoch": 78.61111111111111, "grad_norm": 4.954408268531552e-06, "learning_rate": 4.6239726182166024e-05, "loss": 0.0, "num_input_tokens_seen": 1980800, "step": 7075 }, { "epoch": 78.66666666666667, "grad_norm": 2.7952373784501106e-06, "learning_rate": 4.623454636333147e-05, "loss": 0.0, "num_input_tokens_seen": 1982176, "step": 7080 }, { "epoch": 78.72222222222223, "grad_norm": 1.5987324104571599e-06, "learning_rate": 4.622936326986301e-05, "loss": 0.0, "num_input_tokens_seen": 1983584, "step": 7085 }, { "epoch": 78.77777777777777, "grad_norm": 2.8457432108552894e-06, "learning_rate": 4.6224176902559946e-05, "loss": 0.0, "num_input_tokens_seen": 1984976, "step": 7090 }, { "epoch": 78.83333333333333, "grad_norm": 1.7294104281972977e-06, "learning_rate": 4.621898726222209e-05, "loss": 0.0, "num_input_tokens_seen": 1986384, "step": 7095 }, { "epoch": 78.88888888888889, "grad_norm": 5.637269623548491e-06, "learning_rate": 4.6213794349649744e-05, "loss": 0.0, "num_input_tokens_seen": 1987792, "step": 7100 }, { "epoch": 78.94444444444444, "grad_norm": 5.221773335506441e-06, "learning_rate": 4.6208598165643715e-05, "loss": 0.0, "num_input_tokens_seen": 1989216, "step": 7105 }, { "epoch": 79.0, "grad_norm": 2.038209277088754e-05, "learning_rate": 4.620339871100533e-05, "loss": 0.0, "num_input_tokens_seen": 1990624, "step": 7110 }, { "epoch": 79.05555555555556, "grad_norm": 8.227275429817382e-06, "learning_rate": 4.6198195986536394e-05, "loss": 0.0, "num_input_tokens_seen": 1992048, "step": 7115 }, { "epoch": 79.11111111111111, "grad_norm": 2.5691708742670016e-06, "learning_rate": 4.619298999303926e-05, "loss": 0.0, "num_input_tokens_seen": 1993440, "step": 7120 }, { "epoch": 79.16666666666667, "grad_norm": 1.568726133882592e-06, "learning_rate": 4.618778073131673e-05, "loss": 0.0, "num_input_tokens_seen": 1994816, "step": 7125 }, { "epoch": 79.22222222222223, "grad_norm": 2.0589945052051917e-05, "learning_rate": 4.618256820217215e-05, "loss": 0.0, "num_input_tokens_seen": 1996224, "step": 7130 }, { "epoch": 79.27777777777777, "grad_norm": 4.537449513009051e-06, "learning_rate": 4.617735240640936e-05, "loss": 0.0, "num_input_tokens_seen": 1997616, "step": 7135 }, { "epoch": 79.33333333333333, "grad_norm": 2.748799670371227e-05, "learning_rate": 4.6172133344832705e-05, "loss": 0.0, "num_input_tokens_seen": 1999008, "step": 7140 }, { "epoch": 79.38888888888889, "grad_norm": 1.7730097169987857e-05, "learning_rate": 4.6166911018247004e-05, "loss": 0.0, "num_input_tokens_seen": 2000368, "step": 7145 }, { "epoch": 79.44444444444444, "grad_norm": 2.237923126813257e-06, "learning_rate": 4.616168542745764e-05, "loss": 0.0, "num_input_tokens_seen": 2001824, "step": 7150 }, { "epoch": 79.5, "grad_norm": 2.5022111458383733e-06, "learning_rate": 4.6156456573270446e-05, "loss": 0.0, "num_input_tokens_seen": 2003200, "step": 7155 }, { "epoch": 79.55555555555556, "grad_norm": 4.092390554433223e-06, "learning_rate": 4.615122445649177e-05, "loss": 0.0, "num_input_tokens_seen": 2004576, "step": 7160 }, { "epoch": 79.61111111111111, "grad_norm": 2.9428831567201996e-06, "learning_rate": 4.6145989077928486e-05, "loss": 0.0, "num_input_tokens_seen": 2005952, "step": 7165 }, { "epoch": 79.66666666666667, "grad_norm": 6.306353043328272e-06, "learning_rate": 4.6140750438387953e-05, "loss": 0.0, "num_input_tokens_seen": 2007344, "step": 7170 }, { "epoch": 79.72222222222223, "grad_norm": 6.530132395710098e-06, "learning_rate": 4.613550853867803e-05, "loss": 0.0, "num_input_tokens_seen": 2008784, "step": 7175 }, { "epoch": 79.77777777777777, "grad_norm": 4.184652880212525e-06, "learning_rate": 4.613026337960708e-05, "loss": 0.0, "num_input_tokens_seen": 2010208, "step": 7180 }, { "epoch": 79.83333333333333, "grad_norm": 3.487319190753624e-05, "learning_rate": 4.612501496198398e-05, "loss": 0.0, "num_input_tokens_seen": 2011568, "step": 7185 }, { "epoch": 79.88888888888889, "grad_norm": 7.389072834484978e-06, "learning_rate": 4.61197632866181e-05, "loss": 0.0, "num_input_tokens_seen": 2012976, "step": 7190 }, { "epoch": 79.94444444444444, "grad_norm": 5.1143329073966015e-06, "learning_rate": 4.611450835431931e-05, "loss": 0.0, "num_input_tokens_seen": 2014384, "step": 7195 }, { "epoch": 80.0, "grad_norm": 3.9621668292966206e-06, "learning_rate": 4.6109250165898e-05, "loss": 0.0, "num_input_tokens_seen": 2015792, "step": 7200 }, { "epoch": 80.0, "eval_loss": 1.1283981800079346, "eval_runtime": 1.3872, "eval_samples_per_second": 28.835, "eval_steps_per_second": 14.417, "num_input_tokens_seen": 2015792, "step": 7200 }, { "epoch": 80.05555555555556, "grad_norm": 4.279133463569451e-06, "learning_rate": 4.610398872216503e-05, "loss": 0.0, "num_input_tokens_seen": 2017184, "step": 7205 }, { "epoch": 80.11111111111111, "grad_norm": 1.786712800821988e-06, "learning_rate": 4.6098724023931796e-05, "loss": 0.0, "num_input_tokens_seen": 2018592, "step": 7210 }, { "epoch": 80.16666666666667, "grad_norm": 1.646428017920698e-06, "learning_rate": 4.609345607201017e-05, "loss": 0.0, "num_input_tokens_seen": 2020000, "step": 7215 }, { "epoch": 80.22222222222223, "grad_norm": 5.678210527548799e-06, "learning_rate": 4.608818486721254e-05, "loss": 0.0, "num_input_tokens_seen": 2021440, "step": 7220 }, { "epoch": 80.27777777777777, "grad_norm": 3.2700672818464227e-06, "learning_rate": 4.608291041035179e-05, "loss": 0.0, "num_input_tokens_seen": 2022832, "step": 7225 }, { "epoch": 80.33333333333333, "grad_norm": 3.245453626732342e-06, "learning_rate": 4.607763270224132e-05, "loss": 0.0, "num_input_tokens_seen": 2024224, "step": 7230 }, { "epoch": 80.38888888888889, "grad_norm": 4.8731752031017095e-05, "learning_rate": 4.6072351743695e-05, "loss": 0.0, "num_input_tokens_seen": 2025632, "step": 7235 }, { "epoch": 80.44444444444444, "grad_norm": 4.658005764213158e-06, "learning_rate": 4.606706753552723e-05, "loss": 0.0, "num_input_tokens_seen": 2027008, "step": 7240 }, { "epoch": 80.5, "grad_norm": 1.7337326880806359e-06, "learning_rate": 4.6061780078552906e-05, "loss": 0.0, "num_input_tokens_seen": 2028448, "step": 7245 }, { "epoch": 80.55555555555556, "grad_norm": 1.5419589090015506e-06, "learning_rate": 4.605648937358742e-05, "loss": 0.0, "num_input_tokens_seen": 2029824, "step": 7250 }, { "epoch": 80.61111111111111, "grad_norm": 1.1407158126530703e-05, "learning_rate": 4.605119542144665e-05, "loss": 0.0, "num_input_tokens_seen": 2031200, "step": 7255 }, { "epoch": 80.66666666666667, "grad_norm": 5.3834542086406145e-06, "learning_rate": 4.604589822294701e-05, "loss": 0.0, "num_input_tokens_seen": 2032624, "step": 7260 }, { "epoch": 80.72222222222223, "grad_norm": 8.69825544214109e-06, "learning_rate": 4.604059777890537e-05, "loss": 0.0, "num_input_tokens_seen": 2034016, "step": 7265 }, { "epoch": 80.77777777777777, "grad_norm": 7.0548817348026205e-06, "learning_rate": 4.6035294090139145e-05, "loss": 0.0, "num_input_tokens_seen": 2035440, "step": 7270 }, { "epoch": 80.83333333333333, "grad_norm": 1.866008574324951e-06, "learning_rate": 4.6029987157466226e-05, "loss": 0.0, "num_input_tokens_seen": 2036832, "step": 7275 }, { "epoch": 80.88888888888889, "grad_norm": 2.793278554236167e-06, "learning_rate": 4.602467698170502e-05, "loss": 0.0, "num_input_tokens_seen": 2038192, "step": 7280 }, { "epoch": 80.94444444444444, "grad_norm": 9.377574315294623e-06, "learning_rate": 4.601936356367439e-05, "loss": 0.0, "num_input_tokens_seen": 2039584, "step": 7285 }, { "epoch": 81.0, "grad_norm": 7.094791726558469e-06, "learning_rate": 4.601404690419377e-05, "loss": 0.0, "num_input_tokens_seen": 2040992, "step": 7290 }, { "epoch": 81.05555555555556, "grad_norm": 3.491511961328797e-05, "learning_rate": 4.600872700408303e-05, "loss": 0.0, "num_input_tokens_seen": 2042352, "step": 7295 }, { "epoch": 81.11111111111111, "grad_norm": 3.5980094708065735e-06, "learning_rate": 4.600340386416258e-05, "loss": 0.0, "num_input_tokens_seen": 2043776, "step": 7300 }, { "epoch": 81.16666666666667, "grad_norm": 1.9570320546336006e-06, "learning_rate": 4.5998077485253296e-05, "loss": 0.0, "num_input_tokens_seen": 2045168, "step": 7305 }, { "epoch": 81.22222222222223, "grad_norm": 3.639176156866597e-06, "learning_rate": 4.59927478681766e-05, "loss": 0.0, "num_input_tokens_seen": 2046592, "step": 7310 }, { "epoch": 81.27777777777777, "grad_norm": 2.8308152195677394e-06, "learning_rate": 4.5987415013754366e-05, "loss": 0.0, "num_input_tokens_seen": 2047984, "step": 7315 }, { "epoch": 81.33333333333333, "grad_norm": 7.776557140459772e-06, "learning_rate": 4.598207892280899e-05, "loss": 0.0, "num_input_tokens_seen": 2049408, "step": 7320 }, { "epoch": 81.38888888888889, "grad_norm": 2.5186884613503935e-06, "learning_rate": 4.597673959616337e-05, "loss": 0.0, "num_input_tokens_seen": 2050800, "step": 7325 }, { "epoch": 81.44444444444444, "grad_norm": 2.8272650069993688e-06, "learning_rate": 4.597139703464089e-05, "loss": 0.0, "num_input_tokens_seen": 2052224, "step": 7330 }, { "epoch": 81.5, "grad_norm": 5.435075308923842e-06, "learning_rate": 4.596605123906545e-05, "loss": 0.0, "num_input_tokens_seen": 2053616, "step": 7335 }, { "epoch": 81.55555555555556, "grad_norm": 2.125450464518508e-06, "learning_rate": 4.596070221026143e-05, "loss": 0.0, "num_input_tokens_seen": 2055008, "step": 7340 }, { "epoch": 81.61111111111111, "grad_norm": 2.166146487070364e-06, "learning_rate": 4.595534994905372e-05, "loss": 0.0, "num_input_tokens_seen": 2056416, "step": 7345 }, { "epoch": 81.66666666666667, "grad_norm": 2.739683395702741e-06, "learning_rate": 4.594999445626771e-05, "loss": 0.0, "num_input_tokens_seen": 2057856, "step": 7350 }, { "epoch": 81.72222222222223, "grad_norm": 5.6255880735989194e-06, "learning_rate": 4.5944635732729276e-05, "loss": 0.0, "num_input_tokens_seen": 2059248, "step": 7355 }, { "epoch": 81.77777777777777, "grad_norm": 1.1751554666261654e-05, "learning_rate": 4.5939273779264804e-05, "loss": 0.0, "num_input_tokens_seen": 2060624, "step": 7360 }, { "epoch": 81.83333333333333, "grad_norm": 3.859769549308112e-06, "learning_rate": 4.593390859670118e-05, "loss": 0.0, "num_input_tokens_seen": 2062000, "step": 7365 }, { "epoch": 81.88888888888889, "grad_norm": 7.22139793651877e-06, "learning_rate": 4.5928540185865776e-05, "loss": 0.0, "num_input_tokens_seen": 2063408, "step": 7370 }, { "epoch": 81.94444444444444, "grad_norm": 1.0381295396655332e-05, "learning_rate": 4.592316854758648e-05, "loss": 0.0, "num_input_tokens_seen": 2064816, "step": 7375 }, { "epoch": 82.0, "grad_norm": 2.883530896724551e-06, "learning_rate": 4.5917793682691646e-05, "loss": 0.0, "num_input_tokens_seen": 2066240, "step": 7380 }, { "epoch": 82.05555555555556, "grad_norm": 1.0861645023396704e-05, "learning_rate": 4.5912415592010164e-05, "loss": 0.0, "num_input_tokens_seen": 2067632, "step": 7385 }, { "epoch": 82.11111111111111, "grad_norm": 1.6677399798936676e-06, "learning_rate": 4.5907034276371386e-05, "loss": 0.0, "num_input_tokens_seen": 2069008, "step": 7390 }, { "epoch": 82.16666666666667, "grad_norm": 3.6751575862581376e-06, "learning_rate": 4.5901649736605196e-05, "loss": 0.0, "num_input_tokens_seen": 2070400, "step": 7395 }, { "epoch": 82.22222222222223, "grad_norm": 7.798978003847878e-06, "learning_rate": 4.589626197354195e-05, "loss": 0.0, "num_input_tokens_seen": 2071808, "step": 7400 }, { "epoch": 82.22222222222223, "eval_loss": 1.1409505605697632, "eval_runtime": 1.3837, "eval_samples_per_second": 28.908, "eval_steps_per_second": 14.454, "num_input_tokens_seen": 2071808, "step": 7400 }, { "epoch": 82.27777777777777, "grad_norm": 2.241701804450713e-05, "learning_rate": 4.5890870988012504e-05, "loss": 0.0, "num_input_tokens_seen": 2073216, "step": 7405 }, { "epoch": 82.33333333333333, "grad_norm": 2.0419340216903947e-05, "learning_rate": 4.5885476780848226e-05, "loss": 0.0, "num_input_tokens_seen": 2074624, "step": 7410 }, { "epoch": 82.38888888888889, "grad_norm": 1.028066321850929e-06, "learning_rate": 4.5880079352880964e-05, "loss": 0.0, "num_input_tokens_seen": 2076000, "step": 7415 }, { "epoch": 82.44444444444444, "grad_norm": 2.3343059183389414e-06, "learning_rate": 4.5874678704943065e-05, "loss": 0.0, "num_input_tokens_seen": 2077424, "step": 7420 }, { "epoch": 82.5, "grad_norm": 4.242727300152183e-06, "learning_rate": 4.5869274837867394e-05, "loss": 0.0, "num_input_tokens_seen": 2078832, "step": 7425 }, { "epoch": 82.55555555555556, "grad_norm": 3.021310294570867e-05, "learning_rate": 4.5863867752487275e-05, "loss": 0.0, "num_input_tokens_seen": 2080240, "step": 7430 }, { "epoch": 82.61111111111111, "grad_norm": 7.864628059905954e-06, "learning_rate": 4.5858457449636554e-05, "loss": 0.0, "num_input_tokens_seen": 2081632, "step": 7435 }, { "epoch": 82.66666666666667, "grad_norm": 1.3928366570326034e-05, "learning_rate": 4.5853043930149574e-05, "loss": 0.0, "num_input_tokens_seen": 2083056, "step": 7440 }, { "epoch": 82.72222222222223, "grad_norm": 1.8484398651708034e-06, "learning_rate": 4.584762719486117e-05, "loss": 0.0, "num_input_tokens_seen": 2084416, "step": 7445 }, { "epoch": 82.77777777777777, "grad_norm": 6.229451173567213e-06, "learning_rate": 4.584220724460665e-05, "loss": 0.0, "num_input_tokens_seen": 2085808, "step": 7450 }, { "epoch": 82.83333333333333, "grad_norm": 3.0339845125126885e-06, "learning_rate": 4.5836784080221865e-05, "loss": 0.0, "num_input_tokens_seen": 2087216, "step": 7455 }, { "epoch": 82.88888888888889, "grad_norm": 2.5269348498113686e-06, "learning_rate": 4.583135770254312e-05, "loss": 0.0, "num_input_tokens_seen": 2088544, "step": 7460 }, { "epoch": 82.94444444444444, "grad_norm": 2.181793661293341e-06, "learning_rate": 4.5825928112407236e-05, "loss": 0.0, "num_input_tokens_seen": 2089920, "step": 7465 }, { "epoch": 83.0, "grad_norm": 1.5047896795294946e-06, "learning_rate": 4.582049531065152e-05, "loss": 0.0, "num_input_tokens_seen": 2091376, "step": 7470 }, { "epoch": 83.05555555555556, "grad_norm": 4.0551508391217794e-06, "learning_rate": 4.5815059298113783e-05, "loss": 0.0, "num_input_tokens_seen": 2092784, "step": 7475 }, { "epoch": 83.11111111111111, "grad_norm": 1.112953668780392e-05, "learning_rate": 4.580962007563232e-05, "loss": 0.0, "num_input_tokens_seen": 2094192, "step": 7480 }, { "epoch": 83.16666666666667, "grad_norm": 2.008862111324561e-06, "learning_rate": 4.5804177644045935e-05, "loss": 0.0, "num_input_tokens_seen": 2095568, "step": 7485 }, { "epoch": 83.22222222222223, "grad_norm": 1.7126218381235958e-06, "learning_rate": 4.579873200419391e-05, "loss": 0.0, "num_input_tokens_seen": 2096960, "step": 7490 }, { "epoch": 83.27777777777777, "grad_norm": 1.4811833352723625e-05, "learning_rate": 4.5793283156916046e-05, "loss": 0.0, "num_input_tokens_seen": 2098320, "step": 7495 }, { "epoch": 83.33333333333333, "grad_norm": 4.650085884350119e-06, "learning_rate": 4.578783110305261e-05, "loss": 0.0, "num_input_tokens_seen": 2099744, "step": 7500 }, { "epoch": 83.38888888888889, "grad_norm": 1.1513143363117706e-06, "learning_rate": 4.578237584344438e-05, "loss": 0.0, "num_input_tokens_seen": 2101168, "step": 7505 }, { "epoch": 83.44444444444444, "grad_norm": 1.8193178448200342e-06, "learning_rate": 4.577691737893263e-05, "loss": 0.0, "num_input_tokens_seen": 2102576, "step": 7510 }, { "epoch": 83.5, "grad_norm": 9.857463737716898e-06, "learning_rate": 4.577145571035912e-05, "loss": 0.0, "num_input_tokens_seen": 2103984, "step": 7515 }, { "epoch": 83.55555555555556, "grad_norm": 2.4549499357817695e-06, "learning_rate": 4.576599083856611e-05, "loss": 0.0, "num_input_tokens_seen": 2105392, "step": 7520 }, { "epoch": 83.61111111111111, "grad_norm": 1.9384320694371127e-05, "learning_rate": 4.576052276439635e-05, "loss": 0.0, "num_input_tokens_seen": 2106784, "step": 7525 }, { "epoch": 83.66666666666667, "grad_norm": 1.4623132074120804e-06, "learning_rate": 4.575505148869308e-05, "loss": 0.0, "num_input_tokens_seen": 2108128, "step": 7530 }, { "epoch": 83.72222222222223, "grad_norm": 1.6734223891035072e-06, "learning_rate": 4.574957701230006e-05, "loss": 0.0, "num_input_tokens_seen": 2109536, "step": 7535 }, { "epoch": 83.77777777777777, "grad_norm": 2.9149441616027616e-06, "learning_rate": 4.57440993360615e-05, "loss": 0.0, "num_input_tokens_seen": 2110944, "step": 7540 }, { "epoch": 83.83333333333333, "grad_norm": 3.0759465516894124e-06, "learning_rate": 4.5738618460822134e-05, "loss": 0.0, "num_input_tokens_seen": 2112368, "step": 7545 }, { "epoch": 83.88888888888889, "grad_norm": 2.37113272305578e-06, "learning_rate": 4.573313438742719e-05, "loss": 0.0, "num_input_tokens_seen": 2113792, "step": 7550 }, { "epoch": 83.94444444444444, "grad_norm": 1.2425938393789693e-06, "learning_rate": 4.5727647116722374e-05, "loss": 0.0, "num_input_tokens_seen": 2115168, "step": 7555 }, { "epoch": 84.0, "grad_norm": 4.26111137130647e-06, "learning_rate": 4.5722156649553884e-05, "loss": 0.0, "num_input_tokens_seen": 2116592, "step": 7560 }, { "epoch": 84.05555555555556, "grad_norm": 2.8271890641917707e-06, "learning_rate": 4.571666298676843e-05, "loss": 0.0, "num_input_tokens_seen": 2117984, "step": 7565 }, { "epoch": 84.11111111111111, "grad_norm": 2.4774587927822722e-06, "learning_rate": 4.571116612921321e-05, "loss": 0.0, "num_input_tokens_seen": 2119392, "step": 7570 }, { "epoch": 84.16666666666667, "grad_norm": 5.449242053146008e-06, "learning_rate": 4.57056660777359e-05, "loss": 0.0, "num_input_tokens_seen": 2120800, "step": 7575 }, { "epoch": 84.22222222222223, "grad_norm": 2.3030848751659505e-05, "learning_rate": 4.5700162833184666e-05, "loss": 0.0, "num_input_tokens_seen": 2122192, "step": 7580 }, { "epoch": 84.27777777777777, "grad_norm": 5.533569037652342e-06, "learning_rate": 4.5694656396408195e-05, "loss": 0.0, "num_input_tokens_seen": 2123600, "step": 7585 }, { "epoch": 84.33333333333333, "grad_norm": 2.098036748066079e-05, "learning_rate": 4.5689146768255646e-05, "loss": 0.0, "num_input_tokens_seen": 2125056, "step": 7590 }, { "epoch": 84.38888888888889, "grad_norm": 3.4312790830881568e-06, "learning_rate": 4.568363394957667e-05, "loss": 0.0, "num_input_tokens_seen": 2126448, "step": 7595 }, { "epoch": 84.44444444444444, "grad_norm": 1.126667939388426e-05, "learning_rate": 4.567811794122141e-05, "loss": 0.0, "num_input_tokens_seen": 2127808, "step": 7600 }, { "epoch": 84.44444444444444, "eval_loss": 1.1386152505874634, "eval_runtime": 1.3872, "eval_samples_per_second": 28.836, "eval_steps_per_second": 14.418, "num_input_tokens_seen": 2127808, "step": 7600 }, { "epoch": 84.5, "grad_norm": 1.7455688521295087e-06, "learning_rate": 4.56725987440405e-05, "loss": 0.0, "num_input_tokens_seen": 2129248, "step": 7605 }, { "epoch": 84.55555555555556, "grad_norm": 2.323190710740164e-05, "learning_rate": 4.566707635888508e-05, "loss": 0.0, "num_input_tokens_seen": 2130688, "step": 7610 }, { "epoch": 84.61111111111111, "grad_norm": 2.88171963802597e-06, "learning_rate": 4.566155078660677e-05, "loss": 0.0, "num_input_tokens_seen": 2132064, "step": 7615 }, { "epoch": 84.66666666666667, "grad_norm": 3.6759265640284866e-06, "learning_rate": 4.565602202805768e-05, "loss": 0.0, "num_input_tokens_seen": 2133408, "step": 7620 }, { "epoch": 84.72222222222223, "grad_norm": 2.0443178527784767e-06, "learning_rate": 4.56504900840904e-05, "loss": 0.0, "num_input_tokens_seen": 2134784, "step": 7625 }, { "epoch": 84.77777777777777, "grad_norm": 2.6676516426960006e-06, "learning_rate": 4.564495495555805e-05, "loss": 0.0, "num_input_tokens_seen": 2136224, "step": 7630 }, { "epoch": 84.83333333333333, "grad_norm": 1.8574345403976622e-06, "learning_rate": 4.5639416643314204e-05, "loss": 0.0, "num_input_tokens_seen": 2137616, "step": 7635 }, { "epoch": 84.88888888888889, "grad_norm": 7.522508440160891e-06, "learning_rate": 4.5633875148212946e-05, "loss": 0.0, "num_input_tokens_seen": 2139040, "step": 7640 }, { "epoch": 84.94444444444444, "grad_norm": 3.304965730421827e-06, "learning_rate": 4.562833047110883e-05, "loss": 0.0, "num_input_tokens_seen": 2140416, "step": 7645 }, { "epoch": 85.0, "grad_norm": 4.034471658087568e-06, "learning_rate": 4.5622782612856923e-05, "loss": 0.0, "num_input_tokens_seen": 2141824, "step": 7650 }, { "epoch": 85.05555555555556, "grad_norm": 4.629271643352695e-06, "learning_rate": 4.561723157431278e-05, "loss": 0.0, "num_input_tokens_seen": 2143248, "step": 7655 }, { "epoch": 85.11111111111111, "grad_norm": 6.330441010504728e-06, "learning_rate": 4.5611677356332435e-05, "loss": 0.0, "num_input_tokens_seen": 2144656, "step": 7660 }, { "epoch": 85.16666666666667, "grad_norm": 1.4218685464584269e-06, "learning_rate": 4.560611995977242e-05, "loss": 0.0, "num_input_tokens_seen": 2146080, "step": 7665 }, { "epoch": 85.22222222222223, "grad_norm": 3.3352457649016287e-06, "learning_rate": 4.560055938548975e-05, "loss": 0.0, "num_input_tokens_seen": 2147488, "step": 7670 }, { "epoch": 85.27777777777777, "grad_norm": 2.7024580049328506e-06, "learning_rate": 4.5594995634341944e-05, "loss": 0.0, "num_input_tokens_seen": 2148864, "step": 7675 }, { "epoch": 85.33333333333333, "grad_norm": 4.640938641387038e-06, "learning_rate": 4.5589428707187e-05, "loss": 0.0, "num_input_tokens_seen": 2150272, "step": 7680 }, { "epoch": 85.38888888888889, "grad_norm": 7.836773875169456e-06, "learning_rate": 4.55838586048834e-05, "loss": 0.0, "num_input_tokens_seen": 2151664, "step": 7685 }, { "epoch": 85.44444444444444, "grad_norm": 2.667240323717124e-06, "learning_rate": 4.557828532829013e-05, "loss": 0.0, "num_input_tokens_seen": 2153088, "step": 7690 }, { "epoch": 85.5, "grad_norm": 5.501048235601047e-06, "learning_rate": 4.557270887826667e-05, "loss": 0.0, "num_input_tokens_seen": 2154512, "step": 7695 }, { "epoch": 85.55555555555556, "grad_norm": 1.4799615200900007e-05, "learning_rate": 4.556712925567296e-05, "loss": 0.0, "num_input_tokens_seen": 2155904, "step": 7700 }, { "epoch": 85.61111111111111, "grad_norm": 1.6568220644330722e-06, "learning_rate": 4.5561546461369454e-05, "loss": 0.0, "num_input_tokens_seen": 2157264, "step": 7705 }, { "epoch": 85.66666666666667, "grad_norm": 3.0926355520932702e-06, "learning_rate": 4.55559604962171e-05, "loss": 0.0, "num_input_tokens_seen": 2158688, "step": 7710 }, { "epoch": 85.72222222222223, "grad_norm": 3.0208036605472444e-06, "learning_rate": 4.55503713610773e-05, "loss": 0.0, "num_input_tokens_seen": 2160096, "step": 7715 }, { "epoch": 85.77777777777777, "grad_norm": 1.497486937296344e-05, "learning_rate": 4.5544779056812e-05, "loss": 0.0, "num_input_tokens_seen": 2161520, "step": 7720 }, { "epoch": 85.83333333333333, "grad_norm": 2.0406207113410346e-05, "learning_rate": 4.553918358428358e-05, "loss": 0.0, "num_input_tokens_seen": 2162928, "step": 7725 }, { "epoch": 85.88888888888889, "grad_norm": 3.477128302620258e-06, "learning_rate": 4.553358494435494e-05, "loss": 0.0, "num_input_tokens_seen": 2164336, "step": 7730 }, { "epoch": 85.94444444444444, "grad_norm": 2.721852752074483e-06, "learning_rate": 4.5527983137889464e-05, "loss": 0.0, "num_input_tokens_seen": 2165760, "step": 7735 }, { "epoch": 86.0, "grad_norm": 1.188517580885673e-05, "learning_rate": 4.5522378165751015e-05, "loss": 0.0, "num_input_tokens_seen": 2167152, "step": 7740 }, { "epoch": 86.05555555555556, "grad_norm": 3.4080881050613243e-06, "learning_rate": 4.5516770028803954e-05, "loss": 0.0, "num_input_tokens_seen": 2168544, "step": 7745 }, { "epoch": 86.11111111111111, "grad_norm": 1.2132024949096376e-06, "learning_rate": 4.5511158727913116e-05, "loss": 0.0, "num_input_tokens_seen": 2169904, "step": 7750 }, { "epoch": 86.16666666666667, "grad_norm": 1.1828036804217845e-05, "learning_rate": 4.5505544263943856e-05, "loss": 0.0, "num_input_tokens_seen": 2171296, "step": 7755 }, { "epoch": 86.22222222222223, "grad_norm": 2.0625457182177342e-05, "learning_rate": 4.549992663776197e-05, "loss": 0.0, "num_input_tokens_seen": 2172704, "step": 7760 }, { "epoch": 86.27777777777777, "grad_norm": 1.1255954632360954e-05, "learning_rate": 4.5494305850233786e-05, "loss": 0.0, "num_input_tokens_seen": 2174112, "step": 7765 }, { "epoch": 86.33333333333333, "grad_norm": 1.8624079984874697e-06, "learning_rate": 4.5488681902226094e-05, "loss": 0.0, "num_input_tokens_seen": 2175520, "step": 7770 }, { "epoch": 86.38888888888889, "grad_norm": 2.155575884899008e-06, "learning_rate": 4.5483054794606174e-05, "loss": 0.0, "num_input_tokens_seen": 2176928, "step": 7775 }, { "epoch": 86.44444444444444, "grad_norm": 3.290142331024981e-06, "learning_rate": 4.547742452824179e-05, "loss": 0.0, "num_input_tokens_seen": 2178320, "step": 7780 }, { "epoch": 86.5, "grad_norm": 2.1093264876981266e-06, "learning_rate": 4.5471791104001215e-05, "loss": 0.0, "num_input_tokens_seen": 2179728, "step": 7785 }, { "epoch": 86.55555555555556, "grad_norm": 1.490728300268529e-05, "learning_rate": 4.546615452275319e-05, "loss": 0.0, "num_input_tokens_seen": 2181088, "step": 7790 }, { "epoch": 86.61111111111111, "grad_norm": 7.482145520043559e-06, "learning_rate": 4.5460514785366944e-05, "loss": 0.0, "num_input_tokens_seen": 2182480, "step": 7795 }, { "epoch": 86.66666666666667, "grad_norm": 4.16914281231584e-06, "learning_rate": 4.545487189271219e-05, "loss": 0.0, "num_input_tokens_seen": 2183888, "step": 7800 }, { "epoch": 86.66666666666667, "eval_loss": 1.1319087743759155, "eval_runtime": 1.3949, "eval_samples_per_second": 28.675, "eval_steps_per_second": 14.338, "num_input_tokens_seen": 2183888, "step": 7800 }, { "epoch": 86.72222222222223, "grad_norm": 2.2076014829508495e-06, "learning_rate": 4.544922584565914e-05, "loss": 0.0, "num_input_tokens_seen": 2185296, "step": 7805 }, { "epoch": 86.77777777777777, "grad_norm": 2.2545179945154814e-06, "learning_rate": 4.544357664507848e-05, "loss": 0.0, "num_input_tokens_seen": 2186672, "step": 7810 }, { "epoch": 86.83333333333333, "grad_norm": 5.310554570314707e-06, "learning_rate": 4.54379242918414e-05, "loss": 0.0, "num_input_tokens_seen": 2188128, "step": 7815 }, { "epoch": 86.88888888888889, "grad_norm": 9.123484119299974e-07, "learning_rate": 4.543226878681955e-05, "loss": 0.0, "num_input_tokens_seen": 2189536, "step": 7820 }, { "epoch": 86.94444444444444, "grad_norm": 1.1392756960049155e-06, "learning_rate": 4.5426610130885087e-05, "loss": 0.0, "num_input_tokens_seen": 2190880, "step": 7825 }, { "epoch": 87.0, "grad_norm": 1.4522578339892789e-06, "learning_rate": 4.542094832491064e-05, "loss": 0.0, "num_input_tokens_seen": 2192304, "step": 7830 }, { "epoch": 87.05555555555556, "grad_norm": 1.4814523638051469e-05, "learning_rate": 4.541528336976934e-05, "loss": 0.0, "num_input_tokens_seen": 2193696, "step": 7835 }, { "epoch": 87.11111111111111, "grad_norm": 2.5992601422331063e-06, "learning_rate": 4.540961526633479e-05, "loss": 0.0, "num_input_tokens_seen": 2195104, "step": 7840 }, { "epoch": 87.16666666666667, "grad_norm": 2.66159554485057e-06, "learning_rate": 4.540394401548108e-05, "loss": 0.0, "num_input_tokens_seen": 2196480, "step": 7845 }, { "epoch": 87.22222222222223, "grad_norm": 2.357236780881067e-06, "learning_rate": 4.539826961808279e-05, "loss": 0.0, "num_input_tokens_seen": 2197856, "step": 7850 }, { "epoch": 87.27777777777777, "grad_norm": 9.416741704626475e-06, "learning_rate": 4.5392592075014994e-05, "loss": 0.0, "num_input_tokens_seen": 2199248, "step": 7855 }, { "epoch": 87.33333333333333, "grad_norm": 6.942650088603841e-06, "learning_rate": 4.538691138715322e-05, "loss": 0.0, "num_input_tokens_seen": 2200640, "step": 7860 }, { "epoch": 87.38888888888889, "grad_norm": 3.0933288144296966e-06, "learning_rate": 4.5381227555373516e-05, "loss": 0.0, "num_input_tokens_seen": 2202048, "step": 7865 }, { "epoch": 87.44444444444444, "grad_norm": 1.2545690424303757e-06, "learning_rate": 4.537554058055239e-05, "loss": 0.0, "num_input_tokens_seen": 2203472, "step": 7870 }, { "epoch": 87.5, "grad_norm": 2.1881417069380404e-06, "learning_rate": 4.5369850463566865e-05, "loss": 0.0, "num_input_tokens_seen": 2204864, "step": 7875 }, { "epoch": 87.55555555555556, "grad_norm": 3.65150481229648e-05, "learning_rate": 4.5364157205294404e-05, "loss": 0.0, "num_input_tokens_seen": 2206288, "step": 7880 }, { "epoch": 87.61111111111111, "grad_norm": 2.59628404819523e-06, "learning_rate": 4.5358460806612996e-05, "loss": 0.0, "num_input_tokens_seen": 2207728, "step": 7885 }, { "epoch": 87.66666666666667, "grad_norm": 1.3974224202684127e-05, "learning_rate": 4.535276126840109e-05, "loss": 0.0, "num_input_tokens_seen": 2209136, "step": 7890 }, { "epoch": 87.72222222222223, "grad_norm": 1.3262997526908293e-05, "learning_rate": 4.5347058591537626e-05, "loss": 0.0, "num_input_tokens_seen": 2210528, "step": 7895 }, { "epoch": 87.77777777777777, "grad_norm": 6.7444457272358704e-06, "learning_rate": 4.534135277690203e-05, "loss": 0.0, "num_input_tokens_seen": 2211904, "step": 7900 }, { "epoch": 87.83333333333333, "grad_norm": 1.5795097851878381e-06, "learning_rate": 4.533564382537421e-05, "loss": 0.0, "num_input_tokens_seen": 2213264, "step": 7905 }, { "epoch": 87.88888888888889, "grad_norm": 1.8269259953740402e-06, "learning_rate": 4.532993173783456e-05, "loss": 0.0, "num_input_tokens_seen": 2214688, "step": 7910 }, { "epoch": 87.94444444444444, "grad_norm": 2.3457923816749826e-06, "learning_rate": 4.5324216515163954e-05, "loss": 0.0, "num_input_tokens_seen": 2216080, "step": 7915 }, { "epoch": 88.0, "grad_norm": 1.7744407614372903e-06, "learning_rate": 4.531849815824375e-05, "loss": 0.0, "num_input_tokens_seen": 2217488, "step": 7920 }, { "epoch": 88.05555555555556, "grad_norm": 1.2668157296502613e-06, "learning_rate": 4.5312776667955795e-05, "loss": 0.0, "num_input_tokens_seen": 2218864, "step": 7925 }, { "epoch": 88.11111111111111, "grad_norm": 2.930797791123041e-06, "learning_rate": 4.5307052045182405e-05, "loss": 0.0, "num_input_tokens_seen": 2220240, "step": 7930 }, { "epoch": 88.16666666666667, "grad_norm": 1.1600991456361953e-06, "learning_rate": 4.53013242908064e-05, "loss": 0.0, "num_input_tokens_seen": 2221616, "step": 7935 }, { "epoch": 88.22222222222223, "grad_norm": 3.19863192999037e-06, "learning_rate": 4.529559340571107e-05, "loss": 0.0, "num_input_tokens_seen": 2222992, "step": 7940 }, { "epoch": 88.27777777777777, "grad_norm": 2.9078771603963105e-06, "learning_rate": 4.528985939078018e-05, "loss": 0.0, "num_input_tokens_seen": 2224368, "step": 7945 }, { "epoch": 88.33333333333333, "grad_norm": 8.513550710631534e-06, "learning_rate": 4.5284122246898e-05, "loss": 0.0, "num_input_tokens_seen": 2225776, "step": 7950 }, { "epoch": 88.38888888888889, "grad_norm": 8.011837962840218e-06, "learning_rate": 4.527838197494926e-05, "loss": 0.0, "num_input_tokens_seen": 2227184, "step": 7955 }, { "epoch": 88.44444444444444, "grad_norm": 8.135314601531718e-06, "learning_rate": 4.527263857581918e-05, "loss": 0.0, "num_input_tokens_seen": 2228608, "step": 7960 }, { "epoch": 88.5, "grad_norm": 4.133701622777153e-06, "learning_rate": 4.526689205039347e-05, "loss": 0.0, "num_input_tokens_seen": 2230032, "step": 7965 }, { "epoch": 88.55555555555556, "grad_norm": 4.11427845392609e-06, "learning_rate": 4.5261142399558324e-05, "loss": 0.0, "num_input_tokens_seen": 2231392, "step": 7970 }, { "epoch": 88.61111111111111, "grad_norm": 7.832859409973025e-06, "learning_rate": 4.525538962420041e-05, "loss": 0.0, "num_input_tokens_seen": 2232816, "step": 7975 }, { "epoch": 88.66666666666667, "grad_norm": 2.9853908927179873e-06, "learning_rate": 4.524963372520685e-05, "loss": 0.0, "num_input_tokens_seen": 2234224, "step": 7980 }, { "epoch": 88.72222222222223, "grad_norm": 2.169909748772625e-06, "learning_rate": 4.524387470346531e-05, "loss": 0.0, "num_input_tokens_seen": 2235664, "step": 7985 }, { "epoch": 88.77777777777777, "grad_norm": 1.1919278222194407e-06, "learning_rate": 4.5238112559863885e-05, "loss": 0.0, "num_input_tokens_seen": 2237040, "step": 7990 }, { "epoch": 88.83333333333333, "grad_norm": 4.429362888913602e-06, "learning_rate": 4.5232347295291175e-05, "loss": 0.0, "num_input_tokens_seen": 2238432, "step": 7995 }, { "epoch": 88.88888888888889, "grad_norm": 1.5320454167522257e-06, "learning_rate": 4.522657891063626e-05, "loss": 0.0, "num_input_tokens_seen": 2239840, "step": 8000 }, { "epoch": 88.88888888888889, "eval_loss": 1.152980089187622, "eval_runtime": 1.3899, "eval_samples_per_second": 28.779, "eval_steps_per_second": 14.389, "num_input_tokens_seen": 2239840, "step": 8000 }, { "epoch": 88.94444444444444, "grad_norm": 1.9873159544658847e-06, "learning_rate": 4.52208074067887e-05, "loss": 0.0, "num_input_tokens_seen": 2241200, "step": 8005 }, { "epoch": 89.0, "grad_norm": 3.5245770959591027e-06, "learning_rate": 4.5215032784638516e-05, "loss": 0.0, "num_input_tokens_seen": 2242608, "step": 8010 }, { "epoch": 89.05555555555556, "grad_norm": 5.4334768719854765e-06, "learning_rate": 4.5209255045076245e-05, "loss": 0.0, "num_input_tokens_seen": 2244016, "step": 8015 }, { "epoch": 89.11111111111111, "grad_norm": 6.989976668592135e-07, "learning_rate": 4.5203474188992875e-05, "loss": 0.0, "num_input_tokens_seen": 2245440, "step": 8020 }, { "epoch": 89.16666666666667, "grad_norm": 2.065749185931054e-06, "learning_rate": 4.51976902172799e-05, "loss": 0.0, "num_input_tokens_seen": 2246832, "step": 8025 }, { "epoch": 89.22222222222223, "grad_norm": 3.9881174416223075e-06, "learning_rate": 4.519190313082927e-05, "loss": 0.0, "num_input_tokens_seen": 2248176, "step": 8030 }, { "epoch": 89.27777777777777, "grad_norm": 1.0267578545608558e-05, "learning_rate": 4.518611293053343e-05, "loss": 0.0, "num_input_tokens_seen": 2249568, "step": 8035 }, { "epoch": 89.33333333333333, "grad_norm": 5.4966167226666585e-06, "learning_rate": 4.51803196172853e-05, "loss": 0.0, "num_input_tokens_seen": 2250976, "step": 8040 }, { "epoch": 89.38888888888889, "grad_norm": 3.3878261547215516e-06, "learning_rate": 4.517452319197828e-05, "loss": 0.0, "num_input_tokens_seen": 2252384, "step": 8045 }, { "epoch": 89.44444444444444, "grad_norm": 4.599215571943205e-06, "learning_rate": 4.5168723655506265e-05, "loss": 0.0, "num_input_tokens_seen": 2253792, "step": 8050 }, { "epoch": 89.5, "grad_norm": 3.095091187788057e-06, "learning_rate": 4.51629210087636e-05, "loss": 0.0, "num_input_tokens_seen": 2255168, "step": 8055 }, { "epoch": 89.55555555555556, "grad_norm": 3.5501382171787554e-06, "learning_rate": 4.515711525264513e-05, "loss": 0.0, "num_input_tokens_seen": 2256624, "step": 8060 }, { "epoch": 89.61111111111111, "grad_norm": 1.9959454675699817e-06, "learning_rate": 4.5151306388046175e-05, "loss": 0.0, "num_input_tokens_seen": 2258016, "step": 8065 }, { "epoch": 89.66666666666667, "grad_norm": 2.0739887531817658e-06, "learning_rate": 4.514549441586255e-05, "loss": 0.0, "num_input_tokens_seen": 2259456, "step": 8070 }, { "epoch": 89.72222222222223, "grad_norm": 4.4914941099705175e-05, "learning_rate": 4.513967933699051e-05, "loss": 0.0, "num_input_tokens_seen": 2260880, "step": 8075 }, { "epoch": 89.77777777777777, "grad_norm": 1.2450980193534633e-06, "learning_rate": 4.513386115232684e-05, "loss": 0.0, "num_input_tokens_seen": 2262272, "step": 8080 }, { "epoch": 89.83333333333333, "grad_norm": 5.902261818846455e-06, "learning_rate": 4.5128039862768745e-05, "loss": 0.0, "num_input_tokens_seen": 2263632, "step": 8085 }, { "epoch": 89.88888888888889, "grad_norm": 2.535692829042091e-06, "learning_rate": 4.512221546921397e-05, "loss": 0.0, "num_input_tokens_seen": 2265008, "step": 8090 }, { "epoch": 89.94444444444444, "grad_norm": 2.9780592285533203e-06, "learning_rate": 4.5116387972560694e-05, "loss": 0.0, "num_input_tokens_seen": 2266400, "step": 8095 }, { "epoch": 90.0, "grad_norm": 2.6655754936655285e-06, "learning_rate": 4.511055737370759e-05, "loss": 0.0, "num_input_tokens_seen": 2267824, "step": 8100 }, { "epoch": 90.05555555555556, "grad_norm": 1.8611783161759377e-05, "learning_rate": 4.510472367355383e-05, "loss": 0.0, "num_input_tokens_seen": 2269232, "step": 8105 }, { "epoch": 90.11111111111111, "grad_norm": 6.254592790355673e-06, "learning_rate": 4.509888687299901e-05, "loss": 0.0, "num_input_tokens_seen": 2270608, "step": 8110 }, { "epoch": 90.16666666666667, "grad_norm": 5.669615347869694e-06, "learning_rate": 4.5093046972943266e-05, "loss": 0.0, "num_input_tokens_seen": 2272048, "step": 8115 }, { "epoch": 90.22222222222223, "grad_norm": 4.614035788108595e-06, "learning_rate": 4.508720397428717e-05, "loss": 0.0, "num_input_tokens_seen": 2273440, "step": 8120 }, { "epoch": 90.27777777777777, "grad_norm": 8.340110753124463e-07, "learning_rate": 4.508135787793178e-05, "loss": 0.0, "num_input_tokens_seen": 2274832, "step": 8125 }, { "epoch": 90.33333333333333, "grad_norm": 9.783939276530873e-06, "learning_rate": 4.5075508684778664e-05, "loss": 0.0, "num_input_tokens_seen": 2276256, "step": 8130 }, { "epoch": 90.38888888888889, "grad_norm": 3.5132845823682146e-06, "learning_rate": 4.506965639572982e-05, "loss": 0.0, "num_input_tokens_seen": 2277648, "step": 8135 }, { "epoch": 90.44444444444444, "grad_norm": 2.4221777493949048e-05, "learning_rate": 4.506380101168774e-05, "loss": 0.0, "num_input_tokens_seen": 2279072, "step": 8140 }, { "epoch": 90.5, "grad_norm": 2.0782645151484758e-06, "learning_rate": 4.505794253355542e-05, "loss": 0.0, "num_input_tokens_seen": 2280464, "step": 8145 }, { "epoch": 90.55555555555556, "grad_norm": 1.5553513321719947e-06, "learning_rate": 4.5052080962236286e-05, "loss": 0.0, "num_input_tokens_seen": 2281856, "step": 8150 }, { "epoch": 90.61111111111111, "grad_norm": 6.16386387264356e-06, "learning_rate": 4.504621629863428e-05, "loss": 0.0, "num_input_tokens_seen": 2283280, "step": 8155 }, { "epoch": 90.66666666666667, "grad_norm": 4.476311914913822e-06, "learning_rate": 4.504034854365381e-05, "loss": 0.0, "num_input_tokens_seen": 2284688, "step": 8160 }, { "epoch": 90.72222222222223, "grad_norm": 4.703093054558849e-06, "learning_rate": 4.503447769819974e-05, "loss": 0.0, "num_input_tokens_seen": 2286080, "step": 8165 }, { "epoch": 90.77777777777777, "grad_norm": 4.074724984093336e-06, "learning_rate": 4.502860376317745e-05, "loss": 0.0, "num_input_tokens_seen": 2287456, "step": 8170 }, { "epoch": 90.83333333333333, "grad_norm": 2.0165362002444454e-06, "learning_rate": 4.502272673949276e-05, "loss": 0.0, "num_input_tokens_seen": 2288832, "step": 8175 }, { "epoch": 90.88888888888889, "grad_norm": 2.7075202524429187e-06, "learning_rate": 4.501684662805199e-05, "loss": 0.0, "num_input_tokens_seen": 2290272, "step": 8180 }, { "epoch": 90.94444444444444, "grad_norm": 4.336755409894977e-06, "learning_rate": 4.5010963429761924e-05, "loss": 0.0, "num_input_tokens_seen": 2291664, "step": 8185 }, { "epoch": 91.0, "grad_norm": 1.6000419691408752e-06, "learning_rate": 4.500507714552982e-05, "loss": 0.0, "num_input_tokens_seen": 2293088, "step": 8190 }, { "epoch": 91.05555555555556, "grad_norm": 3.662964445538819e-06, "learning_rate": 4.499918777626342e-05, "loss": 0.0, "num_input_tokens_seen": 2294544, "step": 8195 }, { "epoch": 91.11111111111111, "grad_norm": 1.2485720617405605e-05, "learning_rate": 4.499329532287093e-05, "loss": 0.0, "num_input_tokens_seen": 2295888, "step": 8200 }, { "epoch": 91.11111111111111, "eval_loss": 1.167669653892517, "eval_runtime": 1.3943, "eval_samples_per_second": 28.689, "eval_steps_per_second": 14.344, "num_input_tokens_seen": 2295888, "step": 8200 }, { "epoch": 91.16666666666667, "grad_norm": 1.6062012946349569e-06, "learning_rate": 4.4987399786261064e-05, "loss": 0.0, "num_input_tokens_seen": 2297232, "step": 8205 }, { "epoch": 91.22222222222223, "grad_norm": 8.08071854407899e-06, "learning_rate": 4.498150116734297e-05, "loss": 0.0, "num_input_tokens_seen": 2298608, "step": 8210 }, { "epoch": 91.27777777777777, "grad_norm": 3.6229043871571776e-06, "learning_rate": 4.4975599467026294e-05, "loss": 0.0, "num_input_tokens_seen": 2300016, "step": 8215 }, { "epoch": 91.33333333333333, "grad_norm": 2.9481647288776003e-05, "learning_rate": 4.496969468622114e-05, "loss": 0.0, "num_input_tokens_seen": 2301408, "step": 8220 }, { "epoch": 91.38888888888889, "grad_norm": 3.932348590751644e-06, "learning_rate": 4.496378682583813e-05, "loss": 0.0, "num_input_tokens_seen": 2302816, "step": 8225 }, { "epoch": 91.44444444444444, "grad_norm": 1.110937591874972e-05, "learning_rate": 4.495787588678829e-05, "loss": 0.0, "num_input_tokens_seen": 2304240, "step": 8230 }, { "epoch": 91.5, "grad_norm": 1.7952007738131215e-06, "learning_rate": 4.4951961869983196e-05, "loss": 0.0, "num_input_tokens_seen": 2305664, "step": 8235 }, { "epoch": 91.55555555555556, "grad_norm": 5.10236395712127e-06, "learning_rate": 4.494604477633485e-05, "loss": 0.0, "num_input_tokens_seen": 2307024, "step": 8240 }, { "epoch": 91.61111111111111, "grad_norm": 2.246500343972002e-06, "learning_rate": 4.4940124606755734e-05, "loss": 0.0, "num_input_tokens_seen": 2308416, "step": 8245 }, { "epoch": 91.66666666666667, "grad_norm": 2.438173396512866e-06, "learning_rate": 4.493420136215882e-05, "loss": 0.0, "num_input_tokens_seen": 2309760, "step": 8250 }, { "epoch": 91.72222222222223, "grad_norm": 1.584742449267651e-06, "learning_rate": 4.492827504345756e-05, "loss": 0.0, "num_input_tokens_seen": 2311184, "step": 8255 }, { "epoch": 91.77777777777777, "grad_norm": 3.618750952227856e-06, "learning_rate": 4.492234565156584e-05, "loss": 0.0, "num_input_tokens_seen": 2312576, "step": 8260 }, { "epoch": 91.83333333333333, "grad_norm": 3.2216105410043383e-06, "learning_rate": 4.491641318739807e-05, "loss": 0.0, "num_input_tokens_seen": 2313952, "step": 8265 }, { "epoch": 91.88888888888889, "grad_norm": 2.4774378744041314e-06, "learning_rate": 4.4910477651869096e-05, "loss": 0.0, "num_input_tokens_seen": 2315360, "step": 8270 }, { "epoch": 91.94444444444444, "grad_norm": 5.234131549514132e-06, "learning_rate": 4.4904539045894254e-05, "loss": 0.0, "num_input_tokens_seen": 2316832, "step": 8275 }, { "epoch": 92.0, "grad_norm": 2.6233794869767735e-06, "learning_rate": 4.4898597370389364e-05, "loss": 0.0, "num_input_tokens_seen": 2318256, "step": 8280 }, { "epoch": 92.05555555555556, "grad_norm": 1.9887409052898875e-06, "learning_rate": 4.489265262627069e-05, "loss": 0.0, "num_input_tokens_seen": 2319648, "step": 8285 }, { "epoch": 92.11111111111111, "grad_norm": 2.9641603305208264e-06, "learning_rate": 4.488670481445499e-05, "loss": 0.0, "num_input_tokens_seen": 2321056, "step": 8290 }, { "epoch": 92.16666666666667, "grad_norm": 8.283905685857462e-07, "learning_rate": 4.488075393585951e-05, "loss": 0.0, "num_input_tokens_seen": 2322432, "step": 8295 }, { "epoch": 92.22222222222223, "grad_norm": 3.07462437376671e-06, "learning_rate": 4.487479999140193e-05, "loss": 0.0, "num_input_tokens_seen": 2323840, "step": 8300 }, { "epoch": 92.27777777777777, "grad_norm": 4.703104877989972e-06, "learning_rate": 4.4868842982000425e-05, "loss": 0.0, "num_input_tokens_seen": 2325216, "step": 8305 }, { "epoch": 92.33333333333333, "grad_norm": 1.9249455363024026e-06, "learning_rate": 4.486288290857365e-05, "loss": 0.0, "num_input_tokens_seen": 2326608, "step": 8310 }, { "epoch": 92.38888888888889, "grad_norm": 1.530977783659182e-06, "learning_rate": 4.4856919772040715e-05, "loss": 0.0, "num_input_tokens_seen": 2328048, "step": 8315 }, { "epoch": 92.44444444444444, "grad_norm": 3.8348739508364815e-06, "learning_rate": 4.485095357332122e-05, "loss": 0.0, "num_input_tokens_seen": 2329440, "step": 8320 }, { "epoch": 92.5, "grad_norm": 4.612978045770433e-06, "learning_rate": 4.484498431333521e-05, "loss": 0.0, "num_input_tokens_seen": 2330832, "step": 8325 }, { "epoch": 92.55555555555556, "grad_norm": 3.297763214504812e-06, "learning_rate": 4.4839011993003245e-05, "loss": 0.0, "num_input_tokens_seen": 2332208, "step": 8330 }, { "epoch": 92.61111111111111, "grad_norm": 2.5742895104485797e-06, "learning_rate": 4.4833036613246305e-05, "loss": 0.0, "num_input_tokens_seen": 2333568, "step": 8335 }, { "epoch": 92.66666666666667, "grad_norm": 9.153765859082341e-06, "learning_rate": 4.482705817498589e-05, "loss": 0.0, "num_input_tokens_seen": 2334960, "step": 8340 }, { "epoch": 92.72222222222223, "grad_norm": 9.525333553028759e-06, "learning_rate": 4.4821076679143934e-05, "loss": 0.0, "num_input_tokens_seen": 2336368, "step": 8345 }, { "epoch": 92.77777777777777, "grad_norm": 1.3379705706029199e-05, "learning_rate": 4.481509212664288e-05, "loss": 0.0, "num_input_tokens_seen": 2337808, "step": 8350 }, { "epoch": 92.83333333333333, "grad_norm": 4.0863478716346435e-06, "learning_rate": 4.480910451840559e-05, "loss": 0.0, "num_input_tokens_seen": 2339248, "step": 8355 }, { "epoch": 92.88888888888889, "grad_norm": 4.4139387682662345e-06, "learning_rate": 4.480311385535546e-05, "loss": 0.0, "num_input_tokens_seen": 2340640, "step": 8360 }, { "epoch": 92.94444444444444, "grad_norm": 1.0510686934139812e-06, "learning_rate": 4.47971201384163e-05, "loss": 0.0, "num_input_tokens_seen": 2342032, "step": 8365 }, { "epoch": 93.0, "grad_norm": 3.4597569538163953e-06, "learning_rate": 4.4791123368512446e-05, "loss": 0.0, "num_input_tokens_seen": 2343424, "step": 8370 }, { "epoch": 93.05555555555556, "grad_norm": 2.078121497106622e-06, "learning_rate": 4.478512354656864e-05, "loss": 0.0, "num_input_tokens_seen": 2344848, "step": 8375 }, { "epoch": 93.11111111111111, "grad_norm": 7.790262316120788e-06, "learning_rate": 4.477912067351016e-05, "loss": 0.0, "num_input_tokens_seen": 2346256, "step": 8380 }, { "epoch": 93.16666666666667, "grad_norm": 4.247605374985142e-06, "learning_rate": 4.477311475026271e-05, "loss": 0.0, "num_input_tokens_seen": 2347696, "step": 8385 }, { "epoch": 93.22222222222223, "grad_norm": 2.540731884437264e-06, "learning_rate": 4.476710577775248e-05, "loss": 0.0, "num_input_tokens_seen": 2349072, "step": 8390 }, { "epoch": 93.27777777777777, "grad_norm": 3.614012484831619e-06, "learning_rate": 4.476109375690612e-05, "loss": 0.0, "num_input_tokens_seen": 2350448, "step": 8395 }, { "epoch": 93.33333333333333, "grad_norm": 1.2032220411128947e-06, "learning_rate": 4.4755078688650784e-05, "loss": 0.0, "num_input_tokens_seen": 2351872, "step": 8400 }, { "epoch": 93.33333333333333, "eval_loss": 1.1811110973358154, "eval_runtime": 1.3853, "eval_samples_per_second": 28.875, "eval_steps_per_second": 14.438, "num_input_tokens_seen": 2351872, "step": 8400 }, { "epoch": 93.38888888888889, "grad_norm": 6.263645900617121e-06, "learning_rate": 4.474906057391406e-05, "loss": 0.0, "num_input_tokens_seen": 2353280, "step": 8405 }, { "epoch": 93.44444444444444, "grad_norm": 3.9484366425313056e-06, "learning_rate": 4.4743039413624e-05, "loss": 0.0, "num_input_tokens_seen": 2354672, "step": 8410 }, { "epoch": 93.5, "grad_norm": 1.86445879535313e-06, "learning_rate": 4.473701520870916e-05, "loss": 0.0, "num_input_tokens_seen": 2356080, "step": 8415 }, { "epoch": 93.55555555555556, "grad_norm": 2.878535451600328e-05, "learning_rate": 4.4730987960098544e-05, "loss": 0.0, "num_input_tokens_seen": 2357456, "step": 8420 }, { "epoch": 93.61111111111111, "grad_norm": 2.3146183139033383e-06, "learning_rate": 4.4724957668721635e-05, "loss": 0.0, "num_input_tokens_seen": 2358816, "step": 8425 }, { "epoch": 93.66666666666667, "grad_norm": 3.6900628401781432e-06, "learning_rate": 4.471892433550836e-05, "loss": 0.0, "num_input_tokens_seen": 2360224, "step": 8430 }, { "epoch": 93.72222222222223, "grad_norm": 5.614198926195968e-06, "learning_rate": 4.471288796138916e-05, "loss": 0.0, "num_input_tokens_seen": 2361616, "step": 8435 }, { "epoch": 93.77777777777777, "grad_norm": 1.0346997214583098e-06, "learning_rate": 4.470684854729491e-05, "loss": 0.0, "num_input_tokens_seen": 2363024, "step": 8440 }, { "epoch": 93.83333333333333, "grad_norm": 2.3256195618159836e-06, "learning_rate": 4.4700806094156955e-05, "loss": 0.0, "num_input_tokens_seen": 2364384, "step": 8445 }, { "epoch": 93.88888888888889, "grad_norm": 3.1741969905851875e-06, "learning_rate": 4.469476060290713e-05, "loss": 0.0, "num_input_tokens_seen": 2365792, "step": 8450 }, { "epoch": 93.94444444444444, "grad_norm": 1.310435095547291e-06, "learning_rate": 4.468871207447772e-05, "loss": 0.0, "num_input_tokens_seen": 2367152, "step": 8455 }, { "epoch": 94.0, "grad_norm": 5.4244328566710465e-06, "learning_rate": 4.4682660509801486e-05, "loss": 0.0, "num_input_tokens_seen": 2368560, "step": 8460 }, { "epoch": 94.05555555555556, "grad_norm": 9.070905662156292e-07, "learning_rate": 4.467660590981165e-05, "loss": 0.0, "num_input_tokens_seen": 2369952, "step": 8465 }, { "epoch": 94.11111111111111, "grad_norm": 1.078229888662463e-05, "learning_rate": 4.467054827544191e-05, "loss": 0.0, "num_input_tokens_seen": 2371344, "step": 8470 }, { "epoch": 94.16666666666667, "grad_norm": 3.3457095014455263e-06, "learning_rate": 4.4664487607626434e-05, "loss": 0.0, "num_input_tokens_seen": 2372736, "step": 8475 }, { "epoch": 94.22222222222223, "grad_norm": 1.8493988136469852e-06, "learning_rate": 4.4658423907299845e-05, "loss": 0.0, "num_input_tokens_seen": 2374096, "step": 8480 }, { "epoch": 94.27777777777777, "grad_norm": 3.286889977971441e-06, "learning_rate": 4.465235717539725e-05, "loss": 0.0, "num_input_tokens_seen": 2375472, "step": 8485 }, { "epoch": 94.33333333333333, "grad_norm": 1.2644552498386474e-06, "learning_rate": 4.464628741285421e-05, "loss": 0.0, "num_input_tokens_seen": 2376880, "step": 8490 }, { "epoch": 94.38888888888889, "grad_norm": 2.7814260192826623e-06, "learning_rate": 4.4640214620606754e-05, "loss": 0.0, "num_input_tokens_seen": 2378256, "step": 8495 }, { "epoch": 94.44444444444444, "grad_norm": 1.5491652902710484e-06, "learning_rate": 4.46341387995914e-05, "loss": 0.0, "num_input_tokens_seen": 2379648, "step": 8500 }, { "epoch": 94.5, "grad_norm": 1.7205168205691734e-06, "learning_rate": 4.4628059950745106e-05, "loss": 0.0, "num_input_tokens_seen": 2381088, "step": 8505 }, { "epoch": 94.55555555555556, "grad_norm": 1.7681630197330378e-05, "learning_rate": 4.4621978075005297e-05, "loss": 0.0, "num_input_tokens_seen": 2382464, "step": 8510 }, { "epoch": 94.61111111111111, "grad_norm": 3.2587033729214454e-06, "learning_rate": 4.461589317330989e-05, "loss": 0.0, "num_input_tokens_seen": 2383904, "step": 8515 }, { "epoch": 94.66666666666667, "grad_norm": 6.2166686802811455e-06, "learning_rate": 4.460980524659724e-05, "loss": 0.0, "num_input_tokens_seen": 2385328, "step": 8520 }, { "epoch": 94.72222222222223, "grad_norm": 3.305143081888673e-06, "learning_rate": 4.46037142958062e-05, "loss": 0.0, "num_input_tokens_seen": 2386720, "step": 8525 }, { "epoch": 94.77777777777777, "grad_norm": 4.206624907965306e-06, "learning_rate": 4.4597620321876046e-05, "loss": 0.0, "num_input_tokens_seen": 2388144, "step": 8530 }, { "epoch": 94.83333333333333, "grad_norm": 2.437698867652216e-06, "learning_rate": 4.459152332574656e-05, "loss": 0.0, "num_input_tokens_seen": 2389552, "step": 8535 }, { "epoch": 94.88888888888889, "grad_norm": 8.036598046601284e-06, "learning_rate": 4.4585423308357985e-05, "loss": 0.0, "num_input_tokens_seen": 2390960, "step": 8540 }, { "epoch": 94.94444444444444, "grad_norm": 1.6180048305614037e-06, "learning_rate": 4.457932027065102e-05, "loss": 0.0, "num_input_tokens_seen": 2392368, "step": 8545 }, { "epoch": 95.0, "grad_norm": 2.4285477593366522e-06, "learning_rate": 4.45732142135668e-05, "loss": 0.0, "num_input_tokens_seen": 2393808, "step": 8550 }, { "epoch": 95.05555555555556, "grad_norm": 5.564528692048043e-06, "learning_rate": 4.4567105138046986e-05, "loss": 0.0, "num_input_tokens_seen": 2395232, "step": 8555 }, { "epoch": 95.11111111111111, "grad_norm": 4.962351795256836e-06, "learning_rate": 4.456099304503365e-05, "loss": 0.0, "num_input_tokens_seen": 2396640, "step": 8560 }, { "epoch": 95.16666666666667, "grad_norm": 5.711328867619159e-06, "learning_rate": 4.455487793546939e-05, "loss": 0.0, "num_input_tokens_seen": 2398080, "step": 8565 }, { "epoch": 95.22222222222223, "grad_norm": 2.151439502995345e-06, "learning_rate": 4.454875981029719e-05, "loss": 0.0, "num_input_tokens_seen": 2399456, "step": 8570 }, { "epoch": 95.27777777777777, "grad_norm": 1.126183974520245e-06, "learning_rate": 4.454263867046057e-05, "loss": 0.0, "num_input_tokens_seen": 2400864, "step": 8575 }, { "epoch": 95.33333333333333, "grad_norm": 1.8046380318992306e-06, "learning_rate": 4.4536514516903484e-05, "loss": 0.0, "num_input_tokens_seen": 2402288, "step": 8580 }, { "epoch": 95.38888888888889, "grad_norm": 4.395777523313882e-06, "learning_rate": 4.453038735057034e-05, "loss": 0.0, "num_input_tokens_seen": 2403616, "step": 8585 }, { "epoch": 95.44444444444444, "grad_norm": 2.5674728476587916e-06, "learning_rate": 4.4524257172406034e-05, "loss": 0.0, "num_input_tokens_seen": 2404992, "step": 8590 }, { "epoch": 95.5, "grad_norm": 1.5226277128022048e-06, "learning_rate": 4.451812398335592e-05, "loss": 0.0, "num_input_tokens_seen": 2406400, "step": 8595 }, { "epoch": 95.55555555555556, "grad_norm": 3.7066588447487447e-06, "learning_rate": 4.4511987784365805e-05, "loss": 0.0, "num_input_tokens_seen": 2407824, "step": 8600 }, { "epoch": 95.55555555555556, "eval_loss": 1.1916097402572632, "eval_runtime": 1.399, "eval_samples_per_second": 28.592, "eval_steps_per_second": 14.296, "num_input_tokens_seen": 2407824, "step": 8600 }, { "epoch": 95.61111111111111, "grad_norm": 4.299439751775935e-06, "learning_rate": 4.450584857638197e-05, "loss": 0.0, "num_input_tokens_seen": 2409216, "step": 8605 }, { "epoch": 95.66666666666667, "grad_norm": 1.884124003481702e-06, "learning_rate": 4.449970636035116e-05, "loss": 0.0, "num_input_tokens_seen": 2410592, "step": 8610 }, { "epoch": 95.72222222222223, "grad_norm": 1.2360441132841515e-06, "learning_rate": 4.4493561137220574e-05, "loss": 0.0, "num_input_tokens_seen": 2412000, "step": 8615 }, { "epoch": 95.77777777777777, "grad_norm": 9.61066461968585e-07, "learning_rate": 4.44874129079379e-05, "loss": 0.0, "num_input_tokens_seen": 2413408, "step": 8620 }, { "epoch": 95.83333333333333, "grad_norm": 8.020512723305728e-06, "learning_rate": 4.4481261673451255e-05, "loss": 0.0, "num_input_tokens_seen": 2414800, "step": 8625 }, { "epoch": 95.88888888888889, "grad_norm": 6.860482699266868e-06, "learning_rate": 4.4475107434709245e-05, "loss": 0.0, "num_input_tokens_seen": 2416208, "step": 8630 }, { "epoch": 95.94444444444444, "grad_norm": 1.9625256300059846e-06, "learning_rate": 4.446895019266093e-05, "loss": 0.0, "num_input_tokens_seen": 2417632, "step": 8635 }, { "epoch": 96.0, "grad_norm": 7.380726856354158e-06, "learning_rate": 4.446278994825583e-05, "loss": 0.0, "num_input_tokens_seen": 2419008, "step": 8640 }, { "epoch": 96.05555555555556, "grad_norm": 1.7702024024401908e-06, "learning_rate": 4.445662670244394e-05, "loss": 0.0, "num_input_tokens_seen": 2420432, "step": 8645 }, { "epoch": 96.11111111111111, "grad_norm": 1.8101943624060368e-06, "learning_rate": 4.44504604561757e-05, "loss": 0.0, "num_input_tokens_seen": 2421808, "step": 8650 }, { "epoch": 96.16666666666667, "grad_norm": 1.7906019138536067e-06, "learning_rate": 4.4444291210402035e-05, "loss": 0.0, "num_input_tokens_seen": 2423216, "step": 8655 }, { "epoch": 96.22222222222223, "grad_norm": 4.626158897735877e-06, "learning_rate": 4.443811896607431e-05, "loss": 0.0, "num_input_tokens_seen": 2424592, "step": 8660 }, { "epoch": 96.27777777777777, "grad_norm": 1.2611791362360236e-06, "learning_rate": 4.443194372414436e-05, "loss": 0.0, "num_input_tokens_seen": 2426048, "step": 8665 }, { "epoch": 96.33333333333333, "grad_norm": 6.702178325213026e-06, "learning_rate": 4.442576548556449e-05, "loss": 0.0, "num_input_tokens_seen": 2427424, "step": 8670 }, { "epoch": 96.38888888888889, "grad_norm": 8.732879450690234e-07, "learning_rate": 4.441958425128747e-05, "loss": 0.0, "num_input_tokens_seen": 2428816, "step": 8675 }, { "epoch": 96.44444444444444, "grad_norm": 1.3058437389190658e-06, "learning_rate": 4.4413400022266515e-05, "loss": 0.0, "num_input_tokens_seen": 2430192, "step": 8680 }, { "epoch": 96.5, "grad_norm": 2.4967903300421312e-06, "learning_rate": 4.4407212799455313e-05, "loss": 0.0, "num_input_tokens_seen": 2431584, "step": 8685 }, { "epoch": 96.55555555555556, "grad_norm": 1.6414344372606138e-06, "learning_rate": 4.4401022583808003e-05, "loss": 0.0, "num_input_tokens_seen": 2432976, "step": 8690 }, { "epoch": 96.61111111111111, "grad_norm": 3.9676460801274516e-06, "learning_rate": 4.439482937627921e-05, "loss": 0.0, "num_input_tokens_seen": 2434368, "step": 8695 }, { "epoch": 96.66666666666667, "grad_norm": 3.0729681839147815e-06, "learning_rate": 4.4388633177824004e-05, "loss": 0.0, "num_input_tokens_seen": 2435792, "step": 8700 }, { "epoch": 96.72222222222223, "grad_norm": 1.1833833923446946e-05, "learning_rate": 4.4382433989397895e-05, "loss": 0.0, "num_input_tokens_seen": 2437216, "step": 8705 }, { "epoch": 96.77777777777777, "grad_norm": 5.949845672148513e-06, "learning_rate": 4.4376231811956895e-05, "loss": 0.0, "num_input_tokens_seen": 2438608, "step": 8710 }, { "epoch": 96.83333333333333, "grad_norm": 4.70066197522101e-06, "learning_rate": 4.437002664645745e-05, "loss": 0.0, "num_input_tokens_seen": 2439968, "step": 8715 }, { "epoch": 96.88888888888889, "grad_norm": 3.372005267010536e-06, "learning_rate": 4.436381849385649e-05, "loss": 0.0, "num_input_tokens_seen": 2441392, "step": 8720 }, { "epoch": 96.94444444444444, "grad_norm": 2.4393416424572933e-06, "learning_rate": 4.435760735511136e-05, "loss": 0.0, "num_input_tokens_seen": 2442768, "step": 8725 }, { "epoch": 97.0, "grad_norm": 1.1306319720461033e-05, "learning_rate": 4.435139323117992e-05, "loss": 0.0, "num_input_tokens_seen": 2444160, "step": 8730 }, { "epoch": 97.05555555555556, "grad_norm": 3.907347036147257e-06, "learning_rate": 4.434517612302046e-05, "loss": 0.0, "num_input_tokens_seen": 2445568, "step": 8735 }, { "epoch": 97.11111111111111, "grad_norm": 3.910618943336885e-06, "learning_rate": 4.433895603159174e-05, "loss": 0.0, "num_input_tokens_seen": 2446976, "step": 8740 }, { "epoch": 97.16666666666667, "grad_norm": 3.194976306986064e-05, "learning_rate": 4.433273295785296e-05, "loss": 0.0, "num_input_tokens_seen": 2448416, "step": 8745 }, { "epoch": 97.22222222222223, "grad_norm": 1.1133231964777224e-05, "learning_rate": 4.432650690276382e-05, "loss": 0.0, "num_input_tokens_seen": 2449808, "step": 8750 }, { "epoch": 97.27777777777777, "grad_norm": 3.6912483665219042e-06, "learning_rate": 4.4320277867284435e-05, "loss": 0.0, "num_input_tokens_seen": 2451216, "step": 8755 }, { "epoch": 97.33333333333333, "grad_norm": 1.4355011899169767e-06, "learning_rate": 4.431404585237541e-05, "loss": 0.0, "num_input_tokens_seen": 2452608, "step": 8760 }, { "epoch": 97.38888888888889, "grad_norm": 1.5876097450018278e-06, "learning_rate": 4.43078108589978e-05, "loss": 0.0, "num_input_tokens_seen": 2454000, "step": 8765 }, { "epoch": 97.44444444444444, "grad_norm": 9.229124202647654e-07, "learning_rate": 4.4301572888113116e-05, "loss": 0.0, "num_input_tokens_seen": 2455408, "step": 8770 }, { "epoch": 97.5, "grad_norm": 2.7146154479851248e-06, "learning_rate": 4.4295331940683337e-05, "loss": 0.0, "num_input_tokens_seen": 2456800, "step": 8775 }, { "epoch": 97.55555555555556, "grad_norm": 2.0291906821512384e-06, "learning_rate": 4.428908801767089e-05, "loss": 0.0, "num_input_tokens_seen": 2458208, "step": 8780 }, { "epoch": 97.61111111111111, "grad_norm": 1.357147425551375e-06, "learning_rate": 4.428284112003868e-05, "loss": 0.0, "num_input_tokens_seen": 2459568, "step": 8785 }, { "epoch": 97.66666666666667, "grad_norm": 2.705145334402914e-06, "learning_rate": 4.4276591248750033e-05, "loss": 0.0, "num_input_tokens_seen": 2460976, "step": 8790 }, { "epoch": 97.72222222222223, "grad_norm": 5.0123376240662765e-06, "learning_rate": 4.4270338404768774e-05, "loss": 0.0, "num_input_tokens_seen": 2462384, "step": 8795 }, { "epoch": 97.77777777777777, "grad_norm": 2.484360493326676e-06, "learning_rate": 4.426408258905917e-05, "loss": 0.0, "num_input_tokens_seen": 2463744, "step": 8800 }, { "epoch": 97.77777777777777, "eval_loss": 1.1785892248153687, "eval_runtime": 1.3862, "eval_samples_per_second": 28.856, "eval_steps_per_second": 14.428, "num_input_tokens_seen": 2463744, "step": 8800 }, { "epoch": 97.83333333333333, "grad_norm": 3.1235961159836734e-06, "learning_rate": 4.425782380258594e-05, "loss": 0.0, "num_input_tokens_seen": 2465136, "step": 8805 }, { "epoch": 97.88888888888889, "grad_norm": 3.9482847569161095e-06, "learning_rate": 4.425156204631427e-05, "loss": 0.0, "num_input_tokens_seen": 2466512, "step": 8810 }, { "epoch": 97.94444444444444, "grad_norm": 1.5118322380658356e-06, "learning_rate": 4.424529732120981e-05, "loss": 0.0, "num_input_tokens_seen": 2467904, "step": 8815 }, { "epoch": 98.0, "grad_norm": 3.301710648884182e-06, "learning_rate": 4.423902962823864e-05, "loss": 0.0, "num_input_tokens_seen": 2469296, "step": 8820 }, { "epoch": 98.05555555555556, "grad_norm": 3.244146000724868e-06, "learning_rate": 4.423275896836733e-05, "loss": 0.0, "num_input_tokens_seen": 2470688, "step": 8825 }, { "epoch": 98.11111111111111, "grad_norm": 3.5261491575511172e-06, "learning_rate": 4.42264853425629e-05, "loss": 0.0, "num_input_tokens_seen": 2472080, "step": 8830 }, { "epoch": 98.16666666666667, "grad_norm": 3.597403292587842e-06, "learning_rate": 4.4220208751792816e-05, "loss": 0.0, "num_input_tokens_seen": 2473488, "step": 8835 }, { "epoch": 98.22222222222223, "grad_norm": 5.174958459974732e-06, "learning_rate": 4.421392919702499e-05, "loss": 0.0, "num_input_tokens_seen": 2474896, "step": 8840 }, { "epoch": 98.27777777777777, "grad_norm": 7.490465918635891e-07, "learning_rate": 4.4207646679227846e-05, "loss": 0.0, "num_input_tokens_seen": 2476272, "step": 8845 }, { "epoch": 98.33333333333333, "grad_norm": 4.235977485222975e-06, "learning_rate": 4.42013611993702e-05, "loss": 0.0, "num_input_tokens_seen": 2477696, "step": 8850 }, { "epoch": 98.38888888888889, "grad_norm": 2.0853979094681563e-06, "learning_rate": 4.419507275842135e-05, "loss": 0.0, "num_input_tokens_seen": 2479040, "step": 8855 }, { "epoch": 98.44444444444444, "grad_norm": 1.7316042431048118e-06, "learning_rate": 4.418878135735106e-05, "loss": 0.0, "num_input_tokens_seen": 2480496, "step": 8860 }, { "epoch": 98.5, "grad_norm": 1.1692490261339117e-05, "learning_rate": 4.418248699712955e-05, "loss": 0.0, "num_input_tokens_seen": 2481888, "step": 8865 }, { "epoch": 98.55555555555556, "grad_norm": 2.3351328763965284e-06, "learning_rate": 4.417618967872748e-05, "loss": 0.0, "num_input_tokens_seen": 2483280, "step": 8870 }, { "epoch": 98.61111111111111, "grad_norm": 1.6424562545580557e-06, "learning_rate": 4.4169889403115985e-05, "loss": 0.0, "num_input_tokens_seen": 2484704, "step": 8875 }, { "epoch": 98.66666666666667, "grad_norm": 3.4963134112331318e-06, "learning_rate": 4.4163586171266627e-05, "loss": 0.0, "num_input_tokens_seen": 2486080, "step": 8880 }, { "epoch": 98.72222222222223, "grad_norm": 1.8953284097733558e-06, "learning_rate": 4.415727998415147e-05, "loss": 0.0, "num_input_tokens_seen": 2487520, "step": 8885 }, { "epoch": 98.77777777777777, "grad_norm": 1.620067450858187e-06, "learning_rate": 4.4150970842742985e-05, "loss": 0.0, "num_input_tokens_seen": 2488944, "step": 8890 }, { "epoch": 98.83333333333333, "grad_norm": 2.9948444080218906e-06, "learning_rate": 4.4144658748014134e-05, "loss": 0.0, "num_input_tokens_seen": 2490320, "step": 8895 }, { "epoch": 98.88888888888889, "grad_norm": 2.0269485503376927e-06, "learning_rate": 4.413834370093831e-05, "loss": 0.0, "num_input_tokens_seen": 2491728, "step": 8900 }, { "epoch": 98.94444444444444, "grad_norm": 7.507586019528389e-07, "learning_rate": 4.413202570248939e-05, "loss": 0.0, "num_input_tokens_seen": 2493088, "step": 8905 }, { "epoch": 99.0, "grad_norm": 8.156981493812054e-06, "learning_rate": 4.412570475364167e-05, "loss": 0.0, "num_input_tokens_seen": 2494448, "step": 8910 }, { "epoch": 99.05555555555556, "grad_norm": 2.793288331304211e-06, "learning_rate": 4.411938085536994e-05, "loss": 0.0, "num_input_tokens_seen": 2495888, "step": 8915 }, { "epoch": 99.11111111111111, "grad_norm": 2.482109039192437e-06, "learning_rate": 4.41130540086494e-05, "loss": 0.0, "num_input_tokens_seen": 2497264, "step": 8920 }, { "epoch": 99.16666666666667, "grad_norm": 7.492506028938806e-06, "learning_rate": 4.4106724214455754e-05, "loss": 0.0, "num_input_tokens_seen": 2498624, "step": 8925 }, { "epoch": 99.22222222222223, "grad_norm": 4.408870154293254e-06, "learning_rate": 4.4100391473765115e-05, "loss": 0.0, "num_input_tokens_seen": 2500000, "step": 8930 }, { "epoch": 99.27777777777777, "grad_norm": 2.936085138571798e-06, "learning_rate": 4.409405578755408e-05, "loss": 0.0, "num_input_tokens_seen": 2501392, "step": 8935 }, { "epoch": 99.33333333333333, "grad_norm": 3.292562041679048e-06, "learning_rate": 4.4087717156799705e-05, "loss": 0.0, "num_input_tokens_seen": 2502768, "step": 8940 }, { "epoch": 99.38888888888889, "grad_norm": 1.9785002223215997e-06, "learning_rate": 4.408137558247946e-05, "loss": 0.0, "num_input_tokens_seen": 2504160, "step": 8945 }, { "epoch": 99.44444444444444, "grad_norm": 1.1010686193912989e-06, "learning_rate": 4.4075031065571306e-05, "loss": 0.0, "num_input_tokens_seen": 2505552, "step": 8950 }, { "epoch": 99.5, "grad_norm": 2.3395773496304173e-06, "learning_rate": 4.406868360705366e-05, "loss": 0.0, "num_input_tokens_seen": 2506912, "step": 8955 }, { "epoch": 99.55555555555556, "grad_norm": 3.2231619115918875e-06, "learning_rate": 4.406233320790536e-05, "loss": 0.0, "num_input_tokens_seen": 2508384, "step": 8960 }, { "epoch": 99.61111111111111, "grad_norm": 5.6087815210048575e-06, "learning_rate": 4.4055979869105734e-05, "loss": 0.0, "num_input_tokens_seen": 2509808, "step": 8965 }, { "epoch": 99.66666666666667, "grad_norm": 2.261271902170847e-06, "learning_rate": 4.404962359163454e-05, "loss": 0.0, "num_input_tokens_seen": 2511216, "step": 8970 }, { "epoch": 99.72222222222223, "grad_norm": 3.575579285097774e-06, "learning_rate": 4.404326437647199e-05, "loss": 0.0, "num_input_tokens_seen": 2512624, "step": 8975 }, { "epoch": 99.77777777777777, "grad_norm": 2.493823558324948e-06, "learning_rate": 4.403690222459877e-05, "loss": 0.0, "num_input_tokens_seen": 2514032, "step": 8980 }, { "epoch": 99.83333333333333, "grad_norm": 7.220457064249786e-06, "learning_rate": 4.4030537136995984e-05, "loss": 0.0, "num_input_tokens_seen": 2515456, "step": 8985 }, { "epoch": 99.88888888888889, "grad_norm": 1.6274606196020613e-06, "learning_rate": 4.402416911464523e-05, "loss": 0.0, "num_input_tokens_seen": 2516896, "step": 8990 }, { "epoch": 99.94444444444444, "grad_norm": 4.882162556896219e-06, "learning_rate": 4.4017798158528516e-05, "loss": 0.0, "num_input_tokens_seen": 2518304, "step": 8995 }, { "epoch": 100.0, "grad_norm": 1.2246742926436127e-06, "learning_rate": 4.401142426962834e-05, "loss": 0.0, "num_input_tokens_seen": 2519680, "step": 9000 }, { "epoch": 100.0, "eval_loss": 1.210661768913269, "eval_runtime": 1.3846, "eval_samples_per_second": 28.889, "eval_steps_per_second": 14.444, "num_input_tokens_seen": 2519680, "step": 9000 }, { "epoch": 100.05555555555556, "grad_norm": 1.1736912711057812e-05, "learning_rate": 4.400504744892763e-05, "loss": 0.0, "num_input_tokens_seen": 2521088, "step": 9005 }, { "epoch": 100.11111111111111, "grad_norm": 9.66054813034134e-06, "learning_rate": 4.399866769740975e-05, "loss": 0.0, "num_input_tokens_seen": 2522512, "step": 9010 }, { "epoch": 100.16666666666667, "grad_norm": 1.4731949704582803e-06, "learning_rate": 4.399228501605859e-05, "loss": 0.0, "num_input_tokens_seen": 2523904, "step": 9015 }, { "epoch": 100.22222222222223, "grad_norm": 3.469024932201137e-06, "learning_rate": 4.398589940585839e-05, "loss": 0.0, "num_input_tokens_seen": 2525264, "step": 9020 }, { "epoch": 100.27777777777777, "grad_norm": 1.43826787279977e-06, "learning_rate": 4.3979510867793917e-05, "loss": 0.0, "num_input_tokens_seen": 2526704, "step": 9025 }, { "epoch": 100.33333333333333, "grad_norm": 1.659051918068144e-06, "learning_rate": 4.3973119402850346e-05, "loss": 0.0, "num_input_tokens_seen": 2528112, "step": 9030 }, { "epoch": 100.38888888888889, "grad_norm": 2.0569557364069624e-06, "learning_rate": 4.396672501201334e-05, "loss": 0.0, "num_input_tokens_seen": 2529504, "step": 9035 }, { "epoch": 100.44444444444444, "grad_norm": 2.265043349325424e-06, "learning_rate": 4.396032769626899e-05, "loss": 0.0, "num_input_tokens_seen": 2530880, "step": 9040 }, { "epoch": 100.5, "grad_norm": 4.474431989365257e-06, "learning_rate": 4.395392745660384e-05, "loss": 0.0, "num_input_tokens_seen": 2532256, "step": 9045 }, { "epoch": 100.55555555555556, "grad_norm": 2.728588924583164e-06, "learning_rate": 4.394752429400488e-05, "loss": 0.0, "num_input_tokens_seen": 2533616, "step": 9050 }, { "epoch": 100.61111111111111, "grad_norm": 2.743254299275577e-06, "learning_rate": 4.394111820945957e-05, "loss": 0.0, "num_input_tokens_seen": 2535024, "step": 9055 }, { "epoch": 100.66666666666667, "grad_norm": 1.0265015362165286e-06, "learning_rate": 4.393470920395579e-05, "loss": 0.0, "num_input_tokens_seen": 2536432, "step": 9060 }, { "epoch": 100.72222222222223, "grad_norm": 2.155624997612904e-06, "learning_rate": 4.392829727848192e-05, "loss": 0.0, "num_input_tokens_seen": 2537824, "step": 9065 }, { "epoch": 100.77777777777777, "grad_norm": 2.5483725494268583e-06, "learning_rate": 4.392188243402673e-05, "loss": 0.0, "num_input_tokens_seen": 2539232, "step": 9070 }, { "epoch": 100.83333333333333, "grad_norm": 2.3570535176986596e-06, "learning_rate": 4.391546467157949e-05, "loss": 0.0, "num_input_tokens_seen": 2540608, "step": 9075 }, { "epoch": 100.88888888888889, "grad_norm": 1.2035786767228274e-06, "learning_rate": 4.390904399212988e-05, "loss": 0.0, "num_input_tokens_seen": 2542000, "step": 9080 }, { "epoch": 100.94444444444444, "grad_norm": 7.4253057391615584e-06, "learning_rate": 4.390262039666807e-05, "loss": 0.0, "num_input_tokens_seen": 2543408, "step": 9085 }, { "epoch": 101.0, "grad_norm": 1.6927931483223801e-06, "learning_rate": 4.389619388618464e-05, "loss": 0.0, "num_input_tokens_seen": 2544800, "step": 9090 }, { "epoch": 101.05555555555556, "grad_norm": 2.7677201615006197e-06, "learning_rate": 4.3889764461670655e-05, "loss": 0.0, "num_input_tokens_seen": 2546176, "step": 9095 }, { "epoch": 101.11111111111111, "grad_norm": 2.135876911779633e-06, "learning_rate": 4.38833321241176e-05, "loss": 0.0, "num_input_tokens_seen": 2547568, "step": 9100 }, { "epoch": 101.16666666666667, "grad_norm": 6.642660537181655e-06, "learning_rate": 4.3876896874517434e-05, "loss": 0.0, "num_input_tokens_seen": 2548992, "step": 9105 }, { "epoch": 101.22222222222223, "grad_norm": 1.1834866882054484e-06, "learning_rate": 4.3870458713862554e-05, "loss": 0.0, "num_input_tokens_seen": 2550400, "step": 9110 }, { "epoch": 101.27777777777777, "grad_norm": 3.7178419916017447e-06, "learning_rate": 4.386401764314579e-05, "loss": 0.0, "num_input_tokens_seen": 2551776, "step": 9115 }, { "epoch": 101.33333333333333, "grad_norm": 1.022037736220227e-06, "learning_rate": 4.385757366336045e-05, "loss": 0.0, "num_input_tokens_seen": 2553184, "step": 9120 }, { "epoch": 101.38888888888889, "grad_norm": 3.559480092008016e-06, "learning_rate": 4.385112677550027e-05, "loss": 0.0, "num_input_tokens_seen": 2554576, "step": 9125 }, { "epoch": 101.44444444444444, "grad_norm": 2.599491381261032e-06, "learning_rate": 4.384467698055945e-05, "loss": 0.0, "num_input_tokens_seen": 2555968, "step": 9130 }, { "epoch": 101.5, "grad_norm": 4.380755399324698e-06, "learning_rate": 4.383822427953261e-05, "loss": 0.0, "num_input_tokens_seen": 2557376, "step": 9135 }, { "epoch": 101.55555555555556, "grad_norm": 1.2430969036358874e-06, "learning_rate": 4.3831768673414864e-05, "loss": 0.0, "num_input_tokens_seen": 2558800, "step": 9140 }, { "epoch": 101.61111111111111, "grad_norm": 1.7733051436152891e-06, "learning_rate": 4.382531016320173e-05, "loss": 0.0, "num_input_tokens_seen": 2560208, "step": 9145 }, { "epoch": 101.66666666666667, "grad_norm": 4.037125563627342e-06, "learning_rate": 4.3818848749889184e-05, "loss": 0.0, "num_input_tokens_seen": 2561648, "step": 9150 }, { "epoch": 101.72222222222223, "grad_norm": 1.2301065908104647e-06, "learning_rate": 4.381238443447368e-05, "loss": 0.0, "num_input_tokens_seen": 2563024, "step": 9155 }, { "epoch": 101.77777777777777, "grad_norm": 1.574974362483772e-06, "learning_rate": 4.380591721795208e-05, "loss": 0.0, "num_input_tokens_seen": 2564416, "step": 9160 }, { "epoch": 101.83333333333333, "grad_norm": 1.9669300854729954e-06, "learning_rate": 4.3799447101321723e-05, "loss": 0.0, "num_input_tokens_seen": 2565856, "step": 9165 }, { "epoch": 101.88888888888889, "grad_norm": 3.216715413145721e-06, "learning_rate": 4.379297408558036e-05, "loss": 0.0, "num_input_tokens_seen": 2567232, "step": 9170 }, { "epoch": 101.94444444444444, "grad_norm": 4.75857859782991e-06, "learning_rate": 4.378649817172624e-05, "loss": 0.0, "num_input_tokens_seen": 2568608, "step": 9175 }, { "epoch": 102.0, "grad_norm": 2.2119240838947007e-06, "learning_rate": 4.378001936075801e-05, "loss": 0.0, "num_input_tokens_seen": 2569968, "step": 9180 }, { "epoch": 102.05555555555556, "grad_norm": 2.2043609533284325e-06, "learning_rate": 4.377353765367479e-05, "loss": 0.0, "num_input_tokens_seen": 2571392, "step": 9185 }, { "epoch": 102.11111111111111, "grad_norm": 2.3218839487526566e-05, "learning_rate": 4.376705305147614e-05, "loss": 0.0, "num_input_tokens_seen": 2572832, "step": 9190 }, { "epoch": 102.16666666666667, "grad_norm": 2.0263594251446193e-06, "learning_rate": 4.376056555516206e-05, "loss": 0.0, "num_input_tokens_seen": 2574208, "step": 9195 }, { "epoch": 102.22222222222223, "grad_norm": 1.498016104051203e-06, "learning_rate": 4.375407516573302e-05, "loss": 0.0, "num_input_tokens_seen": 2575584, "step": 9200 }, { "epoch": 102.22222222222223, "eval_loss": 1.1826832294464111, "eval_runtime": 1.3865, "eval_samples_per_second": 28.849, "eval_steps_per_second": 14.425, "num_input_tokens_seen": 2575584, "step": 9200 }, { "epoch": 102.27777777777777, "grad_norm": 1.6524965076314402e-06, "learning_rate": 4.3747581884189913e-05, "loss": 0.0, "num_input_tokens_seen": 2576976, "step": 9205 }, { "epoch": 102.33333333333333, "grad_norm": 2.5257199922634754e-06, "learning_rate": 4.374108571153408e-05, "loss": 0.0, "num_input_tokens_seen": 2578400, "step": 9210 }, { "epoch": 102.38888888888889, "grad_norm": 2.46285026150872e-06, "learning_rate": 4.3734586648767316e-05, "loss": 0.0, "num_input_tokens_seen": 2579824, "step": 9215 }, { "epoch": 102.44444444444444, "grad_norm": 1.035752211464569e-05, "learning_rate": 4.372808469689186e-05, "loss": 0.0, "num_input_tokens_seen": 2581248, "step": 9220 }, { "epoch": 102.5, "grad_norm": 2.1449310224852525e-05, "learning_rate": 4.372157985691039e-05, "loss": 0.0, "num_input_tokens_seen": 2582608, "step": 9225 }, { "epoch": 102.55555555555556, "grad_norm": 4.4255662032810505e-06, "learning_rate": 4.371507212982603e-05, "loss": 0.0, "num_input_tokens_seen": 2584064, "step": 9230 }, { "epoch": 102.61111111111111, "grad_norm": 4.347667982074199e-06, "learning_rate": 4.370856151664236e-05, "loss": 0.0, "num_input_tokens_seen": 2585424, "step": 9235 }, { "epoch": 102.66666666666667, "grad_norm": 1.2938232885062462e-06, "learning_rate": 4.3702048018363404e-05, "loss": 0.0, "num_input_tokens_seen": 2586880, "step": 9240 }, { "epoch": 102.72222222222223, "grad_norm": 1.2544189758045832e-06, "learning_rate": 4.369553163599362e-05, "loss": 0.0, "num_input_tokens_seen": 2588240, "step": 9245 }, { "epoch": 102.77777777777777, "grad_norm": 2.2401065962185385e-06, "learning_rate": 4.3689012370537904e-05, "loss": 0.0, "num_input_tokens_seen": 2589616, "step": 9250 }, { "epoch": 102.83333333333333, "grad_norm": 2.1141095203347504e-06, "learning_rate": 4.368249022300164e-05, "loss": 0.0, "num_input_tokens_seen": 2591008, "step": 9255 }, { "epoch": 102.88888888888889, "grad_norm": 2.629786422403413e-06, "learning_rate": 4.367596519439059e-05, "loss": 0.0, "num_input_tokens_seen": 2592432, "step": 9260 }, { "epoch": 102.94444444444444, "grad_norm": 2.8170907171443105e-06, "learning_rate": 4.366943728571101e-05, "loss": 0.0, "num_input_tokens_seen": 2593824, "step": 9265 }, { "epoch": 103.0, "grad_norm": 2.10228222385922e-06, "learning_rate": 4.366290649796959e-05, "loss": 0.0, "num_input_tokens_seen": 2595200, "step": 9270 }, { "epoch": 103.05555555555556, "grad_norm": 2.0097031665500253e-05, "learning_rate": 4.3656372832173456e-05, "loss": 0.0, "num_input_tokens_seen": 2596576, "step": 9275 }, { "epoch": 103.11111111111111, "grad_norm": 2.1425707927846815e-06, "learning_rate": 4.364983628933017e-05, "loss": 0.0, "num_input_tokens_seen": 2597952, "step": 9280 }, { "epoch": 103.16666666666667, "grad_norm": 2.159888026653789e-06, "learning_rate": 4.364329687044777e-05, "loss": 0.0, "num_input_tokens_seen": 2599376, "step": 9285 }, { "epoch": 103.22222222222223, "grad_norm": 8.429593094660959e-07, "learning_rate": 4.36367545765347e-05, "loss": 0.0, "num_input_tokens_seen": 2600832, "step": 9290 }, { "epoch": 103.27777777777777, "grad_norm": 1.168230141956883e-06, "learning_rate": 4.363020940859988e-05, "loss": 0.0, "num_input_tokens_seen": 2602240, "step": 9295 }, { "epoch": 103.33333333333333, "grad_norm": 9.436093932890799e-06, "learning_rate": 4.362366136765263e-05, "loss": 0.0, "num_input_tokens_seen": 2603696, "step": 9300 }, { "epoch": 103.38888888888889, "grad_norm": 2.875576001315494e-06, "learning_rate": 4.361711045470278e-05, "loss": 0.0, "num_input_tokens_seen": 2605088, "step": 9305 }, { "epoch": 103.44444444444444, "grad_norm": 2.8475869839894585e-06, "learning_rate": 4.3610556670760524e-05, "loss": 0.0, "num_input_tokens_seen": 2606496, "step": 9310 }, { "epoch": 103.5, "grad_norm": 1.232505724146904e-06, "learning_rate": 4.360400001683657e-05, "loss": 0.0, "num_input_tokens_seen": 2607920, "step": 9315 }, { "epoch": 103.55555555555556, "grad_norm": 1.975088025574223e-06, "learning_rate": 4.3597440493942e-05, "loss": 0.0, "num_input_tokens_seen": 2609328, "step": 9320 }, { "epoch": 103.61111111111111, "grad_norm": 3.847099378617713e-06, "learning_rate": 4.3590878103088405e-05, "loss": 0.0, "num_input_tokens_seen": 2610720, "step": 9325 }, { "epoch": 103.66666666666667, "grad_norm": 1.8915360442406381e-06, "learning_rate": 4.358431284528779e-05, "loss": 0.0, "num_input_tokens_seen": 2612144, "step": 9330 }, { "epoch": 103.72222222222223, "grad_norm": 2.695938519536867e-06, "learning_rate": 4.357774472155257e-05, "loss": 0.0, "num_input_tokens_seen": 2613536, "step": 9335 }, { "epoch": 103.77777777777777, "grad_norm": 3.717033678185544e-06, "learning_rate": 4.3571173732895664e-05, "loss": 0.0, "num_input_tokens_seen": 2614944, "step": 9340 }, { "epoch": 103.83333333333333, "grad_norm": 1.851783054007683e-06, "learning_rate": 4.356459988033039e-05, "loss": 0.0, "num_input_tokens_seen": 2616304, "step": 9345 }, { "epoch": 103.88888888888889, "grad_norm": 9.889529337669956e-07, "learning_rate": 4.355802316487051e-05, "loss": 0.0, "num_input_tokens_seen": 2617648, "step": 9350 }, { "epoch": 103.94444444444444, "grad_norm": 9.53698418015847e-07, "learning_rate": 4.355144358753025e-05, "loss": 0.0, "num_input_tokens_seen": 2619056, "step": 9355 }, { "epoch": 104.0, "grad_norm": 3.59433465746406e-06, "learning_rate": 4.354486114932425e-05, "loss": 0.0, "num_input_tokens_seen": 2620432, "step": 9360 }, { "epoch": 104.05555555555556, "grad_norm": 1.1778002772189211e-05, "learning_rate": 4.353827585126762e-05, "loss": 0.0, "num_input_tokens_seen": 2621808, "step": 9365 }, { "epoch": 104.11111111111111, "grad_norm": 6.603612746403087e-06, "learning_rate": 4.353168769437588e-05, "loss": 0.0, "num_input_tokens_seen": 2623264, "step": 9370 }, { "epoch": 104.16666666666667, "grad_norm": 1.1709006457749638e-06, "learning_rate": 4.3525096679665014e-05, "loss": 0.0, "num_input_tokens_seen": 2624624, "step": 9375 }, { "epoch": 104.22222222222223, "grad_norm": 1.675433736636478e-06, "learning_rate": 4.351850280815144e-05, "loss": 0.0, "num_input_tokens_seen": 2626064, "step": 9380 }, { "epoch": 104.27777777777777, "grad_norm": 1.324967342952732e-06, "learning_rate": 4.3511906080852014e-05, "loss": 0.0, "num_input_tokens_seen": 2627472, "step": 9385 }, { "epoch": 104.33333333333333, "grad_norm": 3.0298569981823675e-06, "learning_rate": 4.350530649878404e-05, "loss": 0.0, "num_input_tokens_seen": 2628864, "step": 9390 }, { "epoch": 104.38888888888889, "grad_norm": 8.601919944339897e-06, "learning_rate": 4.3498704062965246e-05, "loss": 0.0, "num_input_tokens_seen": 2630272, "step": 9395 }, { "epoch": 104.44444444444444, "grad_norm": 2.310421450602007e-06, "learning_rate": 4.3492098774413815e-05, "loss": 0.0, "num_input_tokens_seen": 2631680, "step": 9400 }, { "epoch": 104.44444444444444, "eval_loss": 1.1795517206192017, "eval_runtime": 1.3843, "eval_samples_per_second": 28.896, "eval_steps_per_second": 14.448, "num_input_tokens_seen": 2631680, "step": 9400 }, { "epoch": 104.5, "grad_norm": 1.019482260744553e-05, "learning_rate": 4.3485490634148375e-05, "loss": 0.0, "num_input_tokens_seen": 2633088, "step": 9405 }, { "epoch": 104.55555555555556, "grad_norm": 1.8584260033094324e-06, "learning_rate": 4.347887964318797e-05, "loss": 0.0, "num_input_tokens_seen": 2634480, "step": 9410 }, { "epoch": 104.61111111111111, "grad_norm": 1.019077785713307e-06, "learning_rate": 4.34722658025521e-05, "loss": 0.0, "num_input_tokens_seen": 2635824, "step": 9415 }, { "epoch": 104.66666666666667, "grad_norm": 2.384287427048548e-06, "learning_rate": 4.346564911326071e-05, "loss": 0.0, "num_input_tokens_seen": 2637264, "step": 9420 }, { "epoch": 104.72222222222223, "grad_norm": 4.11923883802956e-06, "learning_rate": 4.345902957633418e-05, "loss": 0.0, "num_input_tokens_seen": 2638688, "step": 9425 }, { "epoch": 104.77777777777777, "grad_norm": 4.138335498282686e-06, "learning_rate": 4.345240719279331e-05, "loss": 0.0, "num_input_tokens_seen": 2640080, "step": 9430 }, { "epoch": 104.83333333333333, "grad_norm": 1.1949852023462881e-06, "learning_rate": 4.3445781963659374e-05, "loss": 0.0, "num_input_tokens_seen": 2641456, "step": 9435 }, { "epoch": 104.88888888888889, "grad_norm": 2.161742031603353e-06, "learning_rate": 4.3439153889954045e-05, "loss": 0.0, "num_input_tokens_seen": 2642832, "step": 9440 }, { "epoch": 104.94444444444444, "grad_norm": 2.419210659354576e-06, "learning_rate": 4.343252297269946e-05, "loss": 0.0, "num_input_tokens_seen": 2644240, "step": 9445 }, { "epoch": 105.0, "grad_norm": 1.9872516077157343e-06, "learning_rate": 4.342588921291821e-05, "loss": 0.0, "num_input_tokens_seen": 2645648, "step": 9450 }, { "epoch": 105.05555555555556, "grad_norm": 2.130449502146803e-06, "learning_rate": 4.341925261163328e-05, "loss": 0.0, "num_input_tokens_seen": 2647056, "step": 9455 }, { "epoch": 105.11111111111111, "grad_norm": 1.9423584944888717e-06, "learning_rate": 4.341261316986813e-05, "loss": 0.0, "num_input_tokens_seen": 2648432, "step": 9460 }, { "epoch": 105.16666666666667, "grad_norm": 1.8538318045102642e-06, "learning_rate": 4.340597088864664e-05, "loss": 0.0, "num_input_tokens_seen": 2649840, "step": 9465 }, { "epoch": 105.22222222222223, "grad_norm": 4.325888312450843e-06, "learning_rate": 4.339932576899313e-05, "loss": 0.0, "num_input_tokens_seen": 2651296, "step": 9470 }, { "epoch": 105.27777777777777, "grad_norm": 7.482896307919873e-06, "learning_rate": 4.3392677811932375e-05, "loss": 0.0, "num_input_tokens_seen": 2652672, "step": 9475 }, { "epoch": 105.33333333333333, "grad_norm": 3.7697150219173636e-06, "learning_rate": 4.338602701848956e-05, "loss": 0.0, "num_input_tokens_seen": 2654080, "step": 9480 }, { "epoch": 105.38888888888889, "grad_norm": 5.295028131513391e-07, "learning_rate": 4.337937338969033e-05, "loss": 0.0, "num_input_tokens_seen": 2655504, "step": 9485 }, { "epoch": 105.44444444444444, "grad_norm": 6.841813956270926e-06, "learning_rate": 4.337271692656075e-05, "loss": 0.0, "num_input_tokens_seen": 2656928, "step": 9490 }, { "epoch": 105.5, "grad_norm": 9.92982222669525e-06, "learning_rate": 4.336605763012733e-05, "loss": 0.0, "num_input_tokens_seen": 2658320, "step": 9495 }, { "epoch": 105.55555555555556, "grad_norm": 1.7950484334505745e-06, "learning_rate": 4.3359395501417026e-05, "loss": 0.0, "num_input_tokens_seen": 2659696, "step": 9500 }, { "epoch": 105.61111111111111, "grad_norm": 1.928171514009591e-06, "learning_rate": 4.335273054145722e-05, "loss": 0.0, "num_input_tokens_seen": 2661056, "step": 9505 }, { "epoch": 105.66666666666667, "grad_norm": 1.5727207482996164e-06, "learning_rate": 4.334606275127572e-05, "loss": 0.0, "num_input_tokens_seen": 2662480, "step": 9510 }, { "epoch": 105.72222222222223, "grad_norm": 1.938162313308567e-06, "learning_rate": 4.33393921319008e-05, "loss": 0.0, "num_input_tokens_seen": 2663872, "step": 9515 }, { "epoch": 105.77777777777777, "grad_norm": 1.210617256219848e-06, "learning_rate": 4.3332718684361146e-05, "loss": 0.0, "num_input_tokens_seen": 2665264, "step": 9520 }, { "epoch": 105.83333333333333, "grad_norm": 6.803060841775732e-06, "learning_rate": 4.332604240968588e-05, "loss": 0.0, "num_input_tokens_seen": 2666672, "step": 9525 }, { "epoch": 105.88888888888889, "grad_norm": 1.4793484979236382e-06, "learning_rate": 4.331936330890459e-05, "loss": 0.0, "num_input_tokens_seen": 2668080, "step": 9530 }, { "epoch": 105.94444444444444, "grad_norm": 1.815040150177083e-06, "learning_rate": 4.331268138304725e-05, "loss": 0.0, "num_input_tokens_seen": 2669472, "step": 9535 }, { "epoch": 106.0, "grad_norm": 1.5240964330587303e-06, "learning_rate": 4.330599663314431e-05, "loss": 0.0, "num_input_tokens_seen": 2670864, "step": 9540 }, { "epoch": 106.05555555555556, "grad_norm": 3.139550472042174e-06, "learning_rate": 4.329930906022665e-05, "loss": 0.0, "num_input_tokens_seen": 2672288, "step": 9545 }, { "epoch": 106.11111111111111, "grad_norm": 8.671532327753084e-07, "learning_rate": 4.3292618665325564e-05, "loss": 0.0, "num_input_tokens_seen": 2673712, "step": 9550 }, { "epoch": 106.16666666666667, "grad_norm": 2.1999683212925447e-06, "learning_rate": 4.3285925449472796e-05, "loss": 0.0, "num_input_tokens_seen": 2675104, "step": 9555 }, { "epoch": 106.22222222222223, "grad_norm": 3.107609018115909e-06, "learning_rate": 4.327922941370054e-05, "loss": 0.0, "num_input_tokens_seen": 2676464, "step": 9560 }, { "epoch": 106.27777777777777, "grad_norm": 1.2866635188402142e-06, "learning_rate": 4.3272530559041384e-05, "loss": 0.0, "num_input_tokens_seen": 2677904, "step": 9565 }, { "epoch": 106.33333333333333, "grad_norm": 1.0944421774183866e-05, "learning_rate": 4.32658288865284e-05, "loss": 0.0, "num_input_tokens_seen": 2679280, "step": 9570 }, { "epoch": 106.38888888888889, "grad_norm": 7.706320502620656e-06, "learning_rate": 4.325912439719505e-05, "loss": 0.0, "num_input_tokens_seen": 2680672, "step": 9575 }, { "epoch": 106.44444444444444, "grad_norm": 1.1179973853359115e-06, "learning_rate": 4.3252417092075266e-05, "loss": 0.0, "num_input_tokens_seen": 2682064, "step": 9580 }, { "epoch": 106.5, "grad_norm": 1.4995756600910681e-06, "learning_rate": 4.3245706972203385e-05, "loss": 0.0, "num_input_tokens_seen": 2683488, "step": 9585 }, { "epoch": 106.55555555555556, "grad_norm": 2.175462896047975e-06, "learning_rate": 4.323899403861421e-05, "loss": 0.0, "num_input_tokens_seen": 2684896, "step": 9590 }, { "epoch": 106.61111111111111, "grad_norm": 1.8176099274569424e-06, "learning_rate": 4.3232278292342935e-05, "loss": 0.0, "num_input_tokens_seen": 2686272, "step": 9595 }, { "epoch": 106.66666666666667, "grad_norm": 4.442129011295037e-06, "learning_rate": 4.322555973442524e-05, "loss": 0.0, "num_input_tokens_seen": 2687728, "step": 9600 }, { "epoch": 106.66666666666667, "eval_loss": 1.228224515914917, "eval_runtime": 1.3881, "eval_samples_per_second": 28.816, "eval_steps_per_second": 14.408, "num_input_tokens_seen": 2687728, "step": 9600 }, { "epoch": 106.72222222222223, "grad_norm": 1.8023654320131755e-06, "learning_rate": 4.3218838365897184e-05, "loss": 0.0, "num_input_tokens_seen": 2689136, "step": 9605 }, { "epoch": 106.77777777777777, "grad_norm": 3.4072058952006046e-06, "learning_rate": 4.3212114187795306e-05, "loss": 0.0, "num_input_tokens_seen": 2690544, "step": 9610 }, { "epoch": 106.83333333333333, "grad_norm": 3.1992344702302944e-06, "learning_rate": 4.320538720115656e-05, "loss": 0.0, "num_input_tokens_seen": 2691936, "step": 9615 }, { "epoch": 106.88888888888889, "grad_norm": 2.4692901661182987e-06, "learning_rate": 4.319865740701831e-05, "loss": 0.0, "num_input_tokens_seen": 2693344, "step": 9620 }, { "epoch": 106.94444444444444, "grad_norm": 2.085767164317076e-06, "learning_rate": 4.3191924806418396e-05, "loss": 0.0, "num_input_tokens_seen": 2694720, "step": 9625 }, { "epoch": 107.0, "grad_norm": 1.1869930176544585e-06, "learning_rate": 4.318518940039507e-05, "loss": 0.0, "num_input_tokens_seen": 2696112, "step": 9630 }, { "epoch": 107.05555555555556, "grad_norm": 1.6512062757101376e-06, "learning_rate": 4.3178451189987e-05, "loss": 0.0, "num_input_tokens_seen": 2697536, "step": 9635 }, { "epoch": 107.11111111111111, "grad_norm": 6.201731594046578e-06, "learning_rate": 4.3171710176233315e-05, "loss": 0.0, "num_input_tokens_seen": 2698976, "step": 9640 }, { "epoch": 107.16666666666667, "grad_norm": 3.843466402031481e-06, "learning_rate": 4.316496636017355e-05, "loss": 0.0, "num_input_tokens_seen": 2700384, "step": 9645 }, { "epoch": 107.22222222222223, "grad_norm": 7.239517344714841e-06, "learning_rate": 4.315821974284771e-05, "loss": 0.0, "num_input_tokens_seen": 2701792, "step": 9650 }, { "epoch": 107.27777777777777, "grad_norm": 5.117417458677664e-06, "learning_rate": 4.315147032529619e-05, "loss": 0.0, "num_input_tokens_seen": 2703184, "step": 9655 }, { "epoch": 107.33333333333333, "grad_norm": 2.35040788538754e-05, "learning_rate": 4.3144718108559845e-05, "loss": 0.0, "num_input_tokens_seen": 2704560, "step": 9660 }, { "epoch": 107.38888888888889, "grad_norm": 1.1627189451246522e-06, "learning_rate": 4.3137963093679945e-05, "loss": 0.0, "num_input_tokens_seen": 2705968, "step": 9665 }, { "epoch": 107.44444444444444, "grad_norm": 2.188733560615219e-06, "learning_rate": 4.31312052816982e-05, "loss": 0.0, "num_input_tokens_seen": 2707376, "step": 9670 }, { "epoch": 107.5, "grad_norm": 1.391177420373424e-06, "learning_rate": 4.312444467365675e-05, "loss": 0.0, "num_input_tokens_seen": 2708768, "step": 9675 }, { "epoch": 107.55555555555556, "grad_norm": 6.914457571838284e-06, "learning_rate": 4.311768127059816e-05, "loss": 0.0, "num_input_tokens_seen": 2710160, "step": 9680 }, { "epoch": 107.61111111111111, "grad_norm": 5.5631649047427345e-06, "learning_rate": 4.3110915073565444e-05, "loss": 0.0, "num_input_tokens_seen": 2711584, "step": 9685 }, { "epoch": 107.66666666666667, "grad_norm": 1.3051629821347888e-06, "learning_rate": 4.310414608360203e-05, "loss": 0.0, "num_input_tokens_seen": 2712960, "step": 9690 }, { "epoch": 107.72222222222223, "grad_norm": 6.2891090237826575e-06, "learning_rate": 4.309737430175177e-05, "loss": 0.0, "num_input_tokens_seen": 2714368, "step": 9695 }, { "epoch": 107.77777777777777, "grad_norm": 6.507983016490471e-06, "learning_rate": 4.309059972905897e-05, "loss": 0.0, "num_input_tokens_seen": 2715760, "step": 9700 }, { "epoch": 107.83333333333333, "grad_norm": 3.970613306591986e-06, "learning_rate": 4.308382236656836e-05, "loss": 0.0, "num_input_tokens_seen": 2717200, "step": 9705 }, { "epoch": 107.88888888888889, "grad_norm": 1.5144396456889808e-06, "learning_rate": 4.307704221532507e-05, "loss": 0.0, "num_input_tokens_seen": 2718624, "step": 9710 }, { "epoch": 107.94444444444444, "grad_norm": 1.3382394854488666e-06, "learning_rate": 4.307025927637471e-05, "loss": 0.0, "num_input_tokens_seen": 2720032, "step": 9715 }, { "epoch": 108.0, "grad_norm": 4.683280167228077e-06, "learning_rate": 4.306347355076328e-05, "loss": 0.0, "num_input_tokens_seen": 2721392, "step": 9720 }, { "epoch": 108.05555555555556, "grad_norm": 3.456307922533597e-06, "learning_rate": 4.305668503953724e-05, "loss": 0.0, "num_input_tokens_seen": 2722784, "step": 9725 }, { "epoch": 108.11111111111111, "grad_norm": 1.3184451290726429e-06, "learning_rate": 4.3049893743743436e-05, "loss": 0.0, "num_input_tokens_seen": 2724192, "step": 9730 }, { "epoch": 108.16666666666667, "grad_norm": 1.1431235407144413e-06, "learning_rate": 4.304309966442919e-05, "loss": 0.0, "num_input_tokens_seen": 2725600, "step": 9735 }, { "epoch": 108.22222222222223, "grad_norm": 1.8283963072462939e-06, "learning_rate": 4.303630280264224e-05, "loss": 0.0, "num_input_tokens_seen": 2726992, "step": 9740 }, { "epoch": 108.27777777777777, "grad_norm": 5.024228812544607e-06, "learning_rate": 4.302950315943074e-05, "loss": 0.0, "num_input_tokens_seen": 2728416, "step": 9745 }, { "epoch": 108.33333333333333, "grad_norm": 2.280523858644301e-06, "learning_rate": 4.3022700735843275e-05, "loss": 0.0, "num_input_tokens_seen": 2729792, "step": 9750 }, { "epoch": 108.38888888888889, "grad_norm": 4.946959961671382e-06, "learning_rate": 4.301589553292887e-05, "loss": 0.0, "num_input_tokens_seen": 2731200, "step": 9755 }, { "epoch": 108.44444444444444, "grad_norm": 2.374989890085999e-06, "learning_rate": 4.300908755173697e-05, "loss": 0.0, "num_input_tokens_seen": 2732640, "step": 9760 }, { "epoch": 108.5, "grad_norm": 2.446180587867275e-06, "learning_rate": 4.300227679331745e-05, "loss": 0.0, "num_input_tokens_seen": 2734016, "step": 9765 }, { "epoch": 108.55555555555556, "grad_norm": 1.0247201316815335e-05, "learning_rate": 4.299546325872063e-05, "loss": 0.0, "num_input_tokens_seen": 2735376, "step": 9770 }, { "epoch": 108.61111111111111, "grad_norm": 2.188101007050136e-06, "learning_rate": 4.2988646948997225e-05, "loss": 0.0, "num_input_tokens_seen": 2736800, "step": 9775 }, { "epoch": 108.66666666666667, "grad_norm": 4.502674073592061e-06, "learning_rate": 4.29818278651984e-05, "loss": 0.0, "num_input_tokens_seen": 2738176, "step": 9780 }, { "epoch": 108.72222222222223, "grad_norm": 9.078512448468246e-07, "learning_rate": 4.297500600837574e-05, "loss": 0.0, "num_input_tokens_seen": 2739552, "step": 9785 }, { "epoch": 108.77777777777777, "grad_norm": 8.122233339236118e-07, "learning_rate": 4.2968181379581276e-05, "loss": 0.0, "num_input_tokens_seen": 2740960, "step": 9790 }, { "epoch": 108.83333333333333, "grad_norm": 1.9798821995209437e-06, "learning_rate": 4.296135397986743e-05, "loss": 0.0, "num_input_tokens_seen": 2742368, "step": 9795 }, { "epoch": 108.88888888888889, "grad_norm": 2.708794681893778e-06, "learning_rate": 4.295452381028709e-05, "loss": 0.0, "num_input_tokens_seen": 2743792, "step": 9800 }, { "epoch": 108.88888888888889, "eval_loss": 1.2008123397827148, "eval_runtime": 1.3924, "eval_samples_per_second": 28.728, "eval_steps_per_second": 14.364, "num_input_tokens_seen": 2743792, "step": 9800 }, { "epoch": 108.94444444444444, "grad_norm": 1.5786367839609738e-06, "learning_rate": 4.294769087189354e-05, "loss": 0.0, "num_input_tokens_seen": 2745184, "step": 9805 }, { "epoch": 109.0, "grad_norm": 1.2853531643486349e-06, "learning_rate": 4.294085516574052e-05, "loss": 0.0, "num_input_tokens_seen": 2746592, "step": 9810 }, { "epoch": 109.05555555555556, "grad_norm": 3.2947107229119865e-06, "learning_rate": 4.2934016692882176e-05, "loss": 0.0, "num_input_tokens_seen": 2748016, "step": 9815 }, { "epoch": 109.11111111111111, "grad_norm": 1.3026526630710578e-06, "learning_rate": 4.292717545437308e-05, "loss": 0.0, "num_input_tokens_seen": 2749408, "step": 9820 }, { "epoch": 109.16666666666667, "grad_norm": 3.0888195396983065e-06, "learning_rate": 4.292033145126825e-05, "loss": 0.0, "num_input_tokens_seen": 2750816, "step": 9825 }, { "epoch": 109.22222222222223, "grad_norm": 2.2622302822128404e-06, "learning_rate": 4.29134846846231e-05, "loss": 0.0, "num_input_tokens_seen": 2752224, "step": 9830 }, { "epoch": 109.27777777777777, "grad_norm": 9.476798368268646e-06, "learning_rate": 4.29066351554935e-05, "loss": 0.0, "num_input_tokens_seen": 2753616, "step": 9835 }, { "epoch": 109.33333333333333, "grad_norm": 5.110591246193508e-06, "learning_rate": 4.289978286493574e-05, "loss": 0.0, "num_input_tokens_seen": 2754992, "step": 9840 }, { "epoch": 109.38888888888889, "grad_norm": 1.5910893580439733e-06, "learning_rate": 4.28929278140065e-05, "loss": 0.0, "num_input_tokens_seen": 2756400, "step": 9845 }, { "epoch": 109.44444444444444, "grad_norm": 2.550696763137239e-06, "learning_rate": 4.288607000376295e-05, "loss": 0.0, "num_input_tokens_seen": 2757760, "step": 9850 }, { "epoch": 109.5, "grad_norm": 2.1866483166377293e-06, "learning_rate": 4.2879209435262624e-05, "loss": 0.0, "num_input_tokens_seen": 2759216, "step": 9855 }, { "epoch": 109.55555555555556, "grad_norm": 2.428788320685271e-06, "learning_rate": 4.287234610956353e-05, "loss": 0.0, "num_input_tokens_seen": 2760576, "step": 9860 }, { "epoch": 109.61111111111111, "grad_norm": 1.9777787656494183e-06, "learning_rate": 4.2865480027724056e-05, "loss": 0.0, "num_input_tokens_seen": 2761968, "step": 9865 }, { "epoch": 109.66666666666667, "grad_norm": 1.433772558812052e-05, "learning_rate": 4.285861119080306e-05, "loss": 0.0, "num_input_tokens_seen": 2763376, "step": 9870 }, { "epoch": 109.72222222222223, "grad_norm": 9.518522347207181e-06, "learning_rate": 4.2851739599859784e-05, "loss": 0.0, "num_input_tokens_seen": 2764768, "step": 9875 }, { "epoch": 109.77777777777777, "grad_norm": 3.8789203244959936e-06, "learning_rate": 4.2844865255953934e-05, "loss": 0.0, "num_input_tokens_seen": 2766208, "step": 9880 }, { "epoch": 109.83333333333333, "grad_norm": 2.112830316036707e-06, "learning_rate": 4.2837988160145605e-05, "loss": 0.0, "num_input_tokens_seen": 2767616, "step": 9885 }, { "epoch": 109.88888888888889, "grad_norm": 9.107834557653405e-07, "learning_rate": 4.2831108313495336e-05, "loss": 0.0, "num_input_tokens_seen": 2769008, "step": 9890 }, { "epoch": 109.94444444444444, "grad_norm": 3.4246152154082665e-06, "learning_rate": 4.282422571706408e-05, "loss": 0.0, "num_input_tokens_seen": 2770416, "step": 9895 }, { "epoch": 110.0, "grad_norm": 3.524799922161037e-06, "learning_rate": 4.281734037191323e-05, "loss": 0.0, "num_input_tokens_seen": 2771824, "step": 9900 }, { "epoch": 110.05555555555556, "grad_norm": 2.0209381546010263e-06, "learning_rate": 4.281045227910459e-05, "loss": 0.0, "num_input_tokens_seen": 2773216, "step": 9905 }, { "epoch": 110.11111111111111, "grad_norm": 1.0310938023394556e-06, "learning_rate": 4.280356143970038e-05, "loss": 0.0, "num_input_tokens_seen": 2774608, "step": 9910 }, { "epoch": 110.16666666666667, "grad_norm": 3.5964712878922e-06, "learning_rate": 4.279666785476327e-05, "loss": 0.0, "num_input_tokens_seen": 2776064, "step": 9915 }, { "epoch": 110.22222222222223, "grad_norm": 1.758450707711745e-06, "learning_rate": 4.2789771525356325e-05, "loss": 0.0, "num_input_tokens_seen": 2777472, "step": 9920 }, { "epoch": 110.27777777777777, "grad_norm": 3.285678758402355e-06, "learning_rate": 4.2782872452543056e-05, "loss": 0.0, "num_input_tokens_seen": 2778896, "step": 9925 }, { "epoch": 110.33333333333333, "grad_norm": 7.295133855222957e-06, "learning_rate": 4.2775970637387376e-05, "loss": 0.0, "num_input_tokens_seen": 2780272, "step": 9930 }, { "epoch": 110.38888888888889, "grad_norm": 2.351327793803648e-06, "learning_rate": 4.276906608095363e-05, "loss": 0.0, "num_input_tokens_seen": 2781680, "step": 9935 }, { "epoch": 110.44444444444444, "grad_norm": 7.317950348806335e-06, "learning_rate": 4.276215878430661e-05, "loss": 0.0, "num_input_tokens_seen": 2783072, "step": 9940 }, { "epoch": 110.5, "grad_norm": 7.313858532143058e-06, "learning_rate": 4.275524874851149e-05, "loss": 0.0, "num_input_tokens_seen": 2784480, "step": 9945 }, { "epoch": 110.55555555555556, "grad_norm": 1.2331181551417103e-06, "learning_rate": 4.274833597463388e-05, "loss": 0.0, "num_input_tokens_seen": 2785904, "step": 9950 }, { "epoch": 110.61111111111111, "grad_norm": 1.6768080968176946e-06, "learning_rate": 4.2741420463739824e-05, "loss": 0.0, "num_input_tokens_seen": 2787328, "step": 9955 }, { "epoch": 110.66666666666667, "grad_norm": 2.00359590962762e-06, "learning_rate": 4.273450221689578e-05, "loss": 0.0, "num_input_tokens_seen": 2788720, "step": 9960 }, { "epoch": 110.72222222222223, "grad_norm": 6.655276138189947e-06, "learning_rate": 4.272758123516863e-05, "loss": 0.0, "num_input_tokens_seen": 2790160, "step": 9965 }, { "epoch": 110.77777777777777, "grad_norm": 1.7394984297425253e-06, "learning_rate": 4.272065751962567e-05, "loss": 0.0, "num_input_tokens_seen": 2791552, "step": 9970 }, { "epoch": 110.83333333333333, "grad_norm": 1.1926667866646312e-06, "learning_rate": 4.271373107133464e-05, "loss": 0.0, "num_input_tokens_seen": 2792928, "step": 9975 }, { "epoch": 110.88888888888889, "grad_norm": 6.913949164299993e-06, "learning_rate": 4.270680189136366e-05, "loss": 0.0, "num_input_tokens_seen": 2794304, "step": 9980 }, { "epoch": 110.94444444444444, "grad_norm": 2.7286271233606385e-06, "learning_rate": 4.269986998078132e-05, "loss": 0.0, "num_input_tokens_seen": 2795648, "step": 9985 }, { "epoch": 111.0, "grad_norm": 2.1132168512849603e-06, "learning_rate": 4.2692935340656595e-05, "loss": 0.0, "num_input_tokens_seen": 2797056, "step": 9990 }, { "epoch": 111.05555555555556, "grad_norm": 2.3751003936922643e-06, "learning_rate": 4.26859979720589e-05, "loss": 0.0, "num_input_tokens_seen": 2798432, "step": 9995 }, { "epoch": 111.11111111111111, "grad_norm": 3.1170350212050835e-06, "learning_rate": 4.267905787605806e-05, "loss": 0.0, "num_input_tokens_seen": 2799840, "step": 10000 }, { "epoch": 111.11111111111111, "eval_loss": 1.1934638023376465, "eval_runtime": 1.4041, "eval_samples_per_second": 28.488, "eval_steps_per_second": 14.244, "num_input_tokens_seen": 2799840, "step": 10000 }, { "epoch": 111.16666666666667, "grad_norm": 2.015052359638503e-06, "learning_rate": 4.267211505372433e-05, "loss": 0.0, "num_input_tokens_seen": 2801232, "step": 10005 }, { "epoch": 111.22222222222223, "grad_norm": 2.799805770337116e-06, "learning_rate": 4.266516950612837e-05, "loss": 0.0, "num_input_tokens_seen": 2802608, "step": 10010 }, { "epoch": 111.27777777777777, "grad_norm": 4.075604465469951e-06, "learning_rate": 4.265822123434128e-05, "loss": 0.0, "num_input_tokens_seen": 2804032, "step": 10015 }, { "epoch": 111.33333333333333, "grad_norm": 3.6539536267810035e-06, "learning_rate": 4.265127023943457e-05, "loss": 0.0, "num_input_tokens_seen": 2805440, "step": 10020 }, { "epoch": 111.38888888888889, "grad_norm": 4.502099272940541e-06, "learning_rate": 4.2644316522480176e-05, "loss": 0.0, "num_input_tokens_seen": 2806864, "step": 10025 }, { "epoch": 111.44444444444444, "grad_norm": 1.8583368728286587e-06, "learning_rate": 4.263736008455044e-05, "loss": 0.0, "num_input_tokens_seen": 2808272, "step": 10030 }, { "epoch": 111.5, "grad_norm": 8.889869604900014e-06, "learning_rate": 4.2630400926718125e-05, "loss": 0.0, "num_input_tokens_seen": 2809664, "step": 10035 }, { "epoch": 111.55555555555556, "grad_norm": 1.2425273325789021e-06, "learning_rate": 4.262343905005644e-05, "loss": 0.0, "num_input_tokens_seen": 2811072, "step": 10040 }, { "epoch": 111.61111111111111, "grad_norm": 1.946922793649719e-06, "learning_rate": 4.261647445563897e-05, "loss": 0.0, "num_input_tokens_seen": 2812464, "step": 10045 }, { "epoch": 111.66666666666667, "grad_norm": 9.417989531357307e-06, "learning_rate": 4.260950714453976e-05, "loss": 0.0, "num_input_tokens_seen": 2813856, "step": 10050 }, { "epoch": 111.72222222222223, "grad_norm": 2.3989223336684518e-06, "learning_rate": 4.2602537117833266e-05, "loss": 0.0, "num_input_tokens_seen": 2815232, "step": 10055 }, { "epoch": 111.77777777777777, "grad_norm": 7.654045475646853e-06, "learning_rate": 4.259556437659433e-05, "loss": 0.0, "num_input_tokens_seen": 2816608, "step": 10060 }, { "epoch": 111.83333333333333, "grad_norm": 1.3022460052525275e-06, "learning_rate": 4.258858892189825e-05, "loss": 0.0, "num_input_tokens_seen": 2818016, "step": 10065 }, { "epoch": 111.88888888888889, "grad_norm": 1.9187937141396105e-06, "learning_rate": 4.2581610754820725e-05, "loss": 0.0, "num_input_tokens_seen": 2819376, "step": 10070 }, { "epoch": 111.94444444444444, "grad_norm": 4.248438472131966e-06, "learning_rate": 4.2574629876437876e-05, "loss": 0.0, "num_input_tokens_seen": 2820752, "step": 10075 }, { "epoch": 112.0, "grad_norm": 2.782899173325859e-06, "learning_rate": 4.256764628782625e-05, "loss": 0.0, "num_input_tokens_seen": 2822160, "step": 10080 }, { "epoch": 112.05555555555556, "grad_norm": 4.133388301852392e-06, "learning_rate": 4.256065999006279e-05, "loss": 0.0, "num_input_tokens_seen": 2823520, "step": 10085 }, { "epoch": 112.11111111111111, "grad_norm": 2.0356790173536865e-06, "learning_rate": 4.2553670984224885e-05, "loss": 0.0, "num_input_tokens_seen": 2824960, "step": 10090 }, { "epoch": 112.16666666666667, "grad_norm": 7.876436029619072e-06, "learning_rate": 4.254667927139032e-05, "loss": 0.0, "num_input_tokens_seen": 2826368, "step": 10095 }, { "epoch": 112.22222222222223, "grad_norm": 4.177511982561555e-06, "learning_rate": 4.2539684852637295e-05, "loss": 0.0, "num_input_tokens_seen": 2827760, "step": 10100 }, { "epoch": 112.27777777777777, "grad_norm": 4.578237621899461e-06, "learning_rate": 4.253268772904446e-05, "loss": 0.0, "num_input_tokens_seen": 2829152, "step": 10105 }, { "epoch": 112.33333333333333, "grad_norm": 9.233014566234488e-07, "learning_rate": 4.252568790169085e-05, "loss": 0.0, "num_input_tokens_seen": 2830576, "step": 10110 }, { "epoch": 112.38888888888889, "grad_norm": 1.2734765277855331e-06, "learning_rate": 4.251868537165592e-05, "loss": 0.0, "num_input_tokens_seen": 2831968, "step": 10115 }, { "epoch": 112.44444444444444, "grad_norm": 1.198819177261612e-06, "learning_rate": 4.251168014001955e-05, "loss": 0.0, "num_input_tokens_seen": 2833376, "step": 10120 }, { "epoch": 112.5, "grad_norm": 2.8710096557915676e-06, "learning_rate": 4.250467220786204e-05, "loss": 0.0, "num_input_tokens_seen": 2834752, "step": 10125 }, { "epoch": 112.55555555555556, "grad_norm": 2.5178162559313932e-06, "learning_rate": 4.249766157626409e-05, "loss": 0.0, "num_input_tokens_seen": 2836160, "step": 10130 }, { "epoch": 112.61111111111111, "grad_norm": 2.1507369183382252e-06, "learning_rate": 4.249064824630684e-05, "loss": 0.0, "num_input_tokens_seen": 2837568, "step": 10135 }, { "epoch": 112.66666666666667, "grad_norm": 1.01627836102125e-06, "learning_rate": 4.248363221907183e-05, "loss": 0.0, "num_input_tokens_seen": 2838928, "step": 10140 }, { "epoch": 112.72222222222223, "grad_norm": 5.992128080833936e-06, "learning_rate": 4.2476613495641026e-05, "loss": 0.0, "num_input_tokens_seen": 2840272, "step": 10145 }, { "epoch": 112.77777777777777, "grad_norm": 6.132241651357617e-06, "learning_rate": 4.246959207709679e-05, "loss": 0.0, "num_input_tokens_seen": 2841712, "step": 10150 }, { "epoch": 112.83333333333333, "grad_norm": 7.113807782843651e-07, "learning_rate": 4.246256796452192e-05, "loss": 0.0, "num_input_tokens_seen": 2843136, "step": 10155 }, { "epoch": 112.88888888888889, "grad_norm": 6.302927886281395e-06, "learning_rate": 4.245554115899962e-05, "loss": 0.0, "num_input_tokens_seen": 2844528, "step": 10160 }, { "epoch": 112.94444444444444, "grad_norm": 1.255936331290286e-05, "learning_rate": 4.2448511661613514e-05, "loss": 0.0, "num_input_tokens_seen": 2845904, "step": 10165 }, { "epoch": 113.0, "grad_norm": 1.4722490959684364e-06, "learning_rate": 4.2441479473447635e-05, "loss": 0.0, "num_input_tokens_seen": 2847344, "step": 10170 }, { "epoch": 113.05555555555556, "grad_norm": 2.3270974907063646e-06, "learning_rate": 4.243444459558644e-05, "loss": 0.0, "num_input_tokens_seen": 2848736, "step": 10175 }, { "epoch": 113.11111111111111, "grad_norm": 2.282616378579405e-06, "learning_rate": 4.24274070291148e-05, "loss": 0.0, "num_input_tokens_seen": 2850144, "step": 10180 }, { "epoch": 113.16666666666667, "grad_norm": 4.7734524741827045e-06, "learning_rate": 4.242036677511798e-05, "loss": 0.0, "num_input_tokens_seen": 2851568, "step": 10185 }, { "epoch": 113.22222222222223, "grad_norm": 4.5527353904617485e-06, "learning_rate": 4.241332383468169e-05, "loss": 0.0, "num_input_tokens_seen": 2852976, "step": 10190 }, { "epoch": 113.27777777777777, "grad_norm": 1.2780286624547443e-06, "learning_rate": 4.2406278208892034e-05, "loss": 0.0, "num_input_tokens_seen": 2854384, "step": 10195 }, { "epoch": 113.33333333333333, "grad_norm": 1.4459575368164224e-06, "learning_rate": 4.2399229898835536e-05, "loss": 0.0, "num_input_tokens_seen": 2855808, "step": 10200 }, { "epoch": 113.33333333333333, "eval_loss": 1.19344162940979, "eval_runtime": 1.4036, "eval_samples_per_second": 28.498, "eval_steps_per_second": 14.249, "num_input_tokens_seen": 2855808, "step": 10200 }, { "epoch": 113.38888888888889, "grad_norm": 2.989294671351672e-06, "learning_rate": 4.239217890559914e-05, "loss": 0.0, "num_input_tokens_seen": 2857184, "step": 10205 }, { "epoch": 113.44444444444444, "grad_norm": 4.89086323796073e-06, "learning_rate": 4.238512523027019e-05, "loss": 0.0, "num_input_tokens_seen": 2858544, "step": 10210 }, { "epoch": 113.5, "grad_norm": 1.357099222332181e-06, "learning_rate": 4.237806887393645e-05, "loss": 0.0, "num_input_tokens_seen": 2859936, "step": 10215 }, { "epoch": 113.55555555555556, "grad_norm": 2.2379947495210217e-06, "learning_rate": 4.237100983768611e-05, "loss": 0.0, "num_input_tokens_seen": 2861360, "step": 10220 }, { "epoch": 113.61111111111111, "grad_norm": 2.604912197057274e-06, "learning_rate": 4.2363948122607756e-05, "loss": 0.0, "num_input_tokens_seen": 2862752, "step": 10225 }, { "epoch": 113.66666666666667, "grad_norm": 1.4660091665064101e-06, "learning_rate": 4.235688372979039e-05, "loss": 0.0, "num_input_tokens_seen": 2864064, "step": 10230 }, { "epoch": 113.72222222222223, "grad_norm": 1.726993559714174e-06, "learning_rate": 4.234981666032343e-05, "loss": 0.0, "num_input_tokens_seen": 2865440, "step": 10235 }, { "epoch": 113.77777777777777, "grad_norm": 2.522070872146287e-06, "learning_rate": 4.2342746915296704e-05, "loss": 0.0, "num_input_tokens_seen": 2866800, "step": 10240 }, { "epoch": 113.83333333333333, "grad_norm": 1.517132886874606e-06, "learning_rate": 4.233567449580047e-05, "loss": 0.0, "num_input_tokens_seen": 2868208, "step": 10245 }, { "epoch": 113.88888888888889, "grad_norm": 2.025782123382669e-06, "learning_rate": 4.232859940292537e-05, "loss": 0.0, "num_input_tokens_seen": 2869600, "step": 10250 }, { "epoch": 113.94444444444444, "grad_norm": 9.090244361686928e-07, "learning_rate": 4.232152163776248e-05, "loss": 0.0, "num_input_tokens_seen": 2871040, "step": 10255 }, { "epoch": 114.0, "grad_norm": 2.3208676793728955e-06, "learning_rate": 4.231444120140328e-05, "loss": 0.0, "num_input_tokens_seen": 2872448, "step": 10260 }, { "epoch": 114.05555555555556, "grad_norm": 3.785658009292092e-06, "learning_rate": 4.230735809493967e-05, "loss": 0.0, "num_input_tokens_seen": 2873840, "step": 10265 }, { "epoch": 114.11111111111111, "grad_norm": 9.490560728409037e-07, "learning_rate": 4.2300272319463926e-05, "loss": 0.0, "num_input_tokens_seen": 2875216, "step": 10270 }, { "epoch": 114.16666666666667, "grad_norm": 3.5450841551210033e-06, "learning_rate": 4.2293183876068786e-05, "loss": 0.0, "num_input_tokens_seen": 2876656, "step": 10275 }, { "epoch": 114.22222222222223, "grad_norm": 7.4602721724659204e-06, "learning_rate": 4.228609276584737e-05, "loss": 0.0, "num_input_tokens_seen": 2878016, "step": 10280 }, { "epoch": 114.27777777777777, "grad_norm": 1.3882424809708027e-06, "learning_rate": 4.227899898989323e-05, "loss": 0.0, "num_input_tokens_seen": 2879376, "step": 10285 }, { "epoch": 114.33333333333333, "grad_norm": 7.49718674342148e-06, "learning_rate": 4.2271902549300293e-05, "loss": 0.0, "num_input_tokens_seen": 2880768, "step": 10290 }, { "epoch": 114.38888888888889, "grad_norm": 3.474478717180318e-06, "learning_rate": 4.226480344516294e-05, "loss": 0.0, "num_input_tokens_seen": 2882176, "step": 10295 }, { "epoch": 114.44444444444444, "grad_norm": 2.5189613097609254e-06, "learning_rate": 4.2257701678575925e-05, "loss": 0.0, "num_input_tokens_seen": 2883600, "step": 10300 }, { "epoch": 114.5, "grad_norm": 3.782954991038423e-06, "learning_rate": 4.225059725063444e-05, "loss": 0.0, "num_input_tokens_seen": 2884976, "step": 10305 }, { "epoch": 114.55555555555556, "grad_norm": 3.155225385853555e-06, "learning_rate": 4.2243490162434074e-05, "loss": 0.0, "num_input_tokens_seen": 2886384, "step": 10310 }, { "epoch": 114.61111111111111, "grad_norm": 1.0555633025433053e-06, "learning_rate": 4.223638041507083e-05, "loss": 0.0, "num_input_tokens_seen": 2887776, "step": 10315 }, { "epoch": 114.66666666666667, "grad_norm": 1.3159077525415341e-06, "learning_rate": 4.2229268009641124e-05, "loss": 0.0, "num_input_tokens_seen": 2889200, "step": 10320 }, { "epoch": 114.72222222222223, "grad_norm": 1.7120590882768738e-06, "learning_rate": 4.222215294724177e-05, "loss": 0.0, "num_input_tokens_seen": 2890624, "step": 10325 }, { "epoch": 114.77777777777777, "grad_norm": 1.5592479485349031e-06, "learning_rate": 4.2215035228970005e-05, "loss": 0.0, "num_input_tokens_seen": 2892032, "step": 10330 }, { "epoch": 114.83333333333333, "grad_norm": 1.19202366022364e-06, "learning_rate": 4.2207914855923464e-05, "loss": 0.0, "num_input_tokens_seen": 2893472, "step": 10335 }, { "epoch": 114.88888888888889, "grad_norm": 1.4904146610206226e-06, "learning_rate": 4.220079182920021e-05, "loss": 0.0, "num_input_tokens_seen": 2894880, "step": 10340 }, { "epoch": 114.94444444444444, "grad_norm": 8.044967216846999e-06, "learning_rate": 4.2193666149898705e-05, "loss": 0.0, "num_input_tokens_seen": 2896288, "step": 10345 }, { "epoch": 115.0, "grad_norm": 1.0931364613497863e-06, "learning_rate": 4.21865378191178e-05, "loss": 0.0, "num_input_tokens_seen": 2897648, "step": 10350 }, { "epoch": 115.05555555555556, "grad_norm": 3.701114110299386e-06, "learning_rate": 4.217940683795678e-05, "loss": 0.0, "num_input_tokens_seen": 2899088, "step": 10355 }, { "epoch": 115.11111111111111, "grad_norm": 3.367245199115132e-06, "learning_rate": 4.217227320751534e-05, "loss": 0.0, "num_input_tokens_seen": 2900448, "step": 10360 }, { "epoch": 115.16666666666667, "grad_norm": 8.784185411059298e-06, "learning_rate": 4.216513692889358e-05, "loss": 0.0, "num_input_tokens_seen": 2901856, "step": 10365 }, { "epoch": 115.22222222222223, "grad_norm": 1.1968647868343396e-06, "learning_rate": 4.215799800319199e-05, "loss": 0.0, "num_input_tokens_seen": 2903216, "step": 10370 }, { "epoch": 115.27777777777777, "grad_norm": 1.5079231161507778e-06, "learning_rate": 4.2150856431511485e-05, "loss": 0.0, "num_input_tokens_seen": 2904608, "step": 10375 }, { "epoch": 115.33333333333333, "grad_norm": 7.686693606956396e-06, "learning_rate": 4.214371221495339e-05, "loss": 0.0, "num_input_tokens_seen": 2905984, "step": 10380 }, { "epoch": 115.38888888888889, "grad_norm": 3.494002612569602e-06, "learning_rate": 4.213656535461942e-05, "loss": 0.0, "num_input_tokens_seen": 2907408, "step": 10385 }, { "epoch": 115.44444444444444, "grad_norm": 1.9730136955331545e-06, "learning_rate": 4.2129415851611734e-05, "loss": 0.0, "num_input_tokens_seen": 2908848, "step": 10390 }, { "epoch": 115.5, "grad_norm": 2.8989629754505586e-06, "learning_rate": 4.2122263707032855e-05, "loss": 0.0, "num_input_tokens_seen": 2910256, "step": 10395 }, { "epoch": 115.55555555555556, "grad_norm": 8.449241022390197e-07, "learning_rate": 4.211510892198574e-05, "loss": 0.0, "num_input_tokens_seen": 2911648, "step": 10400 }, { "epoch": 115.55555555555556, "eval_loss": 1.233807921409607, "eval_runtime": 1.3954, "eval_samples_per_second": 28.665, "eval_steps_per_second": 14.332, "num_input_tokens_seen": 2911648, "step": 10400 }, { "epoch": 115.61111111111111, "grad_norm": 4.114930561627261e-06, "learning_rate": 4.210795149757375e-05, "loss": 0.0, "num_input_tokens_seen": 2913056, "step": 10405 }, { "epoch": 115.66666666666667, "grad_norm": 2.460238647472579e-06, "learning_rate": 4.210079143490065e-05, "loss": 0.0, "num_input_tokens_seen": 2914448, "step": 10410 }, { "epoch": 115.72222222222223, "grad_norm": 3.3139472179755103e-06, "learning_rate": 4.2093628735070604e-05, "loss": 0.0, "num_input_tokens_seen": 2915856, "step": 10415 }, { "epoch": 115.77777777777777, "grad_norm": 1.7037465340763447e-06, "learning_rate": 4.208646339918819e-05, "loss": 0.0, "num_input_tokens_seen": 2917216, "step": 10420 }, { "epoch": 115.83333333333333, "grad_norm": 1.2298221463424852e-06, "learning_rate": 4.2079295428358414e-05, "loss": 0.0, "num_input_tokens_seen": 2918624, "step": 10425 }, { "epoch": 115.88888888888889, "grad_norm": 6.946087069081841e-06, "learning_rate": 4.207212482368664e-05, "loss": 0.0, "num_input_tokens_seen": 2920016, "step": 10430 }, { "epoch": 115.94444444444444, "grad_norm": 1.7717003402140108e-06, "learning_rate": 4.206495158627867e-05, "loss": 0.0, "num_input_tokens_seen": 2921440, "step": 10435 }, { "epoch": 116.0, "grad_norm": 1.0230371572106378e-06, "learning_rate": 4.205777571724073e-05, "loss": 0.0, "num_input_tokens_seen": 2922848, "step": 10440 }, { "epoch": 116.05555555555556, "grad_norm": 1.3524932001018897e-06, "learning_rate": 4.20505972176794e-05, "loss": 0.0, "num_input_tokens_seen": 2924272, "step": 10445 }, { "epoch": 116.11111111111111, "grad_norm": 1.3282929103297647e-06, "learning_rate": 4.204341608870171e-05, "loss": 0.0, "num_input_tokens_seen": 2925696, "step": 10450 }, { "epoch": 116.16666666666667, "grad_norm": 3.1737527024233714e-06, "learning_rate": 4.203623233141508e-05, "loss": 0.0, "num_input_tokens_seen": 2927104, "step": 10455 }, { "epoch": 116.22222222222223, "grad_norm": 1.69917461789737e-06, "learning_rate": 4.2029045946927334e-05, "loss": 0.0, "num_input_tokens_seen": 2928560, "step": 10460 }, { "epoch": 116.27777777777777, "grad_norm": 1.5018024441815214e-06, "learning_rate": 4.20218569363467e-05, "loss": 0.0, "num_input_tokens_seen": 2929952, "step": 10465 }, { "epoch": 116.33333333333333, "grad_norm": 9.469159181207942e-07, "learning_rate": 4.2014665300781834e-05, "loss": 0.0, "num_input_tokens_seen": 2931296, "step": 10470 }, { "epoch": 116.38888888888889, "grad_norm": 7.560396397821023e-07, "learning_rate": 4.200747104134174e-05, "loss": 0.0, "num_input_tokens_seen": 2932672, "step": 10475 }, { "epoch": 116.44444444444444, "grad_norm": 4.885009730060119e-06, "learning_rate": 4.200027415913588e-05, "loss": 0.0, "num_input_tokens_seen": 2934064, "step": 10480 }, { "epoch": 116.5, "grad_norm": 2.4320736429217504e-06, "learning_rate": 4.1993074655274126e-05, "loss": 0.0, "num_input_tokens_seen": 2935504, "step": 10485 }, { "epoch": 116.55555555555556, "grad_norm": 4.44781426267582e-06, "learning_rate": 4.198587253086669e-05, "loss": 0.0, "num_input_tokens_seen": 2936864, "step": 10490 }, { "epoch": 116.61111111111111, "grad_norm": 1.7395946088072378e-06, "learning_rate": 4.197866778702426e-05, "loss": 0.0, "num_input_tokens_seen": 2938256, "step": 10495 }, { "epoch": 116.66666666666667, "grad_norm": 2.509392743377248e-06, "learning_rate": 4.197146042485789e-05, "loss": 0.0, "num_input_tokens_seen": 2939632, "step": 10500 }, { "epoch": 116.72222222222223, "grad_norm": 1.6127963817780255e-06, "learning_rate": 4.1964250445479046e-05, "loss": 0.0, "num_input_tokens_seen": 2941056, "step": 10505 }, { "epoch": 116.77777777777777, "grad_norm": 5.587352006841684e-06, "learning_rate": 4.19570378499996e-05, "loss": 0.0, "num_input_tokens_seen": 2942464, "step": 10510 }, { "epoch": 116.83333333333333, "grad_norm": 4.970059308107011e-06, "learning_rate": 4.194982263953182e-05, "loss": 0.0, "num_input_tokens_seen": 2943920, "step": 10515 }, { "epoch": 116.88888888888889, "grad_norm": 2.411971991023165e-06, "learning_rate": 4.194260481518838e-05, "loss": 0.0, "num_input_tokens_seen": 2945296, "step": 10520 }, { "epoch": 116.94444444444444, "grad_norm": 5.688169949280564e-06, "learning_rate": 4.1935384378082366e-05, "loss": 0.0, "num_input_tokens_seen": 2946688, "step": 10525 }, { "epoch": 117.0, "grad_norm": 1.2195580438856268e-06, "learning_rate": 4.1928161329327267e-05, "loss": 0.0, "num_input_tokens_seen": 2948144, "step": 10530 }, { "epoch": 117.05555555555556, "grad_norm": 1.4445857914324733e-06, "learning_rate": 4.1920935670036945e-05, "loss": 0.0, "num_input_tokens_seen": 2949568, "step": 10535 }, { "epoch": 117.11111111111111, "grad_norm": 3.1265346933651017e-06, "learning_rate": 4.1913707401325705e-05, "loss": 0.0, "num_input_tokens_seen": 2950976, "step": 10540 }, { "epoch": 117.16666666666667, "grad_norm": 2.3032259832689306e-06, "learning_rate": 4.1906476524308235e-05, "loss": 0.0, "num_input_tokens_seen": 2952368, "step": 10545 }, { "epoch": 117.22222222222223, "grad_norm": 5.257898919808213e-06, "learning_rate": 4.189924304009962e-05, "loss": 0.0, "num_input_tokens_seen": 2953712, "step": 10550 }, { "epoch": 117.27777777777777, "grad_norm": 2.460647010593675e-06, "learning_rate": 4.189200694981537e-05, "loss": 0.0, "num_input_tokens_seen": 2955104, "step": 10555 }, { "epoch": 117.33333333333333, "grad_norm": 1.140590484283166e-06, "learning_rate": 4.188476825457136e-05, "loss": 0.0, "num_input_tokens_seen": 2956480, "step": 10560 }, { "epoch": 117.38888888888889, "grad_norm": 6.301287157839397e-06, "learning_rate": 4.18775269554839e-05, "loss": 0.0, "num_input_tokens_seen": 2957920, "step": 10565 }, { "epoch": 117.44444444444444, "grad_norm": 1.8805519630404888e-06, "learning_rate": 4.187028305366969e-05, "loss": 0.0, "num_input_tokens_seen": 2959328, "step": 10570 }, { "epoch": 117.5, "grad_norm": 3.344432116136886e-06, "learning_rate": 4.1863036550245824e-05, "loss": 0.0, "num_input_tokens_seen": 2960784, "step": 10575 }, { "epoch": 117.55555555555556, "grad_norm": 1.6638474562569172e-06, "learning_rate": 4.1855787446329806e-05, "loss": 0.0, "num_input_tokens_seen": 2962208, "step": 10580 }, { "epoch": 117.61111111111111, "grad_norm": 1.3601670616480988e-06, "learning_rate": 4.184853574303955e-05, "loss": 0.0, "num_input_tokens_seen": 2963632, "step": 10585 }, { "epoch": 117.66666666666667, "grad_norm": 3.5783543808065588e-06, "learning_rate": 4.184128144149334e-05, "loss": 0.0, "num_input_tokens_seen": 2965056, "step": 10590 }, { "epoch": 117.72222222222223, "grad_norm": 2.3697721189819276e-06, "learning_rate": 4.1834024542809896e-05, "loss": 0.0, "num_input_tokens_seen": 2966480, "step": 10595 }, { "epoch": 117.77777777777777, "grad_norm": 4.045453351864126e-06, "learning_rate": 4.1826765048108315e-05, "loss": 0.0, "num_input_tokens_seen": 2967856, "step": 10600 }, { "epoch": 117.77777777777777, "eval_loss": 1.2304883003234863, "eval_runtime": 1.3843, "eval_samples_per_second": 28.896, "eval_steps_per_second": 14.448, "num_input_tokens_seen": 2967856, "step": 10600 }, { "epoch": 117.83333333333333, "grad_norm": 2.127653260686202e-06, "learning_rate": 4.181950295850811e-05, "loss": 0.0, "num_input_tokens_seen": 2969232, "step": 10605 }, { "epoch": 117.88888888888889, "grad_norm": 7.548927442257991e-06, "learning_rate": 4.181223827512918e-05, "loss": 0.0, "num_input_tokens_seen": 2970608, "step": 10610 }, { "epoch": 117.94444444444444, "grad_norm": 2.0231414055160712e-06, "learning_rate": 4.180497099909183e-05, "loss": 0.0, "num_input_tokens_seen": 2972000, "step": 10615 }, { "epoch": 118.0, "grad_norm": 1.5975642781995703e-06, "learning_rate": 4.179770113151677e-05, "loss": 0.0, "num_input_tokens_seen": 2973360, "step": 10620 }, { "epoch": 118.05555555555556, "grad_norm": 2.0854802187386667e-06, "learning_rate": 4.179042867352511e-05, "loss": 0.0, "num_input_tokens_seen": 2974816, "step": 10625 }, { "epoch": 118.11111111111111, "grad_norm": 1.0344230076952954e-06, "learning_rate": 4.1783153626238334e-05, "loss": 0.0, "num_input_tokens_seen": 2976192, "step": 10630 }, { "epoch": 118.16666666666667, "grad_norm": 1.5000071471149568e-05, "learning_rate": 4.177587599077836e-05, "loss": 0.0, "num_input_tokens_seen": 2977584, "step": 10635 }, { "epoch": 118.22222222222223, "grad_norm": 1.278236254620424e-06, "learning_rate": 4.1768595768267494e-05, "loss": 0.0, "num_input_tokens_seen": 2978992, "step": 10640 }, { "epoch": 118.27777777777777, "grad_norm": 5.290753506415058e-06, "learning_rate": 4.176131295982843e-05, "loss": 0.0, "num_input_tokens_seen": 2980400, "step": 10645 }, { "epoch": 118.33333333333333, "grad_norm": 1.5759869711473584e-06, "learning_rate": 4.1754027566584276e-05, "loss": 0.0, "num_input_tokens_seen": 2981760, "step": 10650 }, { "epoch": 118.38888888888889, "grad_norm": 2.5144599931081757e-06, "learning_rate": 4.174673958965852e-05, "loss": 0.0, "num_input_tokens_seen": 2983168, "step": 10655 }, { "epoch": 118.44444444444444, "grad_norm": 1.8054597603622824e-06, "learning_rate": 4.173944903017507e-05, "loss": 0.0, "num_input_tokens_seen": 2984608, "step": 10660 }, { "epoch": 118.5, "grad_norm": 1.8067770497509628e-06, "learning_rate": 4.173215588925822e-05, "loss": 0.0, "num_input_tokens_seen": 2986000, "step": 10665 }, { "epoch": 118.55555555555556, "grad_norm": 2.1880343865632312e-06, "learning_rate": 4.172486016803266e-05, "loss": 0.0, "num_input_tokens_seen": 2987408, "step": 10670 }, { "epoch": 118.61111111111111, "grad_norm": 1.1740559102690895e-06, "learning_rate": 4.171756186762349e-05, "loss": 0.0, "num_input_tokens_seen": 2988848, "step": 10675 }, { "epoch": 118.66666666666667, "grad_norm": 6.798339313718316e-07, "learning_rate": 4.171026098915619e-05, "loss": 0.0, "num_input_tokens_seen": 2990272, "step": 10680 }, { "epoch": 118.72222222222223, "grad_norm": 1.7530284139866126e-06, "learning_rate": 4.170295753375665e-05, "loss": 0.0, "num_input_tokens_seen": 2991664, "step": 10685 }, { "epoch": 118.77777777777777, "grad_norm": 2.1774721972178668e-06, "learning_rate": 4.169565150255117e-05, "loss": 0.0, "num_input_tokens_seen": 2993072, "step": 10690 }, { "epoch": 118.83333333333333, "grad_norm": 1.7973428612094722e-06, "learning_rate": 4.16883428966664e-05, "loss": 0.0, "num_input_tokens_seen": 2994416, "step": 10695 }, { "epoch": 118.88888888888889, "grad_norm": 1.6821481949591544e-06, "learning_rate": 4.168103171722944e-05, "loss": 0.0, "num_input_tokens_seen": 2995824, "step": 10700 }, { "epoch": 118.94444444444444, "grad_norm": 1.8441562588122906e-06, "learning_rate": 4.167371796536777e-05, "loss": 0.0, "num_input_tokens_seen": 2997232, "step": 10705 }, { "epoch": 119.0, "grad_norm": 2.5016531708388356e-06, "learning_rate": 4.166640164220924e-05, "loss": 0.0, "num_input_tokens_seen": 2998624, "step": 10710 }, { "epoch": 119.05555555555556, "grad_norm": 4.931884177494794e-06, "learning_rate": 4.1659082748882144e-05, "loss": 0.0, "num_input_tokens_seen": 3000032, "step": 10715 }, { "epoch": 119.11111111111111, "grad_norm": 3.6236046980775427e-06, "learning_rate": 4.1651761286515135e-05, "loss": 0.0, "num_input_tokens_seen": 3001424, "step": 10720 }, { "epoch": 119.16666666666667, "grad_norm": 5.751105163653847e-06, "learning_rate": 4.164443725623728e-05, "loss": 0.0, "num_input_tokens_seen": 3002848, "step": 10725 }, { "epoch": 119.22222222222223, "grad_norm": 5.374183729145443e-06, "learning_rate": 4.163711065917802e-05, "loss": 0.0, "num_input_tokens_seen": 3004256, "step": 10730 }, { "epoch": 119.27777777777777, "grad_norm": 5.513039923243923e-06, "learning_rate": 4.1629781496467234e-05, "loss": 0.0, "num_input_tokens_seen": 3005632, "step": 10735 }, { "epoch": 119.33333333333333, "grad_norm": 1.0569596042842022e-06, "learning_rate": 4.1622449769235164e-05, "loss": 0.0, "num_input_tokens_seen": 3007040, "step": 10740 }, { "epoch": 119.38888888888889, "grad_norm": 1.2900804904347751e-06, "learning_rate": 4.161511547861243e-05, "loss": 0.0, "num_input_tokens_seen": 3008432, "step": 10745 }, { "epoch": 119.44444444444444, "grad_norm": 6.84371934767114e-06, "learning_rate": 4.1607778625730104e-05, "loss": 0.0, "num_input_tokens_seen": 3009776, "step": 10750 }, { "epoch": 119.5, "grad_norm": 9.311716894444544e-06, "learning_rate": 4.160043921171961e-05, "loss": 0.0, "num_input_tokens_seen": 3011168, "step": 10755 }, { "epoch": 119.55555555555556, "grad_norm": 2.2554943370778346e-06, "learning_rate": 4.159309723771276e-05, "loss": 0.0, "num_input_tokens_seen": 3012528, "step": 10760 }, { "epoch": 119.61111111111111, "grad_norm": 4.423368409334216e-06, "learning_rate": 4.158575270484181e-05, "loss": 0.0, "num_input_tokens_seen": 3013952, "step": 10765 }, { "epoch": 119.66666666666667, "grad_norm": 1.11185352125176e-06, "learning_rate": 4.157840561423936e-05, "loss": 0.0, "num_input_tokens_seen": 3015360, "step": 10770 }, { "epoch": 119.72222222222223, "grad_norm": 1.262021442016703e-06, "learning_rate": 4.1571055967038416e-05, "loss": 0.0, "num_input_tokens_seen": 3016752, "step": 10775 }, { "epoch": 119.77777777777777, "grad_norm": 1.9179017272108467e-06, "learning_rate": 4.156370376437241e-05, "loss": 0.0, "num_input_tokens_seen": 3018144, "step": 10780 }, { "epoch": 119.83333333333333, "grad_norm": 1.936772150656907e-06, "learning_rate": 4.155634900737513e-05, "loss": 0.0, "num_input_tokens_seen": 3019552, "step": 10785 }, { "epoch": 119.88888888888889, "grad_norm": 2.422298848614446e-06, "learning_rate": 4.1548991697180764e-05, "loss": 0.0, "num_input_tokens_seen": 3020928, "step": 10790 }, { "epoch": 119.94444444444444, "grad_norm": 2.1706996449211147e-06, "learning_rate": 4.1541631834923914e-05, "loss": 0.0, "num_input_tokens_seen": 3022384, "step": 10795 }, { "epoch": 120.0, "grad_norm": 8.216362061830296e-07, "learning_rate": 4.153426942173956e-05, "loss": 0.0, "num_input_tokens_seen": 3023792, "step": 10800 }, { "epoch": 120.0, "eval_loss": 1.1920689344406128, "eval_runtime": 1.3821, "eval_samples_per_second": 28.941, "eval_steps_per_second": 14.47, "num_input_tokens_seen": 3023792, "step": 10800 }, { "epoch": 120.05555555555556, "grad_norm": 5.687428711098619e-06, "learning_rate": 4.152690445876308e-05, "loss": 0.0, "num_input_tokens_seen": 3025184, "step": 10805 }, { "epoch": 120.11111111111111, "grad_norm": 2.433663667034125e-06, "learning_rate": 4.1519536947130245e-05, "loss": 0.0, "num_input_tokens_seen": 3026560, "step": 10810 }, { "epoch": 120.16666666666667, "grad_norm": 4.186802016192814e-06, "learning_rate": 4.151216688797722e-05, "loss": 0.0, "num_input_tokens_seen": 3028000, "step": 10815 }, { "epoch": 120.22222222222223, "grad_norm": 1.4241143162507797e-06, "learning_rate": 4.150479428244054e-05, "loss": 0.0, "num_input_tokens_seen": 3029424, "step": 10820 }, { "epoch": 120.27777777777777, "grad_norm": 1.6798866226963582e-06, "learning_rate": 4.1497419131657176e-05, "loss": 0.0, "num_input_tokens_seen": 3030848, "step": 10825 }, { "epoch": 120.33333333333333, "grad_norm": 7.533054485975299e-06, "learning_rate": 4.149004143676447e-05, "loss": 0.0, "num_input_tokens_seen": 3032288, "step": 10830 }, { "epoch": 120.38888888888889, "grad_norm": 1.3391995707934257e-06, "learning_rate": 4.148266119890015e-05, "loss": 0.0, "num_input_tokens_seen": 3033728, "step": 10835 }, { "epoch": 120.44444444444444, "grad_norm": 2.014708343267557e-06, "learning_rate": 4.1475278419202324e-05, "loss": 0.0, "num_input_tokens_seen": 3035168, "step": 10840 }, { "epoch": 120.5, "grad_norm": 2.204496013291646e-06, "learning_rate": 4.146789309880953e-05, "loss": 0.0, "num_input_tokens_seen": 3036576, "step": 10845 }, { "epoch": 120.55555555555556, "grad_norm": 1.4027922361492529e-06, "learning_rate": 4.146050523886068e-05, "loss": 0.0, "num_input_tokens_seen": 3037984, "step": 10850 }, { "epoch": 120.61111111111111, "grad_norm": 8.430999514530413e-06, "learning_rate": 4.1453114840495055e-05, "loss": 0.0, "num_input_tokens_seen": 3039328, "step": 10855 }, { "epoch": 120.66666666666667, "grad_norm": 3.272891945016454e-06, "learning_rate": 4.1445721904852364e-05, "loss": 0.0, "num_input_tokens_seen": 3040784, "step": 10860 }, { "epoch": 120.72222222222223, "grad_norm": 1.2476968549890444e-05, "learning_rate": 4.143832643307269e-05, "loss": 0.0, "num_input_tokens_seen": 3042176, "step": 10865 }, { "epoch": 120.77777777777777, "grad_norm": 6.4457408370799385e-06, "learning_rate": 4.1430928426296503e-05, "loss": 0.0, "num_input_tokens_seen": 3043552, "step": 10870 }, { "epoch": 120.83333333333333, "grad_norm": 1.8151085896533914e-06, "learning_rate": 4.142352788566466e-05, "loss": 0.0, "num_input_tokens_seen": 3044880, "step": 10875 }, { "epoch": 120.88888888888889, "grad_norm": 1.1135325621580705e-06, "learning_rate": 4.1416124812318424e-05, "loss": 0.0, "num_input_tokens_seen": 3046256, "step": 10880 }, { "epoch": 120.94444444444444, "grad_norm": 1.5366215393441962e-06, "learning_rate": 4.1408719207399453e-05, "loss": 0.0, "num_input_tokens_seen": 3047648, "step": 10885 }, { "epoch": 121.0, "grad_norm": 2.422861143713817e-06, "learning_rate": 4.140131107204978e-05, "loss": 0.0, "num_input_tokens_seen": 3049024, "step": 10890 }, { "epoch": 121.05555555555556, "grad_norm": 8.508155588060617e-06, "learning_rate": 4.139390040741182e-05, "loss": 0.0, "num_input_tokens_seen": 3050416, "step": 10895 }, { "epoch": 121.11111111111111, "grad_norm": 9.520038474875037e-06, "learning_rate": 4.1386487214628396e-05, "loss": 0.0, "num_input_tokens_seen": 3051808, "step": 10900 }, { "epoch": 121.16666666666667, "grad_norm": 1.6539119087610743e-06, "learning_rate": 4.137907149484272e-05, "loss": 0.0, "num_input_tokens_seen": 3053216, "step": 10905 }, { "epoch": 121.22222222222223, "grad_norm": 4.422870006237645e-06, "learning_rate": 4.137165324919839e-05, "loss": 0.0, "num_input_tokens_seen": 3054608, "step": 10910 }, { "epoch": 121.27777777777777, "grad_norm": 1.999882897507632e-06, "learning_rate": 4.136423247883939e-05, "loss": 0.0, "num_input_tokens_seen": 3055984, "step": 10915 }, { "epoch": 121.33333333333333, "grad_norm": 1.6674522385073942e-06, "learning_rate": 4.135680918491009e-05, "loss": 0.0, "num_input_tokens_seen": 3057424, "step": 10920 }, { "epoch": 121.38888888888889, "grad_norm": 3.2377247407566756e-06, "learning_rate": 4.1349383368555265e-05, "loss": 0.0, "num_input_tokens_seen": 3058832, "step": 10925 }, { "epoch": 121.44444444444444, "grad_norm": 6.176741749186476e-07, "learning_rate": 4.1341955030920065e-05, "loss": 0.0, "num_input_tokens_seen": 3060240, "step": 10930 }, { "epoch": 121.5, "grad_norm": 1.627831579753547e-06, "learning_rate": 4.1334524173150036e-05, "loss": 0.0, "num_input_tokens_seen": 3061616, "step": 10935 }, { "epoch": 121.55555555555556, "grad_norm": 9.94189463199291e-07, "learning_rate": 4.13270907963911e-05, "loss": 0.0, "num_input_tokens_seen": 3063008, "step": 10940 }, { "epoch": 121.61111111111111, "grad_norm": 9.109622283176577e-07, "learning_rate": 4.131965490178959e-05, "loss": 0.0, "num_input_tokens_seen": 3064368, "step": 10945 }, { "epoch": 121.66666666666667, "grad_norm": 1.3532970797314192e-06, "learning_rate": 4.131221649049222e-05, "loss": 0.0, "num_input_tokens_seen": 3065792, "step": 10950 }, { "epoch": 121.72222222222223, "grad_norm": 1.7449352753828862e-06, "learning_rate": 4.130477556364606e-05, "loss": 0.0, "num_input_tokens_seen": 3067216, "step": 10955 }, { "epoch": 121.77777777777777, "grad_norm": 1.295620222663274e-05, "learning_rate": 4.129733212239861e-05, "loss": 0.0, "num_input_tokens_seen": 3068608, "step": 10960 }, { "epoch": 121.83333333333333, "grad_norm": 1.9929946120100794e-06, "learning_rate": 4.128988616789774e-05, "loss": 0.0, "num_input_tokens_seen": 3070016, "step": 10965 }, { "epoch": 121.88888888888889, "grad_norm": 1.5391875649584108e-06, "learning_rate": 4.1282437701291724e-05, "loss": 0.0, "num_input_tokens_seen": 3071424, "step": 10970 }, { "epoch": 121.94444444444444, "grad_norm": 1.783517973308335e-06, "learning_rate": 4.1274986723729184e-05, "loss": 0.0, "num_input_tokens_seen": 3072848, "step": 10975 }, { "epoch": 122.0, "grad_norm": 1.505489876763022e-06, "learning_rate": 4.126753323635917e-05, "loss": 0.0, "num_input_tokens_seen": 3074272, "step": 10980 }, { "epoch": 122.05555555555556, "grad_norm": 1.5769816172905848e-06, "learning_rate": 4.12600772403311e-05, "loss": 0.0, "num_input_tokens_seen": 3075664, "step": 10985 }, { "epoch": 122.11111111111111, "grad_norm": 7.659384664293611e-07, "learning_rate": 4.125261873679479e-05, "loss": 0.0, "num_input_tokens_seen": 3077120, "step": 10990 }, { "epoch": 122.16666666666667, "grad_norm": 1.698193273114157e-06, "learning_rate": 4.124515772690042e-05, "loss": 0.0, "num_input_tokens_seen": 3078544, "step": 10995 }, { "epoch": 122.22222222222223, "grad_norm": 2.167995262425393e-06, "learning_rate": 4.123769421179858e-05, "loss": 0.0, "num_input_tokens_seen": 3079920, "step": 11000 }, { "epoch": 122.22222222222223, "eval_loss": 1.2243400812149048, "eval_runtime": 1.3867, "eval_samples_per_second": 28.846, "eval_steps_per_second": 14.423, "num_input_tokens_seen": 3079920, "step": 11000 }, { "epoch": 122.27777777777777, "grad_norm": 2.4363473585253814e-06, "learning_rate": 4.1230228192640236e-05, "loss": 0.0, "num_input_tokens_seen": 3081312, "step": 11005 }, { "epoch": 122.33333333333333, "grad_norm": 1.5569903553114273e-06, "learning_rate": 4.122275967057675e-05, "loss": 0.0, "num_input_tokens_seen": 3082720, "step": 11010 }, { "epoch": 122.38888888888889, "grad_norm": 2.4713256152608665e-06, "learning_rate": 4.1215288646759846e-05, "loss": 0.0, "num_input_tokens_seen": 3084112, "step": 11015 }, { "epoch": 122.44444444444444, "grad_norm": 1.9023648292204598e-06, "learning_rate": 4.120781512234166e-05, "loss": 0.0, "num_input_tokens_seen": 3085456, "step": 11020 }, { "epoch": 122.5, "grad_norm": 1.6858693925314583e-06, "learning_rate": 4.120033909847471e-05, "loss": 0.0, "num_input_tokens_seen": 3086880, "step": 11025 }, { "epoch": 122.55555555555556, "grad_norm": 7.513950777138234e-07, "learning_rate": 4.119286057631187e-05, "loss": 0.0, "num_input_tokens_seen": 3088304, "step": 11030 }, { "epoch": 122.61111111111111, "grad_norm": 6.299645747276372e-07, "learning_rate": 4.118537955700646e-05, "loss": 0.0, "num_input_tokens_seen": 3089680, "step": 11035 }, { "epoch": 122.66666666666667, "grad_norm": 1.937921297212597e-06, "learning_rate": 4.11778960417121e-05, "loss": 0.0, "num_input_tokens_seen": 3091072, "step": 11040 }, { "epoch": 122.72222222222223, "grad_norm": 3.287514800831559e-06, "learning_rate": 4.117041003158288e-05, "loss": 0.0, "num_input_tokens_seen": 3092480, "step": 11045 }, { "epoch": 122.77777777777777, "grad_norm": 2.6376760615676176e-06, "learning_rate": 4.1162921527773215e-05, "loss": 0.0, "num_input_tokens_seen": 3093888, "step": 11050 }, { "epoch": 122.83333333333333, "grad_norm": 2.9006444037804613e-06, "learning_rate": 4.115543053143794e-05, "loss": 0.0, "num_input_tokens_seen": 3095296, "step": 11055 }, { "epoch": 122.88888888888889, "grad_norm": 2.3235695607581874e-06, "learning_rate": 4.114793704373226e-05, "loss": 0.0, "num_input_tokens_seen": 3096688, "step": 11060 }, { "epoch": 122.94444444444444, "grad_norm": 2.714609991016914e-06, "learning_rate": 4.114044106581175e-05, "loss": 0.0, "num_input_tokens_seen": 3098112, "step": 11065 }, { "epoch": 123.0, "grad_norm": 1.8560110675025498e-06, "learning_rate": 4.11329425988324e-05, "loss": 0.0, "num_input_tokens_seen": 3099536, "step": 11070 }, { "epoch": 123.05555555555556, "grad_norm": 1.3984945326228626e-06, "learning_rate": 4.112544164395056e-05, "loss": 0.0, "num_input_tokens_seen": 3100960, "step": 11075 }, { "epoch": 123.11111111111111, "grad_norm": 2.6045931917906273e-06, "learning_rate": 4.111793820232297e-05, "loss": 0.0, "num_input_tokens_seen": 3102336, "step": 11080 }, { "epoch": 123.16666666666667, "grad_norm": 2.3174682155513437e-06, "learning_rate": 4.1110432275106767e-05, "loss": 0.0, "num_input_tokens_seen": 3103728, "step": 11085 }, { "epoch": 123.22222222222223, "grad_norm": 1.6879957911442034e-06, "learning_rate": 4.110292386345944e-05, "loss": 0.0, "num_input_tokens_seen": 3105136, "step": 11090 }, { "epoch": 123.27777777777777, "grad_norm": 2.2710987650498282e-06, "learning_rate": 4.109541296853891e-05, "loss": 0.0, "num_input_tokens_seen": 3106544, "step": 11095 }, { "epoch": 123.33333333333333, "grad_norm": 1.400524865857733e-06, "learning_rate": 4.108789959150341e-05, "loss": 0.0, "num_input_tokens_seen": 3107936, "step": 11100 }, { "epoch": 123.38888888888889, "grad_norm": 5.531043825612869e-06, "learning_rate": 4.108038373351163e-05, "loss": 0.0, "num_input_tokens_seen": 3109376, "step": 11105 }, { "epoch": 123.44444444444444, "grad_norm": 5.8252289818483405e-06, "learning_rate": 4.10728653957226e-05, "loss": 0.0, "num_input_tokens_seen": 3110784, "step": 11110 }, { "epoch": 123.5, "grad_norm": 1.2490106655604905e-06, "learning_rate": 4.106534457929575e-05, "loss": 0.0, "num_input_tokens_seen": 3112224, "step": 11115 }, { "epoch": 123.55555555555556, "grad_norm": 1.0364507261328981e-06, "learning_rate": 4.105782128539086e-05, "loss": 0.0, "num_input_tokens_seen": 3113648, "step": 11120 }, { "epoch": 123.61111111111111, "grad_norm": 2.783498757708003e-06, "learning_rate": 4.1050295515168144e-05, "loss": 0.0, "num_input_tokens_seen": 3114976, "step": 11125 }, { "epoch": 123.66666666666667, "grad_norm": 3.520496420605923e-06, "learning_rate": 4.1042767269788155e-05, "loss": 0.0, "num_input_tokens_seen": 3116368, "step": 11130 }, { "epoch": 123.72222222222223, "grad_norm": 1.5903963230812224e-06, "learning_rate": 4.103523655041185e-05, "loss": 0.0, "num_input_tokens_seen": 3117760, "step": 11135 }, { "epoch": 123.77777777777777, "grad_norm": 2.583189598226454e-06, "learning_rate": 4.102770335820055e-05, "loss": 0.0, "num_input_tokens_seen": 3119168, "step": 11140 }, { "epoch": 123.83333333333333, "grad_norm": 1.318718318543688e-06, "learning_rate": 4.1020167694315984e-05, "loss": 0.0, "num_input_tokens_seen": 3120544, "step": 11145 }, { "epoch": 123.88888888888889, "grad_norm": 1.440362439097953e-06, "learning_rate": 4.101262955992023e-05, "loss": 0.0, "num_input_tokens_seen": 3121952, "step": 11150 }, { "epoch": 123.94444444444444, "grad_norm": 4.317144430387998e-06, "learning_rate": 4.100508895617578e-05, "loss": 0.0, "num_input_tokens_seen": 3123312, "step": 11155 }, { "epoch": 124.0, "grad_norm": 3.5799491797661176e-06, "learning_rate": 4.099754588424547e-05, "loss": 0.0, "num_input_tokens_seen": 3124720, "step": 11160 }, { "epoch": 124.05555555555556, "grad_norm": 2.3898383005871437e-06, "learning_rate": 4.0990000345292546e-05, "loss": 0.0, "num_input_tokens_seen": 3126128, "step": 11165 }, { "epoch": 124.11111111111111, "grad_norm": 9.477388630330097e-06, "learning_rate": 4.098245234048064e-05, "loss": 0.0, "num_input_tokens_seen": 3127552, "step": 11170 }, { "epoch": 124.16666666666667, "grad_norm": 1.962627493412583e-06, "learning_rate": 4.0974901870973726e-05, "loss": 0.0, "num_input_tokens_seen": 3128960, "step": 11175 }, { "epoch": 124.22222222222223, "grad_norm": 1.8307416667084908e-06, "learning_rate": 4.096734893793619e-05, "loss": 0.0, "num_input_tokens_seen": 3130352, "step": 11180 }, { "epoch": 124.27777777777777, "grad_norm": 2.9887326036259765e-06, "learning_rate": 4.095979354253279e-05, "loss": 0.0, "num_input_tokens_seen": 3131712, "step": 11185 }, { "epoch": 124.33333333333333, "grad_norm": 5.516051714948844e-06, "learning_rate": 4.0952235685928656e-05, "loss": 0.0, "num_input_tokens_seen": 3133088, "step": 11190 }, { "epoch": 124.38888888888889, "grad_norm": 3.0643782338302117e-06, "learning_rate": 4.094467536928932e-05, "loss": 0.0, "num_input_tokens_seen": 3134512, "step": 11195 }, { "epoch": 124.44444444444444, "grad_norm": 9.133890671364497e-06, "learning_rate": 4.093711259378067e-05, "loss": 0.0, "num_input_tokens_seen": 3135904, "step": 11200 }, { "epoch": 124.44444444444444, "eval_loss": 1.2162835597991943, "eval_runtime": 1.3975, "eval_samples_per_second": 28.623, "eval_steps_per_second": 14.312, "num_input_tokens_seen": 3135904, "step": 11200 }, { "epoch": 124.5, "grad_norm": 1.7813719068726641e-06, "learning_rate": 4.092954736056897e-05, "loss": 0.0, "num_input_tokens_seen": 3137344, "step": 11205 }, { "epoch": 124.55555555555556, "grad_norm": 4.1500406950945035e-06, "learning_rate": 4.09219796708209e-05, "loss": 0.0, "num_input_tokens_seen": 3138752, "step": 11210 }, { "epoch": 124.61111111111111, "grad_norm": 1.2033177654302563e-06, "learning_rate": 4.0914409525703464e-05, "loss": 0.0, "num_input_tokens_seen": 3140160, "step": 11215 }, { "epoch": 124.66666666666667, "grad_norm": 1.0541486972215353e-06, "learning_rate": 4.090683692638408e-05, "loss": 0.0, "num_input_tokens_seen": 3141536, "step": 11220 }, { "epoch": 124.72222222222223, "grad_norm": 2.327007450730889e-06, "learning_rate": 4.089926187403056e-05, "loss": 0.0, "num_input_tokens_seen": 3142992, "step": 11225 }, { "epoch": 124.77777777777777, "grad_norm": 1.3463418326864485e-06, "learning_rate": 4.0891684369811044e-05, "loss": 0.0, "num_input_tokens_seen": 3144384, "step": 11230 }, { "epoch": 124.83333333333333, "grad_norm": 7.533095072176366e-07, "learning_rate": 4.0884104414894107e-05, "loss": 0.0, "num_input_tokens_seen": 3145760, "step": 11235 }, { "epoch": 124.88888888888889, "grad_norm": 1.1288705081824446e-06, "learning_rate": 4.087652201044864e-05, "loss": 0.0, "num_input_tokens_seen": 3147152, "step": 11240 }, { "epoch": 124.94444444444444, "grad_norm": 1.1975689631071873e-06, "learning_rate": 4.086893715764397e-05, "loss": 0.0, "num_input_tokens_seen": 3148560, "step": 11245 }, { "epoch": 125.0, "grad_norm": 2.142072844435461e-06, "learning_rate": 4.086134985764977e-05, "loss": 0.0, "num_input_tokens_seen": 3149952, "step": 11250 }, { "epoch": 125.05555555555556, "grad_norm": 1.9801077542069834e-06, "learning_rate": 4.0853760111636085e-05, "loss": 0.0, "num_input_tokens_seen": 3151344, "step": 11255 }, { "epoch": 125.11111111111111, "grad_norm": 2.034053068200592e-06, "learning_rate": 4.084616792077337e-05, "loss": 0.0, "num_input_tokens_seen": 3152736, "step": 11260 }, { "epoch": 125.16666666666667, "grad_norm": 1.2379948657326167e-06, "learning_rate": 4.083857328623243e-05, "loss": 0.0, "num_input_tokens_seen": 3154144, "step": 11265 }, { "epoch": 125.22222222222223, "grad_norm": 4.247935521561885e-06, "learning_rate": 4.083097620918444e-05, "loss": 0.0, "num_input_tokens_seen": 3155536, "step": 11270 }, { "epoch": 125.27777777777777, "grad_norm": 1.527586164229433e-06, "learning_rate": 4.082337669080097e-05, "loss": 0.0, "num_input_tokens_seen": 3156912, "step": 11275 }, { "epoch": 125.33333333333333, "grad_norm": 1.8676317949939403e-06, "learning_rate": 4.081577473225398e-05, "loss": 0.0, "num_input_tokens_seen": 3158304, "step": 11280 }, { "epoch": 125.38888888888889, "grad_norm": 1.4656823168479605e-06, "learning_rate": 4.080817033471577e-05, "loss": 0.0, "num_input_tokens_seen": 3159776, "step": 11285 }, { "epoch": 125.44444444444444, "grad_norm": 2.628345555422129e-06, "learning_rate": 4.080056349935903e-05, "loss": 0.0, "num_input_tokens_seen": 3161136, "step": 11290 }, { "epoch": 125.5, "grad_norm": 7.271299296007783e-07, "learning_rate": 4.079295422735684e-05, "loss": 0.0, "num_input_tokens_seen": 3162512, "step": 11295 }, { "epoch": 125.55555555555556, "grad_norm": 2.682109425222734e-06, "learning_rate": 4.078534251988264e-05, "loss": 0.0, "num_input_tokens_seen": 3163920, "step": 11300 }, { "epoch": 125.61111111111111, "grad_norm": 1.2251434782228898e-06, "learning_rate": 4.077772837811025e-05, "loss": 0.0, "num_input_tokens_seen": 3165280, "step": 11305 }, { "epoch": 125.66666666666667, "grad_norm": 2.3383643110719277e-06, "learning_rate": 4.0770111803213874e-05, "loss": 0.0, "num_input_tokens_seen": 3166704, "step": 11310 }, { "epoch": 125.72222222222223, "grad_norm": 2.4169557946152054e-06, "learning_rate": 4.076249279636807e-05, "loss": 0.0, "num_input_tokens_seen": 3168064, "step": 11315 }, { "epoch": 125.77777777777777, "grad_norm": 2.195855813624803e-06, "learning_rate": 4.075487135874781e-05, "loss": 0.0, "num_input_tokens_seen": 3169440, "step": 11320 }, { "epoch": 125.83333333333333, "grad_norm": 3.2880259368539555e-06, "learning_rate": 4.074724749152837e-05, "loss": 0.0, "num_input_tokens_seen": 3170880, "step": 11325 }, { "epoch": 125.88888888888889, "grad_norm": 6.451226454373682e-06, "learning_rate": 4.07396211958855e-05, "loss": 0.0, "num_input_tokens_seen": 3172256, "step": 11330 }, { "epoch": 125.94444444444444, "grad_norm": 1.2907995369459968e-05, "learning_rate": 4.073199247299523e-05, "loss": 0.0, "num_input_tokens_seen": 3173712, "step": 11335 }, { "epoch": 126.0, "grad_norm": 1.1677112752295216e-06, "learning_rate": 4.072436132403403e-05, "loss": 0.0, "num_input_tokens_seen": 3175056, "step": 11340 }, { "epoch": 126.05555555555556, "grad_norm": 1.0703396355893346e-06, "learning_rate": 4.0716727750178704e-05, "loss": 0.0, "num_input_tokens_seen": 3176432, "step": 11345 }, { "epoch": 126.11111111111111, "grad_norm": 1.3492988273355877e-06, "learning_rate": 4.0709091752606455e-05, "loss": 0.0, "num_input_tokens_seen": 3177840, "step": 11350 }, { "epoch": 126.16666666666667, "grad_norm": 7.768127034069039e-06, "learning_rate": 4.070145333249484e-05, "loss": 0.0, "num_input_tokens_seen": 3179232, "step": 11355 }, { "epoch": 126.22222222222223, "grad_norm": 1.982174808290438e-06, "learning_rate": 4.069381249102181e-05, "loss": 0.0, "num_input_tokens_seen": 3180672, "step": 11360 }, { "epoch": 126.27777777777777, "grad_norm": 1.2411169336701278e-06, "learning_rate": 4.0686169229365665e-05, "loss": 0.0, "num_input_tokens_seen": 3182080, "step": 11365 }, { "epoch": 126.33333333333333, "grad_norm": 1.368638436360925e-06, "learning_rate": 4.067852354870511e-05, "loss": 0.0, "num_input_tokens_seen": 3183472, "step": 11370 }, { "epoch": 126.38888888888889, "grad_norm": 2.8815072710131062e-06, "learning_rate": 4.067087545021919e-05, "loss": 0.0, "num_input_tokens_seen": 3184880, "step": 11375 }, { "epoch": 126.44444444444444, "grad_norm": 1.61033119638887e-06, "learning_rate": 4.066322493508734e-05, "loss": 0.0, "num_input_tokens_seen": 3186272, "step": 11380 }, { "epoch": 126.5, "grad_norm": 6.894362059028936e-07, "learning_rate": 4.065557200448937e-05, "loss": 0.0, "num_input_tokens_seen": 3187648, "step": 11385 }, { "epoch": 126.55555555555556, "grad_norm": 1.2699104445346165e-05, "learning_rate": 4.064791665960546e-05, "loss": 0.0, "num_input_tokens_seen": 3189040, "step": 11390 }, { "epoch": 126.61111111111111, "grad_norm": 1.8784245412462042e-06, "learning_rate": 4.064025890161615e-05, "loss": 0.0, "num_input_tokens_seen": 3190432, "step": 11395 }, { "epoch": 126.66666666666667, "grad_norm": 2.188351345466799e-06, "learning_rate": 4.0632598731702373e-05, "loss": 0.0, "num_input_tokens_seen": 3191808, "step": 11400 }, { "epoch": 126.66666666666667, "eval_loss": 1.218216896057129, "eval_runtime": 1.3959, "eval_samples_per_second": 28.655, "eval_steps_per_second": 14.328, "num_input_tokens_seen": 3191808, "step": 11400 }, { "epoch": 126.72222222222223, "grad_norm": 1.0215810561930994e-06, "learning_rate": 4.0624936151045426e-05, "loss": 0.0, "num_input_tokens_seen": 3193168, "step": 11405 }, { "epoch": 126.77777777777777, "grad_norm": 5.999378117849119e-06, "learning_rate": 4.061727116082696e-05, "loss": 0.0, "num_input_tokens_seen": 3194576, "step": 11410 }, { "epoch": 126.83333333333333, "grad_norm": 2.0327975107647944e-06, "learning_rate": 4.060960376222903e-05, "loss": 0.0, "num_input_tokens_seen": 3196000, "step": 11415 }, { "epoch": 126.88888888888889, "grad_norm": 1.8657487999007571e-06, "learning_rate": 4.0601933956434034e-05, "loss": 0.0, "num_input_tokens_seen": 3197376, "step": 11420 }, { "epoch": 126.94444444444444, "grad_norm": 2.833775397448335e-06, "learning_rate": 4.059426174462476e-05, "loss": 0.0, "num_input_tokens_seen": 3198816, "step": 11425 }, { "epoch": 127.0, "grad_norm": 1.2850247230744571e-06, "learning_rate": 4.058658712798435e-05, "loss": 0.0, "num_input_tokens_seen": 3200208, "step": 11430 }, { "epoch": 127.05555555555556, "grad_norm": 1.774438487700536e-06, "learning_rate": 4.0578910107696336e-05, "loss": 0.0, "num_input_tokens_seen": 3201584, "step": 11435 }, { "epoch": 127.11111111111111, "grad_norm": 2.443462790324702e-06, "learning_rate": 4.05712306849446e-05, "loss": 0.0, "num_input_tokens_seen": 3203008, "step": 11440 }, { "epoch": 127.16666666666667, "grad_norm": 1.2861753475590376e-06, "learning_rate": 4.0563548860913415e-05, "loss": 0.0, "num_input_tokens_seen": 3204384, "step": 11445 }, { "epoch": 127.22222222222223, "grad_norm": 1.7064177200154518e-06, "learning_rate": 4.0555864636787414e-05, "loss": 0.0, "num_input_tokens_seen": 3205776, "step": 11450 }, { "epoch": 127.27777777777777, "grad_norm": 4.132501999265514e-06, "learning_rate": 4.054817801375159e-05, "loss": 0.0, "num_input_tokens_seen": 3207200, "step": 11455 }, { "epoch": 127.33333333333333, "grad_norm": 1.468284722250246e-06, "learning_rate": 4.054048899299134e-05, "loss": 0.0, "num_input_tokens_seen": 3208576, "step": 11460 }, { "epoch": 127.38888888888889, "grad_norm": 2.2805734261055477e-06, "learning_rate": 4.0532797575692385e-05, "loss": 0.0, "num_input_tokens_seen": 3209936, "step": 11465 }, { "epoch": 127.44444444444444, "grad_norm": 8.270003490906674e-06, "learning_rate": 4.052510376304085e-05, "loss": 0.0, "num_input_tokens_seen": 3211344, "step": 11470 }, { "epoch": 127.5, "grad_norm": 7.103960228960204e-07, "learning_rate": 4.051740755622321e-05, "loss": 0.0, "num_input_tokens_seen": 3212720, "step": 11475 }, { "epoch": 127.55555555555556, "grad_norm": 2.236193267890485e-06, "learning_rate": 4.050970895642632e-05, "loss": 0.0, "num_input_tokens_seen": 3214128, "step": 11480 }, { "epoch": 127.61111111111111, "grad_norm": 3.347072379256133e-06, "learning_rate": 4.050200796483741e-05, "loss": 0.0, "num_input_tokens_seen": 3215488, "step": 11485 }, { "epoch": 127.66666666666667, "grad_norm": 1.1419308520999039e-06, "learning_rate": 4.049430458264405e-05, "loss": 0.0, "num_input_tokens_seen": 3216896, "step": 11490 }, { "epoch": 127.72222222222223, "grad_norm": 9.185612270812271e-07, "learning_rate": 4.048659881103422e-05, "loss": 0.0, "num_input_tokens_seen": 3218240, "step": 11495 }, { "epoch": 127.77777777777777, "grad_norm": 1.6762024870331516e-06, "learning_rate": 4.0478890651196235e-05, "loss": 0.0, "num_input_tokens_seen": 3219696, "step": 11500 }, { "epoch": 127.83333333333333, "grad_norm": 1.681747676229861e-06, "learning_rate": 4.047118010431879e-05, "loss": 0.0, "num_input_tokens_seen": 3221104, "step": 11505 }, { "epoch": 127.88888888888889, "grad_norm": 2.172598442484741e-06, "learning_rate": 4.046346717159094e-05, "loss": 0.0, "num_input_tokens_seen": 3222544, "step": 11510 }, { "epoch": 127.94444444444444, "grad_norm": 1.833577016441268e-06, "learning_rate": 4.045575185420214e-05, "loss": 0.0, "num_input_tokens_seen": 3223968, "step": 11515 }, { "epoch": 128.0, "grad_norm": 1.9798349057964515e-06, "learning_rate": 4.0448034153342165e-05, "loss": 0.0, "num_input_tokens_seen": 3225376, "step": 11520 }, { "epoch": 128.05555555555554, "grad_norm": 1.6235204611803056e-06, "learning_rate": 4.0440314070201194e-05, "loss": 0.0, "num_input_tokens_seen": 3226752, "step": 11525 }, { "epoch": 128.11111111111111, "grad_norm": 1.7581329530003131e-06, "learning_rate": 4.043259160596976e-05, "loss": 0.0, "num_input_tokens_seen": 3228128, "step": 11530 }, { "epoch": 128.16666666666666, "grad_norm": 1.5759848110974417e-06, "learning_rate": 4.0424866761838767e-05, "loss": 0.0, "num_input_tokens_seen": 3229520, "step": 11535 }, { "epoch": 128.22222222222223, "grad_norm": 2.32017237067339e-06, "learning_rate": 4.041713953899948e-05, "loss": 0.0, "num_input_tokens_seen": 3230976, "step": 11540 }, { "epoch": 128.27777777777777, "grad_norm": 4.376278411655221e-06, "learning_rate": 4.0409409938643515e-05, "loss": 0.0, "num_input_tokens_seen": 3232400, "step": 11545 }, { "epoch": 128.33333333333334, "grad_norm": 1.9310398329253076e-06, "learning_rate": 4.0401677961962904e-05, "loss": 0.0, "num_input_tokens_seen": 3233808, "step": 11550 }, { "epoch": 128.38888888888889, "grad_norm": 2.1792084226035513e-06, "learning_rate": 4.039394361015001e-05, "loss": 0.0, "num_input_tokens_seen": 3235168, "step": 11555 }, { "epoch": 128.44444444444446, "grad_norm": 2.3447000785381533e-06, "learning_rate": 4.038620688439755e-05, "loss": 0.0, "num_input_tokens_seen": 3236528, "step": 11560 }, { "epoch": 128.5, "grad_norm": 1.429261828889139e-06, "learning_rate": 4.037846778589862e-05, "loss": 0.0, "num_input_tokens_seen": 3237936, "step": 11565 }, { "epoch": 128.55555555555554, "grad_norm": 2.3404084004141623e-06, "learning_rate": 4.0370726315846715e-05, "loss": 0.0, "num_input_tokens_seen": 3239392, "step": 11570 }, { "epoch": 128.61111111111111, "grad_norm": 5.281667654344346e-06, "learning_rate": 4.036298247543565e-05, "loss": 0.0, "num_input_tokens_seen": 3240800, "step": 11575 }, { "epoch": 128.66666666666666, "grad_norm": 1.4515809425574844e-06, "learning_rate": 4.035523626585962e-05, "loss": 0.0, "num_input_tokens_seen": 3242176, "step": 11580 }, { "epoch": 128.72222222222223, "grad_norm": 3.938273039238993e-06, "learning_rate": 4.0347487688313194e-05, "loss": 0.0, "num_input_tokens_seen": 3243600, "step": 11585 }, { "epoch": 128.77777777777777, "grad_norm": 1.462638806515315e-06, "learning_rate": 4.0339736743991296e-05, "loss": 0.0, "num_input_tokens_seen": 3244992, "step": 11590 }, { "epoch": 128.83333333333334, "grad_norm": 1.6336497310476261e-06, "learning_rate": 4.0331983434089227e-05, "loss": 0.0, "num_input_tokens_seen": 3246400, "step": 11595 }, { "epoch": 128.88888888888889, "grad_norm": 3.426073817536235e-06, "learning_rate": 4.032422775980264e-05, "loss": 0.0, "num_input_tokens_seen": 3247840, "step": 11600 }, { "epoch": 128.88888888888889, "eval_loss": 1.2363343238830566, "eval_runtime": 1.3893, "eval_samples_per_second": 28.792, "eval_steps_per_second": 14.396, "num_input_tokens_seen": 3247840, "step": 11600 }, { "epoch": 128.94444444444446, "grad_norm": 1.0911985555139836e-06, "learning_rate": 4.031646972232754e-05, "loss": 0.0, "num_input_tokens_seen": 3249232, "step": 11605 }, { "epoch": 129.0, "grad_norm": 2.2638728296442423e-06, "learning_rate": 4.0308709322860344e-05, "loss": 0.0, "num_input_tokens_seen": 3250624, "step": 11610 }, { "epoch": 129.05555555555554, "grad_norm": 2.500459686416434e-06, "learning_rate": 4.0300946562597784e-05, "loss": 0.0, "num_input_tokens_seen": 3252032, "step": 11615 }, { "epoch": 129.11111111111111, "grad_norm": 2.162225655411021e-06, "learning_rate": 4.029318144273698e-05, "loss": 0.0, "num_input_tokens_seen": 3253408, "step": 11620 }, { "epoch": 129.16666666666666, "grad_norm": 1.7663612652540905e-06, "learning_rate": 4.0285413964475415e-05, "loss": 0.0, "num_input_tokens_seen": 3254800, "step": 11625 }, { "epoch": 129.22222222222223, "grad_norm": 1.2041929267070373e-06, "learning_rate": 4.0277644129010927e-05, "loss": 0.0, "num_input_tokens_seen": 3256192, "step": 11630 }, { "epoch": 129.27777777777777, "grad_norm": 8.723299629309622e-07, "learning_rate": 4.0269871937541724e-05, "loss": 0.0, "num_input_tokens_seen": 3257552, "step": 11635 }, { "epoch": 129.33333333333334, "grad_norm": 1.4698123322887113e-06, "learning_rate": 4.026209739126637e-05, "loss": 0.0, "num_input_tokens_seen": 3258976, "step": 11640 }, { "epoch": 129.38888888888889, "grad_norm": 3.589445896068355e-06, "learning_rate": 4.025432049138381e-05, "loss": 0.0, "num_input_tokens_seen": 3260384, "step": 11645 }, { "epoch": 129.44444444444446, "grad_norm": 2.1251485122775193e-06, "learning_rate": 4.0246541239093325e-05, "loss": 0.0, "num_input_tokens_seen": 3261792, "step": 11650 }, { "epoch": 129.5, "grad_norm": 1.0271323844790459e-06, "learning_rate": 4.023875963559459e-05, "loss": 0.0, "num_input_tokens_seen": 3263184, "step": 11655 }, { "epoch": 129.55555555555554, "grad_norm": 8.770231033849996e-06, "learning_rate": 4.023097568208761e-05, "loss": 0.0, "num_input_tokens_seen": 3264560, "step": 11660 }, { "epoch": 129.61111111111111, "grad_norm": 9.09723894437775e-06, "learning_rate": 4.022318937977277e-05, "loss": 0.0, "num_input_tokens_seen": 3265968, "step": 11665 }, { "epoch": 129.66666666666666, "grad_norm": 1.200009137392044e-06, "learning_rate": 4.021540072985084e-05, "loss": 0.0, "num_input_tokens_seen": 3267360, "step": 11670 }, { "epoch": 129.72222222222223, "grad_norm": 3.1559320632368326e-06, "learning_rate": 4.020760973352289e-05, "loss": 0.0, "num_input_tokens_seen": 3268768, "step": 11675 }, { "epoch": 129.77777777777777, "grad_norm": 2.1327043668861734e-06, "learning_rate": 4.019981639199042e-05, "loss": 0.0, "num_input_tokens_seen": 3270160, "step": 11680 }, { "epoch": 129.83333333333334, "grad_norm": 1.7049060261342674e-06, "learning_rate": 4.0192020706455245e-05, "loss": 0.0, "num_input_tokens_seen": 3271536, "step": 11685 }, { "epoch": 129.88888888888889, "grad_norm": 1.643427253839036e-06, "learning_rate": 4.018422267811956e-05, "loss": 0.0, "num_input_tokens_seen": 3272944, "step": 11690 }, { "epoch": 129.94444444444446, "grad_norm": 4.953900770487962e-06, "learning_rate": 4.017642230818592e-05, "loss": 0.0, "num_input_tokens_seen": 3274320, "step": 11695 }, { "epoch": 130.0, "grad_norm": 1.4628466260546702e-06, "learning_rate": 4.0168619597857246e-05, "loss": 0.0, "num_input_tokens_seen": 3275712, "step": 11700 }, { "epoch": 130.05555555555554, "grad_norm": 2.9346110750338994e-06, "learning_rate": 4.016081454833681e-05, "loss": 0.0, "num_input_tokens_seen": 3277152, "step": 11705 }, { "epoch": 130.11111111111111, "grad_norm": 1.936526587087428e-06, "learning_rate": 4.0153007160828245e-05, "loss": 0.0, "num_input_tokens_seen": 3278592, "step": 11710 }, { "epoch": 130.16666666666666, "grad_norm": 3.1367856081487844e-06, "learning_rate": 4.0145197436535555e-05, "loss": 0.0, "num_input_tokens_seen": 3280016, "step": 11715 }, { "epoch": 130.22222222222223, "grad_norm": 2.846721372407046e-06, "learning_rate": 4.0137385376663095e-05, "loss": 0.0, "num_input_tokens_seen": 3281360, "step": 11720 }, { "epoch": 130.27777777777777, "grad_norm": 2.4095484150166158e-06, "learning_rate": 4.012957098241558e-05, "loss": 0.0, "num_input_tokens_seen": 3282816, "step": 11725 }, { "epoch": 130.33333333333334, "grad_norm": 5.316091574059101e-06, "learning_rate": 4.0121754254998076e-05, "loss": 0.0, "num_input_tokens_seen": 3284208, "step": 11730 }, { "epoch": 130.38888888888889, "grad_norm": 8.969512350631703e-07, "learning_rate": 4.011393519561606e-05, "loss": 0.0, "num_input_tokens_seen": 3285584, "step": 11735 }, { "epoch": 130.44444444444446, "grad_norm": 2.0604245491995243e-06, "learning_rate": 4.010611380547529e-05, "loss": 0.0, "num_input_tokens_seen": 3286976, "step": 11740 }, { "epoch": 130.5, "grad_norm": 1.7912640259964974e-06, "learning_rate": 4.009829008578192e-05, "loss": 0.0, "num_input_tokens_seen": 3288352, "step": 11745 }, { "epoch": 130.55555555555554, "grad_norm": 3.043573087779805e-06, "learning_rate": 4.00904640377425e-05, "loss": 0.0, "num_input_tokens_seen": 3289760, "step": 11750 }, { "epoch": 130.61111111111111, "grad_norm": 7.294246415767702e-07, "learning_rate": 4.0082635662563886e-05, "loss": 0.0, "num_input_tokens_seen": 3291200, "step": 11755 }, { "epoch": 130.66666666666666, "grad_norm": 1.0343750318497769e-06, "learning_rate": 4.007480496145331e-05, "loss": 0.0, "num_input_tokens_seen": 3292608, "step": 11760 }, { "epoch": 130.72222222222223, "grad_norm": 1.0024781431638985e-06, "learning_rate": 4.006697193561837e-05, "loss": 0.0, "num_input_tokens_seen": 3294032, "step": 11765 }, { "epoch": 130.77777777777777, "grad_norm": 2.3011612029222306e-06, "learning_rate": 4.005913658626701e-05, "loss": 0.0, "num_input_tokens_seen": 3295360, "step": 11770 }, { "epoch": 130.83333333333334, "grad_norm": 1.4789305851081735e-06, "learning_rate": 4.005129891460754e-05, "loss": 0.0, "num_input_tokens_seen": 3296768, "step": 11775 }, { "epoch": 130.88888888888889, "grad_norm": 1.7418462903151521e-06, "learning_rate": 4.004345892184864e-05, "loss": 0.0, "num_input_tokens_seen": 3298160, "step": 11780 }, { "epoch": 130.94444444444446, "grad_norm": 1.7250237078769715e-06, "learning_rate": 4.003561660919932e-05, "loss": 0.0, "num_input_tokens_seen": 3299536, "step": 11785 }, { "epoch": 131.0, "grad_norm": 5.303767238729051e-07, "learning_rate": 4.002777197786897e-05, "loss": 0.0, "num_input_tokens_seen": 3300896, "step": 11790 }, { "epoch": 131.05555555555554, "grad_norm": 5.03700903209392e-06, "learning_rate": 4.0019925029067326e-05, "loss": 0.0, "num_input_tokens_seen": 3302272, "step": 11795 }, { "epoch": 131.11111111111111, "grad_norm": 6.200560619618045e-06, "learning_rate": 4.0012075764004495e-05, "loss": 0.0, "num_input_tokens_seen": 3303712, "step": 11800 }, { "epoch": 131.11111111111111, "eval_loss": 1.2182016372680664, "eval_runtime": 1.3834, "eval_samples_per_second": 28.914, "eval_steps_per_second": 14.457, "num_input_tokens_seen": 3303712, "step": 11800 }, { "epoch": 131.16666666666666, "grad_norm": 3.064547172471066e-06, "learning_rate": 4.000422418389094e-05, "loss": 0.0, "num_input_tokens_seen": 3305152, "step": 11805 }, { "epoch": 131.22222222222223, "grad_norm": 3.837829353869893e-06, "learning_rate": 3.999637028993744e-05, "loss": 0.0, "num_input_tokens_seen": 3306576, "step": 11810 }, { "epoch": 131.27777777777777, "grad_norm": 1.112200607167324e-06, "learning_rate": 3.99885140833552e-05, "loss": 0.0, "num_input_tokens_seen": 3307984, "step": 11815 }, { "epoch": 131.33333333333334, "grad_norm": 1.8792768514686031e-06, "learning_rate": 3.998065556535572e-05, "loss": 0.0, "num_input_tokens_seen": 3309392, "step": 11820 }, { "epoch": 131.38888888888889, "grad_norm": 2.3359491478913696e-06, "learning_rate": 3.9972794737150895e-05, "loss": 0.0, "num_input_tokens_seen": 3310736, "step": 11825 }, { "epoch": 131.44444444444446, "grad_norm": 5.941371910012094e-06, "learning_rate": 3.996493159995297e-05, "loss": 0.0, "num_input_tokens_seen": 3312096, "step": 11830 }, { "epoch": 131.5, "grad_norm": 3.514573336360627e-06, "learning_rate": 3.995706615497453e-05, "loss": 0.0, "num_input_tokens_seen": 3313536, "step": 11835 }, { "epoch": 131.55555555555554, "grad_norm": 2.5576716780051356e-06, "learning_rate": 3.994919840342852e-05, "loss": 0.0, "num_input_tokens_seen": 3314944, "step": 11840 }, { "epoch": 131.61111111111111, "grad_norm": 2.008926912822062e-06, "learning_rate": 3.994132834652825e-05, "loss": 0.0, "num_input_tokens_seen": 3316304, "step": 11845 }, { "epoch": 131.66666666666666, "grad_norm": 2.4794221644697245e-06, "learning_rate": 3.99334559854874e-05, "loss": 0.0, "num_input_tokens_seen": 3317680, "step": 11850 }, { "epoch": 131.72222222222223, "grad_norm": 3.0660828542750096e-06, "learning_rate": 3.9925581321519955e-05, "loss": 0.0, "num_input_tokens_seen": 3319056, "step": 11855 }, { "epoch": 131.77777777777777, "grad_norm": 1.2122171710871044e-06, "learning_rate": 3.991770435584031e-05, "loss": 0.0, "num_input_tokens_seen": 3320480, "step": 11860 }, { "epoch": 131.83333333333334, "grad_norm": 1.6990786662063329e-06, "learning_rate": 3.990982508966319e-05, "loss": 0.0, "num_input_tokens_seen": 3321840, "step": 11865 }, { "epoch": 131.88888888888889, "grad_norm": 3.40921201313904e-06, "learning_rate": 3.990194352420367e-05, "loss": 0.0, "num_input_tokens_seen": 3323248, "step": 11870 }, { "epoch": 131.94444444444446, "grad_norm": 1.6469141428387957e-06, "learning_rate": 3.9894059660677184e-05, "loss": 0.0, "num_input_tokens_seen": 3324640, "step": 11875 }, { "epoch": 132.0, "grad_norm": 1.5297167692551739e-06, "learning_rate": 3.9886173500299526e-05, "loss": 0.0, "num_input_tokens_seen": 3326064, "step": 11880 }, { "epoch": 132.05555555555554, "grad_norm": 1.9118244836136e-06, "learning_rate": 3.987828504428685e-05, "loss": 0.0, "num_input_tokens_seen": 3327456, "step": 11885 }, { "epoch": 132.11111111111111, "grad_norm": 1.8326029476156691e-06, "learning_rate": 3.987039429385565e-05, "loss": 0.0, "num_input_tokens_seen": 3328928, "step": 11890 }, { "epoch": 132.16666666666666, "grad_norm": 2.1499793092516484e-06, "learning_rate": 3.986250125022277e-05, "loss": 0.0, "num_input_tokens_seen": 3330352, "step": 11895 }, { "epoch": 132.22222222222223, "grad_norm": 1.0027710004578694e-06, "learning_rate": 3.985460591460544e-05, "loss": 0.0, "num_input_tokens_seen": 3331760, "step": 11900 }, { "epoch": 132.27777777777777, "grad_norm": 6.671431037830189e-06, "learning_rate": 3.984670828822118e-05, "loss": 0.0, "num_input_tokens_seen": 3333152, "step": 11905 }, { "epoch": 132.33333333333334, "grad_norm": 5.87133627050207e-06, "learning_rate": 3.983880837228794e-05, "loss": 0.0, "num_input_tokens_seen": 3334560, "step": 11910 }, { "epoch": 132.38888888888889, "grad_norm": 3.014458798134001e-06, "learning_rate": 3.983090616802396e-05, "loss": 0.0, "num_input_tokens_seen": 3335952, "step": 11915 }, { "epoch": 132.44444444444446, "grad_norm": 3.4259126096003456e-06, "learning_rate": 3.982300167664788e-05, "loss": 0.0, "num_input_tokens_seen": 3337328, "step": 11920 }, { "epoch": 132.5, "grad_norm": 1.8522151776778628e-06, "learning_rate": 3.981509489937868e-05, "loss": 0.0, "num_input_tokens_seen": 3338704, "step": 11925 }, { "epoch": 132.55555555555554, "grad_norm": 3.961585662182188e-06, "learning_rate": 3.9807185837435643e-05, "loss": 0.0, "num_input_tokens_seen": 3340128, "step": 11930 }, { "epoch": 132.61111111111111, "grad_norm": 2.433300551274442e-06, "learning_rate": 3.9799274492038484e-05, "loss": 0.0, "num_input_tokens_seen": 3341552, "step": 11935 }, { "epoch": 132.66666666666666, "grad_norm": 1.5402075632664491e-06, "learning_rate": 3.979136086440722e-05, "loss": 0.0, "num_input_tokens_seen": 3342896, "step": 11940 }, { "epoch": 132.72222222222223, "grad_norm": 1.3162216419004835e-06, "learning_rate": 3.9783444955762226e-05, "loss": 0.0, "num_input_tokens_seen": 3344256, "step": 11945 }, { "epoch": 132.77777777777777, "grad_norm": 2.7543474061531015e-06, "learning_rate": 3.977552676732424e-05, "loss": 0.0, "num_input_tokens_seen": 3345632, "step": 11950 }, { "epoch": 132.83333333333334, "grad_norm": 1.2524495787147316e-06, "learning_rate": 3.976760630031435e-05, "loss": 0.0, "num_input_tokens_seen": 3347056, "step": 11955 }, { "epoch": 132.88888888888889, "grad_norm": 2.337673549845931e-06, "learning_rate": 3.975968355595398e-05, "loss": 0.0, "num_input_tokens_seen": 3348480, "step": 11960 }, { "epoch": 132.94444444444446, "grad_norm": 1.781702735570434e-06, "learning_rate": 3.9751758535464935e-05, "loss": 0.0, "num_input_tokens_seen": 3349888, "step": 11965 }, { "epoch": 133.0, "grad_norm": 6.565542207681574e-07, "learning_rate": 3.9743831240069326e-05, "loss": 0.0, "num_input_tokens_seen": 3351296, "step": 11970 }, { "epoch": 133.05555555555554, "grad_norm": 4.1486346162855625e-06, "learning_rate": 3.9735901670989675e-05, "loss": 0.0, "num_input_tokens_seen": 3352624, "step": 11975 }, { "epoch": 133.11111111111111, "grad_norm": 3.5682046473084483e-06, "learning_rate": 3.97279698294488e-05, "loss": 0.0, "num_input_tokens_seen": 3354032, "step": 11980 }, { "epoch": 133.16666666666666, "grad_norm": 1.8256876046507386e-06, "learning_rate": 3.9720035716669876e-05, "loss": 0.0, "num_input_tokens_seen": 3355440, "step": 11985 }, { "epoch": 133.22222222222223, "grad_norm": 1.8461619220033754e-06, "learning_rate": 3.9712099333876474e-05, "loss": 0.0, "num_input_tokens_seen": 3356848, "step": 11990 }, { "epoch": 133.27777777777777, "grad_norm": 1.7956510873773368e-06, "learning_rate": 3.9704160682292475e-05, "loss": 0.0, "num_input_tokens_seen": 3358272, "step": 11995 }, { "epoch": 133.33333333333334, "grad_norm": 1.5003697626525536e-06, "learning_rate": 3.9696219763142106e-05, "loss": 0.0, "num_input_tokens_seen": 3359680, "step": 12000 }, { "epoch": 133.33333333333334, "eval_loss": 1.2517651319503784, "eval_runtime": 1.3881, "eval_samples_per_second": 28.816, "eval_steps_per_second": 14.408, "num_input_tokens_seen": 3359680, "step": 12000 }, { "epoch": 133.38888888888889, "grad_norm": 1.3968834764455096e-06, "learning_rate": 3.968827657764997e-05, "loss": 0.0, "num_input_tokens_seen": 3361088, "step": 12005 }, { "epoch": 133.44444444444446, "grad_norm": 5.190460797166452e-06, "learning_rate": 3.9680331127041e-05, "loss": 0.0, "num_input_tokens_seen": 3362528, "step": 12010 }, { "epoch": 133.5, "grad_norm": 2.960370693472214e-06, "learning_rate": 3.9672383412540495e-05, "loss": 0.0, "num_input_tokens_seen": 3363920, "step": 12015 }, { "epoch": 133.55555555555554, "grad_norm": 1.744668225001078e-06, "learning_rate": 3.966443343537407e-05, "loss": 0.0, "num_input_tokens_seen": 3365312, "step": 12020 }, { "epoch": 133.61111111111111, "grad_norm": 2.7757068892242387e-06, "learning_rate": 3.965648119676772e-05, "loss": 0.0, "num_input_tokens_seen": 3366736, "step": 12025 }, { "epoch": 133.66666666666666, "grad_norm": 1.0024944003816927e-06, "learning_rate": 3.96485266979478e-05, "loss": 0.0, "num_input_tokens_seen": 3368160, "step": 12030 }, { "epoch": 133.72222222222223, "grad_norm": 2.369904677834711e-06, "learning_rate": 3.9640569940140974e-05, "loss": 0.0, "num_input_tokens_seen": 3369584, "step": 12035 }, { "epoch": 133.77777777777777, "grad_norm": 9.63866682468506e-07, "learning_rate": 3.963261092457428e-05, "loss": 0.0, "num_input_tokens_seen": 3371008, "step": 12040 }, { "epoch": 133.83333333333334, "grad_norm": 8.435022209596355e-07, "learning_rate": 3.962464965247509e-05, "loss": 0.0, "num_input_tokens_seen": 3372416, "step": 12045 }, { "epoch": 133.88888888888889, "grad_norm": 1.3443257103062933e-06, "learning_rate": 3.9616686125071135e-05, "loss": 0.0, "num_input_tokens_seen": 3373840, "step": 12050 }, { "epoch": 133.94444444444446, "grad_norm": 1.1248345799685922e-05, "learning_rate": 3.9608720343590506e-05, "loss": 0.0, "num_input_tokens_seen": 3375216, "step": 12055 }, { "epoch": 134.0, "grad_norm": 6.106758064561291e-06, "learning_rate": 3.960075230926161e-05, "loss": 0.0, "num_input_tokens_seen": 3376560, "step": 12060 }, { "epoch": 134.05555555555554, "grad_norm": 5.9605963542708196e-06, "learning_rate": 3.959278202331322e-05, "loss": 0.0, "num_input_tokens_seen": 3377968, "step": 12065 }, { "epoch": 134.11111111111111, "grad_norm": 1.6819008124002721e-06, "learning_rate": 3.958480948697446e-05, "loss": 0.0, "num_input_tokens_seen": 3379312, "step": 12070 }, { "epoch": 134.16666666666666, "grad_norm": 4.192460437479895e-06, "learning_rate": 3.95768347014748e-05, "loss": 0.0, "num_input_tokens_seen": 3380736, "step": 12075 }, { "epoch": 134.22222222222223, "grad_norm": 4.173632987658493e-06, "learning_rate": 3.956885766804404e-05, "loss": 0.0, "num_input_tokens_seen": 3382144, "step": 12080 }, { "epoch": 134.27777777777777, "grad_norm": 1.4013915006216848e-06, "learning_rate": 3.956087838791235e-05, "loss": 0.0, "num_input_tokens_seen": 3383504, "step": 12085 }, { "epoch": 134.33333333333334, "grad_norm": 1.8713410554482834e-06, "learning_rate": 3.955289686231022e-05, "loss": 0.0, "num_input_tokens_seen": 3384928, "step": 12090 }, { "epoch": 134.38888888888889, "grad_norm": 3.8145728922245326e-06, "learning_rate": 3.9544913092468504e-05, "loss": 0.0, "num_input_tokens_seen": 3386336, "step": 12095 }, { "epoch": 134.44444444444446, "grad_norm": 1.1447912129369797e-06, "learning_rate": 3.9536927079618425e-05, "loss": 0.0, "num_input_tokens_seen": 3387728, "step": 12100 }, { "epoch": 134.5, "grad_norm": 2.0054581000295e-06, "learning_rate": 3.9528938824991494e-05, "loss": 0.0, "num_input_tokens_seen": 3389120, "step": 12105 }, { "epoch": 134.55555555555554, "grad_norm": 1.4148838545224862e-06, "learning_rate": 3.952094832981962e-05, "loss": 0.0, "num_input_tokens_seen": 3390544, "step": 12110 }, { "epoch": 134.61111111111111, "grad_norm": 3.4412760214763694e-06, "learning_rate": 3.951295559533503e-05, "loss": 0.0, "num_input_tokens_seen": 3391920, "step": 12115 }, { "epoch": 134.66666666666666, "grad_norm": 1.6147615724548814e-06, "learning_rate": 3.95049606227703e-05, "loss": 0.0, "num_input_tokens_seen": 3393296, "step": 12120 }, { "epoch": 134.72222222222223, "grad_norm": 2.8755250696121948e-06, "learning_rate": 3.949696341335838e-05, "loss": 0.0, "num_input_tokens_seen": 3394720, "step": 12125 }, { "epoch": 134.77777777777777, "grad_norm": 7.98486098574358e-07, "learning_rate": 3.9488963968332503e-05, "loss": 0.0, "num_input_tokens_seen": 3396128, "step": 12130 }, { "epoch": 134.83333333333334, "grad_norm": 6.736080194968963e-06, "learning_rate": 3.948096228892631e-05, "loss": 0.0, "num_input_tokens_seen": 3397536, "step": 12135 }, { "epoch": 134.88888888888889, "grad_norm": 2.313227014383301e-06, "learning_rate": 3.947295837637375e-05, "loss": 0.0, "num_input_tokens_seen": 3398928, "step": 12140 }, { "epoch": 134.94444444444446, "grad_norm": 3.6676806303148624e-06, "learning_rate": 3.9464952231909135e-05, "loss": 0.0, "num_input_tokens_seen": 3400336, "step": 12145 }, { "epoch": 135.0, "grad_norm": 1.6049294799813651e-06, "learning_rate": 3.945694385676711e-05, "loss": 0.0, "num_input_tokens_seen": 3401776, "step": 12150 }, { "epoch": 135.05555555555554, "grad_norm": 2.1467053556989413e-06, "learning_rate": 3.944893325218265e-05, "loss": 0.0, "num_input_tokens_seen": 3403184, "step": 12155 }, { "epoch": 135.11111111111111, "grad_norm": 5.051218522567069e-06, "learning_rate": 3.944092041939112e-05, "loss": 0.0, "num_input_tokens_seen": 3404560, "step": 12160 }, { "epoch": 135.16666666666666, "grad_norm": 5.815019903820939e-06, "learning_rate": 3.943290535962818e-05, "loss": 0.0, "num_input_tokens_seen": 3406000, "step": 12165 }, { "epoch": 135.22222222222223, "grad_norm": 1.776784188223246e-06, "learning_rate": 3.942488807412985e-05, "loss": 0.0, "num_input_tokens_seen": 3407392, "step": 12170 }, { "epoch": 135.27777777777777, "grad_norm": 1.0583876246528234e-06, "learning_rate": 3.941686856413251e-05, "loss": 0.0, "num_input_tokens_seen": 3408800, "step": 12175 }, { "epoch": 135.33333333333334, "grad_norm": 9.17917247988953e-07, "learning_rate": 3.9408846830872874e-05, "loss": 0.0, "num_input_tokens_seen": 3410240, "step": 12180 }, { "epoch": 135.38888888888889, "grad_norm": 2.0141010281804483e-06, "learning_rate": 3.940082287558798e-05, "loss": 0.0, "num_input_tokens_seen": 3411616, "step": 12185 }, { "epoch": 135.44444444444446, "grad_norm": 2.1220466805971228e-06, "learning_rate": 3.939279669951522e-05, "loss": 0.0, "num_input_tokens_seen": 3413040, "step": 12190 }, { "epoch": 135.5, "grad_norm": 2.0511913589871256e-06, "learning_rate": 3.938476830389234e-05, "loss": 0.0, "num_input_tokens_seen": 3414432, "step": 12195 }, { "epoch": 135.55555555555554, "grad_norm": 1.2514237823779695e-05, "learning_rate": 3.937673768995742e-05, "loss": 0.0, "num_input_tokens_seen": 3415824, "step": 12200 }, { "epoch": 135.55555555555554, "eval_loss": 1.1945501565933228, "eval_runtime": 1.4027, "eval_samples_per_second": 28.516, "eval_steps_per_second": 14.258, "num_input_tokens_seen": 3415824, "step": 12200 }, { "epoch": 135.61111111111111, "grad_norm": 1.4605481055696146e-06, "learning_rate": 3.936870485894888e-05, "loss": 0.0, "num_input_tokens_seen": 3417184, "step": 12205 }, { "epoch": 135.66666666666666, "grad_norm": 2.4657047106302343e-06, "learning_rate": 3.9360669812105475e-05, "loss": 0.0, "num_input_tokens_seen": 3418576, "step": 12210 }, { "epoch": 135.72222222222223, "grad_norm": 2.4387877601839136e-06, "learning_rate": 3.9352632550666325e-05, "loss": 0.0, "num_input_tokens_seen": 3419984, "step": 12215 }, { "epoch": 135.77777777777777, "grad_norm": 3.1682554890721804e-06, "learning_rate": 3.9344593075870866e-05, "loss": 0.0, "num_input_tokens_seen": 3421360, "step": 12220 }, { "epoch": 135.83333333333334, "grad_norm": 2.6934887955576414e-06, "learning_rate": 3.933655138895889e-05, "loss": 0.0, "num_input_tokens_seen": 3422720, "step": 12225 }, { "epoch": 135.88888888888889, "grad_norm": 4.654044914786937e-06, "learning_rate": 3.932850749117053e-05, "loss": 0.0, "num_input_tokens_seen": 3424112, "step": 12230 }, { "epoch": 135.94444444444446, "grad_norm": 8.583981070842128e-06, "learning_rate": 3.932046138374624e-05, "loss": 0.0, "num_input_tokens_seen": 3425488, "step": 12235 }, { "epoch": 136.0, "grad_norm": 3.810794396486017e-06, "learning_rate": 3.9312413067926854e-05, "loss": 0.0, "num_input_tokens_seen": 3426912, "step": 12240 }, { "epoch": 136.05555555555554, "grad_norm": 1.665480226620275e-06, "learning_rate": 3.9304362544953506e-05, "loss": 0.0, "num_input_tokens_seen": 3428304, "step": 12245 }, { "epoch": 136.11111111111111, "grad_norm": 2.463334112690063e-06, "learning_rate": 3.929630981606769e-05, "loss": 0.0, "num_input_tokens_seen": 3429712, "step": 12250 }, { "epoch": 136.16666666666666, "grad_norm": 1.4541853943228489e-06, "learning_rate": 3.928825488251124e-05, "loss": 0.0, "num_input_tokens_seen": 3431088, "step": 12255 }, { "epoch": 136.22222222222223, "grad_norm": 2.467524154781131e-06, "learning_rate": 3.9280197745526344e-05, "loss": 0.0, "num_input_tokens_seen": 3432448, "step": 12260 }, { "epoch": 136.27777777777777, "grad_norm": 1.2774327160514076e-06, "learning_rate": 3.9272138406355495e-05, "loss": 0.0, "num_input_tokens_seen": 3433872, "step": 12265 }, { "epoch": 136.33333333333334, "grad_norm": 2.5914253001246834e-06, "learning_rate": 3.926407686624154e-05, "loss": 0.0, "num_input_tokens_seen": 3435248, "step": 12270 }, { "epoch": 136.38888888888889, "grad_norm": 5.333178023647633e-07, "learning_rate": 3.9256013126427684e-05, "loss": 0.0, "num_input_tokens_seen": 3436592, "step": 12275 }, { "epoch": 136.44444444444446, "grad_norm": 1.5252442153723678e-06, "learning_rate": 3.9247947188157455e-05, "loss": 0.0, "num_input_tokens_seen": 3438016, "step": 12280 }, { "epoch": 136.5, "grad_norm": 1.2615356581591186e-06, "learning_rate": 3.9239879052674715e-05, "loss": 0.0, "num_input_tokens_seen": 3439392, "step": 12285 }, { "epoch": 136.55555555555554, "grad_norm": 1.8885689314629417e-06, "learning_rate": 3.9231808721223673e-05, "loss": 0.0, "num_input_tokens_seen": 3440816, "step": 12290 }, { "epoch": 136.61111111111111, "grad_norm": 4.499108854361111e-06, "learning_rate": 3.9223736195048886e-05, "loss": 0.0, "num_input_tokens_seen": 3442208, "step": 12295 }, { "epoch": 136.66666666666666, "grad_norm": 8.749753419579065e-07, "learning_rate": 3.921566147539523e-05, "loss": 0.0, "num_input_tokens_seen": 3443632, "step": 12300 }, { "epoch": 136.72222222222223, "grad_norm": 5.3092035159352235e-06, "learning_rate": 3.920758456350792e-05, "loss": 0.0, "num_input_tokens_seen": 3445040, "step": 12305 }, { "epoch": 136.77777777777777, "grad_norm": 1.531157863610133e-06, "learning_rate": 3.919950546063253e-05, "loss": 0.0, "num_input_tokens_seen": 3446352, "step": 12310 }, { "epoch": 136.83333333333334, "grad_norm": 5.306927505444037e-06, "learning_rate": 3.919142416801496e-05, "loss": 0.0, "num_input_tokens_seen": 3447792, "step": 12315 }, { "epoch": 136.88888888888889, "grad_norm": 3.2951872981357155e-06, "learning_rate": 3.918334068690144e-05, "loss": 0.0, "num_input_tokens_seen": 3449152, "step": 12320 }, { "epoch": 136.94444444444446, "grad_norm": 6.515468612633413e-06, "learning_rate": 3.917525501853855e-05, "loss": 0.0, "num_input_tokens_seen": 3450576, "step": 12325 }, { "epoch": 137.0, "grad_norm": 2.3370350845652865e-06, "learning_rate": 3.916716716417319e-05, "loss": 0.0, "num_input_tokens_seen": 3451984, "step": 12330 }, { "epoch": 137.05555555555554, "grad_norm": 1.5862927966736606e-06, "learning_rate": 3.915907712505263e-05, "loss": 0.0, "num_input_tokens_seen": 3453360, "step": 12335 }, { "epoch": 137.11111111111111, "grad_norm": 1.6596886780462228e-06, "learning_rate": 3.915098490242444e-05, "loss": 0.0, "num_input_tokens_seen": 3454736, "step": 12340 }, { "epoch": 137.16666666666666, "grad_norm": 7.56160807213746e-06, "learning_rate": 3.914289049753654e-05, "loss": 0.0, "num_input_tokens_seen": 3456160, "step": 12345 }, { "epoch": 137.22222222222223, "grad_norm": 2.0257593860151246e-06, "learning_rate": 3.913479391163719e-05, "loss": 0.0, "num_input_tokens_seen": 3457536, "step": 12350 }, { "epoch": 137.27777777777777, "grad_norm": 2.737996055657277e-06, "learning_rate": 3.9126695145975e-05, "loss": 0.0, "num_input_tokens_seen": 3458976, "step": 12355 }, { "epoch": 137.33333333333334, "grad_norm": 3.237602413719287e-06, "learning_rate": 3.911859420179889e-05, "loss": 0.0, "num_input_tokens_seen": 3460336, "step": 12360 }, { "epoch": 137.38888888888889, "grad_norm": 1.1238068964303238e-06, "learning_rate": 3.911049108035813e-05, "loss": 0.0, "num_input_tokens_seen": 3461760, "step": 12365 }, { "epoch": 137.44444444444446, "grad_norm": 2.6089455786859617e-06, "learning_rate": 3.910238578290232e-05, "loss": 0.0, "num_input_tokens_seen": 3463200, "step": 12370 }, { "epoch": 137.5, "grad_norm": 3.511414888635045e-06, "learning_rate": 3.90942783106814e-05, "loss": 0.0, "num_input_tokens_seen": 3464576, "step": 12375 }, { "epoch": 137.55555555555554, "grad_norm": 3.503456355247181e-06, "learning_rate": 3.908616866494564e-05, "loss": 0.0, "num_input_tokens_seen": 3465936, "step": 12380 }, { "epoch": 137.61111111111111, "grad_norm": 2.045121391347493e-06, "learning_rate": 3.907805684694566e-05, "loss": 0.0, "num_input_tokens_seen": 3467344, "step": 12385 }, { "epoch": 137.66666666666666, "grad_norm": 6.1055693549860734e-06, "learning_rate": 3.90699428579324e-05, "loss": 0.0, "num_input_tokens_seen": 3468720, "step": 12390 }, { "epoch": 137.72222222222223, "grad_norm": 1.4049976471142145e-06, "learning_rate": 3.906182669915713e-05, "loss": 0.0, "num_input_tokens_seen": 3470128, "step": 12395 }, { "epoch": 137.77777777777777, "grad_norm": 3.988096977991518e-06, "learning_rate": 3.9053708371871476e-05, "loss": 0.0, "num_input_tokens_seen": 3471520, "step": 12400 }, { "epoch": 137.77777777777777, "eval_loss": 1.2287622690200806, "eval_runtime": 1.3874, "eval_samples_per_second": 28.832, "eval_steps_per_second": 14.416, "num_input_tokens_seen": 3471520, "step": 12400 }, { "epoch": 137.83333333333334, "grad_norm": 1.281358777305286e-06, "learning_rate": 3.904558787732738e-05, "loss": 0.0, "num_input_tokens_seen": 3472944, "step": 12405 }, { "epoch": 137.88888888888889, "grad_norm": 1.8659840179680032e-06, "learning_rate": 3.9037465216777135e-05, "loss": 0.0, "num_input_tokens_seen": 3474352, "step": 12410 }, { "epoch": 137.94444444444446, "grad_norm": 1.7463132735429099e-06, "learning_rate": 3.902934039147334e-05, "loss": 0.0, "num_input_tokens_seen": 3475776, "step": 12415 }, { "epoch": 138.0, "grad_norm": 2.2771987460146192e-06, "learning_rate": 3.902121340266894e-05, "loss": 0.0, "num_input_tokens_seen": 3477184, "step": 12420 }, { "epoch": 138.05555555555554, "grad_norm": 1.960904910447425e-06, "learning_rate": 3.9013084251617246e-05, "loss": 0.0, "num_input_tokens_seen": 3478608, "step": 12425 }, { "epoch": 138.11111111111111, "grad_norm": 2.5222952899639495e-06, "learning_rate": 3.9004952939571865e-05, "loss": 0.0, "num_input_tokens_seen": 3479968, "step": 12430 }, { "epoch": 138.16666666666666, "grad_norm": 2.2580634322366677e-06, "learning_rate": 3.899681946778673e-05, "loss": 0.0, "num_input_tokens_seen": 3481392, "step": 12435 }, { "epoch": 138.22222222222223, "grad_norm": 3.412415026105009e-06, "learning_rate": 3.898868383751615e-05, "loss": 0.0, "num_input_tokens_seen": 3482832, "step": 12440 }, { "epoch": 138.27777777777777, "grad_norm": 1.4583492884412408e-06, "learning_rate": 3.8980546050014724e-05, "loss": 0.0, "num_input_tokens_seen": 3484192, "step": 12445 }, { "epoch": 138.33333333333334, "grad_norm": 6.208379545569187e-06, "learning_rate": 3.897240610653741e-05, "loss": 0.0, "num_input_tokens_seen": 3485584, "step": 12450 }, { "epoch": 138.38888888888889, "grad_norm": 4.9968407438427676e-06, "learning_rate": 3.896426400833948e-05, "loss": 0.0, "num_input_tokens_seen": 3486992, "step": 12455 }, { "epoch": 138.44444444444446, "grad_norm": 2.305345788045088e-06, "learning_rate": 3.895611975667656e-05, "loss": 0.0, "num_input_tokens_seen": 3488368, "step": 12460 }, { "epoch": 138.5, "grad_norm": 2.0547022359096445e-06, "learning_rate": 3.8947973352804584e-05, "loss": 0.0, "num_input_tokens_seen": 3489792, "step": 12465 }, { "epoch": 138.55555555555554, "grad_norm": 9.678057040218846e-07, "learning_rate": 3.893982479797984e-05, "loss": 0.0, "num_input_tokens_seen": 3491200, "step": 12470 }, { "epoch": 138.61111111111111, "grad_norm": 1.1898019920408842e-06, "learning_rate": 3.8931674093458926e-05, "loss": 0.0, "num_input_tokens_seen": 3492608, "step": 12475 }, { "epoch": 138.66666666666666, "grad_norm": 1.5666128092561848e-06, "learning_rate": 3.89235212404988e-05, "loss": 0.0, "num_input_tokens_seen": 3494016, "step": 12480 }, { "epoch": 138.72222222222223, "grad_norm": 1.560352302476531e-06, "learning_rate": 3.891536624035672e-05, "loss": 0.0, "num_input_tokens_seen": 3495424, "step": 12485 }, { "epoch": 138.77777777777777, "grad_norm": 1.8134467154595768e-06, "learning_rate": 3.8907209094290295e-05, "loss": 0.0, "num_input_tokens_seen": 3496864, "step": 12490 }, { "epoch": 138.83333333333334, "grad_norm": 2.940863623734913e-06, "learning_rate": 3.8899049803557466e-05, "loss": 0.0, "num_input_tokens_seen": 3498240, "step": 12495 }, { "epoch": 138.88888888888889, "grad_norm": 1.5318640862460597e-06, "learning_rate": 3.889088836941648e-05, "loss": 0.0, "num_input_tokens_seen": 3499680, "step": 12500 }, { "epoch": 138.94444444444446, "grad_norm": 2.8377980925142765e-06, "learning_rate": 3.8882724793125946e-05, "loss": 0.0, "num_input_tokens_seen": 3501056, "step": 12505 }, { "epoch": 139.0, "grad_norm": 3.7120892102393555e-06, "learning_rate": 3.8874559075944794e-05, "loss": 0.0, "num_input_tokens_seen": 3502448, "step": 12510 }, { "epoch": 139.05555555555554, "grad_norm": 2.5109864054684294e-06, "learning_rate": 3.886639121913227e-05, "loss": 0.0, "num_input_tokens_seen": 3503872, "step": 12515 }, { "epoch": 139.11111111111111, "grad_norm": 2.7456210318632657e-06, "learning_rate": 3.885822122394797e-05, "loss": 0.0, "num_input_tokens_seen": 3505280, "step": 12520 }, { "epoch": 139.16666666666666, "grad_norm": 1.8533746697357856e-06, "learning_rate": 3.8850049091651794e-05, "loss": 0.0, "num_input_tokens_seen": 3506672, "step": 12525 }, { "epoch": 139.22222222222223, "grad_norm": 1.897410243145714e-06, "learning_rate": 3.8841874823504e-05, "loss": 0.0, "num_input_tokens_seen": 3508048, "step": 12530 }, { "epoch": 139.27777777777777, "grad_norm": 2.742127662713756e-06, "learning_rate": 3.8833698420765157e-05, "loss": 0.0, "num_input_tokens_seen": 3509408, "step": 12535 }, { "epoch": 139.33333333333334, "grad_norm": 1.633728174965654e-06, "learning_rate": 3.882551988469618e-05, "loss": 0.0, "num_input_tokens_seen": 3510800, "step": 12540 }, { "epoch": 139.38888888888889, "grad_norm": 1.6438310694866232e-06, "learning_rate": 3.881733921655829e-05, "loss": 0.0, "num_input_tokens_seen": 3512224, "step": 12545 }, { "epoch": 139.44444444444446, "grad_norm": 3.239225861761952e-06, "learning_rate": 3.8809156417613054e-05, "loss": 0.0, "num_input_tokens_seen": 3513664, "step": 12550 }, { "epoch": 139.5, "grad_norm": 3.883900717482902e-06, "learning_rate": 3.8800971489122364e-05, "loss": 0.0, "num_input_tokens_seen": 3515072, "step": 12555 }, { "epoch": 139.55555555555554, "grad_norm": 1.2487157619034406e-06, "learning_rate": 3.8792784432348434e-05, "loss": 0.0, "num_input_tokens_seen": 3516464, "step": 12560 }, { "epoch": 139.61111111111111, "grad_norm": 7.93901472206926e-06, "learning_rate": 3.878459524855381e-05, "loss": 0.0, "num_input_tokens_seen": 3517824, "step": 12565 }, { "epoch": 139.66666666666666, "grad_norm": 8.86618818185525e-06, "learning_rate": 3.8776403939001384e-05, "loss": 0.0, "num_input_tokens_seen": 3519216, "step": 12570 }, { "epoch": 139.72222222222223, "grad_norm": 4.706967501988402e-06, "learning_rate": 3.876821050495433e-05, "loss": 0.0, "num_input_tokens_seen": 3520624, "step": 12575 }, { "epoch": 139.77777777777777, "grad_norm": 4.546532636595657e-06, "learning_rate": 3.87600149476762e-05, "loss": 0.0, "num_input_tokens_seen": 3522048, "step": 12580 }, { "epoch": 139.83333333333334, "grad_norm": 2.551354327806621e-06, "learning_rate": 3.8751817268430843e-05, "loss": 0.0, "num_input_tokens_seen": 3523440, "step": 12585 }, { "epoch": 139.88888888888889, "grad_norm": 4.2617730287020095e-06, "learning_rate": 3.8743617468482464e-05, "loss": 0.0, "num_input_tokens_seen": 3524832, "step": 12590 }, { "epoch": 139.94444444444446, "grad_norm": 1.7361904838253395e-06, "learning_rate": 3.8735415549095535e-05, "loss": 0.0, "num_input_tokens_seen": 3526224, "step": 12595 }, { "epoch": 140.0, "grad_norm": 1.8975516695718397e-06, "learning_rate": 3.8727211511534934e-05, "loss": 0.0, "num_input_tokens_seen": 3527664, "step": 12600 }, { "epoch": 140.0, "eval_loss": 1.2415287494659424, "eval_runtime": 1.3917, "eval_samples_per_second": 28.741, "eval_steps_per_second": 14.371, "num_input_tokens_seen": 3527664, "step": 12600 }, { "epoch": 140.05555555555554, "grad_norm": 3.8120135741337435e-06, "learning_rate": 3.8719005357065804e-05, "loss": 0.0, "num_input_tokens_seen": 3529040, "step": 12605 }, { "epoch": 140.11111111111111, "grad_norm": 2.2635072127741296e-06, "learning_rate": 3.8710797086953645e-05, "loss": 0.0, "num_input_tokens_seen": 3530448, "step": 12610 }, { "epoch": 140.16666666666666, "grad_norm": 1.526325490885938e-06, "learning_rate": 3.870258670246427e-05, "loss": 0.0, "num_input_tokens_seen": 3531824, "step": 12615 }, { "epoch": 140.22222222222223, "grad_norm": 7.756112267998105e-07, "learning_rate": 3.869437420486384e-05, "loss": 0.0, "num_input_tokens_seen": 3533280, "step": 12620 }, { "epoch": 140.27777777777777, "grad_norm": 1.0244702934869565e-06, "learning_rate": 3.8686159595418805e-05, "loss": 0.0, "num_input_tokens_seen": 3534688, "step": 12625 }, { "epoch": 140.33333333333334, "grad_norm": 1.030946236824093e-06, "learning_rate": 3.867794287539597e-05, "loss": 0.0, "num_input_tokens_seen": 3536080, "step": 12630 }, { "epoch": 140.38888888888889, "grad_norm": 2.0956379103154177e-06, "learning_rate": 3.866972404606245e-05, "loss": 0.0, "num_input_tokens_seen": 3537424, "step": 12635 }, { "epoch": 140.44444444444446, "grad_norm": 8.80754157606134e-07, "learning_rate": 3.866150310868571e-05, "loss": 0.0, "num_input_tokens_seen": 3538832, "step": 12640 }, { "epoch": 140.5, "grad_norm": 3.276009374530986e-05, "learning_rate": 3.8653280064533506e-05, "loss": 0.0, "num_input_tokens_seen": 3540256, "step": 12645 }, { "epoch": 140.55555555555554, "grad_norm": 2.4093856154649984e-06, "learning_rate": 3.864505491487394e-05, "loss": 0.0, "num_input_tokens_seen": 3541648, "step": 12650 }, { "epoch": 140.61111111111111, "grad_norm": 6.7389182731858455e-06, "learning_rate": 3.8636827660975414e-05, "loss": 0.0, "num_input_tokens_seen": 3543120, "step": 12655 }, { "epoch": 140.66666666666666, "grad_norm": 1.3396436315815663e-06, "learning_rate": 3.862859830410671e-05, "loss": 0.0, "num_input_tokens_seen": 3544528, "step": 12660 }, { "epoch": 140.72222222222223, "grad_norm": 1.7705252730593202e-06, "learning_rate": 3.862036684553688e-05, "loss": 0.0, "num_input_tokens_seen": 3545904, "step": 12665 }, { "epoch": 140.77777777777777, "grad_norm": 1.701683231658535e-06, "learning_rate": 3.8612133286535314e-05, "loss": 0.0, "num_input_tokens_seen": 3547328, "step": 12670 }, { "epoch": 140.83333333333334, "grad_norm": 2.0157976905466057e-06, "learning_rate": 3.860389762837173e-05, "loss": 0.0, "num_input_tokens_seen": 3548720, "step": 12675 }, { "epoch": 140.88888888888889, "grad_norm": 3.6887649912387133e-06, "learning_rate": 3.859565987231618e-05, "loss": 0.0, "num_input_tokens_seen": 3550080, "step": 12680 }, { "epoch": 140.94444444444446, "grad_norm": 1.3213142437962233e-06, "learning_rate": 3.858742001963902e-05, "loss": 0.0, "num_input_tokens_seen": 3551488, "step": 12685 }, { "epoch": 141.0, "grad_norm": 2.070227083095233e-06, "learning_rate": 3.857917807161094e-05, "loss": 0.0, "num_input_tokens_seen": 3552864, "step": 12690 }, { "epoch": 141.05555555555554, "grad_norm": 1.1980893077634391e-06, "learning_rate": 3.857093402950296e-05, "loss": 0.0, "num_input_tokens_seen": 3554272, "step": 12695 }, { "epoch": 141.11111111111111, "grad_norm": 1.41414966492448e-06, "learning_rate": 3.8562687894586414e-05, "loss": 0.0, "num_input_tokens_seen": 3555696, "step": 12700 }, { "epoch": 141.16666666666666, "grad_norm": 4.383835403132252e-06, "learning_rate": 3.8554439668132946e-05, "loss": 0.0, "num_input_tokens_seen": 3557104, "step": 12705 }, { "epoch": 141.22222222222223, "grad_norm": 2.4705873329367023e-06, "learning_rate": 3.854618935141455e-05, "loss": 0.0, "num_input_tokens_seen": 3558480, "step": 12710 }, { "epoch": 141.27777777777777, "grad_norm": 4.422689471539343e-06, "learning_rate": 3.8537936945703525e-05, "loss": 0.0, "num_input_tokens_seen": 3559872, "step": 12715 }, { "epoch": 141.33333333333334, "grad_norm": 3.6124199596088147e-06, "learning_rate": 3.852968245227249e-05, "loss": 0.0, "num_input_tokens_seen": 3561280, "step": 12720 }, { "epoch": 141.38888888888889, "grad_norm": 8.43163888930576e-06, "learning_rate": 3.85214258723944e-05, "loss": 0.0, "num_input_tokens_seen": 3562688, "step": 12725 }, { "epoch": 141.44444444444446, "grad_norm": 1.891308215817844e-06, "learning_rate": 3.8513167207342524e-05, "loss": 0.0, "num_input_tokens_seen": 3564080, "step": 12730 }, { "epoch": 141.5, "grad_norm": 2.5125152660621097e-06, "learning_rate": 3.850490645839044e-05, "loss": 0.0, "num_input_tokens_seen": 3565440, "step": 12735 }, { "epoch": 141.55555555555554, "grad_norm": 1.6904040194276604e-06, "learning_rate": 3.849664362681207e-05, "loss": 0.0, "num_input_tokens_seen": 3566880, "step": 12740 }, { "epoch": 141.61111111111111, "grad_norm": 5.805914952361491e-06, "learning_rate": 3.848837871388165e-05, "loss": 0.0, "num_input_tokens_seen": 3568288, "step": 12745 }, { "epoch": 141.66666666666666, "grad_norm": 2.4930225208663614e-06, "learning_rate": 3.848011172087371e-05, "loss": 0.0, "num_input_tokens_seen": 3569648, "step": 12750 }, { "epoch": 141.72222222222223, "grad_norm": 4.4145399442641065e-06, "learning_rate": 3.847184264906315e-05, "loss": 0.0, "num_input_tokens_seen": 3571056, "step": 12755 }, { "epoch": 141.77777777777777, "grad_norm": 2.092358954541851e-06, "learning_rate": 3.846357149972516e-05, "loss": 0.0, "num_input_tokens_seen": 3572464, "step": 12760 }, { "epoch": 141.83333333333334, "grad_norm": 2.7300841338728787e-06, "learning_rate": 3.8455298274135246e-05, "loss": 0.0, "num_input_tokens_seen": 3573824, "step": 12765 }, { "epoch": 141.88888888888889, "grad_norm": 4.813879513676511e-06, "learning_rate": 3.8447022973569254e-05, "loss": 0.0, "num_input_tokens_seen": 3575248, "step": 12770 }, { "epoch": 141.94444444444446, "grad_norm": 8.224851626437157e-07, "learning_rate": 3.843874559930332e-05, "loss": 0.0, "num_input_tokens_seen": 3576672, "step": 12775 }, { "epoch": 142.0, "grad_norm": 2.1955133888695855e-06, "learning_rate": 3.843046615261394e-05, "loss": 0.0, "num_input_tokens_seen": 3578144, "step": 12780 }, { "epoch": 142.05555555555554, "grad_norm": 2.496632305337698e-06, "learning_rate": 3.842218463477791e-05, "loss": 0.0, "num_input_tokens_seen": 3579552, "step": 12785 }, { "epoch": 142.11111111111111, "grad_norm": 3.446247774263611e-06, "learning_rate": 3.841390104707233e-05, "loss": 0.0, "num_input_tokens_seen": 3580944, "step": 12790 }, { "epoch": 142.16666666666666, "grad_norm": 2.418228859824012e-06, "learning_rate": 3.8405615390774643e-05, "loss": 0.0, "num_input_tokens_seen": 3582336, "step": 12795 }, { "epoch": 142.22222222222223, "grad_norm": 3.0106455142231425e-06, "learning_rate": 3.839732766716259e-05, "loss": 0.0, "num_input_tokens_seen": 3583696, "step": 12800 }, { "epoch": 142.22222222222223, "eval_loss": 1.198456048965454, "eval_runtime": 1.4113, "eval_samples_per_second": 28.342, "eval_steps_per_second": 14.171, "num_input_tokens_seen": 3583696, "step": 12800 }, { "epoch": 142.27777777777777, "grad_norm": 5.705594958271831e-06, "learning_rate": 3.838903787751425e-05, "loss": 0.0, "num_input_tokens_seen": 3585120, "step": 12805 }, { "epoch": 142.33333333333334, "grad_norm": 4.1679440982989036e-06, "learning_rate": 3.838074602310802e-05, "loss": 0.0, "num_input_tokens_seen": 3586496, "step": 12810 }, { "epoch": 142.38888888888889, "grad_norm": 9.82636174740037e-06, "learning_rate": 3.837245210522258e-05, "loss": 0.0, "num_input_tokens_seen": 3587888, "step": 12815 }, { "epoch": 142.44444444444446, "grad_norm": 1.2193396514703636e-06, "learning_rate": 3.8364156125136996e-05, "loss": 0.0, "num_input_tokens_seen": 3589280, "step": 12820 }, { "epoch": 142.5, "grad_norm": 3.191091991538997e-06, "learning_rate": 3.835585808413059e-05, "loss": 0.0, "num_input_tokens_seen": 3590704, "step": 12825 }, { "epoch": 142.55555555555554, "grad_norm": 3.638622047219542e-06, "learning_rate": 3.8347557983483024e-05, "loss": 0.0, "num_input_tokens_seen": 3592096, "step": 12830 }, { "epoch": 142.61111111111111, "grad_norm": 2.339374532311922e-06, "learning_rate": 3.833925582447428e-05, "loss": 0.0, "num_input_tokens_seen": 3593472, "step": 12835 }, { "epoch": 142.66666666666666, "grad_norm": 1.642785264266422e-06, "learning_rate": 3.8330951608384656e-05, "loss": 0.0, "num_input_tokens_seen": 3594880, "step": 12840 }, { "epoch": 142.72222222222223, "grad_norm": 1.373693748973892e-06, "learning_rate": 3.832264533649477e-05, "loss": 0.0, "num_input_tokens_seen": 3596304, "step": 12845 }, { "epoch": 142.77777777777777, "grad_norm": 1.6139190393005265e-06, "learning_rate": 3.8314337010085555e-05, "loss": 0.0, "num_input_tokens_seen": 3597696, "step": 12850 }, { "epoch": 142.83333333333334, "grad_norm": 2.3943612177390605e-06, "learning_rate": 3.830602663043824e-05, "loss": 0.0, "num_input_tokens_seen": 3599120, "step": 12855 }, { "epoch": 142.88888888888889, "grad_norm": 2.747794269453152e-06, "learning_rate": 3.8297714198834414e-05, "loss": 0.0, "num_input_tokens_seen": 3600528, "step": 12860 }, { "epoch": 142.94444444444446, "grad_norm": 2.4551957267249236e-06, "learning_rate": 3.828939971655595e-05, "loss": 0.0, "num_input_tokens_seen": 3601888, "step": 12865 }, { "epoch": 143.0, "grad_norm": 3.7454547054949217e-06, "learning_rate": 3.828108318488505e-05, "loss": 0.0, "num_input_tokens_seen": 3603312, "step": 12870 }, { "epoch": 143.05555555555554, "grad_norm": 1.042775056703249e-05, "learning_rate": 3.8272764605104216e-05, "loss": 0.0, "num_input_tokens_seen": 3604704, "step": 12875 }, { "epoch": 143.11111111111111, "grad_norm": 6.337650575005682e-06, "learning_rate": 3.826444397849628e-05, "loss": 0.0, "num_input_tokens_seen": 3606096, "step": 12880 }, { "epoch": 143.16666666666666, "grad_norm": 4.889503088634228e-06, "learning_rate": 3.825612130634439e-05, "loss": 0.0, "num_input_tokens_seen": 3607440, "step": 12885 }, { "epoch": 143.22222222222223, "grad_norm": 1.7002099639285007e-06, "learning_rate": 3.824779658993202e-05, "loss": 0.0, "num_input_tokens_seen": 3608848, "step": 12890 }, { "epoch": 143.27777777777777, "grad_norm": 1.5910231923044194e-06, "learning_rate": 3.823946983054292e-05, "loss": 0.0, "num_input_tokens_seen": 3610288, "step": 12895 }, { "epoch": 143.33333333333334, "grad_norm": 3.4579743442009203e-06, "learning_rate": 3.82311410294612e-05, "loss": 0.0, "num_input_tokens_seen": 3611664, "step": 12900 }, { "epoch": 143.38888888888889, "grad_norm": 5.8991631703975145e-06, "learning_rate": 3.822281018797127e-05, "loss": 0.0, "num_input_tokens_seen": 3613056, "step": 12905 }, { "epoch": 143.44444444444446, "grad_norm": 2.1334558368835133e-06, "learning_rate": 3.821447730735783e-05, "loss": 0.0, "num_input_tokens_seen": 3614480, "step": 12910 }, { "epoch": 143.5, "grad_norm": 2.876886128433398e-06, "learning_rate": 3.820614238890592e-05, "loss": 0.0, "num_input_tokens_seen": 3615888, "step": 12915 }, { "epoch": 143.55555555555554, "grad_norm": 4.32774395449087e-05, "learning_rate": 3.819780543390091e-05, "loss": 0.0, "num_input_tokens_seen": 3617312, "step": 12920 }, { "epoch": 143.61111111111111, "grad_norm": 6.246152111089032e-07, "learning_rate": 3.818946644362844e-05, "loss": 0.0, "num_input_tokens_seen": 3618752, "step": 12925 }, { "epoch": 143.66666666666666, "grad_norm": 9.029231478052679e-07, "learning_rate": 3.81811254193745e-05, "loss": 0.0, "num_input_tokens_seen": 3620176, "step": 12930 }, { "epoch": 143.72222222222223, "grad_norm": 1.5718818531240686e-06, "learning_rate": 3.8172782362425366e-05, "loss": 0.0, "num_input_tokens_seen": 3621568, "step": 12935 }, { "epoch": 143.77777777777777, "grad_norm": 3.5006414691451937e-06, "learning_rate": 3.816443727406765e-05, "loss": 0.0, "num_input_tokens_seen": 3623008, "step": 12940 }, { "epoch": 143.83333333333334, "grad_norm": 2.502430561435176e-06, "learning_rate": 3.815609015558829e-05, "loss": 0.0, "num_input_tokens_seen": 3624400, "step": 12945 }, { "epoch": 143.88888888888889, "grad_norm": 1.5369363381978474e-06, "learning_rate": 3.814774100827448e-05, "loss": 0.0, "num_input_tokens_seen": 3625776, "step": 12950 }, { "epoch": 143.94444444444446, "grad_norm": 1.0588656778054428e-06, "learning_rate": 3.813938983341379e-05, "loss": 0.0, "num_input_tokens_seen": 3627168, "step": 12955 }, { "epoch": 144.0, "grad_norm": 1.527320705463353e-06, "learning_rate": 3.813103663229407e-05, "loss": 0.0, "num_input_tokens_seen": 3628544, "step": 12960 }, { "epoch": 144.05555555555554, "grad_norm": 3.1032336664793547e-06, "learning_rate": 3.812268140620349e-05, "loss": 0.0, "num_input_tokens_seen": 3629968, "step": 12965 }, { "epoch": 144.11111111111111, "grad_norm": 3.350159659021301e-06, "learning_rate": 3.811432415643051e-05, "loss": 0.0, "num_input_tokens_seen": 3631344, "step": 12970 }, { "epoch": 144.16666666666666, "grad_norm": 2.3588324893353274e-06, "learning_rate": 3.8105964884263954e-05, "loss": 0.0, "num_input_tokens_seen": 3632736, "step": 12975 }, { "epoch": 144.22222222222223, "grad_norm": 3.65625260201341e-06, "learning_rate": 3.809760359099291e-05, "loss": 0.0, "num_input_tokens_seen": 3634144, "step": 12980 }, { "epoch": 144.27777777777777, "grad_norm": 2.2810811515228124e-06, "learning_rate": 3.8089240277906804e-05, "loss": 0.0, "num_input_tokens_seen": 3635520, "step": 12985 }, { "epoch": 144.33333333333334, "grad_norm": 2.0139045773248654e-06, "learning_rate": 3.808087494629535e-05, "loss": 0.0, "num_input_tokens_seen": 3636864, "step": 12990 }, { "epoch": 144.38888888888889, "grad_norm": 1.9654769403132377e-06, "learning_rate": 3.8072507597448595e-05, "loss": 0.0, "num_input_tokens_seen": 3638240, "step": 12995 }, { "epoch": 144.44444444444446, "grad_norm": 3.3734854696376715e-06, "learning_rate": 3.806413823265689e-05, "loss": 0.0, "num_input_tokens_seen": 3639680, "step": 13000 }, { "epoch": 144.44444444444446, "eval_loss": 1.2050942182540894, "eval_runtime": 1.3889, "eval_samples_per_second": 28.8, "eval_steps_per_second": 14.4, "num_input_tokens_seen": 3639680, "step": 13000 }, { "epoch": 144.5, "grad_norm": 1.1908795158888097e-06, "learning_rate": 3.805576685321089e-05, "loss": 0.0, "num_input_tokens_seen": 3641088, "step": 13005 }, { "epoch": 144.55555555555554, "grad_norm": 5.985294251331652e-07, "learning_rate": 3.804739346040158e-05, "loss": 0.0, "num_input_tokens_seen": 3642464, "step": 13010 }, { "epoch": 144.61111111111111, "grad_norm": 1.677530804045091e-06, "learning_rate": 3.8039018055520234e-05, "loss": 0.0, "num_input_tokens_seen": 3643824, "step": 13015 }, { "epoch": 144.66666666666666, "grad_norm": 2.508120587663143e-06, "learning_rate": 3.803064063985844e-05, "loss": 0.0, "num_input_tokens_seen": 3645232, "step": 13020 }, { "epoch": 144.72222222222223, "grad_norm": 1.5964233170961961e-06, "learning_rate": 3.802226121470811e-05, "loss": 0.0, "num_input_tokens_seen": 3646640, "step": 13025 }, { "epoch": 144.77777777777777, "grad_norm": 1.6622833527435432e-06, "learning_rate": 3.801387978136145e-05, "loss": 0.0, "num_input_tokens_seen": 3648032, "step": 13030 }, { "epoch": 144.83333333333334, "grad_norm": 1.0998116977134487e-06, "learning_rate": 3.800549634111099e-05, "loss": 0.0, "num_input_tokens_seen": 3649456, "step": 13035 }, { "epoch": 144.88888888888889, "grad_norm": 4.054492364957696e-06, "learning_rate": 3.799711089524955e-05, "loss": 0.0, "num_input_tokens_seen": 3650864, "step": 13040 }, { "epoch": 144.94444444444446, "grad_norm": 5.641469215333927e-06, "learning_rate": 3.7988723445070285e-05, "loss": 0.0, "num_input_tokens_seen": 3652272, "step": 13045 }, { "epoch": 145.0, "grad_norm": 3.2968605410133023e-06, "learning_rate": 3.798033399186663e-05, "loss": 0.0, "num_input_tokens_seen": 3653696, "step": 13050 }, { "epoch": 145.05555555555554, "grad_norm": 1.4739846392330946e-06, "learning_rate": 3.797194253693237e-05, "loss": 0.0, "num_input_tokens_seen": 3655120, "step": 13055 }, { "epoch": 145.11111111111111, "grad_norm": 2.8762133297277614e-06, "learning_rate": 3.796354908156153e-05, "loss": 0.0, "num_input_tokens_seen": 3656512, "step": 13060 }, { "epoch": 145.16666666666666, "grad_norm": 3.6284448015067028e-06, "learning_rate": 3.795515362704853e-05, "loss": 0.0, "num_input_tokens_seen": 3657872, "step": 13065 }, { "epoch": 145.22222222222223, "grad_norm": 3.618710934460978e-06, "learning_rate": 3.794675617468803e-05, "loss": 0.0, "num_input_tokens_seen": 3659280, "step": 13070 }, { "epoch": 145.27777777777777, "grad_norm": 1.1333244174238644e-06, "learning_rate": 3.793835672577503e-05, "loss": 0.0, "num_input_tokens_seen": 3660720, "step": 13075 }, { "epoch": 145.33333333333334, "grad_norm": 4.163373887422495e-06, "learning_rate": 3.7929955281604826e-05, "loss": 0.0, "num_input_tokens_seen": 3662144, "step": 13080 }, { "epoch": 145.38888888888889, "grad_norm": 2.236221007478889e-06, "learning_rate": 3.7921551843473036e-05, "loss": 0.0, "num_input_tokens_seen": 3663552, "step": 13085 }, { "epoch": 145.44444444444446, "grad_norm": 2.4561729787819786e-06, "learning_rate": 3.791314641267557e-05, "loss": 0.0, "num_input_tokens_seen": 3664944, "step": 13090 }, { "epoch": 145.5, "grad_norm": 3.323678583910805e-06, "learning_rate": 3.790473899050864e-05, "loss": 0.0, "num_input_tokens_seen": 3666368, "step": 13095 }, { "epoch": 145.55555555555554, "grad_norm": 2.5334268229926238e-06, "learning_rate": 3.7896329578268794e-05, "loss": 0.0, "num_input_tokens_seen": 3667776, "step": 13100 }, { "epoch": 145.61111111111111, "grad_norm": 2.202032192144543e-06, "learning_rate": 3.7887918177252855e-05, "loss": 0.0, "num_input_tokens_seen": 3669168, "step": 13105 }, { "epoch": 145.66666666666666, "grad_norm": 2.7105061235488392e-06, "learning_rate": 3.787950478875798e-05, "loss": 0.0, "num_input_tokens_seen": 3670576, "step": 13110 }, { "epoch": 145.72222222222223, "grad_norm": 1.4996578556747409e-06, "learning_rate": 3.787108941408162e-05, "loss": 0.0, "num_input_tokens_seen": 3671968, "step": 13115 }, { "epoch": 145.77777777777777, "grad_norm": 4.5613146539835725e-06, "learning_rate": 3.786267205452151e-05, "loss": 0.0, "num_input_tokens_seen": 3673344, "step": 13120 }, { "epoch": 145.83333333333334, "grad_norm": 3.3462622468505288e-06, "learning_rate": 3.785425271137573e-05, "loss": 0.0, "num_input_tokens_seen": 3674752, "step": 13125 }, { "epoch": 145.88888888888889, "grad_norm": 9.275198863178957e-06, "learning_rate": 3.7845831385942655e-05, "loss": 0.0, "num_input_tokens_seen": 3676160, "step": 13130 }, { "epoch": 145.94444444444446, "grad_norm": 2.175895815526019e-06, "learning_rate": 3.7837408079520944e-05, "loss": 0.0, "num_input_tokens_seen": 3677568, "step": 13135 }, { "epoch": 146.0, "grad_norm": 1.9711305867531337e-06, "learning_rate": 3.782898279340957e-05, "loss": 0.0, "num_input_tokens_seen": 3678992, "step": 13140 }, { "epoch": 146.05555555555554, "grad_norm": 1.6176830968106515e-06, "learning_rate": 3.782055552890784e-05, "loss": 0.0, "num_input_tokens_seen": 3680368, "step": 13145 }, { "epoch": 146.11111111111111, "grad_norm": 3.3630267353146337e-06, "learning_rate": 3.781212628731534e-05, "loss": 0.0, "num_input_tokens_seen": 3681776, "step": 13150 }, { "epoch": 146.16666666666666, "grad_norm": 1.8369280496699503e-06, "learning_rate": 3.7803695069931946e-05, "loss": 0.0, "num_input_tokens_seen": 3683184, "step": 13155 }, { "epoch": 146.22222222222223, "grad_norm": 2.1666960492439102e-06, "learning_rate": 3.779526187805789e-05, "loss": 0.0, "num_input_tokens_seen": 3684640, "step": 13160 }, { "epoch": 146.27777777777777, "grad_norm": 2.3658446934859967e-06, "learning_rate": 3.778682671299364e-05, "loss": 0.0, "num_input_tokens_seen": 3686000, "step": 13165 }, { "epoch": 146.33333333333334, "grad_norm": 1.6572063259445713e-06, "learning_rate": 3.777838957604003e-05, "loss": 0.0, "num_input_tokens_seen": 3687344, "step": 13170 }, { "epoch": 146.38888888888889, "grad_norm": 3.347803385622683e-06, "learning_rate": 3.776995046849816e-05, "loss": 0.0, "num_input_tokens_seen": 3688720, "step": 13175 }, { "epoch": 146.44444444444446, "grad_norm": 4.918546437693294e-06, "learning_rate": 3.776150939166945e-05, "loss": 0.0, "num_input_tokens_seen": 3690112, "step": 13180 }, { "epoch": 146.5, "grad_norm": 1.213798554999812e-06, "learning_rate": 3.775306634685562e-05, "loss": 0.0, "num_input_tokens_seen": 3691472, "step": 13185 }, { "epoch": 146.55555555555554, "grad_norm": 1.2355156286503188e-05, "learning_rate": 3.7744621335358696e-05, "loss": 0.0, "num_input_tokens_seen": 3692896, "step": 13190 }, { "epoch": 146.61111111111111, "grad_norm": 7.22979575584759e-06, "learning_rate": 3.7736174358481e-05, "loss": 0.0, "num_input_tokens_seen": 3694320, "step": 13195 }, { "epoch": 146.66666666666666, "grad_norm": 1.1819220162578858e-06, "learning_rate": 3.7727725417525175e-05, "loss": 0.0, "num_input_tokens_seen": 3695712, "step": 13200 }, { "epoch": 146.66666666666666, "eval_loss": 1.212993860244751, "eval_runtime": 1.3912, "eval_samples_per_second": 28.752, "eval_steps_per_second": 14.376, "num_input_tokens_seen": 3695712, "step": 13200 }, { "epoch": 146.72222222222223, "grad_norm": 3.056125478906324e-06, "learning_rate": 3.771927451379414e-05, "loss": 0.0, "num_input_tokens_seen": 3697088, "step": 13205 }, { "epoch": 146.77777777777777, "grad_norm": 2.9572590847237734e-06, "learning_rate": 3.7710821648591135e-05, "loss": 0.0, "num_input_tokens_seen": 3698544, "step": 13210 }, { "epoch": 146.83333333333334, "grad_norm": 1.8492070239517489e-06, "learning_rate": 3.7702366823219694e-05, "loss": 0.0, "num_input_tokens_seen": 3699904, "step": 13215 }, { "epoch": 146.88888888888889, "grad_norm": 1.3996559573570266e-06, "learning_rate": 3.769391003898366e-05, "loss": 0.0, "num_input_tokens_seen": 3701296, "step": 13220 }, { "epoch": 146.94444444444446, "grad_norm": 1.2663288089242997e-06, "learning_rate": 3.768545129718718e-05, "loss": 0.0, "num_input_tokens_seen": 3702656, "step": 13225 }, { "epoch": 147.0, "grad_norm": 1.636323418097163e-06, "learning_rate": 3.7676990599134686e-05, "loss": 0.0, "num_input_tokens_seen": 3704064, "step": 13230 }, { "epoch": 147.05555555555554, "grad_norm": 2.1998234842612874e-06, "learning_rate": 3.766852794613095e-05, "loss": 0.0, "num_input_tokens_seen": 3705440, "step": 13235 }, { "epoch": 147.11111111111111, "grad_norm": 3.5694386042450787e-06, "learning_rate": 3.766006333948099e-05, "loss": 0.0, "num_input_tokens_seen": 3706816, "step": 13240 }, { "epoch": 147.16666666666666, "grad_norm": 2.197129560954636e-06, "learning_rate": 3.765159678049017e-05, "loss": 0.0, "num_input_tokens_seen": 3708192, "step": 13245 }, { "epoch": 147.22222222222223, "grad_norm": 1.8050264998237253e-06, "learning_rate": 3.7643128270464134e-05, "loss": 0.0, "num_input_tokens_seen": 3709616, "step": 13250 }, { "epoch": 147.27777777777777, "grad_norm": 1.5827988590899622e-06, "learning_rate": 3.763465781070884e-05, "loss": 0.0, "num_input_tokens_seen": 3710992, "step": 13255 }, { "epoch": 147.33333333333334, "grad_norm": 1.6230486608037609e-06, "learning_rate": 3.762618540253052e-05, "loss": 0.0, "num_input_tokens_seen": 3712400, "step": 13260 }, { "epoch": 147.38888888888889, "grad_norm": 1.934236024681013e-06, "learning_rate": 3.761771104723576e-05, "loss": 0.0, "num_input_tokens_seen": 3713824, "step": 13265 }, { "epoch": 147.44444444444446, "grad_norm": 5.847321517649107e-06, "learning_rate": 3.7609234746131386e-05, "loss": 0.0, "num_input_tokens_seen": 3715264, "step": 13270 }, { "epoch": 147.5, "grad_norm": 3.945642674807459e-06, "learning_rate": 3.7600756500524556e-05, "loss": 0.0, "num_input_tokens_seen": 3716656, "step": 13275 }, { "epoch": 147.55555555555554, "grad_norm": 2.658230187080335e-06, "learning_rate": 3.759227631172271e-05, "loss": 0.0, "num_input_tokens_seen": 3718064, "step": 13280 }, { "epoch": 147.61111111111111, "grad_norm": 3.369846581335878e-06, "learning_rate": 3.758379418103363e-05, "loss": 0.0, "num_input_tokens_seen": 3719456, "step": 13285 }, { "epoch": 147.66666666666666, "grad_norm": 1.9983483525720658e-06, "learning_rate": 3.757531010976534e-05, "loss": 0.0, "num_input_tokens_seen": 3720848, "step": 13290 }, { "epoch": 147.72222222222223, "grad_norm": 3.7577842704195064e-06, "learning_rate": 3.75668240992262e-05, "loss": 0.0, "num_input_tokens_seen": 3722224, "step": 13295 }, { "epoch": 147.77777777777777, "grad_norm": 1.4337580296341912e-06, "learning_rate": 3.7558336150724865e-05, "loss": 0.0, "num_input_tokens_seen": 3723648, "step": 13300 }, { "epoch": 147.83333333333334, "grad_norm": 3.2089119486045092e-06, "learning_rate": 3.754984626557028e-05, "loss": 0.0, "num_input_tokens_seen": 3725024, "step": 13305 }, { "epoch": 147.88888888888889, "grad_norm": 1.969119239220163e-06, "learning_rate": 3.754135444507168e-05, "loss": 0.0, "num_input_tokens_seen": 3726416, "step": 13310 }, { "epoch": 147.94444444444446, "grad_norm": 2.3745992621115874e-06, "learning_rate": 3.753286069053863e-05, "loss": 0.0, "num_input_tokens_seen": 3727840, "step": 13315 }, { "epoch": 148.0, "grad_norm": 2.07993639378401e-06, "learning_rate": 3.7524365003280945e-05, "loss": 0.0, "num_input_tokens_seen": 3729264, "step": 13320 }, { "epoch": 148.05555555555554, "grad_norm": 2.784863909255364e-06, "learning_rate": 3.75158673846088e-05, "loss": 0.0, "num_input_tokens_seen": 3730656, "step": 13325 }, { "epoch": 148.11111111111111, "grad_norm": 1.0168570042878855e-05, "learning_rate": 3.750736783583262e-05, "loss": 0.0, "num_input_tokens_seen": 3732048, "step": 13330 }, { "epoch": 148.16666666666666, "grad_norm": 4.205721324979095e-06, "learning_rate": 3.7498866358263144e-05, "loss": 0.0, "num_input_tokens_seen": 3733472, "step": 13335 }, { "epoch": 148.22222222222223, "grad_norm": 1.7242163039554725e-06, "learning_rate": 3.74903629532114e-05, "loss": 0.0, "num_input_tokens_seen": 3734848, "step": 13340 }, { "epoch": 148.27777777777777, "grad_norm": 2.3693364710197784e-06, "learning_rate": 3.748185762198873e-05, "loss": 0.0, "num_input_tokens_seen": 3736288, "step": 13345 }, { "epoch": 148.33333333333334, "grad_norm": 2.5569877379894024e-06, "learning_rate": 3.747335036590676e-05, "loss": 0.0, "num_input_tokens_seen": 3737712, "step": 13350 }, { "epoch": 148.38888888888889, "grad_norm": 2.3675656848354265e-06, "learning_rate": 3.7464841186277405e-05, "loss": 0.0, "num_input_tokens_seen": 3739120, "step": 13355 }, { "epoch": 148.44444444444446, "grad_norm": 2.9488039672287414e-06, "learning_rate": 3.7456330084412896e-05, "loss": 0.0, "num_input_tokens_seen": 3740560, "step": 13360 }, { "epoch": 148.5, "grad_norm": 1.8077138292937889e-06, "learning_rate": 3.744781706162576e-05, "loss": 0.0, "num_input_tokens_seen": 3741968, "step": 13365 }, { "epoch": 148.55555555555554, "grad_norm": 5.866150786459912e-06, "learning_rate": 3.743930211922879e-05, "loss": 0.0, "num_input_tokens_seen": 3743312, "step": 13370 }, { "epoch": 148.61111111111111, "grad_norm": 2.3474240151699632e-06, "learning_rate": 3.743078525853513e-05, "loss": 0.0, "num_input_tokens_seen": 3744688, "step": 13375 }, { "epoch": 148.66666666666666, "grad_norm": 1.6245375036305632e-06, "learning_rate": 3.7422266480858154e-05, "loss": 0.0, "num_input_tokens_seen": 3746080, "step": 13380 }, { "epoch": 148.72222222222223, "grad_norm": 1.4172914006849169e-06, "learning_rate": 3.741374578751158e-05, "loss": 0.0, "num_input_tokens_seen": 3747504, "step": 13385 }, { "epoch": 148.77777777777777, "grad_norm": 3.5293144264869625e-06, "learning_rate": 3.740522317980941e-05, "loss": 0.0, "num_input_tokens_seen": 3748928, "step": 13390 }, { "epoch": 148.83333333333334, "grad_norm": 5.029149633628549e-06, "learning_rate": 3.739669865906593e-05, "loss": 0.0, "num_input_tokens_seen": 3750320, "step": 13395 }, { "epoch": 148.88888888888889, "grad_norm": 1.17115780540189e-06, "learning_rate": 3.738817222659573e-05, "loss": 0.0, "num_input_tokens_seen": 3751728, "step": 13400 }, { "epoch": 148.88888888888889, "eval_loss": 1.1997250318527222, "eval_runtime": 1.6649, "eval_samples_per_second": 24.025, "eval_steps_per_second": 12.013, "num_input_tokens_seen": 3751728, "step": 13400 }, { "epoch": 148.94444444444446, "grad_norm": 4.506709046836477e-06, "learning_rate": 3.73796438837137e-05, "loss": 0.0, "num_input_tokens_seen": 3753136, "step": 13405 }, { "epoch": 149.0, "grad_norm": 2.4403216229984537e-06, "learning_rate": 3.7371113631735e-05, "loss": 0.0, "num_input_tokens_seen": 3754544, "step": 13410 }, { "epoch": 149.05555555555554, "grad_norm": 7.85964493843494e-06, "learning_rate": 3.736258147197512e-05, "loss": 0.0, "num_input_tokens_seen": 3755920, "step": 13415 }, { "epoch": 149.11111111111111, "grad_norm": 2.3134389266488142e-06, "learning_rate": 3.735404740574981e-05, "loss": 0.0, "num_input_tokens_seen": 3757312, "step": 13420 }, { "epoch": 149.16666666666666, "grad_norm": 1.7575313222550903e-06, "learning_rate": 3.7345511434375145e-05, "loss": 0.0, "num_input_tokens_seen": 3758720, "step": 13425 }, { "epoch": 149.22222222222223, "grad_norm": 3.4143185985158198e-06, "learning_rate": 3.733697355916748e-05, "loss": 0.0, "num_input_tokens_seen": 3760144, "step": 13430 }, { "epoch": 149.27777777777777, "grad_norm": 1.4633259297625045e-06, "learning_rate": 3.732843378144345e-05, "loss": 0.0, "num_input_tokens_seen": 3761504, "step": 13435 }, { "epoch": 149.33333333333334, "grad_norm": 6.4065930018841755e-06, "learning_rate": 3.7319892102519995e-05, "loss": 0.0, "num_input_tokens_seen": 3762912, "step": 13440 }, { "epoch": 149.38888888888889, "grad_norm": 2.1415285118564498e-06, "learning_rate": 3.731134852371436e-05, "loss": 0.0, "num_input_tokens_seen": 3764336, "step": 13445 }, { "epoch": 149.44444444444446, "grad_norm": 2.3694442461419385e-06, "learning_rate": 3.730280304634408e-05, "loss": 0.0, "num_input_tokens_seen": 3765744, "step": 13450 }, { "epoch": 149.5, "grad_norm": 1.652064952395449e-06, "learning_rate": 3.729425567172696e-05, "loss": 0.0, "num_input_tokens_seen": 3767104, "step": 13455 }, { "epoch": 149.55555555555554, "grad_norm": 3.05037792713847e-06, "learning_rate": 3.728570640118111e-05, "loss": 0.0, "num_input_tokens_seen": 3768512, "step": 13460 }, { "epoch": 149.61111111111111, "grad_norm": 2.7571443297347287e-06, "learning_rate": 3.727715523602494e-05, "loss": 0.0, "num_input_tokens_seen": 3769920, "step": 13465 }, { "epoch": 149.66666666666666, "grad_norm": 8.081176019913983e-06, "learning_rate": 3.726860217757715e-05, "loss": 0.0, "num_input_tokens_seen": 3771328, "step": 13470 }, { "epoch": 149.72222222222223, "grad_norm": 2.8819238195865182e-06, "learning_rate": 3.726004722715673e-05, "loss": 0.0, "num_input_tokens_seen": 3772720, "step": 13475 }, { "epoch": 149.77777777777777, "grad_norm": 3.864577138301684e-06, "learning_rate": 3.725149038608296e-05, "loss": 0.0, "num_input_tokens_seen": 3774144, "step": 13480 }, { "epoch": 149.83333333333334, "grad_norm": 2.9324166916921968e-06, "learning_rate": 3.7242931655675404e-05, "loss": 0.0, "num_input_tokens_seen": 3775536, "step": 13485 }, { "epoch": 149.88888888888889, "grad_norm": 2.2999442990112584e-06, "learning_rate": 3.7234371037253937e-05, "loss": 0.0, "num_input_tokens_seen": 3776912, "step": 13490 }, { "epoch": 149.94444444444446, "grad_norm": 4.1860125747916754e-06, "learning_rate": 3.7225808532138705e-05, "loss": 0.0, "num_input_tokens_seen": 3778304, "step": 13495 }, { "epoch": 150.0, "grad_norm": 2.328315986233065e-06, "learning_rate": 3.721724414165016e-05, "loss": 0.0, "num_input_tokens_seen": 3779696, "step": 13500 }, { "epoch": 150.05555555555554, "grad_norm": 2.198052015955909e-06, "learning_rate": 3.720867786710904e-05, "loss": 0.0, "num_input_tokens_seen": 3781056, "step": 13505 }, { "epoch": 150.11111111111111, "grad_norm": 1.6716583104425808e-06, "learning_rate": 3.7200109709836366e-05, "loss": 0.0, "num_input_tokens_seen": 3782448, "step": 13510 }, { "epoch": 150.16666666666666, "grad_norm": 3.7775000691908645e-06, "learning_rate": 3.7191539671153465e-05, "loss": 0.0, "num_input_tokens_seen": 3783824, "step": 13515 }, { "epoch": 150.22222222222223, "grad_norm": 3.403682967473287e-06, "learning_rate": 3.718296775238193e-05, "loss": 0.0, "num_input_tokens_seen": 3785216, "step": 13520 }, { "epoch": 150.27777777777777, "grad_norm": 3.350645329192048e-06, "learning_rate": 3.7174393954843675e-05, "loss": 0.0, "num_input_tokens_seen": 3786656, "step": 13525 }, { "epoch": 150.33333333333334, "grad_norm": 1.1340018318151124e-05, "learning_rate": 3.716581827986087e-05, "loss": 0.0, "num_input_tokens_seen": 3788064, "step": 13530 }, { "epoch": 150.38888888888889, "grad_norm": 3.5724067402043147e-06, "learning_rate": 3.7157240728756004e-05, "loss": 0.0, "num_input_tokens_seen": 3789488, "step": 13535 }, { "epoch": 150.44444444444446, "grad_norm": 2.5092679152294295e-06, "learning_rate": 3.714866130285184e-05, "loss": 0.0, "num_input_tokens_seen": 3790848, "step": 13540 }, { "epoch": 150.5, "grad_norm": 1.77148831426166e-06, "learning_rate": 3.714008000347143e-05, "loss": 0.0, "num_input_tokens_seen": 3792224, "step": 13545 }, { "epoch": 150.55555555555554, "grad_norm": 4.552199243335053e-06, "learning_rate": 3.7131496831938126e-05, "loss": 0.0, "num_input_tokens_seen": 3793664, "step": 13550 }, { "epoch": 150.61111111111111, "grad_norm": 2.3392276489175856e-06, "learning_rate": 3.7122911789575565e-05, "loss": 0.0, "num_input_tokens_seen": 3795072, "step": 13555 }, { "epoch": 150.66666666666666, "grad_norm": 1.3801734439766733e-06, "learning_rate": 3.711432487770765e-05, "loss": 0.0, "num_input_tokens_seen": 3796464, "step": 13560 }, { "epoch": 150.72222222222223, "grad_norm": 2.127520929207094e-06, "learning_rate": 3.710573609765861e-05, "loss": 0.0, "num_input_tokens_seen": 3797904, "step": 13565 }, { "epoch": 150.77777777777777, "grad_norm": 2.8272104373172624e-06, "learning_rate": 3.709714545075292e-05, "loss": 0.0, "num_input_tokens_seen": 3799312, "step": 13570 }, { "epoch": 150.83333333333334, "grad_norm": 2.61975606008491e-06, "learning_rate": 3.708855293831538e-05, "loss": 0.0, "num_input_tokens_seen": 3800736, "step": 13575 }, { "epoch": 150.88888888888889, "grad_norm": 2.170419747926644e-06, "learning_rate": 3.707995856167107e-05, "loss": 0.0, "num_input_tokens_seen": 3802144, "step": 13580 }, { "epoch": 150.94444444444446, "grad_norm": 3.4967292776855174e-06, "learning_rate": 3.707136232214534e-05, "loss": 0.0, "num_input_tokens_seen": 3803520, "step": 13585 }, { "epoch": 151.0, "grad_norm": 2.232741280749906e-06, "learning_rate": 3.7062764221063844e-05, "loss": 0.0, "num_input_tokens_seen": 3804896, "step": 13590 }, { "epoch": 151.05555555555554, "grad_norm": 1.2299339005039656e-06, "learning_rate": 3.705416425975252e-05, "loss": 0.0, "num_input_tokens_seen": 3806304, "step": 13595 }, { "epoch": 151.11111111111111, "grad_norm": 2.3251284346770262e-06, "learning_rate": 3.704556243953758e-05, "loss": 0.0, "num_input_tokens_seen": 3807744, "step": 13600 }, { "epoch": 151.11111111111111, "eval_loss": 1.2459125518798828, "eval_runtime": 1.3838, "eval_samples_per_second": 28.906, "eval_steps_per_second": 14.453, "num_input_tokens_seen": 3807744, "step": 13600 }, { "epoch": 151.16666666666666, "grad_norm": 1.9304447960166726e-06, "learning_rate": 3.7036958761745535e-05, "loss": 0.0, "num_input_tokens_seen": 3809104, "step": 13605 }, { "epoch": 151.22222222222223, "grad_norm": 3.3611122489674017e-06, "learning_rate": 3.702835322770318e-05, "loss": 0.0, "num_input_tokens_seen": 3810480, "step": 13610 }, { "epoch": 151.27777777777777, "grad_norm": 1.8609968037708313e-06, "learning_rate": 3.701974583873761e-05, "loss": 0.0, "num_input_tokens_seen": 3811920, "step": 13615 }, { "epoch": 151.33333333333334, "grad_norm": 1.8410304392091348e-06, "learning_rate": 3.701113659617618e-05, "loss": 0.0, "num_input_tokens_seen": 3813328, "step": 13620 }, { "epoch": 151.38888888888889, "grad_norm": 2.3454763322661165e-06, "learning_rate": 3.7002525501346535e-05, "loss": 0.0, "num_input_tokens_seen": 3814752, "step": 13625 }, { "epoch": 151.44444444444446, "grad_norm": 3.0696869544044603e-06, "learning_rate": 3.699391255557664e-05, "loss": 0.0, "num_input_tokens_seen": 3816096, "step": 13630 }, { "epoch": 151.5, "grad_norm": 2.4032697183429264e-06, "learning_rate": 3.69852977601947e-05, "loss": 0.0, "num_input_tokens_seen": 3817488, "step": 13635 }, { "epoch": 151.55555555555554, "grad_norm": 1.288773319174652e-06, "learning_rate": 3.697668111652922e-05, "loss": 0.0, "num_input_tokens_seen": 3818880, "step": 13640 }, { "epoch": 151.61111111111111, "grad_norm": 2.161425982194487e-06, "learning_rate": 3.6968062625909005e-05, "loss": 0.0, "num_input_tokens_seen": 3820272, "step": 13645 }, { "epoch": 151.66666666666666, "grad_norm": 2.6994594009011053e-06, "learning_rate": 3.6959442289663135e-05, "loss": 0.0, "num_input_tokens_seen": 3821664, "step": 13650 }, { "epoch": 151.72222222222223, "grad_norm": 4.57518444818561e-06, "learning_rate": 3.695082010912098e-05, "loss": 0.0, "num_input_tokens_seen": 3823072, "step": 13655 }, { "epoch": 151.77777777777777, "grad_norm": 6.890884833410382e-06, "learning_rate": 3.694219608561217e-05, "loss": 0.0, "num_input_tokens_seen": 3824464, "step": 13660 }, { "epoch": 151.83333333333334, "grad_norm": 1.0169394045078661e-05, "learning_rate": 3.693357022046665e-05, "loss": 0.0, "num_input_tokens_seen": 3825888, "step": 13665 }, { "epoch": 151.88888888888889, "grad_norm": 2.154969934053952e-06, "learning_rate": 3.6924942515014644e-05, "loss": 0.0, "num_input_tokens_seen": 3827280, "step": 13670 }, { "epoch": 151.94444444444446, "grad_norm": 2.576656697783619e-06, "learning_rate": 3.691631297058664e-05, "loss": 0.0, "num_input_tokens_seen": 3828720, "step": 13675 }, { "epoch": 152.0, "grad_norm": 2.8492659112089314e-06, "learning_rate": 3.6907681588513424e-05, "loss": 0.0, "num_input_tokens_seen": 3830096, "step": 13680 }, { "epoch": 152.05555555555554, "grad_norm": 2.4279631816170877e-06, "learning_rate": 3.689904837012606e-05, "loss": 0.0, "num_input_tokens_seen": 3831520, "step": 13685 }, { "epoch": 152.11111111111111, "grad_norm": 1.7944098544830922e-06, "learning_rate": 3.689041331675591e-05, "loss": 0.0, "num_input_tokens_seen": 3832864, "step": 13690 }, { "epoch": 152.16666666666666, "grad_norm": 2.283122557855677e-05, "learning_rate": 3.688177642973461e-05, "loss": 0.0, "num_input_tokens_seen": 3834224, "step": 13695 }, { "epoch": 152.22222222222223, "grad_norm": 4.511471615842311e-06, "learning_rate": 3.687313771039406e-05, "loss": 0.0, "num_input_tokens_seen": 3835680, "step": 13700 }, { "epoch": 152.27777777777777, "grad_norm": 1.952298816831899e-06, "learning_rate": 3.686449716006647e-05, "loss": 0.0, "num_input_tokens_seen": 3837072, "step": 13705 }, { "epoch": 152.33333333333334, "grad_norm": 1.5309814216379891e-06, "learning_rate": 3.685585478008432e-05, "loss": 0.0, "num_input_tokens_seen": 3838480, "step": 13710 }, { "epoch": 152.38888888888889, "grad_norm": 3.4268996387254447e-06, "learning_rate": 3.6847210571780364e-05, "loss": 0.0, "num_input_tokens_seen": 3839840, "step": 13715 }, { "epoch": 152.44444444444446, "grad_norm": 2.225110620202031e-06, "learning_rate": 3.683856453648767e-05, "loss": 0.0, "num_input_tokens_seen": 3841200, "step": 13720 }, { "epoch": 152.5, "grad_norm": 1.7730686749928282e-06, "learning_rate": 3.682991667553954e-05, "loss": 0.0, "num_input_tokens_seen": 3842576, "step": 13725 }, { "epoch": 152.55555555555554, "grad_norm": 4.125990471948171e-06, "learning_rate": 3.6821266990269606e-05, "loss": 0.0, "num_input_tokens_seen": 3843968, "step": 13730 }, { "epoch": 152.61111111111111, "grad_norm": 1.995627144424361e-06, "learning_rate": 3.681261548201174e-05, "loss": 0.0, "num_input_tokens_seen": 3845376, "step": 13735 }, { "epoch": 152.66666666666666, "grad_norm": 2.171391770389164e-06, "learning_rate": 3.6803962152100125e-05, "loss": 0.0, "num_input_tokens_seen": 3846816, "step": 13740 }, { "epoch": 152.72222222222223, "grad_norm": 3.130637196591124e-06, "learning_rate": 3.67953070018692e-05, "loss": 0.0, "num_input_tokens_seen": 3848208, "step": 13745 }, { "epoch": 152.77777777777777, "grad_norm": 5.077749847259838e-06, "learning_rate": 3.678665003265371e-05, "loss": 0.0, "num_input_tokens_seen": 3849600, "step": 13750 }, { "epoch": 152.83333333333334, "grad_norm": 2.9159214136598166e-06, "learning_rate": 3.677799124578867e-05, "loss": 0.0, "num_input_tokens_seen": 3851040, "step": 13755 }, { "epoch": 152.88888888888889, "grad_norm": 1.6299437675115769e-06, "learning_rate": 3.676933064260937e-05, "loss": 0.0, "num_input_tokens_seen": 3852432, "step": 13760 }, { "epoch": 152.94444444444446, "grad_norm": 1.930124881255324e-06, "learning_rate": 3.6760668224451365e-05, "loss": 0.0, "num_input_tokens_seen": 3853872, "step": 13765 }, { "epoch": 153.0, "grad_norm": 1.8787262661135173e-06, "learning_rate": 3.675200399265054e-05, "loss": 0.0, "num_input_tokens_seen": 3855264, "step": 13770 }, { "epoch": 153.05555555555554, "grad_norm": 2.2424068447435275e-05, "learning_rate": 3.6743337948543014e-05, "loss": 0.0, "num_input_tokens_seen": 3856656, "step": 13775 }, { "epoch": 153.11111111111111, "grad_norm": 3.0209944270609412e-06, "learning_rate": 3.6734670093465204e-05, "loss": 0.0, "num_input_tokens_seen": 3858048, "step": 13780 }, { "epoch": 153.16666666666666, "grad_norm": 1.64191214935272e-06, "learning_rate": 3.672600042875379e-05, "loss": 0.0, "num_input_tokens_seen": 3859456, "step": 13785 }, { "epoch": 153.22222222222223, "grad_norm": 1.3563064840127481e-06, "learning_rate": 3.671732895574575e-05, "loss": 0.0, "num_input_tokens_seen": 3860848, "step": 13790 }, { "epoch": 153.27777777777777, "grad_norm": 1.8198330735685886e-06, "learning_rate": 3.670865567577834e-05, "loss": 0.0, "num_input_tokens_seen": 3862304, "step": 13795 }, { "epoch": 153.33333333333334, "grad_norm": 1.8597764892547275e-06, "learning_rate": 3.669998059018909e-05, "loss": 0.0, "num_input_tokens_seen": 3863664, "step": 13800 }, { "epoch": 153.33333333333334, "eval_loss": 1.2528512477874756, "eval_runtime": 1.3907, "eval_samples_per_second": 28.763, "eval_steps_per_second": 14.381, "num_input_tokens_seen": 3863664, "step": 13800 }, { "epoch": 153.38888888888889, "grad_norm": 3.500412731227698e-06, "learning_rate": 3.6691303700315796e-05, "loss": 0.0, "num_input_tokens_seen": 3865024, "step": 13805 }, { "epoch": 153.44444444444446, "grad_norm": 1.0152898539672606e-06, "learning_rate": 3.668262500749655e-05, "loss": 0.0, "num_input_tokens_seen": 3866448, "step": 13810 }, { "epoch": 153.5, "grad_norm": 2.266791625515907e-06, "learning_rate": 3.667394451306971e-05, "loss": 0.0, "num_input_tokens_seen": 3867856, "step": 13815 }, { "epoch": 153.55555555555554, "grad_norm": 2.268410071337712e-06, "learning_rate": 3.666526221837393e-05, "loss": 0.0, "num_input_tokens_seen": 3869232, "step": 13820 }, { "epoch": 153.61111111111111, "grad_norm": 2.269806373078609e-06, "learning_rate": 3.665657812474812e-05, "loss": 0.0, "num_input_tokens_seen": 3870640, "step": 13825 }, { "epoch": 153.66666666666666, "grad_norm": 1.5615653410350205e-06, "learning_rate": 3.664789223353147e-05, "loss": 0.0, "num_input_tokens_seen": 3872032, "step": 13830 }, { "epoch": 153.72222222222223, "grad_norm": 3.915302841051016e-06, "learning_rate": 3.663920454606347e-05, "loss": 0.0, "num_input_tokens_seen": 3873424, "step": 13835 }, { "epoch": 153.77777777777777, "grad_norm": 4.066358087584376e-06, "learning_rate": 3.6630515063683856e-05, "loss": 0.0, "num_input_tokens_seen": 3874800, "step": 13840 }, { "epoch": 153.83333333333334, "grad_norm": 1.0801001053550863e-06, "learning_rate": 3.662182378773267e-05, "loss": 0.0, "num_input_tokens_seen": 3876176, "step": 13845 }, { "epoch": 153.88888888888889, "grad_norm": 2.494142563591595e-06, "learning_rate": 3.66131307195502e-05, "loss": 0.0, "num_input_tokens_seen": 3877584, "step": 13850 }, { "epoch": 153.94444444444446, "grad_norm": 3.888178980560042e-06, "learning_rate": 3.6604435860477034e-05, "loss": 0.0, "num_input_tokens_seen": 3878944, "step": 13855 }, { "epoch": 154.0, "grad_norm": 3.947841832996346e-06, "learning_rate": 3.6595739211854025e-05, "loss": 0.0, "num_input_tokens_seen": 3880400, "step": 13860 }, { "epoch": 154.05555555555554, "grad_norm": 1.3627824955619872e-05, "learning_rate": 3.658704077502231e-05, "loss": 0.0, "num_input_tokens_seen": 3881808, "step": 13865 }, { "epoch": 154.11111111111111, "grad_norm": 6.544869393110275e-06, "learning_rate": 3.65783405513233e-05, "loss": 0.0, "num_input_tokens_seen": 3883264, "step": 13870 }, { "epoch": 154.16666666666666, "grad_norm": 1.7419780533600715e-06, "learning_rate": 3.656963854209867e-05, "loss": 0.0, "num_input_tokens_seen": 3884608, "step": 13875 }, { "epoch": 154.22222222222223, "grad_norm": 1.6123376553878188e-05, "learning_rate": 3.656093474869038e-05, "loss": 0.0, "num_input_tokens_seen": 3885952, "step": 13880 }, { "epoch": 154.27777777777777, "grad_norm": 8.73246449373255e-07, "learning_rate": 3.655222917244068e-05, "loss": 0.0, "num_input_tokens_seen": 3887360, "step": 13885 }, { "epoch": 154.33333333333334, "grad_norm": 1.1510971944517223e-06, "learning_rate": 3.6543521814692054e-05, "loss": 0.0, "num_input_tokens_seen": 3888736, "step": 13890 }, { "epoch": 154.38888888888889, "grad_norm": 8.705511049811321e-07, "learning_rate": 3.653481267678731e-05, "loss": 0.0, "num_input_tokens_seen": 3890160, "step": 13895 }, { "epoch": 154.44444444444446, "grad_norm": 1.2037143051202293e-06, "learning_rate": 3.652610176006949e-05, "loss": 0.0, "num_input_tokens_seen": 3891520, "step": 13900 }, { "epoch": 154.5, "grad_norm": 2.45388059738616e-06, "learning_rate": 3.6517389065881925e-05, "loss": 0.0, "num_input_tokens_seen": 3892912, "step": 13905 }, { "epoch": 154.55555555555554, "grad_norm": 3.0469805096799973e-06, "learning_rate": 3.650867459556824e-05, "loss": 0.0, "num_input_tokens_seen": 3894320, "step": 13910 }, { "epoch": 154.61111111111111, "grad_norm": 1.63292202159937e-06, "learning_rate": 3.64999583504723e-05, "loss": 0.0, "num_input_tokens_seen": 3895728, "step": 13915 }, { "epoch": 154.66666666666666, "grad_norm": 1.97986491912161e-06, "learning_rate": 3.649124033193827e-05, "loss": 0.0, "num_input_tokens_seen": 3897136, "step": 13920 }, { "epoch": 154.72222222222223, "grad_norm": 1.4468040490100975e-06, "learning_rate": 3.648252054131057e-05, "loss": 0.0, "num_input_tokens_seen": 3898560, "step": 13925 }, { "epoch": 154.77777777777777, "grad_norm": 2.198186848545447e-06, "learning_rate": 3.647379897993391e-05, "loss": 0.0, "num_input_tokens_seen": 3899968, "step": 13930 }, { "epoch": 154.83333333333334, "grad_norm": 1.2814568890462397e-06, "learning_rate": 3.646507564915325e-05, "loss": 0.0, "num_input_tokens_seen": 3901360, "step": 13935 }, { "epoch": 154.88888888888889, "grad_norm": 2.06677827918611e-06, "learning_rate": 3.645635055031385e-05, "loss": 0.0, "num_input_tokens_seen": 3902752, "step": 13940 }, { "epoch": 154.94444444444446, "grad_norm": 9.431649345970072e-07, "learning_rate": 3.6447623684761224e-05, "loss": 0.0, "num_input_tokens_seen": 3904128, "step": 13945 }, { "epoch": 155.0, "grad_norm": 1.3136483403286547e-06, "learning_rate": 3.643889505384117e-05, "loss": 0.0, "num_input_tokens_seen": 3905552, "step": 13950 }, { "epoch": 155.05555555555554, "grad_norm": 1.565064121678006e-05, "learning_rate": 3.6430164658899744e-05, "loss": 0.0, "num_input_tokens_seen": 3906928, "step": 13955 }, { "epoch": 155.11111111111111, "grad_norm": 9.525768973617232e-07, "learning_rate": 3.642143250128329e-05, "loss": 0.0, "num_input_tokens_seen": 3908256, "step": 13960 }, { "epoch": 155.16666666666666, "grad_norm": 3.6416504372027703e-06, "learning_rate": 3.641269858233841e-05, "loss": 0.0, "num_input_tokens_seen": 3909648, "step": 13965 }, { "epoch": 155.22222222222223, "grad_norm": 1.873081373560126e-06, "learning_rate": 3.640396290341199e-05, "loss": 0.0, "num_input_tokens_seen": 3911088, "step": 13970 }, { "epoch": 155.27777777777777, "grad_norm": 2.7562100513023324e-06, "learning_rate": 3.639522546585118e-05, "loss": 0.0, "num_input_tokens_seen": 3912528, "step": 13975 }, { "epoch": 155.33333333333334, "grad_norm": 3.1439240046893246e-06, "learning_rate": 3.6386486271003404e-05, "loss": 0.0, "num_input_tokens_seen": 3913952, "step": 13980 }, { "epoch": 155.38888888888889, "grad_norm": 1.4782834796278621e-06, "learning_rate": 3.6377745320216346e-05, "loss": 0.0, "num_input_tokens_seen": 3915408, "step": 13985 }, { "epoch": 155.44444444444446, "grad_norm": 3.185829882568214e-06, "learning_rate": 3.636900261483798e-05, "loss": 0.0, "num_input_tokens_seen": 3916768, "step": 13990 }, { "epoch": 155.5, "grad_norm": 1.424305310138152e-06, "learning_rate": 3.636025815621654e-05, "loss": 0.0, "num_input_tokens_seen": 3918160, "step": 13995 }, { "epoch": 155.55555555555554, "grad_norm": 6.428896085708402e-06, "learning_rate": 3.635151194570054e-05, "loss": 0.0, "num_input_tokens_seen": 3919584, "step": 14000 }, { "epoch": 155.55555555555554, "eval_loss": 1.240476369857788, "eval_runtime": 1.3894, "eval_samples_per_second": 28.789, "eval_steps_per_second": 14.394, "num_input_tokens_seen": 3919584, "step": 14000 }, { "epoch": 155.61111111111111, "grad_norm": 3.3747342058632057e-06, "learning_rate": 3.634276398463873e-05, "loss": 0.0, "num_input_tokens_seen": 3920992, "step": 14005 }, { "epoch": 155.66666666666666, "grad_norm": 1.359232555842027e-06, "learning_rate": 3.633401427438018e-05, "loss": 0.0, "num_input_tokens_seen": 3922384, "step": 14010 }, { "epoch": 155.72222222222223, "grad_norm": 2.2353594886226347e-06, "learning_rate": 3.63252628162742e-05, "loss": 0.0, "num_input_tokens_seen": 3923776, "step": 14015 }, { "epoch": 155.77777777777777, "grad_norm": 1.0545905979597592e-06, "learning_rate": 3.6316509611670364e-05, "loss": 0.0, "num_input_tokens_seen": 3925184, "step": 14020 }, { "epoch": 155.83333333333334, "grad_norm": 1.8312360907657421e-06, "learning_rate": 3.630775466191854e-05, "loss": 0.0, "num_input_tokens_seen": 3926560, "step": 14025 }, { "epoch": 155.88888888888889, "grad_norm": 1.21890172977146e-06, "learning_rate": 3.629899796836884e-05, "loss": 0.0, "num_input_tokens_seen": 3927936, "step": 14030 }, { "epoch": 155.94444444444446, "grad_norm": 1.923041509144241e-06, "learning_rate": 3.6290239532371666e-05, "loss": 0.0, "num_input_tokens_seen": 3929344, "step": 14035 }, { "epoch": 156.0, "grad_norm": 1.4399131487152772e-06, "learning_rate": 3.628147935527767e-05, "loss": 0.0, "num_input_tokens_seen": 3930784, "step": 14040 }, { "epoch": 156.05555555555554, "grad_norm": 2.3876616523921257e-06, "learning_rate": 3.627271743843779e-05, "loss": 0.0, "num_input_tokens_seen": 3932176, "step": 14045 }, { "epoch": 156.11111111111111, "grad_norm": 1.7127972569142003e-06, "learning_rate": 3.626395378320321e-05, "loss": 0.0, "num_input_tokens_seen": 3933552, "step": 14050 }, { "epoch": 156.16666666666666, "grad_norm": 1.4667774621557328e-06, "learning_rate": 3.625518839092541e-05, "loss": 0.0, "num_input_tokens_seen": 3934912, "step": 14055 }, { "epoch": 156.22222222222223, "grad_norm": 1.3152183555575903e-06, "learning_rate": 3.624642126295612e-05, "loss": 0.0, "num_input_tokens_seen": 3936272, "step": 14060 }, { "epoch": 156.27777777777777, "grad_norm": 2.8578472210938344e-06, "learning_rate": 3.6237652400647345e-05, "loss": 0.0, "num_input_tokens_seen": 3937712, "step": 14065 }, { "epoch": 156.33333333333334, "grad_norm": 1.2920660310555832e-06, "learning_rate": 3.622888180535134e-05, "loss": 0.0, "num_input_tokens_seen": 3939152, "step": 14070 }, { "epoch": 156.38888888888889, "grad_norm": 2.1588696199614787e-06, "learning_rate": 3.6220109478420655e-05, "loss": 0.0, "num_input_tokens_seen": 3940544, "step": 14075 }, { "epoch": 156.44444444444446, "grad_norm": 4.946088665747084e-06, "learning_rate": 3.6211335421208084e-05, "loss": 0.0, "num_input_tokens_seen": 3941952, "step": 14080 }, { "epoch": 156.5, "grad_norm": 1.9811388938251184e-06, "learning_rate": 3.62025596350667e-05, "loss": 0.0, "num_input_tokens_seen": 3943376, "step": 14085 }, { "epoch": 156.55555555555554, "grad_norm": 3.4631198104762007e-06, "learning_rate": 3.619378212134984e-05, "loss": 0.0, "num_input_tokens_seen": 3944768, "step": 14090 }, { "epoch": 156.61111111111111, "grad_norm": 1.2842422165704193e-06, "learning_rate": 3.618500288141111e-05, "loss": 0.0, "num_input_tokens_seen": 3946160, "step": 14095 }, { "epoch": 156.66666666666666, "grad_norm": 1.56364910708362e-06, "learning_rate": 3.617622191660438e-05, "loss": 0.0, "num_input_tokens_seen": 3947552, "step": 14100 }, { "epoch": 156.72222222222223, "grad_norm": 1.512732524133753e-06, "learning_rate": 3.616743922828377e-05, "loss": 0.0, "num_input_tokens_seen": 3948992, "step": 14105 }, { "epoch": 156.77777777777777, "grad_norm": 9.483838994128746e-07, "learning_rate": 3.615865481780371e-05, "loss": 0.0, "num_input_tokens_seen": 3950368, "step": 14110 }, { "epoch": 156.83333333333334, "grad_norm": 2.7145208605361404e-06, "learning_rate": 3.614986868651883e-05, "loss": 0.0, "num_input_tokens_seen": 3951824, "step": 14115 }, { "epoch": 156.88888888888889, "grad_norm": 2.2627170892519644e-06, "learning_rate": 3.614108083578409e-05, "loss": 0.0, "num_input_tokens_seen": 3953200, "step": 14120 }, { "epoch": 156.94444444444446, "grad_norm": 2.0832835616602097e-06, "learning_rate": 3.613229126695467e-05, "loss": 0.0, "num_input_tokens_seen": 3954592, "step": 14125 }, { "epoch": 157.0, "grad_norm": 1.6404610505560413e-05, "learning_rate": 3.612349998138605e-05, "loss": 0.0, "num_input_tokens_seen": 3956000, "step": 14130 }, { "epoch": 157.05555555555554, "grad_norm": 1.5744027450637077e-06, "learning_rate": 3.6114706980433946e-05, "loss": 0.0, "num_input_tokens_seen": 3957360, "step": 14135 }, { "epoch": 157.11111111111111, "grad_norm": 1.7076397398341214e-06, "learning_rate": 3.610591226545435e-05, "loss": 0.0, "num_input_tokens_seen": 3958736, "step": 14140 }, { "epoch": 157.16666666666666, "grad_norm": 3.0398725812119665e-06, "learning_rate": 3.6097115837803505e-05, "loss": 0.0, "num_input_tokens_seen": 3960112, "step": 14145 }, { "epoch": 157.22222222222223, "grad_norm": 2.0844870505243307e-06, "learning_rate": 3.608831769883795e-05, "loss": 0.0, "num_input_tokens_seen": 3961536, "step": 14150 }, { "epoch": 157.27777777777777, "grad_norm": 2.291205873916624e-06, "learning_rate": 3.607951784991446e-05, "loss": 0.0, "num_input_tokens_seen": 3962944, "step": 14155 }, { "epoch": 157.33333333333334, "grad_norm": 2.9951249871373875e-06, "learning_rate": 3.6070716292390085e-05, "loss": 0.0, "num_input_tokens_seen": 3964368, "step": 14160 }, { "epoch": 157.38888888888889, "grad_norm": 1.517092414360377e-06, "learning_rate": 3.606191302762213e-05, "loss": 0.0, "num_input_tokens_seen": 3965728, "step": 14165 }, { "epoch": 157.44444444444446, "grad_norm": 1.5999681863831938e-06, "learning_rate": 3.605310805696818e-05, "loss": 0.0, "num_input_tokens_seen": 3967168, "step": 14170 }, { "epoch": 157.5, "grad_norm": 9.159197134067654e-07, "learning_rate": 3.6044301381786067e-05, "loss": 0.0, "num_input_tokens_seen": 3968528, "step": 14175 }, { "epoch": 157.55555555555554, "grad_norm": 3.797593080889783e-06, "learning_rate": 3.6035493003433883e-05, "loss": 0.0, "num_input_tokens_seen": 3969952, "step": 14180 }, { "epoch": 157.61111111111111, "grad_norm": 1.2538528153527295e-06, "learning_rate": 3.6026682923269994e-05, "loss": 0.0, "num_input_tokens_seen": 3971328, "step": 14185 }, { "epoch": 157.66666666666666, "grad_norm": 1.855563823482953e-06, "learning_rate": 3.6017871142653034e-05, "loss": 0.0, "num_input_tokens_seen": 3972752, "step": 14190 }, { "epoch": 157.72222222222223, "grad_norm": 1.535833007437759e-06, "learning_rate": 3.600905766294189e-05, "loss": 0.0, "num_input_tokens_seen": 3974144, "step": 14195 }, { "epoch": 157.77777777777777, "grad_norm": 1.6656493926348048e-06, "learning_rate": 3.60002424854957e-05, "loss": 0.0, "num_input_tokens_seen": 3975568, "step": 14200 }, { "epoch": 157.77777777777777, "eval_loss": 1.238028645515442, "eval_runtime": 1.388, "eval_samples_per_second": 28.819, "eval_steps_per_second": 14.41, "num_input_tokens_seen": 3975568, "step": 14200 }, { "epoch": 157.83333333333334, "grad_norm": 2.7428652629168937e-06, "learning_rate": 3.5991425611673876e-05, "loss": 0.0, "num_input_tokens_seen": 3977008, "step": 14205 }, { "epoch": 157.88888888888889, "grad_norm": 1.2791736025974387e-06, "learning_rate": 3.5982607042836105e-05, "loss": 0.0, "num_input_tokens_seen": 3978432, "step": 14210 }, { "epoch": 157.94444444444446, "grad_norm": 1.2425385875758366e-06, "learning_rate": 3.597378678034231e-05, "loss": 0.0, "num_input_tokens_seen": 3979840, "step": 14215 }, { "epoch": 158.0, "grad_norm": 1.599854954292823e-06, "learning_rate": 3.596496482555269e-05, "loss": 0.0, "num_input_tokens_seen": 3981216, "step": 14220 }, { "epoch": 158.05555555555554, "grad_norm": 1.717088252917165e-06, "learning_rate": 3.595614117982769e-05, "loss": 0.0, "num_input_tokens_seen": 3982608, "step": 14225 }, { "epoch": 158.11111111111111, "grad_norm": 1.5903500525382697e-06, "learning_rate": 3.594731584452805e-05, "loss": 0.0, "num_input_tokens_seen": 3983968, "step": 14230 }, { "epoch": 158.16666666666666, "grad_norm": 1.635543412703555e-06, "learning_rate": 3.593848882101472e-05, "loss": 0.0, "num_input_tokens_seen": 3985328, "step": 14235 }, { "epoch": 158.22222222222223, "grad_norm": 3.418174856051337e-06, "learning_rate": 3.592966011064896e-05, "loss": 0.0, "num_input_tokens_seen": 3986720, "step": 14240 }, { "epoch": 158.27777777777777, "grad_norm": 7.781929980410496e-07, "learning_rate": 3.592082971479226e-05, "loss": 0.0, "num_input_tokens_seen": 3988128, "step": 14245 }, { "epoch": 158.33333333333334, "grad_norm": 3.43448186868045e-06, "learning_rate": 3.5911997634806385e-05, "loss": 0.0, "num_input_tokens_seen": 3989536, "step": 14250 }, { "epoch": 158.38888888888889, "grad_norm": 2.683945922399289e-06, "learning_rate": 3.5903163872053336e-05, "loss": 0.0, "num_input_tokens_seen": 3990944, "step": 14255 }, { "epoch": 158.44444444444446, "grad_norm": 1.4350670198837179e-06, "learning_rate": 3.58943284278954e-05, "loss": 0.0, "num_input_tokens_seen": 3992368, "step": 14260 }, { "epoch": 158.5, "grad_norm": 1.3634984270538553e-06, "learning_rate": 3.588549130369512e-05, "loss": 0.0, "num_input_tokens_seen": 3993760, "step": 14265 }, { "epoch": 158.55555555555554, "grad_norm": 1.6099684216897003e-06, "learning_rate": 3.5876652500815274e-05, "loss": 0.0, "num_input_tokens_seen": 3995152, "step": 14270 }, { "epoch": 158.61111111111111, "grad_norm": 1.780744241841603e-06, "learning_rate": 3.586781202061894e-05, "loss": 0.0, "num_input_tokens_seen": 3996560, "step": 14275 }, { "epoch": 158.66666666666666, "grad_norm": 1.7286305364905274e-06, "learning_rate": 3.585896986446942e-05, "loss": 0.0, "num_input_tokens_seen": 3998016, "step": 14280 }, { "epoch": 158.72222222222223, "grad_norm": 1.4339412928165984e-06, "learning_rate": 3.585012603373028e-05, "loss": 0.0, "num_input_tokens_seen": 3999408, "step": 14285 }, { "epoch": 158.77777777777777, "grad_norm": 4.007003553851973e-06, "learning_rate": 3.584128052976535e-05, "loss": 0.0, "num_input_tokens_seen": 4000832, "step": 14290 }, { "epoch": 158.83333333333334, "grad_norm": 6.220427621883573e-07, "learning_rate": 3.5832433353938724e-05, "loss": 0.0, "num_input_tokens_seen": 4002224, "step": 14295 }, { "epoch": 158.88888888888889, "grad_norm": 1.87147395536158e-06, "learning_rate": 3.5823584507614746e-05, "loss": 0.0, "num_input_tokens_seen": 4003632, "step": 14300 }, { "epoch": 158.94444444444446, "grad_norm": 1.9172414340573596e-06, "learning_rate": 3.581473399215802e-05, "loss": 0.0, "num_input_tokens_seen": 4005056, "step": 14305 }, { "epoch": 159.0, "grad_norm": 1.1961600421273033e-06, "learning_rate": 3.580588180893341e-05, "loss": 0.0, "num_input_tokens_seen": 4006448, "step": 14310 }, { "epoch": 159.05555555555554, "grad_norm": 1.2508709232861293e-06, "learning_rate": 3.579702795930602e-05, "loss": 0.0, "num_input_tokens_seen": 4007840, "step": 14315 }, { "epoch": 159.11111111111111, "grad_norm": 1.8637169887369964e-06, "learning_rate": 3.578817244464125e-05, "loss": 0.0, "num_input_tokens_seen": 4009248, "step": 14320 }, { "epoch": 159.16666666666666, "grad_norm": 1.2613286344276275e-06, "learning_rate": 3.577931526630471e-05, "loss": 0.0, "num_input_tokens_seen": 4010640, "step": 14325 }, { "epoch": 159.22222222222223, "grad_norm": 1.3826449958287412e-06, "learning_rate": 3.577045642566229e-05, "loss": 0.0, "num_input_tokens_seen": 4012032, "step": 14330 }, { "epoch": 159.27777777777777, "grad_norm": 1.3656451756105525e-06, "learning_rate": 3.576159592408014e-05, "loss": 0.0, "num_input_tokens_seen": 4013456, "step": 14335 }, { "epoch": 159.33333333333334, "grad_norm": 1.229428562510293e-06, "learning_rate": 3.575273376292466e-05, "loss": 0.0, "num_input_tokens_seen": 4014816, "step": 14340 }, { "epoch": 159.38888888888889, "grad_norm": 7.063425528031075e-06, "learning_rate": 3.574386994356251e-05, "loss": 0.0, "num_input_tokens_seen": 4016192, "step": 14345 }, { "epoch": 159.44444444444446, "grad_norm": 1.4266272501117783e-06, "learning_rate": 3.573500446736059e-05, "loss": 0.0, "num_input_tokens_seen": 4017600, "step": 14350 }, { "epoch": 159.5, "grad_norm": 4.343571163190063e-06, "learning_rate": 3.5726137335686094e-05, "loss": 0.0, "num_input_tokens_seen": 4019008, "step": 14355 }, { "epoch": 159.55555555555554, "grad_norm": 1.5409319757964113e-06, "learning_rate": 3.571726854990642e-05, "loss": 0.0, "num_input_tokens_seen": 4020416, "step": 14360 }, { "epoch": 159.61111111111111, "grad_norm": 1.6918841083679581e-06, "learning_rate": 3.570839811138925e-05, "loss": 0.0, "num_input_tokens_seen": 4021808, "step": 14365 }, { "epoch": 159.66666666666666, "grad_norm": 1.986378947549383e-06, "learning_rate": 3.569952602150252e-05, "loss": 0.0, "num_input_tokens_seen": 4023216, "step": 14370 }, { "epoch": 159.72222222222223, "grad_norm": 9.008700203594344e-07, "learning_rate": 3.569065228161442e-05, "loss": 0.0, "num_input_tokens_seen": 4024608, "step": 14375 }, { "epoch": 159.77777777777777, "grad_norm": 2.2942260784475366e-06, "learning_rate": 3.5681776893093395e-05, "loss": 0.0, "num_input_tokens_seen": 4026032, "step": 14380 }, { "epoch": 159.83333333333334, "grad_norm": 1.6537953797524096e-06, "learning_rate": 3.5672899857308134e-05, "loss": 0.0, "num_input_tokens_seen": 4027408, "step": 14385 }, { "epoch": 159.88888888888889, "grad_norm": 1.064416323970363e-06, "learning_rate": 3.566402117562759e-05, "loss": 0.0, "num_input_tokens_seen": 4028800, "step": 14390 }, { "epoch": 159.94444444444446, "grad_norm": 1.7928250599652529e-06, "learning_rate": 3.565514084942097e-05, "loss": 0.0, "num_input_tokens_seen": 4030208, "step": 14395 }, { "epoch": 160.0, "grad_norm": 2.852599209290929e-06, "learning_rate": 3.564625888005773e-05, "loss": 0.0, "num_input_tokens_seen": 4031632, "step": 14400 }, { "epoch": 160.0, "eval_loss": 1.2565587759017944, "eval_runtime": 1.384, "eval_samples_per_second": 28.903, "eval_steps_per_second": 14.451, "num_input_tokens_seen": 4031632, "step": 14400 }, { "epoch": 160.05555555555554, "grad_norm": 1.269655740543385e-06, "learning_rate": 3.563737526890759e-05, "loss": 0.0, "num_input_tokens_seen": 4032992, "step": 14405 }, { "epoch": 160.11111111111111, "grad_norm": 1.2537715292637586e-06, "learning_rate": 3.562849001734049e-05, "loss": 0.0, "num_input_tokens_seen": 4034416, "step": 14410 }, { "epoch": 160.16666666666666, "grad_norm": 2.6597053874866106e-06, "learning_rate": 3.561960312672667e-05, "loss": 0.0, "num_input_tokens_seen": 4035808, "step": 14415 }, { "epoch": 160.22222222222223, "grad_norm": 5.592598881776212e-07, "learning_rate": 3.5610714598436596e-05, "loss": 0.0, "num_input_tokens_seen": 4037184, "step": 14420 }, { "epoch": 160.27777777777777, "grad_norm": 1.4993077002145583e-06, "learning_rate": 3.5601824433840986e-05, "loss": 0.0, "num_input_tokens_seen": 4038608, "step": 14425 }, { "epoch": 160.33333333333334, "grad_norm": 2.744887524386286e-06, "learning_rate": 3.559293263431082e-05, "loss": 0.0, "num_input_tokens_seen": 4040016, "step": 14430 }, { "epoch": 160.38888888888889, "grad_norm": 1.3190467598178657e-06, "learning_rate": 3.558403920121732e-05, "loss": 0.0, "num_input_tokens_seen": 4041408, "step": 14435 }, { "epoch": 160.44444444444446, "grad_norm": 3.471562877166434e-06, "learning_rate": 3.557514413593197e-05, "loss": 0.0, "num_input_tokens_seen": 4042784, "step": 14440 }, { "epoch": 160.5, "grad_norm": 2.031035592153785e-06, "learning_rate": 3.55662474398265e-05, "loss": 0.0, "num_input_tokens_seen": 4044160, "step": 14445 }, { "epoch": 160.55555555555554, "grad_norm": 2.7169801342097344e-06, "learning_rate": 3.555734911427288e-05, "loss": 0.0, "num_input_tokens_seen": 4045520, "step": 14450 }, { "epoch": 160.61111111111111, "grad_norm": 2.7662933916872134e-06, "learning_rate": 3.5548449160643363e-05, "loss": 0.0, "num_input_tokens_seen": 4046928, "step": 14455 }, { "epoch": 160.66666666666666, "grad_norm": 1.4636409559898311e-06, "learning_rate": 3.553954758031043e-05, "loss": 0.0, "num_input_tokens_seen": 4048304, "step": 14460 }, { "epoch": 160.72222222222223, "grad_norm": 3.5953009955846937e-06, "learning_rate": 3.5530644374646815e-05, "loss": 0.0, "num_input_tokens_seen": 4049728, "step": 14465 }, { "epoch": 160.77777777777777, "grad_norm": 1.2362136203591945e-06, "learning_rate": 3.552173954502549e-05, "loss": 0.0, "num_input_tokens_seen": 4051136, "step": 14470 }, { "epoch": 160.83333333333334, "grad_norm": 6.709561830575694e-07, "learning_rate": 3.55128330928197e-05, "loss": 0.0, "num_input_tokens_seen": 4052544, "step": 14475 }, { "epoch": 160.88888888888889, "grad_norm": 8.352035933967272e-07, "learning_rate": 3.550392501940294e-05, "loss": 0.0, "num_input_tokens_seen": 4054000, "step": 14480 }, { "epoch": 160.94444444444446, "grad_norm": 1.0104139391842182e-06, "learning_rate": 3.5495015326148945e-05, "loss": 0.0, "num_input_tokens_seen": 4055408, "step": 14485 }, { "epoch": 161.0, "grad_norm": 1.0813957942445995e-06, "learning_rate": 3.548610401443169e-05, "loss": 0.0, "num_input_tokens_seen": 4056816, "step": 14490 }, { "epoch": 161.05555555555554, "grad_norm": 9.257313990929106e-07, "learning_rate": 3.547719108562543e-05, "loss": 0.0, "num_input_tokens_seen": 4058224, "step": 14495 }, { "epoch": 161.11111111111111, "grad_norm": 2.173654138459824e-06, "learning_rate": 3.546827654110464e-05, "loss": 0.0, "num_input_tokens_seen": 4059584, "step": 14500 }, { "epoch": 161.16666666666666, "grad_norm": 1.8038232383332797e-06, "learning_rate": 3.545936038224405e-05, "loss": 0.0, "num_input_tokens_seen": 4061008, "step": 14505 }, { "epoch": 161.22222222222223, "grad_norm": 3.687077651193249e-06, "learning_rate": 3.545044261041864e-05, "loss": 0.0, "num_input_tokens_seen": 4062448, "step": 14510 }, { "epoch": 161.27777777777777, "grad_norm": 1.2901666650577681e-06, "learning_rate": 3.5441523227003657e-05, "loss": 0.0, "num_input_tokens_seen": 4063824, "step": 14515 }, { "epoch": 161.33333333333334, "grad_norm": 5.7030147218029015e-06, "learning_rate": 3.543260223337459e-05, "loss": 0.0, "num_input_tokens_seen": 4065248, "step": 14520 }, { "epoch": 161.38888888888889, "grad_norm": 1.3622658343592775e-06, "learning_rate": 3.542367963090714e-05, "loss": 0.0, "num_input_tokens_seen": 4066624, "step": 14525 }, { "epoch": 161.44444444444446, "grad_norm": 1.1618562894000206e-06, "learning_rate": 3.5414755420977295e-05, "loss": 0.0, "num_input_tokens_seen": 4068048, "step": 14530 }, { "epoch": 161.5, "grad_norm": 1.2296060276639764e-06, "learning_rate": 3.54058296049613e-05, "loss": 0.0, "num_input_tokens_seen": 4069472, "step": 14535 }, { "epoch": 161.55555555555554, "grad_norm": 1.0195324193773558e-06, "learning_rate": 3.53969021842356e-05, "loss": 0.0, "num_input_tokens_seen": 4070848, "step": 14540 }, { "epoch": 161.61111111111111, "grad_norm": 1.3483971770256176e-06, "learning_rate": 3.5387973160176926e-05, "loss": 0.0, "num_input_tokens_seen": 4072256, "step": 14545 }, { "epoch": 161.66666666666666, "grad_norm": 1.7219424535142025e-06, "learning_rate": 3.537904253416224e-05, "loss": 0.0, "num_input_tokens_seen": 4073632, "step": 14550 }, { "epoch": 161.72222222222223, "grad_norm": 1.6022319186959066e-06, "learning_rate": 3.537011030756878e-05, "loss": 0.0, "num_input_tokens_seen": 4075024, "step": 14555 }, { "epoch": 161.77777777777777, "grad_norm": 1.2640389286389109e-06, "learning_rate": 3.536117648177399e-05, "loss": 0.0, "num_input_tokens_seen": 4076400, "step": 14560 }, { "epoch": 161.83333333333334, "grad_norm": 3.4479653550079092e-06, "learning_rate": 3.535224105815558e-05, "loss": 0.0, "num_input_tokens_seen": 4077840, "step": 14565 }, { "epoch": 161.88888888888889, "grad_norm": 1.5846927681195666e-06, "learning_rate": 3.5343304038091494e-05, "loss": 0.0, "num_input_tokens_seen": 4079248, "step": 14570 }, { "epoch": 161.94444444444446, "grad_norm": 3.930206275981618e-06, "learning_rate": 3.5334365422959955e-05, "loss": 0.0, "num_input_tokens_seen": 4080656, "step": 14575 }, { "epoch": 162.0, "grad_norm": 1.6122990018629935e-06, "learning_rate": 3.5325425214139396e-05, "loss": 0.0, "num_input_tokens_seen": 4082032, "step": 14580 }, { "epoch": 162.05555555555554, "grad_norm": 1.278159061257611e-06, "learning_rate": 3.531648341300851e-05, "loss": 0.0, "num_input_tokens_seen": 4083424, "step": 14585 }, { "epoch": 162.11111111111111, "grad_norm": 1.1883394108735956e-06, "learning_rate": 3.530754002094623e-05, "loss": 0.0, "num_input_tokens_seen": 4084768, "step": 14590 }, { "epoch": 162.16666666666666, "grad_norm": 3.2551042750128545e-06, "learning_rate": 3.529859503933175e-05, "loss": 0.0, "num_input_tokens_seen": 4086160, "step": 14595 }, { "epoch": 162.22222222222223, "grad_norm": 2.319708528375486e-06, "learning_rate": 3.52896484695445e-05, "loss": 0.0, "num_input_tokens_seen": 4087632, "step": 14600 }, { "epoch": 162.22222222222223, "eval_loss": 1.2692619562149048, "eval_runtime": 1.398, "eval_samples_per_second": 28.611, "eval_steps_per_second": 14.306, "num_input_tokens_seen": 4087632, "step": 14600 }, { "epoch": 162.27777777777777, "grad_norm": 2.5460394681431353e-06, "learning_rate": 3.528070031296414e-05, "loss": 0.0, "num_input_tokens_seen": 4089008, "step": 14605 }, { "epoch": 162.33333333333334, "grad_norm": 1.6086133882708964e-06, "learning_rate": 3.5271750570970605e-05, "loss": 0.0, "num_input_tokens_seen": 4090416, "step": 14610 }, { "epoch": 162.38888888888889, "grad_norm": 1.6134722500282805e-06, "learning_rate": 3.526279924494405e-05, "loss": 0.0, "num_input_tokens_seen": 4091824, "step": 14615 }, { "epoch": 162.44444444444446, "grad_norm": 1.967620846698992e-06, "learning_rate": 3.5253846336264874e-05, "loss": 0.0, "num_input_tokens_seen": 4093248, "step": 14620 }, { "epoch": 162.5, "grad_norm": 4.794391770701623e-06, "learning_rate": 3.5244891846313736e-05, "loss": 0.0, "num_input_tokens_seen": 4094672, "step": 14625 }, { "epoch": 162.55555555555554, "grad_norm": 1.8188027297583176e-06, "learning_rate": 3.5235935776471527e-05, "loss": 0.0, "num_input_tokens_seen": 4096096, "step": 14630 }, { "epoch": 162.61111111111111, "grad_norm": 2.4554133233323228e-06, "learning_rate": 3.522697812811939e-05, "loss": 0.0, "num_input_tokens_seen": 4097456, "step": 14635 }, { "epoch": 162.66666666666666, "grad_norm": 1.233807665812492e-06, "learning_rate": 3.521801890263871e-05, "loss": 0.0, "num_input_tokens_seen": 4098848, "step": 14640 }, { "epoch": 162.72222222222223, "grad_norm": 7.187660230556503e-06, "learning_rate": 3.5209058101411114e-05, "loss": 0.0, "num_input_tokens_seen": 4100272, "step": 14645 }, { "epoch": 162.77777777777777, "grad_norm": 1.8755505379886017e-06, "learning_rate": 3.520009572581845e-05, "loss": 0.0, "num_input_tokens_seen": 4101696, "step": 14650 }, { "epoch": 162.83333333333334, "grad_norm": 4.47119145974284e-06, "learning_rate": 3.519113177724285e-05, "loss": 0.0, "num_input_tokens_seen": 4103120, "step": 14655 }, { "epoch": 162.88888888888889, "grad_norm": 1.9299977793707512e-06, "learning_rate": 3.5182166257066656e-05, "loss": 0.0, "num_input_tokens_seen": 4104512, "step": 14660 }, { "epoch": 162.94444444444446, "grad_norm": 1.8064745290757855e-06, "learning_rate": 3.517319916667247e-05, "loss": 0.0, "num_input_tokens_seen": 4105920, "step": 14665 }, { "epoch": 163.0, "grad_norm": 1.1388764278308372e-06, "learning_rate": 3.516423050744313e-05, "loss": 0.0, "num_input_tokens_seen": 4107312, "step": 14670 }, { "epoch": 163.05555555555554, "grad_norm": 7.293318162737705e-07, "learning_rate": 3.5155260280761704e-05, "loss": 0.0, "num_input_tokens_seen": 4108768, "step": 14675 }, { "epoch": 163.11111111111111, "grad_norm": 1.068058281816775e-06, "learning_rate": 3.514628848801154e-05, "loss": 0.0, "num_input_tokens_seen": 4110144, "step": 14680 }, { "epoch": 163.16666666666666, "grad_norm": 1.383361791340576e-06, "learning_rate": 3.5137315130576174e-05, "loss": 0.0, "num_input_tokens_seen": 4111520, "step": 14685 }, { "epoch": 163.22222222222223, "grad_norm": 1.3050953384663444e-06, "learning_rate": 3.512834020983942e-05, "loss": 0.0, "num_input_tokens_seen": 4112944, "step": 14690 }, { "epoch": 163.27777777777777, "grad_norm": 2.5725380510266405e-06, "learning_rate": 3.5119363727185334e-05, "loss": 0.0, "num_input_tokens_seen": 4114384, "step": 14695 }, { "epoch": 163.33333333333334, "grad_norm": 9.4244177262226e-07, "learning_rate": 3.511038568399819e-05, "loss": 0.0, "num_input_tokens_seen": 4115776, "step": 14700 }, { "epoch": 163.38888888888889, "grad_norm": 1.6866782743818476e-06, "learning_rate": 3.510140608166251e-05, "loss": 0.0, "num_input_tokens_seen": 4117152, "step": 14705 }, { "epoch": 163.44444444444446, "grad_norm": 2.365767841183697e-06, "learning_rate": 3.509242492156308e-05, "loss": 0.0, "num_input_tokens_seen": 4118576, "step": 14710 }, { "epoch": 163.5, "grad_norm": 2.0797028810193297e-06, "learning_rate": 3.5083442205084896e-05, "loss": 0.0, "num_input_tokens_seen": 4119968, "step": 14715 }, { "epoch": 163.55555555555554, "grad_norm": 1.9051487925025867e-06, "learning_rate": 3.507445793361321e-05, "loss": 0.0, "num_input_tokens_seen": 4121328, "step": 14720 }, { "epoch": 163.61111111111111, "grad_norm": 2.4456078335788334e-06, "learning_rate": 3.5065472108533505e-05, "loss": 0.0, "num_input_tokens_seen": 4122736, "step": 14725 }, { "epoch": 163.66666666666666, "grad_norm": 1.211452854477102e-06, "learning_rate": 3.5056484731231504e-05, "loss": 0.0, "num_input_tokens_seen": 4124080, "step": 14730 }, { "epoch": 163.72222222222223, "grad_norm": 1.6333673329427256e-06, "learning_rate": 3.504749580309319e-05, "loss": 0.0, "num_input_tokens_seen": 4125488, "step": 14735 }, { "epoch": 163.77777777777777, "grad_norm": 3.7356576285674237e-06, "learning_rate": 3.5038505325504753e-05, "loss": 0.0, "num_input_tokens_seen": 4126896, "step": 14740 }, { "epoch": 163.83333333333334, "grad_norm": 1.7441379895899445e-06, "learning_rate": 3.502951329985264e-05, "loss": 0.0, "num_input_tokens_seen": 4128304, "step": 14745 }, { "epoch": 163.88888888888889, "grad_norm": 3.857349383906694e-06, "learning_rate": 3.502051972752354e-05, "loss": 0.0, "num_input_tokens_seen": 4129680, "step": 14750 }, { "epoch": 163.94444444444446, "grad_norm": 7.033868314465508e-07, "learning_rate": 3.5011524609904374e-05, "loss": 0.0, "num_input_tokens_seen": 4131088, "step": 14755 }, { "epoch": 164.0, "grad_norm": 1.4027356201040675e-06, "learning_rate": 3.50025279483823e-05, "loss": 0.0, "num_input_tokens_seen": 4132496, "step": 14760 }, { "epoch": 164.05555555555554, "grad_norm": 4.5346314436756074e-06, "learning_rate": 3.499352974434472e-05, "loss": 0.0, "num_input_tokens_seen": 4133872, "step": 14765 }, { "epoch": 164.11111111111111, "grad_norm": 1.2592405482791946e-06, "learning_rate": 3.498452999917926e-05, "loss": 0.0, "num_input_tokens_seen": 4135248, "step": 14770 }, { "epoch": 164.16666666666666, "grad_norm": 1.4426868801820092e-06, "learning_rate": 3.4975528714273795e-05, "loss": 0.0, "num_input_tokens_seen": 4136608, "step": 14775 }, { "epoch": 164.22222222222223, "grad_norm": 2.309103592779138e-06, "learning_rate": 3.4966525891016454e-05, "loss": 0.0, "num_input_tokens_seen": 4138048, "step": 14780 }, { "epoch": 164.27777777777777, "grad_norm": 2.9131872452126117e-06, "learning_rate": 3.495752153079557e-05, "loss": 0.0, "num_input_tokens_seen": 4139456, "step": 14785 }, { "epoch": 164.33333333333334, "grad_norm": 1.8941152575280285e-06, "learning_rate": 3.494851563499974e-05, "loss": 0.0, "num_input_tokens_seen": 4140864, "step": 14790 }, { "epoch": 164.38888888888889, "grad_norm": 1.5426315940203494e-06, "learning_rate": 3.493950820501777e-05, "loss": 0.0, "num_input_tokens_seen": 4142272, "step": 14795 }, { "epoch": 164.44444444444446, "grad_norm": 2.8796800961572444e-06, "learning_rate": 3.493049924223872e-05, "loss": 0.0, "num_input_tokens_seen": 4143664, "step": 14800 }, { "epoch": 164.44444444444446, "eval_loss": 1.2769689559936523, "eval_runtime": 1.389, "eval_samples_per_second": 28.799, "eval_steps_per_second": 14.399, "num_input_tokens_seen": 4143664, "step": 14800 }, { "epoch": 164.5, "grad_norm": 1.0535341061768122e-06, "learning_rate": 3.49214887480519e-05, "loss": 0.0, "num_input_tokens_seen": 4145024, "step": 14805 }, { "epoch": 164.55555555555554, "grad_norm": 1.7898216810863232e-06, "learning_rate": 3.4912476723846834e-05, "loss": 0.0, "num_input_tokens_seen": 4146384, "step": 14810 }, { "epoch": 164.61111111111111, "grad_norm": 4.990283287042985e-06, "learning_rate": 3.490346317101328e-05, "loss": 0.0, "num_input_tokens_seen": 4147824, "step": 14815 }, { "epoch": 164.66666666666666, "grad_norm": 6.70475287734007e-07, "learning_rate": 3.4894448090941266e-05, "loss": 0.0, "num_input_tokens_seen": 4149184, "step": 14820 }, { "epoch": 164.72222222222223, "grad_norm": 2.055973254755372e-06, "learning_rate": 3.488543148502101e-05, "loss": 0.0, "num_input_tokens_seen": 4150592, "step": 14825 }, { "epoch": 164.77777777777777, "grad_norm": 9.49758771184861e-07, "learning_rate": 3.487641335464299e-05, "loss": 0.0, "num_input_tokens_seen": 4152000, "step": 14830 }, { "epoch": 164.83333333333334, "grad_norm": 6.656627192569431e-06, "learning_rate": 3.4867393701197914e-05, "loss": 0.0, "num_input_tokens_seen": 4153408, "step": 14835 }, { "epoch": 164.88888888888889, "grad_norm": 2.6878919925366063e-06, "learning_rate": 3.485837252607673e-05, "loss": 0.0, "num_input_tokens_seen": 4154816, "step": 14840 }, { "epoch": 164.94444444444446, "grad_norm": 1.925233618749189e-06, "learning_rate": 3.4849349830670615e-05, "loss": 0.0, "num_input_tokens_seen": 4156208, "step": 14845 }, { "epoch": 165.0, "grad_norm": 7.804132451383339e-07, "learning_rate": 3.4840325616370976e-05, "loss": 0.0, "num_input_tokens_seen": 4157600, "step": 14850 }, { "epoch": 165.05555555555554, "grad_norm": 1.116128032663255e-06, "learning_rate": 3.483129988456947e-05, "loss": 0.0, "num_input_tokens_seen": 4158992, "step": 14855 }, { "epoch": 165.11111111111111, "grad_norm": 4.360854291007854e-07, "learning_rate": 3.482227263665797e-05, "loss": 0.0, "num_input_tokens_seen": 4160432, "step": 14860 }, { "epoch": 165.16666666666666, "grad_norm": 1.581529659233638e-06, "learning_rate": 3.48132438740286e-05, "loss": 0.0, "num_input_tokens_seen": 4161856, "step": 14865 }, { "epoch": 165.22222222222223, "grad_norm": 1.9210835944249993e-06, "learning_rate": 3.48042135980737e-05, "loss": 0.0, "num_input_tokens_seen": 4163232, "step": 14870 }, { "epoch": 165.27777777777777, "grad_norm": 1.0574477755653788e-06, "learning_rate": 3.479518181018586e-05, "loss": 0.0, "num_input_tokens_seen": 4164608, "step": 14875 }, { "epoch": 165.33333333333334, "grad_norm": 3.0718288144271355e-06, "learning_rate": 3.4786148511757886e-05, "loss": 0.0, "num_input_tokens_seen": 4165984, "step": 14880 }, { "epoch": 165.38888888888889, "grad_norm": 1.609886908227054e-06, "learning_rate": 3.477711370418284e-05, "loss": 0.0, "num_input_tokens_seen": 4167360, "step": 14885 }, { "epoch": 165.44444444444446, "grad_norm": 1.094358481168456e-06, "learning_rate": 3.476807738885399e-05, "loss": 0.0, "num_input_tokens_seen": 4168784, "step": 14890 }, { "epoch": 165.5, "grad_norm": 8.146415098053694e-07, "learning_rate": 3.475903956716485e-05, "loss": 0.0, "num_input_tokens_seen": 4170192, "step": 14895 }, { "epoch": 165.55555555555554, "grad_norm": 4.134224127483321e-06, "learning_rate": 3.475000024050917e-05, "loss": 0.0, "num_input_tokens_seen": 4171600, "step": 14900 }, { "epoch": 165.61111111111111, "grad_norm": 1.2798578836736851e-06, "learning_rate": 3.4740959410280926e-05, "loss": 0.0, "num_input_tokens_seen": 4172992, "step": 14905 }, { "epoch": 165.66666666666666, "grad_norm": 1.741464529914083e-06, "learning_rate": 3.4731917077874324e-05, "loss": 0.0, "num_input_tokens_seen": 4174384, "step": 14910 }, { "epoch": 165.72222222222223, "grad_norm": 1.71629062606371e-06, "learning_rate": 3.4722873244683816e-05, "loss": 0.0, "num_input_tokens_seen": 4175776, "step": 14915 }, { "epoch": 165.77777777777777, "grad_norm": 2.785189053611248e-06, "learning_rate": 3.4713827912104065e-05, "loss": 0.0, "num_input_tokens_seen": 4177168, "step": 14920 }, { "epoch": 165.83333333333334, "grad_norm": 2.5691133487271145e-06, "learning_rate": 3.470478108152998e-05, "loss": 0.0, "num_input_tokens_seen": 4178544, "step": 14925 }, { "epoch": 165.88888888888889, "grad_norm": 1.5235084447340341e-06, "learning_rate": 3.4695732754356695e-05, "loss": 0.0, "num_input_tokens_seen": 4179984, "step": 14930 }, { "epoch": 165.94444444444446, "grad_norm": 3.2827219911268912e-06, "learning_rate": 3.4686682931979576e-05, "loss": 0.0, "num_input_tokens_seen": 4181376, "step": 14935 }, { "epoch": 166.0, "grad_norm": 7.413904768327484e-06, "learning_rate": 3.467763161579422e-05, "loss": 0.0, "num_input_tokens_seen": 4182816, "step": 14940 }, { "epoch": 166.05555555555554, "grad_norm": 1.2143424328314723e-06, "learning_rate": 3.466857880719645e-05, "loss": 0.0, "num_input_tokens_seen": 4184208, "step": 14945 }, { "epoch": 166.11111111111111, "grad_norm": 1.5474458905373467e-06, "learning_rate": 3.465952450758233e-05, "loss": 0.0, "num_input_tokens_seen": 4185568, "step": 14950 }, { "epoch": 166.16666666666666, "grad_norm": 1.5533016721747117e-06, "learning_rate": 3.4650468718348126e-05, "loss": 0.0, "num_input_tokens_seen": 4186944, "step": 14955 }, { "epoch": 166.22222222222223, "grad_norm": 1.1876398957610945e-06, "learning_rate": 3.464141144089038e-05, "loss": 0.0, "num_input_tokens_seen": 4188336, "step": 14960 }, { "epoch": 166.27777777777777, "grad_norm": 1.6686747130734148e-06, "learning_rate": 3.463235267660583e-05, "loss": 0.0, "num_input_tokens_seen": 4189760, "step": 14965 }, { "epoch": 166.33333333333334, "grad_norm": 1.2993623386137187e-06, "learning_rate": 3.462329242689145e-05, "loss": 0.0, "num_input_tokens_seen": 4191104, "step": 14970 }, { "epoch": 166.38888888888889, "grad_norm": 2.5264796477131313e-06, "learning_rate": 3.461423069314444e-05, "loss": 0.0, "num_input_tokens_seen": 4192544, "step": 14975 }, { "epoch": 166.44444444444446, "grad_norm": 1.8070373926093453e-06, "learning_rate": 3.460516747676224e-05, "loss": 0.0, "num_input_tokens_seen": 4193952, "step": 14980 }, { "epoch": 166.5, "grad_norm": 2.4759724510659e-06, "learning_rate": 3.459610277914251e-05, "loss": 0.0, "num_input_tokens_seen": 4195360, "step": 14985 }, { "epoch": 166.55555555555554, "grad_norm": 5.127755684952717e-06, "learning_rate": 3.458703660168314e-05, "loss": 0.0, "num_input_tokens_seen": 4196768, "step": 14990 }, { "epoch": 166.61111111111111, "grad_norm": 4.45283649241901e-06, "learning_rate": 3.457796894578224e-05, "loss": 0.0, "num_input_tokens_seen": 4198176, "step": 14995 }, { "epoch": 166.66666666666666, "grad_norm": 5.467644541568006e-07, "learning_rate": 3.456889981283817e-05, "loss": 0.0, "num_input_tokens_seen": 4199552, "step": 15000 }, { "epoch": 166.66666666666666, "eval_loss": 1.271511435508728, "eval_runtime": 1.4059, "eval_samples_per_second": 28.452, "eval_steps_per_second": 14.226, "num_input_tokens_seen": 4199552, "step": 15000 }, { "epoch": 166.72222222222223, "grad_norm": 2.5090457711485215e-06, "learning_rate": 3.45598292042495e-05, "loss": 0.0, "num_input_tokens_seen": 4200960, "step": 15005 }, { "epoch": 166.77777777777777, "grad_norm": 2.503888481442118e-06, "learning_rate": 3.4550757121415035e-05, "loss": 0.0, "num_input_tokens_seen": 4202400, "step": 15010 }, { "epoch": 166.83333333333334, "grad_norm": 1.3207394431447028e-06, "learning_rate": 3.454168356573378e-05, "loss": 0.0, "num_input_tokens_seen": 4203760, "step": 15015 }, { "epoch": 166.88888888888889, "grad_norm": 3.436828819758375e-06, "learning_rate": 3.453260853860503e-05, "loss": 0.0, "num_input_tokens_seen": 4205120, "step": 15020 }, { "epoch": 166.94444444444446, "grad_norm": 6.329939310489863e-07, "learning_rate": 3.452353204142824e-05, "loss": 0.0, "num_input_tokens_seen": 4206496, "step": 15025 }, { "epoch": 167.0, "grad_norm": 1.2505579434218816e-06, "learning_rate": 3.4514454075603136e-05, "loss": 0.0, "num_input_tokens_seen": 4207936, "step": 15030 }, { "epoch": 167.05555555555554, "grad_norm": 1.530595113763411e-06, "learning_rate": 3.450537464252964e-05, "loss": 0.0, "num_input_tokens_seen": 4209312, "step": 15035 }, { "epoch": 167.11111111111111, "grad_norm": 1.5636931038898183e-06, "learning_rate": 3.4496293743607925e-05, "loss": 0.0, "num_input_tokens_seen": 4210736, "step": 15040 }, { "epoch": 167.16666666666666, "grad_norm": 2.611288664411404e-06, "learning_rate": 3.448721138023838e-05, "loss": 0.0, "num_input_tokens_seen": 4212112, "step": 15045 }, { "epoch": 167.22222222222223, "grad_norm": 1.7524439499538857e-06, "learning_rate": 3.447812755382162e-05, "loss": 0.0, "num_input_tokens_seen": 4213488, "step": 15050 }, { "epoch": 167.27777777777777, "grad_norm": 2.8625283903238596e-06, "learning_rate": 3.446904226575847e-05, "loss": 0.0, "num_input_tokens_seen": 4214880, "step": 15055 }, { "epoch": 167.33333333333334, "grad_norm": 1.8555583665147424e-06, "learning_rate": 3.445995551745002e-05, "loss": 0.0, "num_input_tokens_seen": 4216288, "step": 15060 }, { "epoch": 167.38888888888889, "grad_norm": 2.3285692805075087e-06, "learning_rate": 3.445086731029753e-05, "loss": 0.0, "num_input_tokens_seen": 4217696, "step": 15065 }, { "epoch": 167.44444444444446, "grad_norm": 1.365311732115515e-06, "learning_rate": 3.444177764570255e-05, "loss": 0.0, "num_input_tokens_seen": 4219056, "step": 15070 }, { "epoch": 167.5, "grad_norm": 1.7420067024431773e-06, "learning_rate": 3.44326865250668e-05, "loss": 0.0, "num_input_tokens_seen": 4220496, "step": 15075 }, { "epoch": 167.55555555555554, "grad_norm": 1.3890448826714419e-06, "learning_rate": 3.442359394979225e-05, "loss": 0.0, "num_input_tokens_seen": 4221888, "step": 15080 }, { "epoch": 167.61111111111111, "grad_norm": 6.772925189579837e-07, "learning_rate": 3.441449992128108e-05, "loss": 0.0, "num_input_tokens_seen": 4223296, "step": 15085 }, { "epoch": 167.66666666666666, "grad_norm": 1.5289904240489705e-06, "learning_rate": 3.440540444093573e-05, "loss": 0.0, "num_input_tokens_seen": 4224736, "step": 15090 }, { "epoch": 167.72222222222223, "grad_norm": 9.062729304787354e-07, "learning_rate": 3.43963075101588e-05, "loss": 0.0, "num_input_tokens_seen": 4226112, "step": 15095 }, { "epoch": 167.77777777777777, "grad_norm": 3.1290744573198026e-06, "learning_rate": 3.438720913035318e-05, "loss": 0.0, "num_input_tokens_seen": 4227520, "step": 15100 }, { "epoch": 167.83333333333334, "grad_norm": 3.868080511892913e-06, "learning_rate": 3.437810930292195e-05, "loss": 0.0, "num_input_tokens_seen": 4228880, "step": 15105 }, { "epoch": 167.88888888888889, "grad_norm": 2.284564061483252e-06, "learning_rate": 3.43690080292684e-05, "loss": 0.0, "num_input_tokens_seen": 4230320, "step": 15110 }, { "epoch": 167.94444444444446, "grad_norm": 7.648514497304859e-07, "learning_rate": 3.435990531079608e-05, "loss": 0.0, "num_input_tokens_seen": 4231728, "step": 15115 }, { "epoch": 168.0, "grad_norm": 1.1715169421222527e-06, "learning_rate": 3.435080114890874e-05, "loss": 0.0, "num_input_tokens_seen": 4233152, "step": 15120 }, { "epoch": 168.05555555555554, "grad_norm": 4.317718321544817e-06, "learning_rate": 3.434169554501035e-05, "loss": 0.0, "num_input_tokens_seen": 4234544, "step": 15125 }, { "epoch": 168.11111111111111, "grad_norm": 1.302400164604478e-06, "learning_rate": 3.433258850050511e-05, "loss": 0.0, "num_input_tokens_seen": 4235904, "step": 15130 }, { "epoch": 168.16666666666666, "grad_norm": 7.219809958769474e-07, "learning_rate": 3.4323480016797446e-05, "loss": 0.0, "num_input_tokens_seen": 4237280, "step": 15135 }, { "epoch": 168.22222222222223, "grad_norm": 2.36833193412167e-06, "learning_rate": 3.4314370095291995e-05, "loss": 0.0, "num_input_tokens_seen": 4238688, "step": 15140 }, { "epoch": 168.27777777777777, "grad_norm": 5.533618150366237e-07, "learning_rate": 3.430525873739363e-05, "loss": 0.0, "num_input_tokens_seen": 4240112, "step": 15145 }, { "epoch": 168.33333333333334, "grad_norm": 1.270366738026496e-06, "learning_rate": 3.429614594450743e-05, "loss": 0.0, "num_input_tokens_seen": 4241520, "step": 15150 }, { "epoch": 168.38888888888889, "grad_norm": 2.1903547349211294e-06, "learning_rate": 3.428703171803869e-05, "loss": 0.0, "num_input_tokens_seen": 4242912, "step": 15155 }, { "epoch": 168.44444444444446, "grad_norm": 1.3417451327768504e-06, "learning_rate": 3.4277916059392964e-05, "loss": 0.0, "num_input_tokens_seen": 4244352, "step": 15160 }, { "epoch": 168.5, "grad_norm": 1.4461220416706055e-06, "learning_rate": 3.426879896997598e-05, "loss": 0.0, "num_input_tokens_seen": 4245744, "step": 15165 }, { "epoch": 168.55555555555554, "grad_norm": 1.8350033315073233e-06, "learning_rate": 3.425968045119372e-05, "loss": 0.0, "num_input_tokens_seen": 4247168, "step": 15170 }, { "epoch": 168.61111111111111, "grad_norm": 7.241602588692331e-07, "learning_rate": 3.425056050445237e-05, "loss": 0.0, "num_input_tokens_seen": 4248560, "step": 15175 }, { "epoch": 168.66666666666666, "grad_norm": 6.776327836632845e-07, "learning_rate": 3.4241439131158336e-05, "loss": 0.0, "num_input_tokens_seen": 4249968, "step": 15180 }, { "epoch": 168.72222222222223, "grad_norm": 5.210476047068369e-06, "learning_rate": 3.423231633271825e-05, "loss": 0.0, "num_input_tokens_seen": 4251392, "step": 15185 }, { "epoch": 168.77777777777777, "grad_norm": 1.949384113686392e-06, "learning_rate": 3.4223192110538985e-05, "loss": 0.0, "num_input_tokens_seen": 4252736, "step": 15190 }, { "epoch": 168.83333333333334, "grad_norm": 1.1198131915080012e-06, "learning_rate": 3.4214066466027575e-05, "loss": 0.0, "num_input_tokens_seen": 4254144, "step": 15195 }, { "epoch": 168.88888888888889, "grad_norm": 1.0889341410802444e-06, "learning_rate": 3.4204939400591325e-05, "loss": 0.0, "num_input_tokens_seen": 4255584, "step": 15200 }, { "epoch": 168.88888888888889, "eval_loss": 1.273184061050415, "eval_runtime": 1.389, "eval_samples_per_second": 28.797, "eval_steps_per_second": 14.399, "num_input_tokens_seen": 4255584, "step": 15200 }, { "epoch": 168.94444444444446, "grad_norm": 4.8616534513712395e-06, "learning_rate": 3.419581091563775e-05, "loss": 0.0, "num_input_tokens_seen": 4256992, "step": 15205 }, { "epoch": 169.0, "grad_norm": 9.70596829574788e-07, "learning_rate": 3.418668101257456e-05, "loss": 0.0, "num_input_tokens_seen": 4258384, "step": 15210 }, { "epoch": 169.05555555555554, "grad_norm": 7.197956506388437e-07, "learning_rate": 3.417754969280971e-05, "loss": 0.0, "num_input_tokens_seen": 4259808, "step": 15215 }, { "epoch": 169.11111111111111, "grad_norm": 2.5986475975514622e-06, "learning_rate": 3.416841695775137e-05, "loss": 0.0, "num_input_tokens_seen": 4261216, "step": 15220 }, { "epoch": 169.16666666666666, "grad_norm": 8.392095764975238e-07, "learning_rate": 3.415928280880792e-05, "loss": 0.0, "num_input_tokens_seen": 4262608, "step": 15225 }, { "epoch": 169.22222222222223, "grad_norm": 1.1948596920774435e-06, "learning_rate": 3.4150147247387965e-05, "loss": 0.0, "num_input_tokens_seen": 4264032, "step": 15230 }, { "epoch": 169.27777777777777, "grad_norm": 3.4433933251420967e-06, "learning_rate": 3.4141010274900306e-05, "loss": 0.0, "num_input_tokens_seen": 4265424, "step": 15235 }, { "epoch": 169.33333333333334, "grad_norm": 2.5037013529072283e-06, "learning_rate": 3.413187189275399e-05, "loss": 0.0, "num_input_tokens_seen": 4266880, "step": 15240 }, { "epoch": 169.38888888888889, "grad_norm": 8.419783625868149e-07, "learning_rate": 3.4122732102358265e-05, "loss": 0.0, "num_input_tokens_seen": 4268224, "step": 15245 }, { "epoch": 169.44444444444446, "grad_norm": 1.2620969300769502e-06, "learning_rate": 3.411359090512261e-05, "loss": 0.0, "num_input_tokens_seen": 4269600, "step": 15250 }, { "epoch": 169.5, "grad_norm": 4.145214916206896e-06, "learning_rate": 3.410444830245672e-05, "loss": 0.0, "num_input_tokens_seen": 4270992, "step": 15255 }, { "epoch": 169.55555555555554, "grad_norm": 1.287427267016028e-06, "learning_rate": 3.409530429577048e-05, "loss": 0.0, "num_input_tokens_seen": 4272368, "step": 15260 }, { "epoch": 169.61111111111111, "grad_norm": 1.5332184375438374e-06, "learning_rate": 3.408615888647402e-05, "loss": 0.0, "num_input_tokens_seen": 4273792, "step": 15265 }, { "epoch": 169.66666666666666, "grad_norm": 2.475624796716147e-06, "learning_rate": 3.4077012075977675e-05, "loss": 0.0, "num_input_tokens_seen": 4275200, "step": 15270 }, { "epoch": 169.72222222222223, "grad_norm": 1.2200993069200194e-06, "learning_rate": 3.4067863865692e-05, "loss": 0.0, "num_input_tokens_seen": 4276592, "step": 15275 }, { "epoch": 169.77777777777777, "grad_norm": 1.4567317521141376e-06, "learning_rate": 3.4058714257027755e-05, "loss": 0.0, "num_input_tokens_seen": 4277952, "step": 15280 }, { "epoch": 169.83333333333334, "grad_norm": 2.872681761800777e-06, "learning_rate": 3.404956325139594e-05, "loss": 0.0, "num_input_tokens_seen": 4279360, "step": 15285 }, { "epoch": 169.88888888888889, "grad_norm": 1.7234633560292423e-06, "learning_rate": 3.404041085020775e-05, "loss": 0.0, "num_input_tokens_seen": 4280752, "step": 15290 }, { "epoch": 169.94444444444446, "grad_norm": 1.5459622773050796e-06, "learning_rate": 3.403125705487459e-05, "loss": 0.0, "num_input_tokens_seen": 4282160, "step": 15295 }, { "epoch": 170.0, "grad_norm": 1.7393207372151664e-06, "learning_rate": 3.402210186680811e-05, "loss": 0.0, "num_input_tokens_seen": 4283584, "step": 15300 }, { "epoch": 170.05555555555554, "grad_norm": 2.215035465269466e-06, "learning_rate": 3.4012945287420137e-05, "loss": 0.0, "num_input_tokens_seen": 4284912, "step": 15305 }, { "epoch": 170.11111111111111, "grad_norm": 1.1536110378074227e-06, "learning_rate": 3.400378731812274e-05, "loss": 0.0, "num_input_tokens_seen": 4286304, "step": 15310 }, { "epoch": 170.16666666666666, "grad_norm": 1.625137883820571e-06, "learning_rate": 3.399462796032817e-05, "loss": 0.0, "num_input_tokens_seen": 4287696, "step": 15315 }, { "epoch": 170.22222222222223, "grad_norm": 2.1543694401771063e-06, "learning_rate": 3.3985467215448954e-05, "loss": 0.0, "num_input_tokens_seen": 4289104, "step": 15320 }, { "epoch": 170.27777777777777, "grad_norm": 1.8232500451631495e-06, "learning_rate": 3.3976305084897776e-05, "loss": 0.0, "num_input_tokens_seen": 4290496, "step": 15325 }, { "epoch": 170.33333333333334, "grad_norm": 1.8634908656167681e-06, "learning_rate": 3.3967141570087544e-05, "loss": 0.0, "num_input_tokens_seen": 4291872, "step": 15330 }, { "epoch": 170.38888888888889, "grad_norm": 2.1190223833400523e-06, "learning_rate": 3.39579766724314e-05, "loss": 0.0, "num_input_tokens_seen": 4293248, "step": 15335 }, { "epoch": 170.44444444444446, "grad_norm": 1.8633620584296295e-06, "learning_rate": 3.3948810393342677e-05, "loss": 0.0, "num_input_tokens_seen": 4294640, "step": 15340 }, { "epoch": 170.5, "grad_norm": 1.513438633082842e-06, "learning_rate": 3.3939642734234936e-05, "loss": 0.0, "num_input_tokens_seen": 4296064, "step": 15345 }, { "epoch": 170.55555555555554, "grad_norm": 1.3344355238587013e-06, "learning_rate": 3.393047369652194e-05, "loss": 0.0, "num_input_tokens_seen": 4297472, "step": 15350 }, { "epoch": 170.61111111111111, "grad_norm": 3.0420953862630995e-06, "learning_rate": 3.3921303281617664e-05, "loss": 0.0, "num_input_tokens_seen": 4298912, "step": 15355 }, { "epoch": 170.66666666666666, "grad_norm": 1.659047029534122e-06, "learning_rate": 3.391213149093632e-05, "loss": 0.0, "num_input_tokens_seen": 4300320, "step": 15360 }, { "epoch": 170.72222222222223, "grad_norm": 7.493132670788327e-06, "learning_rate": 3.3902958325892303e-05, "loss": 0.0, "num_input_tokens_seen": 4301712, "step": 15365 }, { "epoch": 170.77777777777777, "grad_norm": 2.593774070192012e-06, "learning_rate": 3.389378378790023e-05, "loss": 0.0, "num_input_tokens_seen": 4303136, "step": 15370 }, { "epoch": 170.83333333333334, "grad_norm": 1.7671943624009145e-06, "learning_rate": 3.388460787837493e-05, "loss": 0.0, "num_input_tokens_seen": 4304512, "step": 15375 }, { "epoch": 170.88888888888889, "grad_norm": 2.6836569304578006e-06, "learning_rate": 3.387543059873145e-05, "loss": 0.0, "num_input_tokens_seen": 4305920, "step": 15380 }, { "epoch": 170.94444444444446, "grad_norm": 1.0461793635840877e-06, "learning_rate": 3.386625195038503e-05, "loss": 0.0, "num_input_tokens_seen": 4307312, "step": 15385 }, { "epoch": 171.0, "grad_norm": 9.899200676954933e-07, "learning_rate": 3.3857071934751136e-05, "loss": 0.0, "num_input_tokens_seen": 4308752, "step": 15390 }, { "epoch": 171.05555555555554, "grad_norm": 2.3981644972081995e-06, "learning_rate": 3.384789055324544e-05, "loss": 0.0, "num_input_tokens_seen": 4310144, "step": 15395 }, { "epoch": 171.11111111111111, "grad_norm": 1.8779029460347374e-06, "learning_rate": 3.3838707807283843e-05, "loss": 0.0, "num_input_tokens_seen": 4311504, "step": 15400 }, { "epoch": 171.11111111111111, "eval_loss": 1.2953566312789917, "eval_runtime": 1.3865, "eval_samples_per_second": 28.849, "eval_steps_per_second": 14.425, "num_input_tokens_seen": 4311504, "step": 15400 }, { "epoch": 171.16666666666666, "grad_norm": 1.9934300325985532e-06, "learning_rate": 3.382952369828243e-05, "loss": 0.0, "num_input_tokens_seen": 4312880, "step": 15405 }, { "epoch": 171.22222222222223, "grad_norm": 1.0376541013101814e-06, "learning_rate": 3.38203382276575e-05, "loss": 0.0, "num_input_tokens_seen": 4314272, "step": 15410 }, { "epoch": 171.27777777777777, "grad_norm": 3.091746521022287e-07, "learning_rate": 3.381115139682557e-05, "loss": 0.0, "num_input_tokens_seen": 4315632, "step": 15415 }, { "epoch": 171.33333333333334, "grad_norm": 9.062922003977292e-07, "learning_rate": 3.3801963207203366e-05, "loss": 0.0, "num_input_tokens_seen": 4317008, "step": 15420 }, { "epoch": 171.38888888888889, "grad_norm": 1.1630418157437816e-06, "learning_rate": 3.379277366020782e-05, "loss": 0.0, "num_input_tokens_seen": 4318448, "step": 15425 }, { "epoch": 171.44444444444446, "grad_norm": 5.314933559930068e-07, "learning_rate": 3.3783582757256085e-05, "loss": 0.0, "num_input_tokens_seen": 4319872, "step": 15430 }, { "epoch": 171.5, "grad_norm": 1.3670949101651786e-06, "learning_rate": 3.3774390499765504e-05, "loss": 0.0, "num_input_tokens_seen": 4321264, "step": 15435 }, { "epoch": 171.55555555555554, "grad_norm": 2.2259664547163993e-06, "learning_rate": 3.376519688915364e-05, "loss": 0.0, "num_input_tokens_seen": 4322656, "step": 15440 }, { "epoch": 171.61111111111111, "grad_norm": 1.4458718169407803e-06, "learning_rate": 3.3756001926838273e-05, "loss": 0.0, "num_input_tokens_seen": 4324048, "step": 15445 }, { "epoch": 171.66666666666666, "grad_norm": 4.532858383754501e-06, "learning_rate": 3.374680561423737e-05, "loss": 0.0, "num_input_tokens_seen": 4325424, "step": 15450 }, { "epoch": 171.72222222222223, "grad_norm": 6.998064350227651e-07, "learning_rate": 3.373760795276912e-05, "loss": 0.0, "num_input_tokens_seen": 4326848, "step": 15455 }, { "epoch": 171.77777777777777, "grad_norm": 3.7171694202697836e-06, "learning_rate": 3.372840894385192e-05, "loss": 0.0, "num_input_tokens_seen": 4328240, "step": 15460 }, { "epoch": 171.83333333333334, "grad_norm": 1.2569855698529864e-06, "learning_rate": 3.3719208588904375e-05, "loss": 0.0, "num_input_tokens_seen": 4329616, "step": 15465 }, { "epoch": 171.88888888888889, "grad_norm": 5.91926379911456e-07, "learning_rate": 3.371000688934529e-05, "loss": 0.0, "num_input_tokens_seen": 4330992, "step": 15470 }, { "epoch": 171.94444444444446, "grad_norm": 1.6694662008376326e-06, "learning_rate": 3.370080384659369e-05, "loss": 0.0, "num_input_tokens_seen": 4332400, "step": 15475 }, { "epoch": 172.0, "grad_norm": 1.4279984270615387e-06, "learning_rate": 3.36915994620688e-05, "loss": 0.0, "num_input_tokens_seen": 4333808, "step": 15480 }, { "epoch": 172.05555555555554, "grad_norm": 3.153428906443878e-06, "learning_rate": 3.3682393737190035e-05, "loss": 0.0, "num_input_tokens_seen": 4335216, "step": 15485 }, { "epoch": 172.11111111111111, "grad_norm": 3.471417585387826e-06, "learning_rate": 3.3673186673377054e-05, "loss": 0.0, "num_input_tokens_seen": 4336624, "step": 15490 }, { "epoch": 172.16666666666666, "grad_norm": 7.996779913810315e-07, "learning_rate": 3.366397827204969e-05, "loss": 0.0, "num_input_tokens_seen": 4338000, "step": 15495 }, { "epoch": 172.22222222222223, "grad_norm": 1.1205925147805829e-06, "learning_rate": 3.3654768534628e-05, "loss": 0.0, "num_input_tokens_seen": 4339408, "step": 15500 }, { "epoch": 172.27777777777777, "grad_norm": 2.740026047831634e-06, "learning_rate": 3.3645557462532245e-05, "loss": 0.0, "num_input_tokens_seen": 4340864, "step": 15505 }, { "epoch": 172.33333333333334, "grad_norm": 1.7954923805518774e-06, "learning_rate": 3.363634505718288e-05, "loss": 0.0, "num_input_tokens_seen": 4342272, "step": 15510 }, { "epoch": 172.38888888888889, "grad_norm": 1.3261685580800986e-06, "learning_rate": 3.362713132000057e-05, "loss": 0.0, "num_input_tokens_seen": 4343680, "step": 15515 }, { "epoch": 172.44444444444446, "grad_norm": 9.602323416402214e-07, "learning_rate": 3.36179162524062e-05, "loss": 0.0, "num_input_tokens_seen": 4345104, "step": 15520 }, { "epoch": 172.5, "grad_norm": 1.3531845297620748e-06, "learning_rate": 3.3608699855820846e-05, "loss": 0.0, "num_input_tokens_seen": 4346496, "step": 15525 }, { "epoch": 172.55555555555554, "grad_norm": 1.8479546497474075e-06, "learning_rate": 3.359948213166578e-05, "loss": 0.0, "num_input_tokens_seen": 4347872, "step": 15530 }, { "epoch": 172.61111111111111, "grad_norm": 1.2780440101778368e-06, "learning_rate": 3.359026308136252e-05, "loss": 0.0, "num_input_tokens_seen": 4349248, "step": 15535 }, { "epoch": 172.66666666666666, "grad_norm": 1.1169585150128114e-06, "learning_rate": 3.358104270633272e-05, "loss": 0.0, "num_input_tokens_seen": 4350656, "step": 15540 }, { "epoch": 172.72222222222223, "grad_norm": 3.5101736557408003e-06, "learning_rate": 3.357182100799831e-05, "loss": 0.0, "num_input_tokens_seen": 4352032, "step": 15545 }, { "epoch": 172.77777777777777, "grad_norm": 7.070476840453921e-06, "learning_rate": 3.3562597987781384e-05, "loss": 0.0, "num_input_tokens_seen": 4353456, "step": 15550 }, { "epoch": 172.83333333333334, "grad_norm": 2.5037645627890015e-06, "learning_rate": 3.355337364710424e-05, "loss": 0.0, "num_input_tokens_seen": 4354800, "step": 15555 }, { "epoch": 172.88888888888889, "grad_norm": 1.8667711856323876e-06, "learning_rate": 3.354414798738939e-05, "loss": 0.0, "num_input_tokens_seen": 4356160, "step": 15560 }, { "epoch": 172.94444444444446, "grad_norm": 1.425628852302907e-06, "learning_rate": 3.353492101005955e-05, "loss": 0.0, "num_input_tokens_seen": 4357616, "step": 15565 }, { "epoch": 173.0, "grad_norm": 1.4379675121745095e-06, "learning_rate": 3.352569271653763e-05, "loss": 0.0, "num_input_tokens_seen": 4358992, "step": 15570 }, { "epoch": 173.05555555555554, "grad_norm": 1.7556626517034601e-06, "learning_rate": 3.351646310824675e-05, "loss": 0.0, "num_input_tokens_seen": 4360464, "step": 15575 }, { "epoch": 173.11111111111111, "grad_norm": 5.499942403730529e-07, "learning_rate": 3.350723218661023e-05, "loss": 0.0, "num_input_tokens_seen": 4361856, "step": 15580 }, { "epoch": 173.16666666666666, "grad_norm": 2.089621148115839e-06, "learning_rate": 3.349799995305162e-05, "loss": 0.0, "num_input_tokens_seen": 4363248, "step": 15585 }, { "epoch": 173.22222222222223, "grad_norm": 5.040173505221901e-07, "learning_rate": 3.348876640899461e-05, "loss": 0.0, "num_input_tokens_seen": 4364624, "step": 15590 }, { "epoch": 173.27777777777777, "grad_norm": 5.018320621275052e-07, "learning_rate": 3.3479531555863144e-05, "loss": 0.0, "num_input_tokens_seen": 4366048, "step": 15595 }, { "epoch": 173.33333333333334, "grad_norm": 1.243583483301336e-06, "learning_rate": 3.3470295395081344e-05, "loss": 0.0, "num_input_tokens_seen": 4367408, "step": 15600 }, { "epoch": 173.33333333333334, "eval_loss": 1.2587699890136719, "eval_runtime": 1.399, "eval_samples_per_second": 28.593, "eval_steps_per_second": 14.296, "num_input_tokens_seen": 4367408, "step": 15600 }, { "epoch": 173.38888888888889, "grad_norm": 1.6039655292843236e-06, "learning_rate": 3.3461057928073556e-05, "loss": 0.0, "num_input_tokens_seen": 4368832, "step": 15605 }, { "epoch": 173.44444444444446, "grad_norm": 6.296208994172048e-06, "learning_rate": 3.345181915626431e-05, "loss": 0.0, "num_input_tokens_seen": 4370224, "step": 15610 }, { "epoch": 173.5, "grad_norm": 7.541838385805022e-06, "learning_rate": 3.344257908107834e-05, "loss": 0.0, "num_input_tokens_seen": 4371648, "step": 15615 }, { "epoch": 173.55555555555554, "grad_norm": 3.977973392466083e-06, "learning_rate": 3.343333770394058e-05, "loss": 0.0, "num_input_tokens_seen": 4373008, "step": 15620 }, { "epoch": 173.61111111111111, "grad_norm": 3.669985346732574e-07, "learning_rate": 3.342409502627616e-05, "loss": 0.0, "num_input_tokens_seen": 4374400, "step": 15625 }, { "epoch": 173.66666666666666, "grad_norm": 1.1821646239695838e-06, "learning_rate": 3.341485104951043e-05, "loss": 0.0, "num_input_tokens_seen": 4375808, "step": 15630 }, { "epoch": 173.72222222222223, "grad_norm": 8.626940939393535e-07, "learning_rate": 3.340560577506892e-05, "loss": 0.0, "num_input_tokens_seen": 4377200, "step": 15635 }, { "epoch": 173.77777777777777, "grad_norm": 1.250577838618483e-06, "learning_rate": 3.339635920437735e-05, "loss": 0.0, "num_input_tokens_seen": 4378560, "step": 15640 }, { "epoch": 173.83333333333334, "grad_norm": 5.683831432179431e-07, "learning_rate": 3.338711133886169e-05, "loss": 0.0, "num_input_tokens_seen": 4379936, "step": 15645 }, { "epoch": 173.88888888888889, "grad_norm": 1.0077629895022255e-06, "learning_rate": 3.3377862179948064e-05, "loss": 0.0, "num_input_tokens_seen": 4381360, "step": 15650 }, { "epoch": 173.94444444444446, "grad_norm": 8.566638825868722e-06, "learning_rate": 3.336861172906281e-05, "loss": 0.0, "num_input_tokens_seen": 4382784, "step": 15655 }, { "epoch": 174.0, "grad_norm": 2.472924961693934e-06, "learning_rate": 3.335935998763245e-05, "loss": 0.0, "num_input_tokens_seen": 4384176, "step": 15660 }, { "epoch": 174.05555555555554, "grad_norm": 2.06217669074249e-06, "learning_rate": 3.3350106957083744e-05, "loss": 0.0, "num_input_tokens_seen": 4385536, "step": 15665 }, { "epoch": 174.11111111111111, "grad_norm": 9.969585335056763e-07, "learning_rate": 3.33408526388436e-05, "loss": 0.0, "num_input_tokens_seen": 4386928, "step": 15670 }, { "epoch": 174.16666666666666, "grad_norm": 3.585426156860194e-06, "learning_rate": 3.3331597034339166e-05, "loss": 0.0, "num_input_tokens_seen": 4388352, "step": 15675 }, { "epoch": 174.22222222222223, "grad_norm": 8.59656324792013e-07, "learning_rate": 3.3322340144997764e-05, "loss": 0.0, "num_input_tokens_seen": 4389760, "step": 15680 }, { "epoch": 174.27777777777777, "grad_norm": 1.4605526530431234e-06, "learning_rate": 3.331308197224693e-05, "loss": 0.0, "num_input_tokens_seen": 4391136, "step": 15685 }, { "epoch": 174.33333333333334, "grad_norm": 1.59477349370718e-06, "learning_rate": 3.330382251751438e-05, "loss": 0.0, "num_input_tokens_seen": 4392528, "step": 15690 }, { "epoch": 174.38888888888889, "grad_norm": 1.206685396937246e-06, "learning_rate": 3.3294561782228054e-05, "loss": 0.0, "num_input_tokens_seen": 4393856, "step": 15695 }, { "epoch": 174.44444444444446, "grad_norm": 1.5443972642970039e-06, "learning_rate": 3.328529976781607e-05, "loss": 0.0, "num_input_tokens_seen": 4395232, "step": 15700 }, { "epoch": 174.5, "grad_norm": 1.2647864195969305e-06, "learning_rate": 3.327603647570673e-05, "loss": 0.0, "num_input_tokens_seen": 4396688, "step": 15705 }, { "epoch": 174.55555555555554, "grad_norm": 1.2038348131682142e-06, "learning_rate": 3.326677190732857e-05, "loss": 0.0, "num_input_tokens_seen": 4398096, "step": 15710 }, { "epoch": 174.61111111111111, "grad_norm": 1.2429886737663765e-06, "learning_rate": 3.325750606411029e-05, "loss": 0.0, "num_input_tokens_seen": 4399520, "step": 15715 }, { "epoch": 174.66666666666666, "grad_norm": 1.2157039464000263e-06, "learning_rate": 3.3248238947480804e-05, "loss": 0.0, "num_input_tokens_seen": 4400912, "step": 15720 }, { "epoch": 174.72222222222223, "grad_norm": 4.651497818031203e-07, "learning_rate": 3.323897055886922e-05, "loss": 0.0, "num_input_tokens_seen": 4402336, "step": 15725 }, { "epoch": 174.77777777777777, "grad_norm": 2.483973730704747e-06, "learning_rate": 3.322970089970484e-05, "loss": 0.0, "num_input_tokens_seen": 4403712, "step": 15730 }, { "epoch": 174.83333333333334, "grad_norm": 2.1068040041427594e-06, "learning_rate": 3.3220429971417165e-05, "loss": 0.0, "num_input_tokens_seen": 4405152, "step": 15735 }, { "epoch": 174.88888888888889, "grad_norm": 1.920227759910631e-06, "learning_rate": 3.321115777543588e-05, "loss": 0.0, "num_input_tokens_seen": 4406544, "step": 15740 }, { "epoch": 174.94444444444446, "grad_norm": 6.311259426183824e-07, "learning_rate": 3.320188431319088e-05, "loss": 0.0, "num_input_tokens_seen": 4408016, "step": 15745 }, { "epoch": 175.0, "grad_norm": 3.068548494411516e-06, "learning_rate": 3.319260958611224e-05, "loss": 0.0, "num_input_tokens_seen": 4409408, "step": 15750 }, { "epoch": 175.05555555555554, "grad_norm": 6.033992008269706e-07, "learning_rate": 3.3183333595630256e-05, "loss": 0.0, "num_input_tokens_seen": 4410848, "step": 15755 }, { "epoch": 175.11111111111111, "grad_norm": 1.1873022458530613e-06, "learning_rate": 3.317405634317538e-05, "loss": 0.0, "num_input_tokens_seen": 4412272, "step": 15760 }, { "epoch": 175.16666666666666, "grad_norm": 2.9184579943830613e-06, "learning_rate": 3.3164777830178315e-05, "loss": 0.0, "num_input_tokens_seen": 4413632, "step": 15765 }, { "epoch": 175.22222222222223, "grad_norm": 1.5482497701668763e-06, "learning_rate": 3.315549805806989e-05, "loss": 0.0, "num_input_tokens_seen": 4415008, "step": 15770 }, { "epoch": 175.27777777777777, "grad_norm": 8.414568810621859e-07, "learning_rate": 3.314621702828118e-05, "loss": 0.0, "num_input_tokens_seen": 4416400, "step": 15775 }, { "epoch": 175.33333333333334, "grad_norm": 4.6613763515779283e-07, "learning_rate": 3.313693474224342e-05, "loss": 0.0, "num_input_tokens_seen": 4417808, "step": 15780 }, { "epoch": 175.38888888888889, "grad_norm": 2.3933105239848373e-06, "learning_rate": 3.312765120138809e-05, "loss": 0.0, "num_input_tokens_seen": 4419184, "step": 15785 }, { "epoch": 175.44444444444446, "grad_norm": 1.0546339126449311e-06, "learning_rate": 3.311836640714679e-05, "loss": 0.0, "num_input_tokens_seen": 4420608, "step": 15790 }, { "epoch": 175.5, "grad_norm": 1.2237627515787608e-06, "learning_rate": 3.310908036095137e-05, "loss": 0.0, "num_input_tokens_seen": 4422000, "step": 15795 }, { "epoch": 175.55555555555554, "grad_norm": 1.490293243477936e-06, "learning_rate": 3.309979306423386e-05, "loss": 0.0, "num_input_tokens_seen": 4423376, "step": 15800 }, { "epoch": 175.55555555555554, "eval_loss": 1.2669756412506104, "eval_runtime": 1.3868, "eval_samples_per_second": 28.843, "eval_steps_per_second": 14.421, "num_input_tokens_seen": 4423376, "step": 15800 }, { "epoch": 175.61111111111111, "grad_norm": 4.363375410321169e-06, "learning_rate": 3.309050451842647e-05, "loss": 0.0, "num_input_tokens_seen": 4424800, "step": 15805 }, { "epoch": 175.66666666666666, "grad_norm": 1.5694478179284488e-06, "learning_rate": 3.3081214724961604e-05, "loss": 0.0, "num_input_tokens_seen": 4426208, "step": 15810 }, { "epoch": 175.72222222222223, "grad_norm": 1.8155977841161075e-06, "learning_rate": 3.307192368527188e-05, "loss": 0.0, "num_input_tokens_seen": 4427600, "step": 15815 }, { "epoch": 175.77777777777777, "grad_norm": 7.131734491849784e-07, "learning_rate": 3.306263140079008e-05, "loss": 0.0, "num_input_tokens_seen": 4428976, "step": 15820 }, { "epoch": 175.83333333333334, "grad_norm": 1.126998768086196e-06, "learning_rate": 3.30533378729492e-05, "loss": 0.0, "num_input_tokens_seen": 4430368, "step": 15825 }, { "epoch": 175.88888888888889, "grad_norm": 9.136032304013497e-07, "learning_rate": 3.304404310318242e-05, "loss": 0.0, "num_input_tokens_seen": 4431792, "step": 15830 }, { "epoch": 175.94444444444446, "grad_norm": 7.670489594602259e-07, "learning_rate": 3.3034747092923105e-05, "loss": 0.0, "num_input_tokens_seen": 4433168, "step": 15835 }, { "epoch": 176.0, "grad_norm": 1.9908964077330893e-06, "learning_rate": 3.3025449843604806e-05, "loss": 0.0, "num_input_tokens_seen": 4434592, "step": 15840 }, { "epoch": 176.05555555555554, "grad_norm": 1.5151279058045475e-06, "learning_rate": 3.30161513566613e-05, "loss": 0.0, "num_input_tokens_seen": 4435984, "step": 15845 }, { "epoch": 176.11111111111111, "grad_norm": 8.331619483215036e-07, "learning_rate": 3.3006851633526506e-05, "loss": 0.0, "num_input_tokens_seen": 4437376, "step": 15850 }, { "epoch": 176.16666666666666, "grad_norm": 3.6222607491254166e-07, "learning_rate": 3.2997550675634584e-05, "loss": 0.0, "num_input_tokens_seen": 4438784, "step": 15855 }, { "epoch": 176.22222222222223, "grad_norm": 6.462859687417222e-07, "learning_rate": 3.2988248484419825e-05, "loss": 0.0, "num_input_tokens_seen": 4440224, "step": 15860 }, { "epoch": 176.27777777777777, "grad_norm": 1.221701609210868e-06, "learning_rate": 3.2978945061316776e-05, "loss": 0.0, "num_input_tokens_seen": 4441632, "step": 15865 }, { "epoch": 176.33333333333334, "grad_norm": 1.1308544571875245e-06, "learning_rate": 3.296964040776013e-05, "loss": 0.0, "num_input_tokens_seen": 4443088, "step": 15870 }, { "epoch": 176.38888888888889, "grad_norm": 5.96844927258644e-07, "learning_rate": 3.296033452518478e-05, "loss": 0.0, "num_input_tokens_seen": 4444496, "step": 15875 }, { "epoch": 176.44444444444446, "grad_norm": 5.01416820952727e-07, "learning_rate": 3.2951027415025806e-05, "loss": 0.0, "num_input_tokens_seen": 4445888, "step": 15880 }, { "epoch": 176.5, "grad_norm": 9.779988658920047e-07, "learning_rate": 3.294171907871849e-05, "loss": 0.0, "num_input_tokens_seen": 4447328, "step": 15885 }, { "epoch": 176.55555555555554, "grad_norm": 1.5741921970402473e-06, "learning_rate": 3.293240951769828e-05, "loss": 0.0, "num_input_tokens_seen": 4448704, "step": 15890 }, { "epoch": 176.61111111111111, "grad_norm": 3.2665013804944465e-06, "learning_rate": 3.2923098733400846e-05, "loss": 0.0, "num_input_tokens_seen": 4450128, "step": 15895 }, { "epoch": 176.66666666666666, "grad_norm": 5.339302333595697e-06, "learning_rate": 3.291378672726202e-05, "loss": 0.0, "num_input_tokens_seen": 4451584, "step": 15900 }, { "epoch": 176.72222222222223, "grad_norm": 1.8485044392946293e-06, "learning_rate": 3.2904473500717824e-05, "loss": 0.0, "num_input_tokens_seen": 4452944, "step": 15905 }, { "epoch": 176.77777777777777, "grad_norm": 6.890662689329474e-07, "learning_rate": 3.289515905520449e-05, "loss": 0.0, "num_input_tokens_seen": 4454320, "step": 15910 }, { "epoch": 176.83333333333334, "grad_norm": 1.6110788010337274e-06, "learning_rate": 3.288584339215841e-05, "loss": 0.0, "num_input_tokens_seen": 4455664, "step": 15915 }, { "epoch": 176.88888888888889, "grad_norm": 8.802102797744737e-07, "learning_rate": 3.287652651301617e-05, "loss": 0.0, "num_input_tokens_seen": 4457056, "step": 15920 }, { "epoch": 176.94444444444446, "grad_norm": 2.681757450773148e-06, "learning_rate": 3.286720841921457e-05, "loss": 0.0, "num_input_tokens_seen": 4458448, "step": 15925 }, { "epoch": 177.0, "grad_norm": 1.2321944495852222e-06, "learning_rate": 3.285788911219056e-05, "loss": 0.0, "num_input_tokens_seen": 4459856, "step": 15930 }, { "epoch": 177.05555555555554, "grad_norm": 2.720878683248884e-06, "learning_rate": 3.284856859338131e-05, "loss": 0.0, "num_input_tokens_seen": 4461264, "step": 15935 }, { "epoch": 177.11111111111111, "grad_norm": 2.4206210582633503e-06, "learning_rate": 3.283924686422414e-05, "loss": 0.0, "num_input_tokens_seen": 4462656, "step": 15940 }, { "epoch": 177.16666666666666, "grad_norm": 3.7873914493502525e-07, "learning_rate": 3.282992392615659e-05, "loss": 0.0, "num_input_tokens_seen": 4464080, "step": 15945 }, { "epoch": 177.22222222222223, "grad_norm": 7.42038173484616e-06, "learning_rate": 3.282059978061638e-05, "loss": 0.0, "num_input_tokens_seen": 4465520, "step": 15950 }, { "epoch": 177.27777777777777, "grad_norm": 7.605889322803705e-07, "learning_rate": 3.28112744290414e-05, "loss": 0.0, "num_input_tokens_seen": 4466896, "step": 15955 }, { "epoch": 177.33333333333334, "grad_norm": 6.921086423972156e-07, "learning_rate": 3.280194787286974e-05, "loss": 0.0, "num_input_tokens_seen": 4468336, "step": 15960 }, { "epoch": 177.38888888888889, "grad_norm": 2.478604756106506e-06, "learning_rate": 3.2792620113539674e-05, "loss": 0.0, "num_input_tokens_seen": 4469728, "step": 15965 }, { "epoch": 177.44444444444446, "grad_norm": 1.2656204262384563e-06, "learning_rate": 3.278329115248966e-05, "loss": 0.0, "num_input_tokens_seen": 4471104, "step": 15970 }, { "epoch": 177.5, "grad_norm": 1.1799846788562718e-06, "learning_rate": 3.277396099115834e-05, "loss": 0.0, "num_input_tokens_seen": 4472464, "step": 15975 }, { "epoch": 177.55555555555554, "grad_norm": 9.923292054736521e-07, "learning_rate": 3.276462963098454e-05, "loss": 0.0, "num_input_tokens_seen": 4473872, "step": 15980 }, { "epoch": 177.61111111111111, "grad_norm": 2.3739626158203464e-06, "learning_rate": 3.275529707340728e-05, "loss": 0.0, "num_input_tokens_seen": 4475264, "step": 15985 }, { "epoch": 177.66666666666666, "grad_norm": 1.0731881729952875e-06, "learning_rate": 3.274596331986574e-05, "loss": 0.0, "num_input_tokens_seen": 4476624, "step": 15990 }, { "epoch": 177.72222222222223, "grad_norm": 1.375874944642419e-06, "learning_rate": 3.273662837179932e-05, "loss": 0.0, "num_input_tokens_seen": 4478032, "step": 15995 }, { "epoch": 177.77777777777777, "grad_norm": 1.2073443258486805e-06, "learning_rate": 3.272729223064758e-05, "loss": 0.0, "num_input_tokens_seen": 4479456, "step": 16000 }, { "epoch": 177.77777777777777, "eval_loss": 1.2695128917694092, "eval_runtime": 1.3871, "eval_samples_per_second": 28.837, "eval_steps_per_second": 14.418, "num_input_tokens_seen": 4479456, "step": 16000 }, { "epoch": 177.83333333333334, "grad_norm": 1.1067751302107354e-06, "learning_rate": 3.2717954897850264e-05, "loss": 0.0, "num_input_tokens_seen": 4480816, "step": 16005 }, { "epoch": 177.88888888888889, "grad_norm": 1.805754322958819e-06, "learning_rate": 3.270861637484733e-05, "loss": 0.0, "num_input_tokens_seen": 4482176, "step": 16010 }, { "epoch": 177.94444444444446, "grad_norm": 1.2829690376747749e-06, "learning_rate": 3.2699276663078867e-05, "loss": 0.0, "num_input_tokens_seen": 4483600, "step": 16015 }, { "epoch": 178.0, "grad_norm": 1.3499363831215305e-06, "learning_rate": 3.268993576398519e-05, "loss": 0.0, "num_input_tokens_seen": 4485008, "step": 16020 }, { "epoch": 178.05555555555554, "grad_norm": 1.4438992366194725e-06, "learning_rate": 3.268059367900678e-05, "loss": 0.0, "num_input_tokens_seen": 4486416, "step": 16025 }, { "epoch": 178.11111111111111, "grad_norm": 6.956490778975422e-07, "learning_rate": 3.26712504095843e-05, "loss": 0.0, "num_input_tokens_seen": 4487840, "step": 16030 }, { "epoch": 178.16666666666666, "grad_norm": 9.621285244065803e-07, "learning_rate": 3.2661905957158615e-05, "loss": 0.0, "num_input_tokens_seen": 4489280, "step": 16035 }, { "epoch": 178.22222222222223, "grad_norm": 8.19996387235733e-07, "learning_rate": 3.2652560323170734e-05, "loss": 0.0, "num_input_tokens_seen": 4490704, "step": 16040 }, { "epoch": 178.27777777777777, "grad_norm": 1.397909159095434e-06, "learning_rate": 3.264321350906189e-05, "loss": 0.0, "num_input_tokens_seen": 4492096, "step": 16045 }, { "epoch": 178.33333333333334, "grad_norm": 1.041227619680285e-06, "learning_rate": 3.263386551627346e-05, "loss": 0.0, "num_input_tokens_seen": 4493504, "step": 16050 }, { "epoch": 178.38888888888889, "grad_norm": 3.3871406230900902e-06, "learning_rate": 3.2624516346247055e-05, "loss": 0.0, "num_input_tokens_seen": 4494912, "step": 16055 }, { "epoch": 178.44444444444446, "grad_norm": 1.9184531652172154e-07, "learning_rate": 3.2615166000424404e-05, "loss": 0.0, "num_input_tokens_seen": 4496272, "step": 16060 }, { "epoch": 178.5, "grad_norm": 1.182332766802574e-06, "learning_rate": 3.260581448024745e-05, "loss": 0.0, "num_input_tokens_seen": 4497712, "step": 16065 }, { "epoch": 178.55555555555554, "grad_norm": 2.693201111014787e-07, "learning_rate": 3.2596461787158335e-05, "loss": 0.0, "num_input_tokens_seen": 4499088, "step": 16070 }, { "epoch": 178.61111111111111, "grad_norm": 1.3555583109337022e-06, "learning_rate": 3.258710792259934e-05, "loss": 0.0, "num_input_tokens_seen": 4500480, "step": 16075 }, { "epoch": 178.66666666666666, "grad_norm": 1.3394827647061902e-06, "learning_rate": 3.257775288801296e-05, "loss": 0.0, "num_input_tokens_seen": 4501872, "step": 16080 }, { "epoch": 178.72222222222223, "grad_norm": 1.0223960771327256e-06, "learning_rate": 3.256839668484186e-05, "loss": 0.0, "num_input_tokens_seen": 4503264, "step": 16085 }, { "epoch": 178.77777777777777, "grad_norm": 9.337447295365564e-07, "learning_rate": 3.255903931452888e-05, "loss": 0.0, "num_input_tokens_seen": 4504672, "step": 16090 }, { "epoch": 178.83333333333334, "grad_norm": 2.641863829921931e-06, "learning_rate": 3.2549680778517045e-05, "loss": 0.0, "num_input_tokens_seen": 4506016, "step": 16095 }, { "epoch": 178.88888888888889, "grad_norm": 1.3506869436241686e-06, "learning_rate": 3.2540321078249556e-05, "loss": 0.0, "num_input_tokens_seen": 4507472, "step": 16100 }, { "epoch": 178.94444444444446, "grad_norm": 4.910954771730758e-07, "learning_rate": 3.2530960215169795e-05, "loss": 0.0, "num_input_tokens_seen": 4508848, "step": 16105 }, { "epoch": 179.0, "grad_norm": 1.2261739357199986e-06, "learning_rate": 3.2521598190721345e-05, "loss": 0.0, "num_input_tokens_seen": 4510240, "step": 16110 }, { "epoch": 179.05555555555554, "grad_norm": 7.55548057895794e-07, "learning_rate": 3.251223500634792e-05, "loss": 0.0, "num_input_tokens_seen": 4511600, "step": 16115 }, { "epoch": 179.11111111111111, "grad_norm": 4.458037892618449e-06, "learning_rate": 3.2502870663493445e-05, "loss": 0.0, "num_input_tokens_seen": 4512976, "step": 16120 }, { "epoch": 179.16666666666666, "grad_norm": 8.11088341379218e-07, "learning_rate": 3.249350516360203e-05, "loss": 0.0, "num_input_tokens_seen": 4514368, "step": 16125 }, { "epoch": 179.22222222222223, "grad_norm": 1.1268311936873943e-06, "learning_rate": 3.248413850811797e-05, "loss": 0.0, "num_input_tokens_seen": 4515760, "step": 16130 }, { "epoch": 179.27777777777777, "grad_norm": 1.1227162985960604e-06, "learning_rate": 3.2474770698485677e-05, "loss": 0.0, "num_input_tokens_seen": 4517168, "step": 16135 }, { "epoch": 179.33333333333334, "grad_norm": 1.1666030559354112e-06, "learning_rate": 3.246540173614983e-05, "loss": 0.0, "num_input_tokens_seen": 4518576, "step": 16140 }, { "epoch": 179.38888888888889, "grad_norm": 5.283825430524303e-07, "learning_rate": 3.2456031622555197e-05, "loss": 0.0, "num_input_tokens_seen": 4519968, "step": 16145 }, { "epoch": 179.44444444444446, "grad_norm": 8.988865829451242e-07, "learning_rate": 3.2446660359146794e-05, "loss": 0.0, "num_input_tokens_seen": 4521392, "step": 16150 }, { "epoch": 179.5, "grad_norm": 1.1955969512200681e-06, "learning_rate": 3.2437287947369786e-05, "loss": 0.0, "num_input_tokens_seen": 4522800, "step": 16155 }, { "epoch": 179.55555555555554, "grad_norm": 5.879131776964641e-07, "learning_rate": 3.2427914388669525e-05, "loss": 0.0, "num_input_tokens_seen": 4524224, "step": 16160 }, { "epoch": 179.61111111111111, "grad_norm": 1.3716824014409212e-06, "learning_rate": 3.241853968449151e-05, "loss": 0.0, "num_input_tokens_seen": 4525648, "step": 16165 }, { "epoch": 179.66666666666666, "grad_norm": 1.2678008260991191e-06, "learning_rate": 3.240916383628144e-05, "loss": 0.0, "num_input_tokens_seen": 4527040, "step": 16170 }, { "epoch": 179.72222222222223, "grad_norm": 1.4549345905834343e-06, "learning_rate": 3.239978684548521e-05, "loss": 0.0, "num_input_tokens_seen": 4528464, "step": 16175 }, { "epoch": 179.77777777777777, "grad_norm": 1.046762008627411e-06, "learning_rate": 3.239040871354885e-05, "loss": 0.0, "num_input_tokens_seen": 4529872, "step": 16180 }, { "epoch": 179.83333333333334, "grad_norm": 4.152929250267334e-06, "learning_rate": 3.2381029441918596e-05, "loss": 0.0, "num_input_tokens_seen": 4531248, "step": 16185 }, { "epoch": 179.88888888888889, "grad_norm": 4.3016657969019434e-07, "learning_rate": 3.2371649032040845e-05, "loss": 0.0, "num_input_tokens_seen": 4532688, "step": 16190 }, { "epoch": 179.94444444444446, "grad_norm": 1.893575131362013e-06, "learning_rate": 3.2362267485362174e-05, "loss": 0.0, "num_input_tokens_seen": 4534112, "step": 16195 }, { "epoch": 180.0, "grad_norm": 4.94763071401394e-07, "learning_rate": 3.235288480332934e-05, "loss": 0.0, "num_input_tokens_seen": 4535504, "step": 16200 }, { "epoch": 180.0, "eval_loss": 1.3130041360855103, "eval_runtime": 1.3857, "eval_samples_per_second": 28.867, "eval_steps_per_second": 14.433, "num_input_tokens_seen": 4535504, "step": 16200 }, { "epoch": 180.05555555555554, "grad_norm": 2.895471197916777e-06, "learning_rate": 3.234350098738927e-05, "loss": 0.0, "num_input_tokens_seen": 4536896, "step": 16205 }, { "epoch": 180.11111111111111, "grad_norm": 1.0732028385973535e-06, "learning_rate": 3.233411603898906e-05, "loss": 0.0, "num_input_tokens_seen": 4538272, "step": 16210 }, { "epoch": 180.16666666666666, "grad_norm": 1.0243056749459356e-06, "learning_rate": 3.232472995957599e-05, "loss": 0.0, "num_input_tokens_seen": 4539680, "step": 16215 }, { "epoch": 180.22222222222223, "grad_norm": 1.536227500764653e-06, "learning_rate": 3.231534275059751e-05, "loss": 0.0, "num_input_tokens_seen": 4541104, "step": 16220 }, { "epoch": 180.27777777777777, "grad_norm": 4.0268949419441924e-07, "learning_rate": 3.230595441350125e-05, "loss": 0.0, "num_input_tokens_seen": 4542512, "step": 16225 }, { "epoch": 180.33333333333334, "grad_norm": 2.3537611468782416e-06, "learning_rate": 3.2296564949735e-05, "loss": 0.0, "num_input_tokens_seen": 4543920, "step": 16230 }, { "epoch": 180.38888888888889, "grad_norm": 9.287645639233233e-07, "learning_rate": 3.228717436074675e-05, "loss": 0.0, "num_input_tokens_seen": 4545296, "step": 16235 }, { "epoch": 180.44444444444446, "grad_norm": 5.079789957562753e-07, "learning_rate": 3.227778264798463e-05, "loss": 0.0, "num_input_tokens_seen": 4546688, "step": 16240 }, { "epoch": 180.5, "grad_norm": 9.306265837949468e-07, "learning_rate": 3.226838981289698e-05, "loss": 0.0, "num_input_tokens_seen": 4548064, "step": 16245 }, { "epoch": 180.55555555555554, "grad_norm": 2.0561781184369465e-06, "learning_rate": 3.225899585693227e-05, "loss": 0.0, "num_input_tokens_seen": 4549472, "step": 16250 }, { "epoch": 180.61111111111111, "grad_norm": 2.788555320876185e-06, "learning_rate": 3.224960078153918e-05, "loss": 0.0, "num_input_tokens_seen": 4550832, "step": 16255 }, { "epoch": 180.66666666666666, "grad_norm": 3.38591064519278e-07, "learning_rate": 3.224020458816655e-05, "loss": 0.0, "num_input_tokens_seen": 4552240, "step": 16260 }, { "epoch": 180.72222222222223, "grad_norm": 1.1965055364271393e-06, "learning_rate": 3.223080727826337e-05, "loss": 0.0, "num_input_tokens_seen": 4553696, "step": 16265 }, { "epoch": 180.77777777777777, "grad_norm": 5.301327519191545e-07, "learning_rate": 3.222140885327885e-05, "loss": 0.0, "num_input_tokens_seen": 4555088, "step": 16270 }, { "epoch": 180.83333333333334, "grad_norm": 5.846267754350265e-07, "learning_rate": 3.221200931466234e-05, "loss": 0.0, "num_input_tokens_seen": 4556528, "step": 16275 }, { "epoch": 180.88888888888889, "grad_norm": 6.058783696971659e-07, "learning_rate": 3.220260866386336e-05, "loss": 0.0, "num_input_tokens_seen": 4557920, "step": 16280 }, { "epoch": 180.94444444444446, "grad_norm": 2.108544322254602e-06, "learning_rate": 3.21932069023316e-05, "loss": 0.0, "num_input_tokens_seen": 4559312, "step": 16285 }, { "epoch": 181.0, "grad_norm": 1.690006001808797e-06, "learning_rate": 3.218380403151695e-05, "loss": 0.0, "num_input_tokens_seen": 4560736, "step": 16290 }, { "epoch": 181.05555555555554, "grad_norm": 7.429366064570786e-07, "learning_rate": 3.217440005286943e-05, "loss": 0.0, "num_input_tokens_seen": 4562112, "step": 16295 }, { "epoch": 181.11111111111111, "grad_norm": 3.1700265026302077e-06, "learning_rate": 3.216499496783928e-05, "loss": 0.0, "num_input_tokens_seen": 4563536, "step": 16300 }, { "epoch": 181.16666666666666, "grad_norm": 4.0601730688649695e-06, "learning_rate": 3.2155588777876856e-05, "loss": 0.0, "num_input_tokens_seen": 4564976, "step": 16305 }, { "epoch": 181.22222222222223, "grad_norm": 2.4769972242211225e-06, "learning_rate": 3.214618148443273e-05, "loss": 0.0, "num_input_tokens_seen": 4566352, "step": 16310 }, { "epoch": 181.27777777777777, "grad_norm": 1.9148128558299504e-06, "learning_rate": 3.2136773088957595e-05, "loss": 0.0, "num_input_tokens_seen": 4567728, "step": 16315 }, { "epoch": 181.33333333333334, "grad_norm": 2.4923981527535943e-06, "learning_rate": 3.2127363592902374e-05, "loss": 0.0, "num_input_tokens_seen": 4569104, "step": 16320 }, { "epoch": 181.38888888888889, "grad_norm": 4.6499653194587154e-07, "learning_rate": 3.211795299771812e-05, "loss": 0.0, "num_input_tokens_seen": 4570528, "step": 16325 }, { "epoch": 181.44444444444446, "grad_norm": 1.1599724984989734e-06, "learning_rate": 3.210854130485605e-05, "loss": 0.0, "num_input_tokens_seen": 4571936, "step": 16330 }, { "epoch": 181.5, "grad_norm": 2.4680764454387827e-06, "learning_rate": 3.209912851576759e-05, "loss": 0.0, "num_input_tokens_seen": 4573328, "step": 16335 }, { "epoch": 181.55555555555554, "grad_norm": 4.452761004358763e-07, "learning_rate": 3.208971463190431e-05, "loss": 0.0, "num_input_tokens_seen": 4574656, "step": 16340 }, { "epoch": 181.61111111111111, "grad_norm": 5.633381192637898e-07, "learning_rate": 3.208029965471793e-05, "loss": 0.0, "num_input_tokens_seen": 4576048, "step": 16345 }, { "epoch": 181.66666666666666, "grad_norm": 7.477985946025001e-07, "learning_rate": 3.2070883585660364e-05, "loss": 0.0, "num_input_tokens_seen": 4577456, "step": 16350 }, { "epoch": 181.72222222222223, "grad_norm": 1.3150961422070395e-06, "learning_rate": 3.20614664261837e-05, "loss": 0.0, "num_input_tokens_seen": 4578896, "step": 16355 }, { "epoch": 181.77777777777777, "grad_norm": 6.927230060682632e-07, "learning_rate": 3.205204817774016e-05, "loss": 0.0, "num_input_tokens_seen": 4580320, "step": 16360 }, { "epoch": 181.83333333333334, "grad_norm": 5.014355224375322e-07, "learning_rate": 3.204262884178218e-05, "loss": 0.0, "num_input_tokens_seen": 4581696, "step": 16365 }, { "epoch": 181.88888888888889, "grad_norm": 1.3085059435979929e-06, "learning_rate": 3.2033208419762314e-05, "loss": 0.0, "num_input_tokens_seen": 4583104, "step": 16370 }, { "epoch": 181.94444444444446, "grad_norm": 5.177387265575817e-07, "learning_rate": 3.2023786913133344e-05, "loss": 0.0, "num_input_tokens_seen": 4584528, "step": 16375 }, { "epoch": 182.0, "grad_norm": 1.6920855614444008e-06, "learning_rate": 3.201436432334816e-05, "loss": 0.0, "num_input_tokens_seen": 4585936, "step": 16380 }, { "epoch": 182.05555555555554, "grad_norm": 1.5237260413414333e-06, "learning_rate": 3.2004940651859844e-05, "loss": 0.0, "num_input_tokens_seen": 4587296, "step": 16385 }, { "epoch": 182.11111111111111, "grad_norm": 5.042609245720087e-07, "learning_rate": 3.1995515900121655e-05, "loss": 0.0, "num_input_tokens_seen": 4588704, "step": 16390 }, { "epoch": 182.16666666666666, "grad_norm": 1.1847031373690697e-06, "learning_rate": 3.1986090069587e-05, "loss": 0.0, "num_input_tokens_seen": 4590080, "step": 16395 }, { "epoch": 182.22222222222223, "grad_norm": 1.0237233709631255e-06, "learning_rate": 3.1976663161709466e-05, "loss": 0.0, "num_input_tokens_seen": 4591504, "step": 16400 }, { "epoch": 182.22222222222223, "eval_loss": 1.2976735830307007, "eval_runtime": 1.3869, "eval_samples_per_second": 28.842, "eval_steps_per_second": 14.421, "num_input_tokens_seen": 4591504, "step": 16400 }, { "epoch": 182.27777777777777, "grad_norm": 1.6572764707234455e-06, "learning_rate": 3.196723517794279e-05, "loss": 0.0, "num_input_tokens_seen": 4592912, "step": 16405 }, { "epoch": 182.33333333333334, "grad_norm": 1.9337487628945382e-06, "learning_rate": 3.19578061197409e-05, "loss": 0.0, "num_input_tokens_seen": 4594304, "step": 16410 }, { "epoch": 182.38888888888889, "grad_norm": 8.4049713677814e-07, "learning_rate": 3.194837598855787e-05, "loss": 0.0, "num_input_tokens_seen": 4595696, "step": 16415 }, { "epoch": 182.44444444444446, "grad_norm": 8.980125585367205e-07, "learning_rate": 3.193894478584794e-05, "loss": 0.0, "num_input_tokens_seen": 4597120, "step": 16420 }, { "epoch": 182.5, "grad_norm": 5.319263323144696e-07, "learning_rate": 3.192951251306553e-05, "loss": 0.0, "num_input_tokens_seen": 4598512, "step": 16425 }, { "epoch": 182.55555555555554, "grad_norm": 1.030027306114789e-06, "learning_rate": 3.192007917166521e-05, "loss": 0.0, "num_input_tokens_seen": 4599920, "step": 16430 }, { "epoch": 182.61111111111111, "grad_norm": 1.400034420839802e-06, "learning_rate": 3.191064476310171e-05, "loss": 0.0, "num_input_tokens_seen": 4601312, "step": 16435 }, { "epoch": 182.66666666666666, "grad_norm": 1.5188334145932458e-06, "learning_rate": 3.1901209288829944e-05, "loss": 0.0, "num_input_tokens_seen": 4602672, "step": 16440 }, { "epoch": 182.72222222222223, "grad_norm": 5.134414777785423e-07, "learning_rate": 3.1891772750304985e-05, "loss": 0.0, "num_input_tokens_seen": 4604064, "step": 16445 }, { "epoch": 182.77777777777777, "grad_norm": 6.028271855029743e-07, "learning_rate": 3.188233514898206e-05, "loss": 0.0, "num_input_tokens_seen": 4605424, "step": 16450 }, { "epoch": 182.83333333333334, "grad_norm": 1.4044962881598622e-06, "learning_rate": 3.187289648631657e-05, "loss": 0.0, "num_input_tokens_seen": 4606832, "step": 16455 }, { "epoch": 182.88888888888889, "grad_norm": 2.214466121586156e-07, "learning_rate": 3.186345676376406e-05, "loss": 0.0, "num_input_tokens_seen": 4608240, "step": 16460 }, { "epoch": 182.94444444444446, "grad_norm": 2.4374492113565793e-06, "learning_rate": 3.1854015982780275e-05, "loss": 0.0, "num_input_tokens_seen": 4609632, "step": 16465 }, { "epoch": 183.0, "grad_norm": 3.7382730511126283e-07, "learning_rate": 3.1844574144821084e-05, "loss": 0.0, "num_input_tokens_seen": 4611072, "step": 16470 }, { "epoch": 183.05555555555554, "grad_norm": 1.1901914831469185e-06, "learning_rate": 3.1835131251342554e-05, "loss": 0.0, "num_input_tokens_seen": 4612496, "step": 16475 }, { "epoch": 183.11111111111111, "grad_norm": 8.633995207674161e-07, "learning_rate": 3.182568730380089e-05, "loss": 0.0, "num_input_tokens_seen": 4613920, "step": 16480 }, { "epoch": 183.16666666666666, "grad_norm": 1.4160631280901725e-06, "learning_rate": 3.181624230365245e-05, "loss": 0.0, "num_input_tokens_seen": 4615344, "step": 16485 }, { "epoch": 183.22222222222223, "grad_norm": 1.7706222479318967e-06, "learning_rate": 3.180679625235381e-05, "loss": 0.0, "num_input_tokens_seen": 4616736, "step": 16490 }, { "epoch": 183.27777777777777, "grad_norm": 7.609136787323223e-07, "learning_rate": 3.1797349151361646e-05, "loss": 0.0, "num_input_tokens_seen": 4618160, "step": 16495 }, { "epoch": 183.33333333333334, "grad_norm": 6.895346018609416e-07, "learning_rate": 3.178790100213281e-05, "loss": 0.0, "num_input_tokens_seen": 4619568, "step": 16500 }, { "epoch": 183.38888888888889, "grad_norm": 9.929705129252397e-07, "learning_rate": 3.1778451806124346e-05, "loss": 0.0, "num_input_tokens_seen": 4620928, "step": 16505 }, { "epoch": 183.44444444444446, "grad_norm": 7.431137305502489e-07, "learning_rate": 3.176900156479342e-05, "loss": 0.0, "num_input_tokens_seen": 4622320, "step": 16510 }, { "epoch": 183.5, "grad_norm": 2.258463609905448e-06, "learning_rate": 3.17595502795974e-05, "loss": 0.0, "num_input_tokens_seen": 4623712, "step": 16515 }, { "epoch": 183.55555555555554, "grad_norm": 2.6849620553548448e-06, "learning_rate": 3.175009795199377e-05, "loss": 0.0, "num_input_tokens_seen": 4625152, "step": 16520 }, { "epoch": 183.61111111111111, "grad_norm": 8.202018761949148e-07, "learning_rate": 3.1740644583440224e-05, "loss": 0.0, "num_input_tokens_seen": 4626512, "step": 16525 }, { "epoch": 183.66666666666666, "grad_norm": 9.6603355359548e-07, "learning_rate": 3.173119017539457e-05, "loss": 0.0, "num_input_tokens_seen": 4627952, "step": 16530 }, { "epoch": 183.72222222222223, "grad_norm": 1.5699021105319844e-06, "learning_rate": 3.172173472931479e-05, "loss": 0.0, "num_input_tokens_seen": 4629360, "step": 16535 }, { "epoch": 183.77777777777777, "grad_norm": 5.570515213548788e-07, "learning_rate": 3.1712278246659055e-05, "loss": 0.0, "num_input_tokens_seen": 4630720, "step": 16540 }, { "epoch": 183.83333333333334, "grad_norm": 7.007532190073107e-07, "learning_rate": 3.170282072888566e-05, "loss": 0.0, "num_input_tokens_seen": 4632112, "step": 16545 }, { "epoch": 183.88888888888889, "grad_norm": 7.490304483326327e-07, "learning_rate": 3.169336217745307e-05, "loss": 0.0, "num_input_tokens_seen": 4633504, "step": 16550 }, { "epoch": 183.94444444444446, "grad_norm": 6.008353921060916e-07, "learning_rate": 3.1683902593819924e-05, "loss": 0.0, "num_input_tokens_seen": 4634896, "step": 16555 }, { "epoch": 184.0, "grad_norm": 3.833042114820273e-07, "learning_rate": 3.1674441979445e-05, "loss": 0.0, "num_input_tokens_seen": 4636288, "step": 16560 }, { "epoch": 184.05555555555554, "grad_norm": 5.400818281486863e-07, "learning_rate": 3.166498033578725e-05, "loss": 0.0, "num_input_tokens_seen": 4637696, "step": 16565 }, { "epoch": 184.11111111111111, "grad_norm": 4.618211733031785e-07, "learning_rate": 3.165551766430578e-05, "loss": 0.0, "num_input_tokens_seen": 4639072, "step": 16570 }, { "epoch": 184.16666666666666, "grad_norm": 9.99279450297763e-07, "learning_rate": 3.164605396645984e-05, "loss": 0.0, "num_input_tokens_seen": 4640496, "step": 16575 }, { "epoch": 184.22222222222223, "grad_norm": 5.616390694740403e-07, "learning_rate": 3.163658924370886e-05, "loss": 0.0, "num_input_tokens_seen": 4641872, "step": 16580 }, { "epoch": 184.27777777777777, "grad_norm": 7.446253675880143e-06, "learning_rate": 3.1627123497512415e-05, "loss": 0.0, "num_input_tokens_seen": 4643248, "step": 16585 }, { "epoch": 184.33333333333334, "grad_norm": 1.094618596653163e-06, "learning_rate": 3.1617656729330245e-05, "loss": 0.0, "num_input_tokens_seen": 4644608, "step": 16590 }, { "epoch": 184.38888888888889, "grad_norm": 7.125373144845071e-07, "learning_rate": 3.1608188940622255e-05, "loss": 0.0, "num_input_tokens_seen": 4646016, "step": 16595 }, { "epoch": 184.44444444444446, "grad_norm": 5.243702503321401e-07, "learning_rate": 3.159872013284847e-05, "loss": 0.0, "num_input_tokens_seen": 4647424, "step": 16600 }, { "epoch": 184.44444444444446, "eval_loss": 1.2992746829986572, "eval_runtime": 1.3912, "eval_samples_per_second": 28.752, "eval_steps_per_second": 14.376, "num_input_tokens_seen": 4647424, "step": 16600 }, { "epoch": 184.5, "grad_norm": 1.44428656767559e-06, "learning_rate": 3.1589250307469134e-05, "loss": 0.0, "num_input_tokens_seen": 4648832, "step": 16605 }, { "epoch": 184.55555555555554, "grad_norm": 1.6182355011551408e-06, "learning_rate": 3.1579779465944586e-05, "loss": 0.0, "num_input_tokens_seen": 4650240, "step": 16610 }, { "epoch": 184.61111111111111, "grad_norm": 7.726863486823277e-07, "learning_rate": 3.1570307609735363e-05, "loss": 0.0, "num_input_tokens_seen": 4651616, "step": 16615 }, { "epoch": 184.66666666666666, "grad_norm": 1.664540604906506e-06, "learning_rate": 3.156083474030213e-05, "loss": 0.0, "num_input_tokens_seen": 4653040, "step": 16620 }, { "epoch": 184.72222222222223, "grad_norm": 1.1889442248502746e-06, "learning_rate": 3.155136085910573e-05, "loss": 0.0, "num_input_tokens_seen": 4654432, "step": 16625 }, { "epoch": 184.77777777777777, "grad_norm": 1.3952849542420154e-07, "learning_rate": 3.154188596760717e-05, "loss": 0.0, "num_input_tokens_seen": 4655808, "step": 16630 }, { "epoch": 184.83333333333334, "grad_norm": 1.6895010048756376e-06, "learning_rate": 3.153241006726757e-05, "loss": 0.0, "num_input_tokens_seen": 4657248, "step": 16635 }, { "epoch": 184.88888888888889, "grad_norm": 1.2606201380549464e-06, "learning_rate": 3.152293315954825e-05, "loss": 0.0, "num_input_tokens_seen": 4658656, "step": 16640 }, { "epoch": 184.94444444444446, "grad_norm": 5.476479145727353e-07, "learning_rate": 3.1513455245910666e-05, "loss": 0.0, "num_input_tokens_seen": 4660064, "step": 16645 }, { "epoch": 185.0, "grad_norm": 4.6861683244969754e-07, "learning_rate": 3.150397632781643e-05, "loss": 0.0, "num_input_tokens_seen": 4661424, "step": 16650 }, { "epoch": 185.05555555555554, "grad_norm": 1.1814512390628806e-06, "learning_rate": 3.149449640672731e-05, "loss": 0.0, "num_input_tokens_seen": 4662800, "step": 16655 }, { "epoch": 185.11111111111111, "grad_norm": 4.925501571051427e-07, "learning_rate": 3.148501548410523e-05, "loss": 0.0, "num_input_tokens_seen": 4664208, "step": 16660 }, { "epoch": 185.16666666666666, "grad_norm": 1.531827592771151e-06, "learning_rate": 3.1475533561412256e-05, "loss": 0.0, "num_input_tokens_seen": 4665648, "step": 16665 }, { "epoch": 185.22222222222223, "grad_norm": 3.3272117434535176e-06, "learning_rate": 3.146605064011065e-05, "loss": 0.0, "num_input_tokens_seen": 4667024, "step": 16670 }, { "epoch": 185.27777777777777, "grad_norm": 5.74925081764377e-07, "learning_rate": 3.145656672166277e-05, "loss": 0.0, "num_input_tokens_seen": 4668416, "step": 16675 }, { "epoch": 185.33333333333334, "grad_norm": 2.3554487427190907e-07, "learning_rate": 3.144708180753116e-05, "loss": 0.0, "num_input_tokens_seen": 4669808, "step": 16680 }, { "epoch": 185.38888888888889, "grad_norm": 2.4016421775741037e-06, "learning_rate": 3.143759589917851e-05, "loss": 0.0, "num_input_tokens_seen": 4671200, "step": 16685 }, { "epoch": 185.44444444444446, "grad_norm": 1.314842734245758e-06, "learning_rate": 3.142810899806768e-05, "loss": 0.0, "num_input_tokens_seen": 4672608, "step": 16690 }, { "epoch": 185.5, "grad_norm": 1.3417821946859476e-06, "learning_rate": 3.141862110566166e-05, "loss": 0.0, "num_input_tokens_seen": 4674016, "step": 16695 }, { "epoch": 185.55555555555554, "grad_norm": 7.295254249584104e-07, "learning_rate": 3.1409132223423606e-05, "loss": 0.0, "num_input_tokens_seen": 4675408, "step": 16700 }, { "epoch": 185.61111111111111, "grad_norm": 1.7878505786939058e-07, "learning_rate": 3.139964235281682e-05, "loss": 0.0, "num_input_tokens_seen": 4676768, "step": 16705 }, { "epoch": 185.66666666666666, "grad_norm": 7.128180072868417e-07, "learning_rate": 3.139015149530476e-05, "loss": 0.0, "num_input_tokens_seen": 4678208, "step": 16710 }, { "epoch": 185.72222222222223, "grad_norm": 2.795323439386266e-07, "learning_rate": 3.1380659652351034e-05, "loss": 0.0, "num_input_tokens_seen": 4679616, "step": 16715 }, { "epoch": 185.77777777777777, "grad_norm": 7.36347203655896e-07, "learning_rate": 3.137116682541941e-05, "loss": 0.0, "num_input_tokens_seen": 4681056, "step": 16720 }, { "epoch": 185.83333333333334, "grad_norm": 1.2671035847233725e-06, "learning_rate": 3.136167301597379e-05, "loss": 0.0, "num_input_tokens_seen": 4682464, "step": 16725 }, { "epoch": 185.88888888888889, "grad_norm": 1.1152086472066003e-06, "learning_rate": 3.1352178225478254e-05, "loss": 0.0, "num_input_tokens_seen": 4683824, "step": 16730 }, { "epoch": 185.94444444444446, "grad_norm": 2.1027422008046415e-06, "learning_rate": 3.1342682455396996e-05, "loss": 0.0, "num_input_tokens_seen": 4685184, "step": 16735 }, { "epoch": 186.0, "grad_norm": 1.341728534498543e-06, "learning_rate": 3.133318570719441e-05, "loss": 0.0, "num_input_tokens_seen": 4686608, "step": 16740 }, { "epoch": 186.05555555555554, "grad_norm": 1.2487955700635212e-06, "learning_rate": 3.132368798233499e-05, "loss": 0.0, "num_input_tokens_seen": 4687984, "step": 16745 }, { "epoch": 186.11111111111111, "grad_norm": 1.2382911336317193e-06, "learning_rate": 3.131418928228342e-05, "loss": 0.0, "num_input_tokens_seen": 4689376, "step": 16750 }, { "epoch": 186.16666666666666, "grad_norm": 3.0238109047786565e-06, "learning_rate": 3.1304689608504514e-05, "loss": 0.0, "num_input_tokens_seen": 4690800, "step": 16755 }, { "epoch": 186.22222222222223, "grad_norm": 7.94397408299119e-07, "learning_rate": 3.129518896246324e-05, "loss": 0.0, "num_input_tokens_seen": 4692192, "step": 16760 }, { "epoch": 186.27777777777777, "grad_norm": 1.188178316624544e-06, "learning_rate": 3.128568734562472e-05, "loss": 0.0, "num_input_tokens_seen": 4693568, "step": 16765 }, { "epoch": 186.33333333333334, "grad_norm": 1.4772383565286873e-06, "learning_rate": 3.127618475945421e-05, "loss": 0.0, "num_input_tokens_seen": 4694992, "step": 16770 }, { "epoch": 186.38888888888889, "grad_norm": 5.176902391212934e-07, "learning_rate": 3.126668120541715e-05, "loss": 0.0, "num_input_tokens_seen": 4696384, "step": 16775 }, { "epoch": 186.44444444444446, "grad_norm": 4.268077873348375e-07, "learning_rate": 3.1257176684979096e-05, "loss": 0.0, "num_input_tokens_seen": 4697760, "step": 16780 }, { "epoch": 186.5, "grad_norm": 1.9163614979333943e-06, "learning_rate": 3.124767119960576e-05, "loss": 0.0, "num_input_tokens_seen": 4699136, "step": 16785 }, { "epoch": 186.55555555555554, "grad_norm": 9.90182002169604e-07, "learning_rate": 3.123816475076301e-05, "loss": 0.0, "num_input_tokens_seen": 4700512, "step": 16790 }, { "epoch": 186.61111111111111, "grad_norm": 1.027171492751222e-06, "learning_rate": 3.122865733991687e-05, "loss": 0.0, "num_input_tokens_seen": 4701936, "step": 16795 }, { "epoch": 186.66666666666666, "grad_norm": 1.9410404092923272e-06, "learning_rate": 3.1219148968533486e-05, "loss": 0.0, "num_input_tokens_seen": 4703376, "step": 16800 }, { "epoch": 186.66666666666666, "eval_loss": 1.3115383386611938, "eval_runtime": 1.3953, "eval_samples_per_second": 28.668, "eval_steps_per_second": 14.334, "num_input_tokens_seen": 4703376, "step": 16800 }, { "epoch": 186.72222222222223, "grad_norm": 3.82548023480922e-07, "learning_rate": 3.120963963807918e-05, "loss": 0.0, "num_input_tokens_seen": 4704800, "step": 16805 }, { "epoch": 186.77777777777777, "grad_norm": 3.540191073625465e-06, "learning_rate": 3.12001293500204e-05, "loss": 0.0, "num_input_tokens_seen": 4706176, "step": 16810 }, { "epoch": 186.83333333333334, "grad_norm": 1.099725409403618e-06, "learning_rate": 3.1190618105823765e-05, "loss": 0.0, "num_input_tokens_seen": 4707584, "step": 16815 }, { "epoch": 186.88888888888889, "grad_norm": 6.339028004731517e-07, "learning_rate": 3.118110590695603e-05, "loss": 0.0, "num_input_tokens_seen": 4709024, "step": 16820 }, { "epoch": 186.94444444444446, "grad_norm": 4.096742145520693e-07, "learning_rate": 3.117159275488407e-05, "loss": 0.0, "num_input_tokens_seen": 4710432, "step": 16825 }, { "epoch": 187.0, "grad_norm": 4.526467591858818e-07, "learning_rate": 3.1162078651074956e-05, "loss": 0.0, "num_input_tokens_seen": 4711808, "step": 16830 }, { "epoch": 187.05555555555554, "grad_norm": 1.409532956131443e-06, "learning_rate": 3.1152563596995885e-05, "loss": 0.0, "num_input_tokens_seen": 4713184, "step": 16835 }, { "epoch": 187.11111111111111, "grad_norm": 6.098484846006613e-07, "learning_rate": 3.1143047594114186e-05, "loss": 0.0, "num_input_tokens_seen": 4714576, "step": 16840 }, { "epoch": 187.16666666666666, "grad_norm": 4.820050207854365e-07, "learning_rate": 3.113353064389734e-05, "loss": 0.0, "num_input_tokens_seen": 4715968, "step": 16845 }, { "epoch": 187.22222222222223, "grad_norm": 6.430081498365325e-07, "learning_rate": 3.1124012747812993e-05, "loss": 0.0, "num_input_tokens_seen": 4717344, "step": 16850 }, { "epoch": 187.27777777777777, "grad_norm": 4.4216923811291053e-07, "learning_rate": 3.1114493907328936e-05, "loss": 0.0, "num_input_tokens_seen": 4718752, "step": 16855 }, { "epoch": 187.33333333333334, "grad_norm": 1.1091359510828624e-06, "learning_rate": 3.110497412391306e-05, "loss": 0.0, "num_input_tokens_seen": 4720192, "step": 16860 }, { "epoch": 187.38888888888889, "grad_norm": 1.5017708392406348e-06, "learning_rate": 3.1095453399033466e-05, "loss": 0.0, "num_input_tokens_seen": 4721616, "step": 16865 }, { "epoch": 187.44444444444446, "grad_norm": 3.609405609950045e-07, "learning_rate": 3.108593173415835e-05, "loss": 0.0, "num_input_tokens_seen": 4723024, "step": 16870 }, { "epoch": 187.5, "grad_norm": 5.855333711224375e-07, "learning_rate": 3.107640913075609e-05, "loss": 0.0, "num_input_tokens_seen": 4724464, "step": 16875 }, { "epoch": 187.55555555555554, "grad_norm": 9.610188271835796e-07, "learning_rate": 3.106688559029517e-05, "loss": 0.0, "num_input_tokens_seen": 4725840, "step": 16880 }, { "epoch": 187.61111111111111, "grad_norm": 1.0983125093844137e-06, "learning_rate": 3.105736111424425e-05, "loss": 0.0, "num_input_tokens_seen": 4727264, "step": 16885 }, { "epoch": 187.66666666666666, "grad_norm": 2.5673357413324993e-06, "learning_rate": 3.1047835704072136e-05, "loss": 0.0, "num_input_tokens_seen": 4728656, "step": 16890 }, { "epoch": 187.72222222222223, "grad_norm": 8.782046734268079e-07, "learning_rate": 3.103830936124775e-05, "loss": 0.0, "num_input_tokens_seen": 4730032, "step": 16895 }, { "epoch": 187.77777777777777, "grad_norm": 4.196227564534638e-06, "learning_rate": 3.102878208724018e-05, "loss": 0.0, "num_input_tokens_seen": 4731392, "step": 16900 }, { "epoch": 187.83333333333334, "grad_norm": 4.101066224393435e-06, "learning_rate": 3.101925388351865e-05, "loss": 0.0, "num_input_tokens_seen": 4732816, "step": 16905 }, { "epoch": 187.88888888888889, "grad_norm": 8.120757115648303e-07, "learning_rate": 3.1009724751552515e-05, "loss": 0.0, "num_input_tokens_seen": 4734240, "step": 16910 }, { "epoch": 187.94444444444446, "grad_norm": 6.962747534089431e-07, "learning_rate": 3.100019469281131e-05, "loss": 0.0, "num_input_tokens_seen": 4735600, "step": 16915 }, { "epoch": 188.0, "grad_norm": 1.240076244357624e-06, "learning_rate": 3.0990663708764685e-05, "loss": 0.0, "num_input_tokens_seen": 4737040, "step": 16920 }, { "epoch": 188.05555555555554, "grad_norm": 1.8274123476658133e-06, "learning_rate": 3.098113180088243e-05, "loss": 0.0, "num_input_tokens_seen": 4738480, "step": 16925 }, { "epoch": 188.11111111111111, "grad_norm": 4.662243782149744e-07, "learning_rate": 3.097159897063448e-05, "loss": 0.0, "num_input_tokens_seen": 4739936, "step": 16930 }, { "epoch": 188.16666666666666, "grad_norm": 6.644714289905096e-07, "learning_rate": 3.096206521949094e-05, "loss": 0.0, "num_input_tokens_seen": 4741296, "step": 16935 }, { "epoch": 188.22222222222223, "grad_norm": 6.811923753957672e-07, "learning_rate": 3.0952530548922006e-05, "loss": 0.0, "num_input_tokens_seen": 4742720, "step": 16940 }, { "epoch": 188.27777777777777, "grad_norm": 2.2565423023479525e-06, "learning_rate": 3.0942994960398064e-05, "loss": 0.0, "num_input_tokens_seen": 4744128, "step": 16945 }, { "epoch": 188.33333333333334, "grad_norm": 1.1870081380038755e-06, "learning_rate": 3.093345845538961e-05, "loss": 0.0, "num_input_tokens_seen": 4745600, "step": 16950 }, { "epoch": 188.38888888888889, "grad_norm": 7.80397670041566e-07, "learning_rate": 3.09239210353673e-05, "loss": 0.0, "num_input_tokens_seen": 4746960, "step": 16955 }, { "epoch": 188.44444444444446, "grad_norm": 4.883727342530619e-07, "learning_rate": 3.0914382701801926e-05, "loss": 0.0, "num_input_tokens_seen": 4748352, "step": 16960 }, { "epoch": 188.5, "grad_norm": 2.4977312023111153e-06, "learning_rate": 3.090484345616441e-05, "loss": 0.0, "num_input_tokens_seen": 4749728, "step": 16965 }, { "epoch": 188.55555555555554, "grad_norm": 8.814875513962761e-07, "learning_rate": 3.0895303299925825e-05, "loss": 0.0, "num_input_tokens_seen": 4751136, "step": 16970 }, { "epoch": 188.61111111111111, "grad_norm": 6.469772984019073e-07, "learning_rate": 3.0885762234557393e-05, "loss": 0.0, "num_input_tokens_seen": 4752528, "step": 16975 }, { "epoch": 188.66666666666666, "grad_norm": 7.871937555137265e-07, "learning_rate": 3.087622026153045e-05, "loss": 0.0, "num_input_tokens_seen": 4753968, "step": 16980 }, { "epoch": 188.72222222222223, "grad_norm": 4.784363341059361e-07, "learning_rate": 3.086667738231651e-05, "loss": 0.0, "num_input_tokens_seen": 4755344, "step": 16985 }, { "epoch": 188.77777777777777, "grad_norm": 1.3460913805829477e-06, "learning_rate": 3.085713359838718e-05, "loss": 0.0, "num_input_tokens_seen": 4756752, "step": 16990 }, { "epoch": 188.83333333333334, "grad_norm": 1.3499806073014042e-06, "learning_rate": 3.084758891121425e-05, "loss": 0.0, "num_input_tokens_seen": 4758144, "step": 16995 }, { "epoch": 188.88888888888889, "grad_norm": 3.384652131899202e-07, "learning_rate": 3.083804332226963e-05, "loss": 0.0, "num_input_tokens_seen": 4759552, "step": 17000 }, { "epoch": 188.88888888888889, "eval_loss": 1.3025184869766235, "eval_runtime": 1.3931, "eval_samples_per_second": 28.714, "eval_steps_per_second": 14.357, "num_input_tokens_seen": 4759552, "step": 17000 }, { "epoch": 188.94444444444446, "grad_norm": 6.199942390594515e-07, "learning_rate": 3.082849683302536e-05, "loss": 0.0, "num_input_tokens_seen": 4760912, "step": 17005 }, { "epoch": 189.0, "grad_norm": 7.317091785807861e-07, "learning_rate": 3.081894944495363e-05, "loss": 0.0, "num_input_tokens_seen": 4762336, "step": 17010 }, { "epoch": 189.05555555555554, "grad_norm": 9.343722808807797e-07, "learning_rate": 3.080940115952677e-05, "loss": 0.0, "num_input_tokens_seen": 4763760, "step": 17015 }, { "epoch": 189.11111111111111, "grad_norm": 4.872445060755126e-07, "learning_rate": 3.0799851978217245e-05, "loss": 0.0, "num_input_tokens_seen": 4765216, "step": 17020 }, { "epoch": 189.16666666666666, "grad_norm": 7.18057833637431e-07, "learning_rate": 3.0790301902497666e-05, "loss": 0.0, "num_input_tokens_seen": 4766640, "step": 17025 }, { "epoch": 189.22222222222223, "grad_norm": 3.897636986494035e-07, "learning_rate": 3.078075093384076e-05, "loss": 0.0, "num_input_tokens_seen": 4768048, "step": 17030 }, { "epoch": 189.27777777777777, "grad_norm": 2.3955415144882863e-06, "learning_rate": 3.077119907371942e-05, "loss": 0.0, "num_input_tokens_seen": 4769456, "step": 17035 }, { "epoch": 189.33333333333334, "grad_norm": 8.791435561761318e-07, "learning_rate": 3.076164632360666e-05, "loss": 0.0, "num_input_tokens_seen": 4770880, "step": 17040 }, { "epoch": 189.38888888888889, "grad_norm": 1.3975806041344185e-06, "learning_rate": 3.075209268497563e-05, "loss": 0.0, "num_input_tokens_seen": 4772304, "step": 17045 }, { "epoch": 189.44444444444446, "grad_norm": 7.482696560145996e-07, "learning_rate": 3.074253815929961e-05, "loss": 0.0, "num_input_tokens_seen": 4773680, "step": 17050 }, { "epoch": 189.5, "grad_norm": 1.4887567658661283e-06, "learning_rate": 3.0732982748052054e-05, "loss": 0.0, "num_input_tokens_seen": 4775072, "step": 17055 }, { "epoch": 189.55555555555554, "grad_norm": 1.3193256336307968e-06, "learning_rate": 3.072342645270651e-05, "loss": 0.0, "num_input_tokens_seen": 4776512, "step": 17060 }, { "epoch": 189.61111111111111, "grad_norm": 3.9848430333222495e-07, "learning_rate": 3.071386927473668e-05, "loss": 0.0, "num_input_tokens_seen": 4777904, "step": 17065 }, { "epoch": 189.66666666666666, "grad_norm": 4.4565521761796845e-07, "learning_rate": 3.0704311215616404e-05, "loss": 0.0, "num_input_tokens_seen": 4779296, "step": 17070 }, { "epoch": 189.72222222222223, "grad_norm": 2.497758714525844e-06, "learning_rate": 3.0694752276819656e-05, "loss": 0.0, "num_input_tokens_seen": 4780672, "step": 17075 }, { "epoch": 189.77777777777777, "grad_norm": 3.790668188230484e-07, "learning_rate": 3.068519245982054e-05, "loss": 0.0, "num_input_tokens_seen": 4782048, "step": 17080 }, { "epoch": 189.83333333333334, "grad_norm": 5.506306024471996e-07, "learning_rate": 3.0675631766093304e-05, "loss": 0.0, "num_input_tokens_seen": 4783376, "step": 17085 }, { "epoch": 189.88888888888889, "grad_norm": 5.705574949388392e-06, "learning_rate": 3.066607019711232e-05, "loss": 0.0, "num_input_tokens_seen": 4784800, "step": 17090 }, { "epoch": 189.94444444444446, "grad_norm": 8.020519430829154e-07, "learning_rate": 3.065650775435211e-05, "loss": 0.0, "num_input_tokens_seen": 4786208, "step": 17095 }, { "epoch": 190.0, "grad_norm": 5.628196504403604e-07, "learning_rate": 3.0646944439287326e-05, "loss": 0.0, "num_input_tokens_seen": 4787584, "step": 17100 }, { "epoch": 190.05555555555554, "grad_norm": 1.463017838432279e-06, "learning_rate": 3.0637380253392736e-05, "loss": 0.0, "num_input_tokens_seen": 4788976, "step": 17105 }, { "epoch": 190.11111111111111, "grad_norm": 1.692069986347633e-06, "learning_rate": 3.062781519814327e-05, "loss": 0.0, "num_input_tokens_seen": 4790400, "step": 17110 }, { "epoch": 190.16666666666666, "grad_norm": 3.682624310386018e-06, "learning_rate": 3.0618249275013985e-05, "loss": 0.0, "num_input_tokens_seen": 4791792, "step": 17115 }, { "epoch": 190.22222222222223, "grad_norm": 1.2068461501257843e-06, "learning_rate": 3.060868248548005e-05, "loss": 0.0, "num_input_tokens_seen": 4793152, "step": 17120 }, { "epoch": 190.27777777777777, "grad_norm": 7.995129749360785e-07, "learning_rate": 3.0599114831016796e-05, "loss": 0.0, "num_input_tokens_seen": 4794576, "step": 17125 }, { "epoch": 190.33333333333334, "grad_norm": 4.6334912440215703e-07, "learning_rate": 3.0589546313099666e-05, "loss": 0.0, "num_input_tokens_seen": 4795968, "step": 17130 }, { "epoch": 190.38888888888889, "grad_norm": 5.375979412747256e-07, "learning_rate": 3.0579976933204255e-05, "loss": 0.0, "num_input_tokens_seen": 4797408, "step": 17135 }, { "epoch": 190.44444444444446, "grad_norm": 9.71000304161862e-07, "learning_rate": 3.0570406692806284e-05, "loss": 0.0, "num_input_tokens_seen": 4798832, "step": 17140 }, { "epoch": 190.5, "grad_norm": 1.2879706901003374e-06, "learning_rate": 3.05608355933816e-05, "loss": 0.0, "num_input_tokens_seen": 4800240, "step": 17145 }, { "epoch": 190.55555555555554, "grad_norm": 8.755968110563117e-07, "learning_rate": 3.055126363640618e-05, "loss": 0.0, "num_input_tokens_seen": 4801632, "step": 17150 }, { "epoch": 190.61111111111111, "grad_norm": 3.5945407717008493e-07, "learning_rate": 3.0541690823356146e-05, "loss": 0.0, "num_input_tokens_seen": 4803040, "step": 17155 }, { "epoch": 190.66666666666666, "grad_norm": 5.573932071456511e-07, "learning_rate": 3.053211715570775e-05, "loss": 0.0, "num_input_tokens_seen": 4804448, "step": 17160 }, { "epoch": 190.72222222222223, "grad_norm": 6.391522902049473e-07, "learning_rate": 3.052254263493736e-05, "loss": 0.0, "num_input_tokens_seen": 4805872, "step": 17165 }, { "epoch": 190.77777777777777, "grad_norm": 4.3573044194999966e-07, "learning_rate": 3.0512967262521498e-05, "loss": 0.0, "num_input_tokens_seen": 4807264, "step": 17170 }, { "epoch": 190.83333333333334, "grad_norm": 2.0993019234083476e-07, "learning_rate": 3.0503391039936803e-05, "loss": 0.0, "num_input_tokens_seen": 4808656, "step": 17175 }, { "epoch": 190.88888888888889, "grad_norm": 4.1723490085132653e-07, "learning_rate": 3.0493813968660056e-05, "loss": 0.0, "num_input_tokens_seen": 4810032, "step": 17180 }, { "epoch": 190.94444444444446, "grad_norm": 9.99864710138354e-07, "learning_rate": 3.0484236050168153e-05, "loss": 0.0, "num_input_tokens_seen": 4811424, "step": 17185 }, { "epoch": 191.0, "grad_norm": 3.1265568622984574e-07, "learning_rate": 3.0474657285938123e-05, "loss": 0.0, "num_input_tokens_seen": 4812784, "step": 17190 }, { "epoch": 191.05555555555554, "grad_norm": 1.362479224553681e-06, "learning_rate": 3.046507767744715e-05, "loss": 0.0, "num_input_tokens_seen": 4814160, "step": 17195 }, { "epoch": 191.11111111111111, "grad_norm": 6.560102860930783e-07, "learning_rate": 3.045549722617252e-05, "loss": 0.0, "num_input_tokens_seen": 4815552, "step": 17200 }, { "epoch": 191.11111111111111, "eval_loss": 1.2984585762023926, "eval_runtime": 1.3933, "eval_samples_per_second": 28.709, "eval_steps_per_second": 14.354, "num_input_tokens_seen": 4815552, "step": 17200 }, { "epoch": 191.16666666666666, "grad_norm": 5.078185267848312e-07, "learning_rate": 3.0445915933591658e-05, "loss": 0.0, "num_input_tokens_seen": 4816912, "step": 17205 }, { "epoch": 191.22222222222223, "grad_norm": 1.4842138398307725e-06, "learning_rate": 3.0436333801182114e-05, "loss": 0.0, "num_input_tokens_seen": 4818320, "step": 17210 }, { "epoch": 191.27777777777777, "grad_norm": 7.718193728578626e-07, "learning_rate": 3.0426750830421596e-05, "loss": 0.0, "num_input_tokens_seen": 4819728, "step": 17215 }, { "epoch": 191.33333333333334, "grad_norm": 6.21750871232507e-07, "learning_rate": 3.0417167022787897e-05, "loss": 0.0, "num_input_tokens_seen": 4821136, "step": 17220 }, { "epoch": 191.38888888888889, "grad_norm": 4.235332653479418e-07, "learning_rate": 3.0407582379758966e-05, "loss": 0.0, "num_input_tokens_seen": 4822544, "step": 17225 }, { "epoch": 191.44444444444446, "grad_norm": 7.47355556995899e-07, "learning_rate": 3.039799690281287e-05, "loss": 0.0, "num_input_tokens_seen": 4823936, "step": 17230 }, { "epoch": 191.5, "grad_norm": 1.1439553873060504e-06, "learning_rate": 3.0388410593427823e-05, "loss": 0.0, "num_input_tokens_seen": 4825344, "step": 17235 }, { "epoch": 191.55555555555554, "grad_norm": 1.7146654727184796e-06, "learning_rate": 3.0378823453082146e-05, "loss": 0.0, "num_input_tokens_seen": 4826752, "step": 17240 }, { "epoch": 191.61111111111111, "grad_norm": 1.302065697927901e-06, "learning_rate": 3.03692354832543e-05, "loss": 0.0, "num_input_tokens_seen": 4828144, "step": 17245 }, { "epoch": 191.66666666666666, "grad_norm": 2.168593482565484e-07, "learning_rate": 3.0359646685422865e-05, "loss": 0.0, "num_input_tokens_seen": 4829568, "step": 17250 }, { "epoch": 191.72222222222223, "grad_norm": 8.696450777279097e-07, "learning_rate": 3.035005706106656e-05, "loss": 0.0, "num_input_tokens_seen": 4830896, "step": 17255 }, { "epoch": 191.77777777777777, "grad_norm": 2.9783004151795467e-07, "learning_rate": 3.034046661166422e-05, "loss": 0.0, "num_input_tokens_seen": 4832304, "step": 17260 }, { "epoch": 191.83333333333334, "grad_norm": 1.5330188034567982e-06, "learning_rate": 3.033087533869482e-05, "loss": 0.0, "num_input_tokens_seen": 4833696, "step": 17265 }, { "epoch": 191.88888888888889, "grad_norm": 2.732654706960602e-07, "learning_rate": 3.0321283243637444e-05, "loss": 0.0, "num_input_tokens_seen": 4835152, "step": 17270 }, { "epoch": 191.94444444444446, "grad_norm": 7.68260179029312e-07, "learning_rate": 3.0311690327971326e-05, "loss": 0.0, "num_input_tokens_seen": 4836512, "step": 17275 }, { "epoch": 192.0, "grad_norm": 1.3618317780128564e-06, "learning_rate": 3.030209659317581e-05, "loss": 0.0, "num_input_tokens_seen": 4837936, "step": 17280 }, { "epoch": 192.05555555555554, "grad_norm": 8.53728693073208e-07, "learning_rate": 3.0292502040730362e-05, "loss": 0.0, "num_input_tokens_seen": 4839312, "step": 17285 }, { "epoch": 192.11111111111111, "grad_norm": 2.8480354785642703e-07, "learning_rate": 3.0282906672114597e-05, "loss": 0.0, "num_input_tokens_seen": 4840672, "step": 17290 }, { "epoch": 192.16666666666666, "grad_norm": 9.775696980796056e-07, "learning_rate": 3.027331048880823e-05, "loss": 0.0, "num_input_tokens_seen": 4842080, "step": 17295 }, { "epoch": 192.22222222222223, "grad_norm": 1.5323855677706888e-06, "learning_rate": 3.0263713492291123e-05, "loss": 0.0, "num_input_tokens_seen": 4843504, "step": 17300 }, { "epoch": 192.27777777777777, "grad_norm": 1.2225913224028773e-06, "learning_rate": 3.0254115684043242e-05, "loss": 0.0, "num_input_tokens_seen": 4844912, "step": 17305 }, { "epoch": 192.33333333333334, "grad_norm": 6.941251058378839e-07, "learning_rate": 3.024451706554469e-05, "loss": 0.0, "num_input_tokens_seen": 4846320, "step": 17310 }, { "epoch": 192.38888888888889, "grad_norm": 2.663558689164347e-06, "learning_rate": 3.0234917638275705e-05, "loss": 0.0, "num_input_tokens_seen": 4847696, "step": 17315 }, { "epoch": 192.44444444444446, "grad_norm": 4.531723902800877e-07, "learning_rate": 3.0225317403716635e-05, "loss": 0.0, "num_input_tokens_seen": 4849088, "step": 17320 }, { "epoch": 192.5, "grad_norm": 5.356994847716123e-07, "learning_rate": 3.0215716363347956e-05, "loss": 0.0, "num_input_tokens_seen": 4850512, "step": 17325 }, { "epoch": 192.55555555555554, "grad_norm": 4.374883530999796e-07, "learning_rate": 3.0206114518650275e-05, "loss": 0.0, "num_input_tokens_seen": 4851904, "step": 17330 }, { "epoch": 192.61111111111111, "grad_norm": 2.7195267193746986e-07, "learning_rate": 3.0196511871104304e-05, "loss": 0.0, "num_input_tokens_seen": 4853344, "step": 17335 }, { "epoch": 192.66666666666666, "grad_norm": 4.49647728828495e-07, "learning_rate": 3.01869084221909e-05, "loss": 0.0, "num_input_tokens_seen": 4854768, "step": 17340 }, { "epoch": 192.72222222222223, "grad_norm": 1.5264719195329235e-06, "learning_rate": 3.0177304173391037e-05, "loss": 0.0, "num_input_tokens_seen": 4856160, "step": 17345 }, { "epoch": 192.77777777777777, "grad_norm": 7.942088586787577e-07, "learning_rate": 3.01676991261858e-05, "loss": 0.0, "num_input_tokens_seen": 4857536, "step": 17350 }, { "epoch": 192.83333333333334, "grad_norm": 1.000617771751422e-06, "learning_rate": 3.015809328205642e-05, "loss": 0.0, "num_input_tokens_seen": 4858976, "step": 17355 }, { "epoch": 192.88888888888889, "grad_norm": 4.3881658484679065e-07, "learning_rate": 3.0148486642484248e-05, "loss": 0.0, "num_input_tokens_seen": 4860384, "step": 17360 }, { "epoch": 192.94444444444446, "grad_norm": 6.012812150402169e-07, "learning_rate": 3.0138879208950722e-05, "loss": 0.0, "num_input_tokens_seen": 4861776, "step": 17365 }, { "epoch": 193.0, "grad_norm": 5.145340651324659e-07, "learning_rate": 3.012927098293744e-05, "loss": 0.0, "num_input_tokens_seen": 4863168, "step": 17370 }, { "epoch": 193.05555555555554, "grad_norm": 6.624216553063889e-07, "learning_rate": 3.0119661965926123e-05, "loss": 0.0, "num_input_tokens_seen": 4864544, "step": 17375 }, { "epoch": 193.11111111111111, "grad_norm": 3.1575467573929927e-07, "learning_rate": 3.0110052159398587e-05, "loss": 0.0, "num_input_tokens_seen": 4865952, "step": 17380 }, { "epoch": 193.16666666666666, "grad_norm": 3.94370658796106e-07, "learning_rate": 3.0100441564836802e-05, "loss": 0.0, "num_input_tokens_seen": 4867360, "step": 17385 }, { "epoch": 193.22222222222223, "grad_norm": 5.8980695030186325e-06, "learning_rate": 3.0090830183722817e-05, "loss": 0.0, "num_input_tokens_seen": 4868784, "step": 17390 }, { "epoch": 193.27777777777777, "grad_norm": 4.906285653305531e-07, "learning_rate": 3.0081218017538852e-05, "loss": 0.0, "num_input_tokens_seen": 4870208, "step": 17395 }, { "epoch": 193.33333333333334, "grad_norm": 2.470158619871654e-07, "learning_rate": 3.0071605067767212e-05, "loss": 0.0, "num_input_tokens_seen": 4871600, "step": 17400 }, { "epoch": 193.33333333333334, "eval_loss": 1.3269317150115967, "eval_runtime": 1.3917, "eval_samples_per_second": 28.742, "eval_steps_per_second": 14.371, "num_input_tokens_seen": 4871600, "step": 17400 }, { "epoch": 193.38888888888889, "grad_norm": 4.6450006152554124e-07, "learning_rate": 3.006199133589034e-05, "loss": 0.0, "num_input_tokens_seen": 4872976, "step": 17405 }, { "epoch": 193.44444444444446, "grad_norm": 1.9880175727848837e-07, "learning_rate": 3.005237682339079e-05, "loss": 0.0, "num_input_tokens_seen": 4874400, "step": 17410 }, { "epoch": 193.5, "grad_norm": 5.842030077474192e-07, "learning_rate": 3.0042761531751228e-05, "loss": 0.0, "num_input_tokens_seen": 4875792, "step": 17415 }, { "epoch": 193.55555555555554, "grad_norm": 1.6231439303737716e-06, "learning_rate": 3.0033145462454482e-05, "loss": 0.0, "num_input_tokens_seen": 4877200, "step": 17420 }, { "epoch": 193.61111111111111, "grad_norm": 6.961520853110414e-07, "learning_rate": 3.002352861698345e-05, "loss": 0.0, "num_input_tokens_seen": 4878624, "step": 17425 }, { "epoch": 193.66666666666666, "grad_norm": 1.85508994832162e-07, "learning_rate": 3.0013910996821178e-05, "loss": 0.0, "num_input_tokens_seen": 4880048, "step": 17430 }, { "epoch": 193.72222222222223, "grad_norm": 1.978156660698005e-06, "learning_rate": 3.0004292603450817e-05, "loss": 0.0, "num_input_tokens_seen": 4881440, "step": 17435 }, { "epoch": 193.77777777777777, "grad_norm": 5.86852252126846e-07, "learning_rate": 2.9994673438355653e-05, "loss": 0.0, "num_input_tokens_seen": 4882880, "step": 17440 }, { "epoch": 193.83333333333334, "grad_norm": 8.572923775318486e-07, "learning_rate": 2.9985053503019078e-05, "loss": 0.0, "num_input_tokens_seen": 4884272, "step": 17445 }, { "epoch": 193.88888888888889, "grad_norm": 6.473977123278019e-07, "learning_rate": 2.99754327989246e-05, "loss": 0.0, "num_input_tokens_seen": 4885664, "step": 17450 }, { "epoch": 193.94444444444446, "grad_norm": 4.998100848752074e-07, "learning_rate": 2.9965811327555864e-05, "loss": 0.0, "num_input_tokens_seen": 4887072, "step": 17455 }, { "epoch": 194.0, "grad_norm": 4.6876004944351735e-07, "learning_rate": 2.995618909039662e-05, "loss": 0.0, "num_input_tokens_seen": 4888464, "step": 17460 }, { "epoch": 194.05555555555554, "grad_norm": 1.6969837588476366e-06, "learning_rate": 2.9946566088930727e-05, "loss": 0.0, "num_input_tokens_seen": 4889856, "step": 17465 }, { "epoch": 194.11111111111111, "grad_norm": 4.17709458133686e-07, "learning_rate": 2.9936942324642192e-05, "loss": 0.0, "num_input_tokens_seen": 4891216, "step": 17470 }, { "epoch": 194.16666666666666, "grad_norm": 4.018883146272856e-07, "learning_rate": 2.9927317799015097e-05, "loss": 0.0, "num_input_tokens_seen": 4892640, "step": 17475 }, { "epoch": 194.22222222222223, "grad_norm": 3.3710441016410186e-07, "learning_rate": 2.9917692513533685e-05, "loss": 0.0, "num_input_tokens_seen": 4894048, "step": 17480 }, { "epoch": 194.27777777777777, "grad_norm": 1.0469049271932818e-07, "learning_rate": 2.990806646968229e-05, "loss": 0.0, "num_input_tokens_seen": 4895472, "step": 17485 }, { "epoch": 194.33333333333334, "grad_norm": 2.0904695929857553e-07, "learning_rate": 2.989843966894536e-05, "loss": 0.0, "num_input_tokens_seen": 4896848, "step": 17490 }, { "epoch": 194.38888888888889, "grad_norm": 4.7268827074731234e-07, "learning_rate": 2.9888812112807472e-05, "loss": 0.0, "num_input_tokens_seen": 4898224, "step": 17495 }, { "epoch": 194.44444444444446, "grad_norm": 2.396476759258803e-07, "learning_rate": 2.987918380275333e-05, "loss": 0.0, "num_input_tokens_seen": 4899616, "step": 17500 }, { "epoch": 194.5, "grad_norm": 1.7095080693252385e-06, "learning_rate": 2.9869554740267724e-05, "loss": 0.0, "num_input_tokens_seen": 4901024, "step": 17505 }, { "epoch": 194.55555555555554, "grad_norm": 5.176308945920027e-07, "learning_rate": 2.9859924926835585e-05, "loss": 0.0, "num_input_tokens_seen": 4902432, "step": 17510 }, { "epoch": 194.61111111111111, "grad_norm": 6.867405204502575e-07, "learning_rate": 2.9850294363941944e-05, "loss": 0.0, "num_input_tokens_seen": 4903872, "step": 17515 }, { "epoch": 194.66666666666666, "grad_norm": 1.125934090850933e-06, "learning_rate": 2.9840663053071967e-05, "loss": 0.0, "num_input_tokens_seen": 4905248, "step": 17520 }, { "epoch": 194.72222222222223, "grad_norm": 5.808834657727857e-07, "learning_rate": 2.983103099571091e-05, "loss": 0.0, "num_input_tokens_seen": 4906656, "step": 17525 }, { "epoch": 194.77777777777777, "grad_norm": 5.45049942957121e-07, "learning_rate": 2.9821398193344164e-05, "loss": 0.0, "num_input_tokens_seen": 4908080, "step": 17530 }, { "epoch": 194.83333333333334, "grad_norm": 5.071753435004211e-07, "learning_rate": 2.9811764647457226e-05, "loss": 0.0, "num_input_tokens_seen": 4909488, "step": 17535 }, { "epoch": 194.88888888888889, "grad_norm": 4.817692911274207e-07, "learning_rate": 2.9802130359535714e-05, "loss": 0.0, "num_input_tokens_seen": 4910880, "step": 17540 }, { "epoch": 194.94444444444446, "grad_norm": 2.0319667726198531e-07, "learning_rate": 2.979249533106535e-05, "loss": 0.0, "num_input_tokens_seen": 4912224, "step": 17545 }, { "epoch": 195.0, "grad_norm": 5.349837692847359e-07, "learning_rate": 2.9782859563531986e-05, "loss": 0.0, "num_input_tokens_seen": 4913616, "step": 17550 }, { "epoch": 195.05555555555554, "grad_norm": 3.013417426700471e-06, "learning_rate": 2.977322305842156e-05, "loss": 0.0, "num_input_tokens_seen": 4915024, "step": 17555 }, { "epoch": 195.11111111111111, "grad_norm": 5.340944539966586e-07, "learning_rate": 2.9763585817220162e-05, "loss": 0.0, "num_input_tokens_seen": 4916496, "step": 17560 }, { "epoch": 195.16666666666666, "grad_norm": 4.1749254364731314e-07, "learning_rate": 2.975394784141397e-05, "loss": 0.0, "num_input_tokens_seen": 4917936, "step": 17565 }, { "epoch": 195.22222222222223, "grad_norm": 1.8703340174397454e-06, "learning_rate": 2.974430913248928e-05, "loss": 0.0, "num_input_tokens_seen": 4919392, "step": 17570 }, { "epoch": 195.27777777777777, "grad_norm": 1.3510618828149745e-06, "learning_rate": 2.9734669691932497e-05, "loss": 0.0, "num_input_tokens_seen": 4920752, "step": 17575 }, { "epoch": 195.33333333333334, "grad_norm": 4.771629278366163e-07, "learning_rate": 2.9725029521230147e-05, "loss": 0.0, "num_input_tokens_seen": 4922144, "step": 17580 }, { "epoch": 195.38888888888889, "grad_norm": 6.06843741479679e-07, "learning_rate": 2.9715388621868873e-05, "loss": 0.0, "num_input_tokens_seen": 4923536, "step": 17585 }, { "epoch": 195.44444444444446, "grad_norm": 7.085856736921414e-07, "learning_rate": 2.970574699533541e-05, "loss": 0.0, "num_input_tokens_seen": 4924928, "step": 17590 }, { "epoch": 195.5, "grad_norm": 4.1879764012264786e-07, "learning_rate": 2.969610464311662e-05, "loss": 0.0, "num_input_tokens_seen": 4926288, "step": 17595 }, { "epoch": 195.55555555555554, "grad_norm": 6.622829573643685e-07, "learning_rate": 2.9686461566699487e-05, "loss": 0.0, "num_input_tokens_seen": 4927696, "step": 17600 }, { "epoch": 195.55555555555554, "eval_loss": 1.2976744174957275, "eval_runtime": 1.3857, "eval_samples_per_second": 28.865, "eval_steps_per_second": 14.433, "num_input_tokens_seen": 4927696, "step": 17600 }, { "epoch": 195.61111111111111, "grad_norm": 1.5064661056385376e-06, "learning_rate": 2.9676817767571086e-05, "loss": 0.0, "num_input_tokens_seen": 4929040, "step": 17605 }, { "epoch": 195.66666666666666, "grad_norm": 1.3789305057798629e-06, "learning_rate": 2.966717324721861e-05, "loss": 0.0, "num_input_tokens_seen": 4930432, "step": 17610 }, { "epoch": 195.72222222222223, "grad_norm": 7.653180773559143e-07, "learning_rate": 2.9657528007129366e-05, "loss": 0.0, "num_input_tokens_seen": 4931840, "step": 17615 }, { "epoch": 195.77777777777777, "grad_norm": 4.763550123243476e-07, "learning_rate": 2.9647882048790777e-05, "loss": 0.0, "num_input_tokens_seen": 4933216, "step": 17620 }, { "epoch": 195.83333333333334, "grad_norm": 2.421123781459755e-07, "learning_rate": 2.963823537369037e-05, "loss": 0.0, "num_input_tokens_seen": 4934576, "step": 17625 }, { "epoch": 195.88888888888889, "grad_norm": 3.100422190982499e-06, "learning_rate": 2.9628587983315775e-05, "loss": 0.0, "num_input_tokens_seen": 4936016, "step": 17630 }, { "epoch": 195.94444444444446, "grad_norm": 8.739907570998184e-07, "learning_rate": 2.9618939879154746e-05, "loss": 0.0, "num_input_tokens_seen": 4937424, "step": 17635 }, { "epoch": 196.0, "grad_norm": 1.8577102309791371e-06, "learning_rate": 2.9609291062695143e-05, "loss": 0.0, "num_input_tokens_seen": 4938816, "step": 17640 }, { "epoch": 196.05555555555554, "grad_norm": 4.333805463829776e-06, "learning_rate": 2.9599641535424938e-05, "loss": 0.0, "num_input_tokens_seen": 4940224, "step": 17645 }, { "epoch": 196.11111111111111, "grad_norm": 6.4461414694960695e-06, "learning_rate": 2.9589991298832202e-05, "loss": 0.0, "num_input_tokens_seen": 4941584, "step": 17650 }, { "epoch": 196.16666666666666, "grad_norm": 3.2367115636589006e-06, "learning_rate": 2.958034035440513e-05, "loss": 0.0, "num_input_tokens_seen": 4942944, "step": 17655 }, { "epoch": 196.22222222222223, "grad_norm": 6.692825422760507e-07, "learning_rate": 2.957068870363201e-05, "loss": 0.0, "num_input_tokens_seen": 4944384, "step": 17660 }, { "epoch": 196.27777777777777, "grad_norm": 3.125352634469891e-07, "learning_rate": 2.956103634800126e-05, "loss": 0.0, "num_input_tokens_seen": 4945792, "step": 17665 }, { "epoch": 196.33333333333334, "grad_norm": 6.50401318580407e-07, "learning_rate": 2.9551383289001384e-05, "loss": 0.0, "num_input_tokens_seen": 4947200, "step": 17670 }, { "epoch": 196.38888888888889, "grad_norm": 3.6122986557529657e-07, "learning_rate": 2.9541729528121005e-05, "loss": 0.0, "num_input_tokens_seen": 4948608, "step": 17675 }, { "epoch": 196.44444444444446, "grad_norm": 1.0642099823598983e-06, "learning_rate": 2.9532075066848856e-05, "loss": 0.0, "num_input_tokens_seen": 4949984, "step": 17680 }, { "epoch": 196.5, "grad_norm": 2.0859432936504163e-07, "learning_rate": 2.9522419906673786e-05, "loss": 0.0, "num_input_tokens_seen": 4951328, "step": 17685 }, { "epoch": 196.55555555555554, "grad_norm": 3.505410859361291e-07, "learning_rate": 2.951276404908474e-05, "loss": 0.0, "num_input_tokens_seen": 4952720, "step": 17690 }, { "epoch": 196.61111111111111, "grad_norm": 8.750640176913294e-07, "learning_rate": 2.9503107495570752e-05, "loss": 0.0, "num_input_tokens_seen": 4954160, "step": 17695 }, { "epoch": 196.66666666666666, "grad_norm": 2.080652024005758e-07, "learning_rate": 2.9493450247621003e-05, "loss": 0.0, "num_input_tokens_seen": 4955568, "step": 17700 }, { "epoch": 196.72222222222223, "grad_norm": 1.6700073501851875e-06, "learning_rate": 2.948379230672476e-05, "loss": 0.0, "num_input_tokens_seen": 4956944, "step": 17705 }, { "epoch": 196.77777777777777, "grad_norm": 3.1104090680855734e-07, "learning_rate": 2.9474133674371396e-05, "loss": 0.0, "num_input_tokens_seen": 4958288, "step": 17710 }, { "epoch": 196.83333333333334, "grad_norm": 5.436684773485467e-07, "learning_rate": 2.9464474352050387e-05, "loss": 0.0, "num_input_tokens_seen": 4959616, "step": 17715 }, { "epoch": 196.88888888888889, "grad_norm": 2.977568556161714e-07, "learning_rate": 2.9454814341251336e-05, "loss": 0.0, "num_input_tokens_seen": 4961024, "step": 17720 }, { "epoch": 196.94444444444446, "grad_norm": 1.701085011518444e-06, "learning_rate": 2.9445153643463942e-05, "loss": 0.0, "num_input_tokens_seen": 4962448, "step": 17725 }, { "epoch": 197.0, "grad_norm": 9.783974519450567e-07, "learning_rate": 2.943549226017798e-05, "loss": 0.0, "num_input_tokens_seen": 4963872, "step": 17730 }, { "epoch": 197.05555555555554, "grad_norm": 3.780621966598119e-07, "learning_rate": 2.942583019288337e-05, "loss": 0.0, "num_input_tokens_seen": 4965280, "step": 17735 }, { "epoch": 197.11111111111111, "grad_norm": 2.0964380382793024e-06, "learning_rate": 2.9416167443070132e-05, "loss": 0.0, "num_input_tokens_seen": 4966688, "step": 17740 }, { "epoch": 197.16666666666666, "grad_norm": 3.3027893664439034e-07, "learning_rate": 2.9406504012228375e-05, "loss": 0.0, "num_input_tokens_seen": 4968064, "step": 17745 }, { "epoch": 197.22222222222223, "grad_norm": 1.8854070731322281e-06, "learning_rate": 2.939683990184832e-05, "loss": 0.0, "num_input_tokens_seen": 4969408, "step": 17750 }, { "epoch": 197.27777777777777, "grad_norm": 2.0973901371235115e-07, "learning_rate": 2.93871751134203e-05, "loss": 0.0, "num_input_tokens_seen": 4970752, "step": 17755 }, { "epoch": 197.33333333333334, "grad_norm": 4.205057280159963e-07, "learning_rate": 2.9377509648434752e-05, "loss": 0.0, "num_input_tokens_seen": 4972176, "step": 17760 }, { "epoch": 197.38888888888889, "grad_norm": 5.236457809587591e-07, "learning_rate": 2.9367843508382203e-05, "loss": 0.0, "num_input_tokens_seen": 4973600, "step": 17765 }, { "epoch": 197.44444444444446, "grad_norm": 1.1400891253288137e-07, "learning_rate": 2.9358176694753293e-05, "loss": 0.0, "num_input_tokens_seen": 4974992, "step": 17770 }, { "epoch": 197.5, "grad_norm": 6.124282094788214e-07, "learning_rate": 2.9348509209038766e-05, "loss": 0.0, "num_input_tokens_seen": 4976384, "step": 17775 }, { "epoch": 197.55555555555554, "grad_norm": 3.3505185115245695e-07, "learning_rate": 2.933884105272947e-05, "loss": 0.0, "num_input_tokens_seen": 4977808, "step": 17780 }, { "epoch": 197.61111111111111, "grad_norm": 1.5432256361691543e-07, "learning_rate": 2.9329172227316366e-05, "loss": 0.0, "num_input_tokens_seen": 4979216, "step": 17785 }, { "epoch": 197.66666666666666, "grad_norm": 6.205245881574228e-07, "learning_rate": 2.93195027342905e-05, "loss": 0.0, "num_input_tokens_seen": 4980592, "step": 17790 }, { "epoch": 197.72222222222223, "grad_norm": 3.34731254270082e-07, "learning_rate": 2.9309832575143024e-05, "loss": 0.0, "num_input_tokens_seen": 4982016, "step": 17795 }, { "epoch": 197.77777777777777, "grad_norm": 2.9170976745263033e-07, "learning_rate": 2.930016175136521e-05, "loss": 0.0, "num_input_tokens_seen": 4983424, "step": 17800 }, { "epoch": 197.77777777777777, "eval_loss": 1.3143818378448486, "eval_runtime": 1.4002, "eval_samples_per_second": 28.567, "eval_steps_per_second": 14.283, "num_input_tokens_seen": 4983424, "step": 17800 }, { "epoch": 197.83333333333334, "grad_norm": 2.609556304378202e-07, "learning_rate": 2.9290490264448412e-05, "loss": 0.0, "num_input_tokens_seen": 4984816, "step": 17805 }, { "epoch": 197.88888888888889, "grad_norm": 6.640873380092671e-07, "learning_rate": 2.9280818115884094e-05, "loss": 0.0, "num_input_tokens_seen": 4986240, "step": 17810 }, { "epoch": 197.94444444444446, "grad_norm": 5.674580734194024e-07, "learning_rate": 2.9271145307163828e-05, "loss": 0.0, "num_input_tokens_seen": 4987664, "step": 17815 }, { "epoch": 198.0, "grad_norm": 1.2622006124729523e-06, "learning_rate": 2.9261471839779287e-05, "loss": 0.0, "num_input_tokens_seen": 4989088, "step": 17820 }, { "epoch": 198.05555555555554, "grad_norm": 4.891361413683626e-07, "learning_rate": 2.925179771522223e-05, "loss": 0.0, "num_input_tokens_seen": 4990496, "step": 17825 }, { "epoch": 198.11111111111111, "grad_norm": 4.06556750931486e-07, "learning_rate": 2.9242122934984535e-05, "loss": 0.0, "num_input_tokens_seen": 4991888, "step": 17830 }, { "epoch": 198.16666666666666, "grad_norm": 3.234214545955183e-06, "learning_rate": 2.9232447500558176e-05, "loss": 0.0, "num_input_tokens_seen": 4993280, "step": 17835 }, { "epoch": 198.22222222222223, "grad_norm": 2.4654727326378634e-07, "learning_rate": 2.9222771413435225e-05, "loss": 0.0, "num_input_tokens_seen": 4994688, "step": 17840 }, { "epoch": 198.27777777777777, "grad_norm": 8.925726433517411e-07, "learning_rate": 2.9213094675107848e-05, "loss": 0.0, "num_input_tokens_seen": 4996112, "step": 17845 }, { "epoch": 198.33333333333334, "grad_norm": 3.1716086823507794e-07, "learning_rate": 2.9203417287068335e-05, "loss": 0.0, "num_input_tokens_seen": 4997520, "step": 17850 }, { "epoch": 198.38888888888889, "grad_norm": 1.0221572210866725e-06, "learning_rate": 2.9193739250809042e-05, "loss": 0.0, "num_input_tokens_seen": 4998928, "step": 17855 }, { "epoch": 198.44444444444446, "grad_norm": 2.2706902313984756e-07, "learning_rate": 2.9184060567822463e-05, "loss": 0.0, "num_input_tokens_seen": 5000288, "step": 17860 }, { "epoch": 198.5, "grad_norm": 3.3280974776062067e-07, "learning_rate": 2.9174381239601166e-05, "loss": 0.0, "num_input_tokens_seen": 5001712, "step": 17865 }, { "epoch": 198.55555555555554, "grad_norm": 2.6645705020200694e-07, "learning_rate": 2.916470126763783e-05, "loss": 0.0, "num_input_tokens_seen": 5003104, "step": 17870 }, { "epoch": 198.61111111111111, "grad_norm": 1.7663641074250336e-06, "learning_rate": 2.9155020653425203e-05, "loss": 0.0, "num_input_tokens_seen": 5004496, "step": 17875 }, { "epoch": 198.66666666666666, "grad_norm": 8.789848493506724e-07, "learning_rate": 2.9145339398456184e-05, "loss": 0.0, "num_input_tokens_seen": 5005936, "step": 17880 }, { "epoch": 198.72222222222223, "grad_norm": 7.063182465572027e-07, "learning_rate": 2.913565750422374e-05, "loss": 0.0, "num_input_tokens_seen": 5007328, "step": 17885 }, { "epoch": 198.77777777777777, "grad_norm": 8.554910095881496e-07, "learning_rate": 2.9125974972220938e-05, "loss": 0.0, "num_input_tokens_seen": 5008720, "step": 17890 }, { "epoch": 198.83333333333334, "grad_norm": 7.184329433584935e-07, "learning_rate": 2.9116291803940932e-05, "loss": 0.0, "num_input_tokens_seen": 5010112, "step": 17895 }, { "epoch": 198.88888888888889, "grad_norm": 5.588166231973446e-07, "learning_rate": 2.910660800087701e-05, "loss": 0.0, "num_input_tokens_seen": 5011520, "step": 17900 }, { "epoch": 198.94444444444446, "grad_norm": 1.3025138514421997e-06, "learning_rate": 2.909692356452254e-05, "loss": 0.0, "num_input_tokens_seen": 5012960, "step": 17905 }, { "epoch": 199.0, "grad_norm": 2.483573098288616e-07, "learning_rate": 2.9087238496370962e-05, "loss": 0.0, "num_input_tokens_seen": 5014352, "step": 17910 }, { "epoch": 199.05555555555554, "grad_norm": 3.5202745607421093e-07, "learning_rate": 2.907755279791583e-05, "loss": 0.0, "num_input_tokens_seen": 5015728, "step": 17915 }, { "epoch": 199.11111111111111, "grad_norm": 3.379418274107593e-07, "learning_rate": 2.906786647065083e-05, "loss": 0.0, "num_input_tokens_seen": 5017152, "step": 17920 }, { "epoch": 199.16666666666666, "grad_norm": 6.462101964643807e-07, "learning_rate": 2.9058179516069695e-05, "loss": 0.0, "num_input_tokens_seen": 5018560, "step": 17925 }, { "epoch": 199.22222222222223, "grad_norm": 3.828203603006841e-07, "learning_rate": 2.9048491935666282e-05, "loss": 0.0, "num_input_tokens_seen": 5019936, "step": 17930 }, { "epoch": 199.27777777777777, "grad_norm": 2.176000890585783e-07, "learning_rate": 2.9038803730934534e-05, "loss": 0.0, "num_input_tokens_seen": 5021344, "step": 17935 }, { "epoch": 199.33333333333334, "grad_norm": 3.865186215534777e-07, "learning_rate": 2.9029114903368503e-05, "loss": 0.0, "num_input_tokens_seen": 5022704, "step": 17940 }, { "epoch": 199.38888888888889, "grad_norm": 7.838004876248306e-07, "learning_rate": 2.9019425454462318e-05, "loss": 0.0, "num_input_tokens_seen": 5024112, "step": 17945 }, { "epoch": 199.44444444444446, "grad_norm": 3.092251006364677e-07, "learning_rate": 2.9009735385710212e-05, "loss": 0.0, "num_input_tokens_seen": 5025504, "step": 17950 }, { "epoch": 199.5, "grad_norm": 6.921434305695584e-07, "learning_rate": 2.900004469860652e-05, "loss": 0.0, "num_input_tokens_seen": 5026912, "step": 17955 }, { "epoch": 199.55555555555554, "grad_norm": 5.3646172091248445e-06, "learning_rate": 2.8990353394645668e-05, "loss": 0.0, "num_input_tokens_seen": 5028304, "step": 17960 }, { "epoch": 199.61111111111111, "grad_norm": 4.872640033681819e-07, "learning_rate": 2.8980661475322186e-05, "loss": 0.0, "num_input_tokens_seen": 5029728, "step": 17965 }, { "epoch": 199.66666666666666, "grad_norm": 3.453455292401486e-06, "learning_rate": 2.897096894213067e-05, "loss": 0.0, "num_input_tokens_seen": 5031120, "step": 17970 }, { "epoch": 199.72222222222223, "grad_norm": 5.741646873502759e-07, "learning_rate": 2.8961275796565845e-05, "loss": 0.0, "num_input_tokens_seen": 5032512, "step": 17975 }, { "epoch": 199.77777777777777, "grad_norm": 3.7176758382884145e-07, "learning_rate": 2.8951582040122517e-05, "loss": 0.0, "num_input_tokens_seen": 5033920, "step": 17980 }, { "epoch": 199.83333333333334, "grad_norm": 2.481688170519192e-07, "learning_rate": 2.894188767429557e-05, "loss": 0.0, "num_input_tokens_seen": 5035312, "step": 17985 }, { "epoch": 199.88888888888889, "grad_norm": 9.207785751641495e-07, "learning_rate": 2.8932192700580014e-05, "loss": 0.0, "num_input_tokens_seen": 5036720, "step": 17990 }, { "epoch": 199.94444444444446, "grad_norm": 4.7520839530079684e-07, "learning_rate": 2.8922497120470916e-05, "loss": 0.0, "num_input_tokens_seen": 5038112, "step": 17995 }, { "epoch": 200.0, "grad_norm": 3.2234396485364414e-07, "learning_rate": 2.891280093546348e-05, "loss": 0.0, "num_input_tokens_seen": 5039536, "step": 18000 }, { "epoch": 200.0, "eval_loss": 1.2850013971328735, "eval_runtime": 1.3902, "eval_samples_per_second": 28.773, "eval_steps_per_second": 14.386, "num_input_tokens_seen": 5039536, "step": 18000 }, { "epoch": 200.05555555555554, "grad_norm": 5.762768751083058e-07, "learning_rate": 2.890310414705297e-05, "loss": 0.0, "num_input_tokens_seen": 5040944, "step": 18005 }, { "epoch": 200.11111111111111, "grad_norm": 1.8955246332552633e-06, "learning_rate": 2.8893406756734742e-05, "loss": 0.0, "num_input_tokens_seen": 5042320, "step": 18010 }, { "epoch": 200.16666666666666, "grad_norm": 3.747467474113364e-07, "learning_rate": 2.888370876600427e-05, "loss": 0.0, "num_input_tokens_seen": 5043728, "step": 18015 }, { "epoch": 200.22222222222223, "grad_norm": 7.5579475833365e-07, "learning_rate": 2.8874010176357104e-05, "loss": 0.0, "num_input_tokens_seen": 5045104, "step": 18020 }, { "epoch": 200.27777777777777, "grad_norm": 1.7901290902955225e-06, "learning_rate": 2.886431098928888e-05, "loss": 0.0, "num_input_tokens_seen": 5046512, "step": 18025 }, { "epoch": 200.33333333333334, "grad_norm": 2.3813529992366966e-07, "learning_rate": 2.885461120629534e-05, "loss": 0.0, "num_input_tokens_seen": 5047920, "step": 18030 }, { "epoch": 200.38888888888889, "grad_norm": 5.901927693230391e-07, "learning_rate": 2.8844910828872317e-05, "loss": 0.0, "num_input_tokens_seen": 5049328, "step": 18035 }, { "epoch": 200.44444444444446, "grad_norm": 5.548443482439325e-07, "learning_rate": 2.8835209858515715e-05, "loss": 0.0, "num_input_tokens_seen": 5050688, "step": 18040 }, { "epoch": 200.5, "grad_norm": 4.274816092220135e-07, "learning_rate": 2.8825508296721566e-05, "loss": 0.0, "num_input_tokens_seen": 5052080, "step": 18045 }, { "epoch": 200.55555555555554, "grad_norm": 5.070608040114166e-07, "learning_rate": 2.881580614498596e-05, "loss": 0.0, "num_input_tokens_seen": 5053504, "step": 18050 }, { "epoch": 200.61111111111111, "grad_norm": 5.644430416396062e-07, "learning_rate": 2.8806103404805103e-05, "loss": 0.0, "num_input_tokens_seen": 5054864, "step": 18055 }, { "epoch": 200.66666666666666, "grad_norm": 2.495073090358346e-07, "learning_rate": 2.8796400077675257e-05, "loss": 0.0, "num_input_tokens_seen": 5056224, "step": 18060 }, { "epoch": 200.72222222222223, "grad_norm": 2.5047251028809114e-07, "learning_rate": 2.8786696165092812e-05, "loss": 0.0, "num_input_tokens_seen": 5057600, "step": 18065 }, { "epoch": 200.77777777777777, "grad_norm": 4.84048371163226e-07, "learning_rate": 2.8776991668554236e-05, "loss": 0.0, "num_input_tokens_seen": 5058992, "step": 18070 }, { "epoch": 200.83333333333334, "grad_norm": 1.2406137557263719e-06, "learning_rate": 2.876728658955608e-05, "loss": 0.0, "num_input_tokens_seen": 5060432, "step": 18075 }, { "epoch": 200.88888888888889, "grad_norm": 1.1381887361494591e-06, "learning_rate": 2.8757580929594986e-05, "loss": 0.0, "num_input_tokens_seen": 5061872, "step": 18080 }, { "epoch": 200.94444444444446, "grad_norm": 2.0593611793628952e-07, "learning_rate": 2.87478746901677e-05, "loss": 0.0, "num_input_tokens_seen": 5063264, "step": 18085 }, { "epoch": 201.0, "grad_norm": 4.169740464021743e-07, "learning_rate": 2.873816787277103e-05, "loss": 0.0, "num_input_tokens_seen": 5064672, "step": 18090 }, { "epoch": 201.05555555555554, "grad_norm": 1.723095266470409e-07, "learning_rate": 2.8728460478901903e-05, "loss": 0.0, "num_input_tokens_seen": 5066080, "step": 18095 }, { "epoch": 201.11111111111111, "grad_norm": 5.739986477237835e-07, "learning_rate": 2.8718752510057307e-05, "loss": 0.0, "num_input_tokens_seen": 5067424, "step": 18100 }, { "epoch": 201.16666666666666, "grad_norm": 4.437328584572242e-07, "learning_rate": 2.870904396773435e-05, "loss": 0.0, "num_input_tokens_seen": 5068816, "step": 18105 }, { "epoch": 201.22222222222223, "grad_norm": 1.7312576972017268e-07, "learning_rate": 2.86993348534302e-05, "loss": 0.0, "num_input_tokens_seen": 5070192, "step": 18110 }, { "epoch": 201.27777777777777, "grad_norm": 9.481260008215031e-07, "learning_rate": 2.868962516864212e-05, "loss": 0.0, "num_input_tokens_seen": 5071600, "step": 18115 }, { "epoch": 201.33333333333334, "grad_norm": 3.147909183098818e-06, "learning_rate": 2.8679914914867477e-05, "loss": 0.0, "num_input_tokens_seen": 5072992, "step": 18120 }, { "epoch": 201.38888888888889, "grad_norm": 1.6285308390706632e-07, "learning_rate": 2.8670204093603713e-05, "loss": 0.0, "num_input_tokens_seen": 5074368, "step": 18125 }, { "epoch": 201.44444444444446, "grad_norm": 4.7219413090715534e-07, "learning_rate": 2.8660492706348357e-05, "loss": 0.0, "num_input_tokens_seen": 5075744, "step": 18130 }, { "epoch": 201.5, "grad_norm": 6.91084665049857e-07, "learning_rate": 2.8650780754599022e-05, "loss": 0.0, "num_input_tokens_seen": 5077168, "step": 18135 }, { "epoch": 201.55555555555554, "grad_norm": 1.1682154621439622e-07, "learning_rate": 2.8641068239853407e-05, "loss": 0.0, "num_input_tokens_seen": 5078560, "step": 18140 }, { "epoch": 201.61111111111111, "grad_norm": 1.879255648873368e-07, "learning_rate": 2.863135516360932e-05, "loss": 0.0, "num_input_tokens_seen": 5079968, "step": 18145 }, { "epoch": 201.66666666666666, "grad_norm": 8.173767582775326e-07, "learning_rate": 2.8621641527364633e-05, "loss": 0.0, "num_input_tokens_seen": 5081392, "step": 18150 }, { "epoch": 201.72222222222223, "grad_norm": 4.843470833293395e-07, "learning_rate": 2.8611927332617313e-05, "loss": 0.0, "num_input_tokens_seen": 5082784, "step": 18155 }, { "epoch": 201.77777777777777, "grad_norm": 7.194791464826267e-07, "learning_rate": 2.8602212580865405e-05, "loss": 0.0, "num_input_tokens_seen": 5084192, "step": 18160 }, { "epoch": 201.83333333333334, "grad_norm": 2.477079590335052e-07, "learning_rate": 2.859249727360705e-05, "loss": 0.0, "num_input_tokens_seen": 5085584, "step": 18165 }, { "epoch": 201.88888888888889, "grad_norm": 4.871527039540524e-07, "learning_rate": 2.8582781412340465e-05, "loss": 0.0, "num_input_tokens_seen": 5086992, "step": 18170 }, { "epoch": 201.94444444444446, "grad_norm": 2.985724734116957e-07, "learning_rate": 2.857306499856397e-05, "loss": 0.0, "num_input_tokens_seen": 5088448, "step": 18175 }, { "epoch": 202.0, "grad_norm": 2.1982330622449808e-07, "learning_rate": 2.856334803377594e-05, "loss": 0.0, "num_input_tokens_seen": 5089808, "step": 18180 }, { "epoch": 202.05555555555554, "grad_norm": 4.907968786938e-07, "learning_rate": 2.8553630519474867e-05, "loss": 0.0, "num_input_tokens_seen": 5091152, "step": 18185 }, { "epoch": 202.11111111111111, "grad_norm": 4.6562976763198094e-07, "learning_rate": 2.8543912457159317e-05, "loss": 0.0, "num_input_tokens_seen": 5092544, "step": 18190 }, { "epoch": 202.16666666666666, "grad_norm": 3.7478000081136997e-07, "learning_rate": 2.853419384832792e-05, "loss": 0.0, "num_input_tokens_seen": 5093968, "step": 18195 }, { "epoch": 202.22222222222223, "grad_norm": 3.607636358538002e-07, "learning_rate": 2.8524474694479423e-05, "loss": 0.0, "num_input_tokens_seen": 5095376, "step": 18200 }, { "epoch": 202.22222222222223, "eval_loss": 1.3127597570419312, "eval_runtime": 1.3832, "eval_samples_per_second": 28.918, "eval_steps_per_second": 14.459, "num_input_tokens_seen": 5095376, "step": 18200 }, { "epoch": 202.27777777777777, "grad_norm": 4.0530628098167654e-07, "learning_rate": 2.851475499711264e-05, "loss": 0.0, "num_input_tokens_seen": 5096784, "step": 18205 }, { "epoch": 202.33333333333334, "grad_norm": 4.1359797364748374e-07, "learning_rate": 2.8505034757726468e-05, "loss": 0.0, "num_input_tokens_seen": 5098160, "step": 18210 }, { "epoch": 202.38888888888889, "grad_norm": 2.0013717403344344e-06, "learning_rate": 2.8495313977819886e-05, "loss": 0.0, "num_input_tokens_seen": 5099600, "step": 18215 }, { "epoch": 202.44444444444446, "grad_norm": 2.1333826794034394e-07, "learning_rate": 2.8485592658891956e-05, "loss": 0.0, "num_input_tokens_seen": 5101008, "step": 18220 }, { "epoch": 202.5, "grad_norm": 2.966644672142138e-07, "learning_rate": 2.8475870802441844e-05, "loss": 0.0, "num_input_tokens_seen": 5102464, "step": 18225 }, { "epoch": 202.55555555555554, "grad_norm": 5.52009964849276e-07, "learning_rate": 2.8466148409968774e-05, "loss": 0.0, "num_input_tokens_seen": 5103904, "step": 18230 }, { "epoch": 202.61111111111111, "grad_norm": 3.3954199807340046e-07, "learning_rate": 2.8456425482972067e-05, "loss": 0.0, "num_input_tokens_seen": 5105312, "step": 18235 }, { "epoch": 202.66666666666666, "grad_norm": 2.6760562832350843e-07, "learning_rate": 2.84467020229511e-05, "loss": 0.0, "num_input_tokens_seen": 5106704, "step": 18240 }, { "epoch": 202.72222222222223, "grad_norm": 2.1170463071484846e-07, "learning_rate": 2.8436978031405375e-05, "loss": 0.0, "num_input_tokens_seen": 5108064, "step": 18245 }, { "epoch": 202.77777777777777, "grad_norm": 4.861478828388499e-07, "learning_rate": 2.842725350983445e-05, "loss": 0.0, "num_input_tokens_seen": 5109472, "step": 18250 }, { "epoch": 202.83333333333334, "grad_norm": 4.5035690732220246e-07, "learning_rate": 2.8417528459737957e-05, "loss": 0.0, "num_input_tokens_seen": 5110864, "step": 18255 }, { "epoch": 202.88888888888889, "grad_norm": 5.615399913949659e-07, "learning_rate": 2.8407802882615624e-05, "loss": 0.0, "num_input_tokens_seen": 5112224, "step": 18260 }, { "epoch": 202.94444444444446, "grad_norm": 7.51273546484299e-07, "learning_rate": 2.8398076779967277e-05, "loss": 0.0, "num_input_tokens_seen": 5113568, "step": 18265 }, { "epoch": 203.0, "grad_norm": 2.2368326426658314e-06, "learning_rate": 2.8388350153292774e-05, "loss": 0.0, "num_input_tokens_seen": 5114960, "step": 18270 }, { "epoch": 203.05555555555554, "grad_norm": 3.6892953403366846e-07, "learning_rate": 2.8378623004092103e-05, "loss": 0.0, "num_input_tokens_seen": 5116320, "step": 18275 }, { "epoch": 203.11111111111111, "grad_norm": 2.1808577912452165e-06, "learning_rate": 2.8368895333865302e-05, "loss": 0.0, "num_input_tokens_seen": 5117728, "step": 18280 }, { "epoch": 203.16666666666666, "grad_norm": 4.420914123670627e-08, "learning_rate": 2.835916714411251e-05, "loss": 0.0, "num_input_tokens_seen": 5119168, "step": 18285 }, { "epoch": 203.22222222222223, "grad_norm": 6.356119115480396e-07, "learning_rate": 2.8349438436333926e-05, "loss": 0.0, "num_input_tokens_seen": 5120608, "step": 18290 }, { "epoch": 203.27777777777777, "grad_norm": 3.47107857123774e-06, "learning_rate": 2.833970921202984e-05, "loss": 0.0, "num_input_tokens_seen": 5122000, "step": 18295 }, { "epoch": 203.33333333333334, "grad_norm": 6.345450174194411e-07, "learning_rate": 2.8329979472700628e-05, "loss": 0.0, "num_input_tokens_seen": 5123424, "step": 18300 }, { "epoch": 203.38888888888889, "grad_norm": 1.7103484424296767e-06, "learning_rate": 2.832024921984674e-05, "loss": 0.0, "num_input_tokens_seen": 5124864, "step": 18305 }, { "epoch": 203.44444444444446, "grad_norm": 5.628435815197008e-07, "learning_rate": 2.8310518454968693e-05, "loss": 0.0, "num_input_tokens_seen": 5126272, "step": 18310 }, { "epoch": 203.5, "grad_norm": 9.082905449986356e-08, "learning_rate": 2.8300787179567095e-05, "loss": 0.0, "num_input_tokens_seen": 5127680, "step": 18315 }, { "epoch": 203.55555555555554, "grad_norm": 1.45874352597275e-07, "learning_rate": 2.8291055395142636e-05, "loss": 0.0, "num_input_tokens_seen": 5129104, "step": 18320 }, { "epoch": 203.61111111111111, "grad_norm": 1.0775223699965863e-06, "learning_rate": 2.8281323103196073e-05, "loss": 0.0, "num_input_tokens_seen": 5130496, "step": 18325 }, { "epoch": 203.66666666666666, "grad_norm": 6.118959277046088e-07, "learning_rate": 2.8271590305228256e-05, "loss": 0.0, "num_input_tokens_seen": 5131904, "step": 18330 }, { "epoch": 203.72222222222223, "grad_norm": 3.20363369610277e-07, "learning_rate": 2.82618570027401e-05, "loss": 0.0, "num_input_tokens_seen": 5133312, "step": 18335 }, { "epoch": 203.77777777777777, "grad_norm": 3.5498902661856846e-07, "learning_rate": 2.8252123197232604e-05, "loss": 0.0, "num_input_tokens_seen": 5134704, "step": 18340 }, { "epoch": 203.83333333333334, "grad_norm": 8.14756845102238e-07, "learning_rate": 2.8242388890206843e-05, "loss": 0.0, "num_input_tokens_seen": 5136096, "step": 18345 }, { "epoch": 203.88888888888889, "grad_norm": 4.916208240501874e-07, "learning_rate": 2.8232654083163967e-05, "loss": 0.0, "num_input_tokens_seen": 5137472, "step": 18350 }, { "epoch": 203.94444444444446, "grad_norm": 5.125745701661799e-06, "learning_rate": 2.822291877760521e-05, "loss": 0.0, "num_input_tokens_seen": 5138896, "step": 18355 }, { "epoch": 204.0, "grad_norm": 1.904602299873659e-06, "learning_rate": 2.8213182975031864e-05, "loss": 0.0, "num_input_tokens_seen": 5140272, "step": 18360 }, { "epoch": 204.05555555555554, "grad_norm": 2.767583282548003e-07, "learning_rate": 2.8203446676945337e-05, "loss": 0.0, "num_input_tokens_seen": 5141616, "step": 18365 }, { "epoch": 204.11111111111111, "grad_norm": 1.9203329770789423e-07, "learning_rate": 2.8193709884847075e-05, "loss": 0.0, "num_input_tokens_seen": 5142976, "step": 18370 }, { "epoch": 204.16666666666666, "grad_norm": 6.32403001077364e-08, "learning_rate": 2.8183972600238605e-05, "loss": 0.0, "num_input_tokens_seen": 5144400, "step": 18375 }, { "epoch": 204.22222222222223, "grad_norm": 1.7004526853270363e-07, "learning_rate": 2.817423482462156e-05, "loss": 0.0, "num_input_tokens_seen": 5145808, "step": 18380 }, { "epoch": 204.27777777777777, "grad_norm": 3.168665330122167e-07, "learning_rate": 2.8164496559497605e-05, "loss": 0.0, "num_input_tokens_seen": 5147216, "step": 18385 }, { "epoch": 204.33333333333334, "grad_norm": 1.42905520306158e-07, "learning_rate": 2.815475780636852e-05, "loss": 0.0, "num_input_tokens_seen": 5148624, "step": 18390 }, { "epoch": 204.38888888888889, "grad_norm": 5.135393053024018e-07, "learning_rate": 2.814501856673613e-05, "loss": 0.0, "num_input_tokens_seen": 5150032, "step": 18395 }, { "epoch": 204.44444444444446, "grad_norm": 2.1959601781418314e-07, "learning_rate": 2.8135278842102353e-05, "loss": 0.0, "num_input_tokens_seen": 5151440, "step": 18400 }, { "epoch": 204.44444444444446, "eval_loss": 1.3168312311172485, "eval_runtime": 1.3833, "eval_samples_per_second": 28.916, "eval_steps_per_second": 14.458, "num_input_tokens_seen": 5151440, "step": 18400 }, { "epoch": 204.5, "grad_norm": 5.558677571571025e-07, "learning_rate": 2.8125538633969183e-05, "loss": 0.0, "num_input_tokens_seen": 5152864, "step": 18405 }, { "epoch": 204.55555555555554, "grad_norm": 2.8651069783336425e-07, "learning_rate": 2.8115797943838677e-05, "loss": 0.0, "num_input_tokens_seen": 5154288, "step": 18410 }, { "epoch": 204.61111111111111, "grad_norm": 6.213873575688922e-07, "learning_rate": 2.810605677321298e-05, "loss": 0.0, "num_input_tokens_seen": 5155664, "step": 18415 }, { "epoch": 204.66666666666666, "grad_norm": 2.258181979186702e-07, "learning_rate": 2.809631512359428e-05, "loss": 0.0, "num_input_tokens_seen": 5157040, "step": 18420 }, { "epoch": 204.72222222222223, "grad_norm": 1.581621177138004e-06, "learning_rate": 2.8086572996484884e-05, "loss": 0.0, "num_input_tokens_seen": 5158464, "step": 18425 }, { "epoch": 204.77777777777777, "grad_norm": 4.6048145918575756e-07, "learning_rate": 2.8076830393387143e-05, "loss": 0.0, "num_input_tokens_seen": 5159872, "step": 18430 }, { "epoch": 204.83333333333334, "grad_norm": 2.2708074993715854e-06, "learning_rate": 2.8067087315803497e-05, "loss": 0.0, "num_input_tokens_seen": 5161280, "step": 18435 }, { "epoch": 204.88888888888889, "grad_norm": 3.658804246242653e-07, "learning_rate": 2.8057343765236433e-05, "loss": 0.0, "num_input_tokens_seen": 5162656, "step": 18440 }, { "epoch": 204.94444444444446, "grad_norm": 2.441975084366277e-07, "learning_rate": 2.804759974318854e-05, "loss": 0.0, "num_input_tokens_seen": 5164064, "step": 18445 }, { "epoch": 205.0, "grad_norm": 1.0617569046189601e-07, "learning_rate": 2.8037855251162482e-05, "loss": 0.0, "num_input_tokens_seen": 5165456, "step": 18450 }, { "epoch": 205.05555555555554, "grad_norm": 3.2946800843092205e-07, "learning_rate": 2.802811029066096e-05, "loss": 0.0, "num_input_tokens_seen": 5166880, "step": 18455 }, { "epoch": 205.11111111111111, "grad_norm": 2.1344122558275558e-07, "learning_rate": 2.8018364863186764e-05, "loss": 0.0, "num_input_tokens_seen": 5168224, "step": 18460 }, { "epoch": 205.16666666666666, "grad_norm": 4.241390456627414e-07, "learning_rate": 2.800861897024279e-05, "loss": 0.0, "num_input_tokens_seen": 5169664, "step": 18465 }, { "epoch": 205.22222222222223, "grad_norm": 3.1443900638805644e-07, "learning_rate": 2.799887261333196e-05, "loss": 0.0, "num_input_tokens_seen": 5171136, "step": 18470 }, { "epoch": 205.27777777777777, "grad_norm": 6.941779133740056e-07, "learning_rate": 2.798912579395728e-05, "loss": 0.0, "num_input_tokens_seen": 5172544, "step": 18475 }, { "epoch": 205.33333333333334, "grad_norm": 1.439190384644462e-07, "learning_rate": 2.797937851362185e-05, "loss": 0.0, "num_input_tokens_seen": 5173920, "step": 18480 }, { "epoch": 205.38888888888889, "grad_norm": 2.623373234200699e-07, "learning_rate": 2.7969630773828802e-05, "loss": 0.0, "num_input_tokens_seen": 5175312, "step": 18485 }, { "epoch": 205.44444444444446, "grad_norm": 5.951608841314737e-07, "learning_rate": 2.7959882576081382e-05, "loss": 0.0, "num_input_tokens_seen": 5176736, "step": 18490 }, { "epoch": 205.5, "grad_norm": 1.8260744809595053e-06, "learning_rate": 2.795013392188286e-05, "loss": 0.0, "num_input_tokens_seen": 5178128, "step": 18495 }, { "epoch": 205.55555555555554, "grad_norm": 3.7696133858844405e-07, "learning_rate": 2.7940384812736614e-05, "loss": 0.0, "num_input_tokens_seen": 5179520, "step": 18500 }, { "epoch": 205.61111111111111, "grad_norm": 8.61964906562207e-07, "learning_rate": 2.7930635250146087e-05, "loss": 0.0, "num_input_tokens_seen": 5180928, "step": 18505 }, { "epoch": 205.66666666666666, "grad_norm": 1.3643806084928656e-07, "learning_rate": 2.792088523561477e-05, "loss": 0.0, "num_input_tokens_seen": 5182272, "step": 18510 }, { "epoch": 205.72222222222223, "grad_norm": 2.840241961621359e-07, "learning_rate": 2.7911134770646246e-05, "loss": 0.0, "num_input_tokens_seen": 5183680, "step": 18515 }, { "epoch": 205.77777777777777, "grad_norm": 4.845530270358722e-07, "learning_rate": 2.7901383856744157e-05, "loss": 0.0, "num_input_tokens_seen": 5185088, "step": 18520 }, { "epoch": 205.83333333333334, "grad_norm": 2.1662880556050368e-07, "learning_rate": 2.7891632495412217e-05, "loss": 0.0, "num_input_tokens_seen": 5186464, "step": 18525 }, { "epoch": 205.88888888888889, "grad_norm": 8.00700377112662e-07, "learning_rate": 2.7881880688154205e-05, "loss": 0.0, "num_input_tokens_seen": 5187840, "step": 18530 }, { "epoch": 205.94444444444446, "grad_norm": 1.9548167529137572e-06, "learning_rate": 2.7872128436473977e-05, "loss": 0.0, "num_input_tokens_seen": 5189248, "step": 18535 }, { "epoch": 206.0, "grad_norm": 4.1573167663955246e-07, "learning_rate": 2.7862375741875448e-05, "loss": 0.0, "num_input_tokens_seen": 5190656, "step": 18540 }, { "epoch": 206.05555555555554, "grad_norm": 4.79608047498914e-07, "learning_rate": 2.785262260586261e-05, "loss": 0.0, "num_input_tokens_seen": 5192048, "step": 18545 }, { "epoch": 206.11111111111111, "grad_norm": 2.5521831048536114e-07, "learning_rate": 2.7842869029939517e-05, "loss": 0.0, "num_input_tokens_seen": 5193440, "step": 18550 }, { "epoch": 206.16666666666666, "grad_norm": 1.7163465315661597e-07, "learning_rate": 2.7833115015610296e-05, "loss": 0.0, "num_input_tokens_seen": 5194848, "step": 18555 }, { "epoch": 206.22222222222223, "grad_norm": 8.566494358319687e-08, "learning_rate": 2.7823360564379136e-05, "loss": 0.0, "num_input_tokens_seen": 5196240, "step": 18560 }, { "epoch": 206.27777777777777, "grad_norm": 1.6921231349442678e-07, "learning_rate": 2.7813605677750297e-05, "loss": 0.0, "num_input_tokens_seen": 5197664, "step": 18565 }, { "epoch": 206.33333333333334, "grad_norm": 1.8829861403446557e-07, "learning_rate": 2.7803850357228102e-05, "loss": 0.0, "num_input_tokens_seen": 5199056, "step": 18570 }, { "epoch": 206.38888888888889, "grad_norm": 3.089524511779018e-07, "learning_rate": 2.779409460431695e-05, "loss": 0.0, "num_input_tokens_seen": 5200480, "step": 18575 }, { "epoch": 206.44444444444446, "grad_norm": 6.428587084883475e-07, "learning_rate": 2.778433842052129e-05, "loss": 0.0, "num_input_tokens_seen": 5201856, "step": 18580 }, { "epoch": 206.5, "grad_norm": 8.539435611965018e-07, "learning_rate": 2.7774581807345664e-05, "loss": 0.0, "num_input_tokens_seen": 5203264, "step": 18585 }, { "epoch": 206.55555555555554, "grad_norm": 6.416606623815824e-08, "learning_rate": 2.776482476629465e-05, "loss": 0.0, "num_input_tokens_seen": 5204672, "step": 18590 }, { "epoch": 206.61111111111111, "grad_norm": 6.075978831177054e-07, "learning_rate": 2.7755067298872924e-05, "loss": 0.0, "num_input_tokens_seen": 5206096, "step": 18595 }, { "epoch": 206.66666666666666, "grad_norm": 2.277718294863007e-06, "learning_rate": 2.774530940658518e-05, "loss": 0.0, "num_input_tokens_seen": 5207488, "step": 18600 }, { "epoch": 206.66666666666666, "eval_loss": 1.3171831369400024, "eval_runtime": 1.3885, "eval_samples_per_second": 28.808, "eval_steps_per_second": 14.404, "num_input_tokens_seen": 5207488, "step": 18600 }, { "epoch": 206.72222222222223, "grad_norm": 3.9986844058148563e-07, "learning_rate": 2.7735551090936236e-05, "loss": 0.0, "num_input_tokens_seen": 5208864, "step": 18605 }, { "epoch": 206.77777777777777, "grad_norm": 8.484022941956937e-07, "learning_rate": 2.7725792353430934e-05, "loss": 0.0, "num_input_tokens_seen": 5210288, "step": 18610 }, { "epoch": 206.83333333333334, "grad_norm": 4.899144556702595e-08, "learning_rate": 2.77160331955742e-05, "loss": 0.0, "num_input_tokens_seen": 5211680, "step": 18615 }, { "epoch": 206.88888888888889, "grad_norm": 4.659488013203372e-07, "learning_rate": 2.7706273618871008e-05, "loss": 0.0, "num_input_tokens_seen": 5213072, "step": 18620 }, { "epoch": 206.94444444444446, "grad_norm": 1.1043748315842095e-07, "learning_rate": 2.769651362482642e-05, "loss": 0.0, "num_input_tokens_seen": 5214464, "step": 18625 }, { "epoch": 207.0, "grad_norm": 8.456252658106678e-07, "learning_rate": 2.768675321494555e-05, "loss": 0.0, "num_input_tokens_seen": 5215840, "step": 18630 }, { "epoch": 207.05555555555554, "grad_norm": 4.205244295008015e-06, "learning_rate": 2.7676992390733565e-05, "loss": 0.0, "num_input_tokens_seen": 5217232, "step": 18635 }, { "epoch": 207.11111111111111, "grad_norm": 1.6017708048821078e-07, "learning_rate": 2.766723115369571e-05, "loss": 0.0, "num_input_tokens_seen": 5218656, "step": 18640 }, { "epoch": 207.16666666666666, "grad_norm": 4.4236853113943653e-07, "learning_rate": 2.765746950533729e-05, "loss": 0.0, "num_input_tokens_seen": 5220032, "step": 18645 }, { "epoch": 207.22222222222223, "grad_norm": 6.740499998159066e-07, "learning_rate": 2.7647707447163684e-05, "loss": 0.0, "num_input_tokens_seen": 5221424, "step": 18650 }, { "epoch": 207.27777777777777, "grad_norm": 2.5911617740348447e-07, "learning_rate": 2.7637944980680315e-05, "loss": 0.0, "num_input_tokens_seen": 5222880, "step": 18655 }, { "epoch": 207.33333333333334, "grad_norm": 3.2840645758369647e-07, "learning_rate": 2.762818210739268e-05, "loss": 0.0, "num_input_tokens_seen": 5224256, "step": 18660 }, { "epoch": 207.38888888888889, "grad_norm": 4.5735100684396457e-07, "learning_rate": 2.7618418828806332e-05, "loss": 0.0, "num_input_tokens_seen": 5225664, "step": 18665 }, { "epoch": 207.44444444444446, "grad_norm": 2.214408141298918e-06, "learning_rate": 2.76086551464269e-05, "loss": 0.0, "num_input_tokens_seen": 5227040, "step": 18670 }, { "epoch": 207.5, "grad_norm": 4.4744041360900155e-07, "learning_rate": 2.759889106176006e-05, "loss": 0.0, "num_input_tokens_seen": 5228480, "step": 18675 }, { "epoch": 207.55555555555554, "grad_norm": 2.1922463133705605e-07, "learning_rate": 2.758912657631156e-05, "loss": 0.0, "num_input_tokens_seen": 5229888, "step": 18680 }, { "epoch": 207.61111111111111, "grad_norm": 1.9345567636719352e-07, "learning_rate": 2.7579361691587198e-05, "loss": 0.0, "num_input_tokens_seen": 5231296, "step": 18685 }, { "epoch": 207.66666666666666, "grad_norm": 3.102252605913236e-07, "learning_rate": 2.756959640909285e-05, "loss": 0.0, "num_input_tokens_seen": 5232688, "step": 18690 }, { "epoch": 207.72222222222223, "grad_norm": 2.2051990526961163e-06, "learning_rate": 2.7559830730334452e-05, "loss": 0.0, "num_input_tokens_seen": 5234096, "step": 18695 }, { "epoch": 207.77777777777777, "grad_norm": 1.5388779672775854e-07, "learning_rate": 2.7550064656817988e-05, "loss": 0.0, "num_input_tokens_seen": 5235520, "step": 18700 }, { "epoch": 207.83333333333334, "grad_norm": 5.814876402610025e-08, "learning_rate": 2.7540298190049503e-05, "loss": 0.0, "num_input_tokens_seen": 5236928, "step": 18705 }, { "epoch": 207.88888888888889, "grad_norm": 6.561305667673878e-07, "learning_rate": 2.7530531331535107e-05, "loss": 0.0, "num_input_tokens_seen": 5238272, "step": 18710 }, { "epoch": 207.94444444444446, "grad_norm": 3.4536682846919575e-07, "learning_rate": 2.752076408278099e-05, "loss": 0.0, "num_input_tokens_seen": 5239680, "step": 18715 }, { "epoch": 208.0, "grad_norm": 4.846575052397384e-07, "learning_rate": 2.751099644529337e-05, "loss": 0.0, "num_input_tokens_seen": 5241056, "step": 18720 }, { "epoch": 208.05555555555554, "grad_norm": 2.313470872650214e-07, "learning_rate": 2.7501228420578533e-05, "loss": 0.0, "num_input_tokens_seen": 5242432, "step": 18725 }, { "epoch": 208.11111111111111, "grad_norm": 3.043412277747848e-07, "learning_rate": 2.7491460010142857e-05, "loss": 0.0, "num_input_tokens_seen": 5243808, "step": 18730 }, { "epoch": 208.16666666666666, "grad_norm": 4.4991213599132607e-07, "learning_rate": 2.7481691215492727e-05, "loss": 0.0, "num_input_tokens_seen": 5245152, "step": 18735 }, { "epoch": 208.22222222222223, "grad_norm": 1.0314901146557531e-06, "learning_rate": 2.747192203813463e-05, "loss": 0.0, "num_input_tokens_seen": 5246544, "step": 18740 }, { "epoch": 208.27777777777777, "grad_norm": 1.548779096083308e-07, "learning_rate": 2.7462152479575087e-05, "loss": 0.0, "num_input_tokens_seen": 5247936, "step": 18745 }, { "epoch": 208.33333333333334, "grad_norm": 3.9182523892122845e-07, "learning_rate": 2.7452382541320697e-05, "loss": 0.0, "num_input_tokens_seen": 5249328, "step": 18750 }, { "epoch": 208.38888888888889, "grad_norm": 2.7006259983863856e-07, "learning_rate": 2.7442612224878096e-05, "loss": 0.0, "num_input_tokens_seen": 5250784, "step": 18755 }, { "epoch": 208.44444444444446, "grad_norm": 2.797553122491081e-07, "learning_rate": 2.7432841531753994e-05, "loss": 0.0, "num_input_tokens_seen": 5252192, "step": 18760 }, { "epoch": 208.5, "grad_norm": 7.645499522368482e-07, "learning_rate": 2.7423070463455147e-05, "loss": 0.0, "num_input_tokens_seen": 5253616, "step": 18765 }, { "epoch": 208.55555555555554, "grad_norm": 4.5320280150917824e-07, "learning_rate": 2.7413299021488397e-05, "loss": 0.0, "num_input_tokens_seen": 5254992, "step": 18770 }, { "epoch": 208.61111111111111, "grad_norm": 1.9400029316329892e-07, "learning_rate": 2.7403527207360615e-05, "loss": 0.0, "num_input_tokens_seen": 5256368, "step": 18775 }, { "epoch": 208.66666666666666, "grad_norm": 2.8512897642940516e-07, "learning_rate": 2.7393755022578722e-05, "loss": 0.0, "num_input_tokens_seen": 5257808, "step": 18780 }, { "epoch": 208.72222222222223, "grad_norm": 1.9221427294269233e-07, "learning_rate": 2.7383982468649714e-05, "loss": 0.0, "num_input_tokens_seen": 5259184, "step": 18785 }, { "epoch": 208.77777777777777, "grad_norm": 4.047042239108123e-07, "learning_rate": 2.7374209547080665e-05, "loss": 0.0, "num_input_tokens_seen": 5260592, "step": 18790 }, { "epoch": 208.83333333333334, "grad_norm": 1.9418062890963483e-07, "learning_rate": 2.7364436259378663e-05, "loss": 0.0, "num_input_tokens_seen": 5261952, "step": 18795 }, { "epoch": 208.88888888888889, "grad_norm": 2.9675845780730015e-07, "learning_rate": 2.735466260705088e-05, "loss": 0.0, "num_input_tokens_seen": 5263360, "step": 18800 }, { "epoch": 208.88888888888889, "eval_loss": 1.2778010368347168, "eval_runtime": 1.4019, "eval_samples_per_second": 28.533, "eval_steps_per_second": 14.266, "num_input_tokens_seen": 5263360, "step": 18800 }, { "epoch": 208.94444444444446, "grad_norm": 4.5633038325831876e-07, "learning_rate": 2.7344888591604524e-05, "loss": 0.0, "num_input_tokens_seen": 5264800, "step": 18805 }, { "epoch": 209.0, "grad_norm": 6.081777996769233e-07, "learning_rate": 2.7335114214546893e-05, "loss": 0.0, "num_input_tokens_seen": 5266240, "step": 18810 }, { "epoch": 209.05555555555554, "grad_norm": 5.391746640270867e-07, "learning_rate": 2.7325339477385293e-05, "loss": 0.0, "num_input_tokens_seen": 5267680, "step": 18815 }, { "epoch": 209.11111111111111, "grad_norm": 1.0321967920390307e-06, "learning_rate": 2.7315564381627128e-05, "loss": 0.0, "num_input_tokens_seen": 5269088, "step": 18820 }, { "epoch": 209.16666666666666, "grad_norm": 5.492560717357264e-07, "learning_rate": 2.7305788928779835e-05, "loss": 0.0, "num_input_tokens_seen": 5270464, "step": 18825 }, { "epoch": 209.22222222222223, "grad_norm": 2.1496732927062112e-07, "learning_rate": 2.729601312035091e-05, "loss": 0.0, "num_input_tokens_seen": 5271856, "step": 18830 }, { "epoch": 209.27777777777777, "grad_norm": 8.546851404389599e-08, "learning_rate": 2.7286236957847915e-05, "loss": 0.0, "num_input_tokens_seen": 5273280, "step": 18835 }, { "epoch": 209.33333333333334, "grad_norm": 1.2175139829651016e-07, "learning_rate": 2.7276460442778446e-05, "loss": 0.0, "num_input_tokens_seen": 5274624, "step": 18840 }, { "epoch": 209.38888888888889, "grad_norm": 6.982155582591076e-07, "learning_rate": 2.726668357665017e-05, "loss": 0.0, "num_input_tokens_seen": 5276016, "step": 18845 }, { "epoch": 209.44444444444446, "grad_norm": 5.127692929818295e-06, "learning_rate": 2.7256906360970808e-05, "loss": 0.0, "num_input_tokens_seen": 5277424, "step": 18850 }, { "epoch": 209.5, "grad_norm": 4.465549920951162e-07, "learning_rate": 2.7247128797248117e-05, "loss": 0.0, "num_input_tokens_seen": 5278864, "step": 18855 }, { "epoch": 209.55555555555554, "grad_norm": 2.0847639348176017e-07, "learning_rate": 2.7237350886989925e-05, "loss": 0.0, "num_input_tokens_seen": 5280240, "step": 18860 }, { "epoch": 209.61111111111111, "grad_norm": 2.9308253601811884e-07, "learning_rate": 2.7227572631704107e-05, "loss": 0.0, "num_input_tokens_seen": 5281632, "step": 18865 }, { "epoch": 209.66666666666666, "grad_norm": 1.4075088472509378e-07, "learning_rate": 2.7217794032898596e-05, "loss": 0.0, "num_input_tokens_seen": 5283040, "step": 18870 }, { "epoch": 209.72222222222223, "grad_norm": 3.231796483760263e-07, "learning_rate": 2.7208015092081384e-05, "loss": 0.0, "num_input_tokens_seen": 5284464, "step": 18875 }, { "epoch": 209.77777777777777, "grad_norm": 2.1151589635337587e-07, "learning_rate": 2.719823581076049e-05, "loss": 0.0, "num_input_tokens_seen": 5285824, "step": 18880 }, { "epoch": 209.83333333333334, "grad_norm": 3.1417906143360597e-07, "learning_rate": 2.718845619044401e-05, "loss": 0.0, "num_input_tokens_seen": 5287216, "step": 18885 }, { "epoch": 209.88888888888889, "grad_norm": 2.949429358523048e-07, "learning_rate": 2.7178676232640088e-05, "loss": 0.0, "num_input_tokens_seen": 5288576, "step": 18890 }, { "epoch": 209.94444444444446, "grad_norm": 1.5380318529878423e-07, "learning_rate": 2.716889593885691e-05, "loss": 0.0, "num_input_tokens_seen": 5290000, "step": 18895 }, { "epoch": 210.0, "grad_norm": 7.940563762076636e-08, "learning_rate": 2.7159115310602716e-05, "loss": 0.0, "num_input_tokens_seen": 5291424, "step": 18900 }, { "epoch": 210.05555555555554, "grad_norm": 3.5800579212263983e-07, "learning_rate": 2.7149334349385814e-05, "loss": 0.0, "num_input_tokens_seen": 5292832, "step": 18905 }, { "epoch": 210.11111111111111, "grad_norm": 1.8579284244424343e-07, "learning_rate": 2.713955305671454e-05, "loss": 0.0, "num_input_tokens_seen": 5294240, "step": 18910 }, { "epoch": 210.16666666666666, "grad_norm": 3.6557909766088414e-07, "learning_rate": 2.71297714340973e-05, "loss": 0.0, "num_input_tokens_seen": 5295632, "step": 18915 }, { "epoch": 210.22222222222223, "grad_norm": 1.3116078889652272e-07, "learning_rate": 2.7119989483042545e-05, "loss": 0.0, "num_input_tokens_seen": 5297136, "step": 18920 }, { "epoch": 210.27777777777777, "grad_norm": 2.6055107582578785e-07, "learning_rate": 2.7110207205058768e-05, "loss": 0.0, "num_input_tokens_seen": 5298528, "step": 18925 }, { "epoch": 210.33333333333334, "grad_norm": 4.5882293875365576e-07, "learning_rate": 2.7100424601654517e-05, "loss": 0.0, "num_input_tokens_seen": 5299920, "step": 18930 }, { "epoch": 210.38888888888889, "grad_norm": 6.055714152353175e-07, "learning_rate": 2.7090641674338403e-05, "loss": 0.0, "num_input_tokens_seen": 5301328, "step": 18935 }, { "epoch": 210.44444444444446, "grad_norm": 2.504892790966551e-07, "learning_rate": 2.7080858424619072e-05, "loss": 0.0, "num_input_tokens_seen": 5302688, "step": 18940 }, { "epoch": 210.5, "grad_norm": 2.2512217867642903e-07, "learning_rate": 2.707107485400521e-05, "loss": 0.0, "num_input_tokens_seen": 5304064, "step": 18945 }, { "epoch": 210.55555555555554, "grad_norm": 8.408906069234945e-07, "learning_rate": 2.7061290964005586e-05, "loss": 0.0, "num_input_tokens_seen": 5305440, "step": 18950 }, { "epoch": 210.61111111111111, "grad_norm": 5.794300363959337e-07, "learning_rate": 2.7051506756129e-05, "loss": 0.0, "num_input_tokens_seen": 5306800, "step": 18955 }, { "epoch": 210.66666666666666, "grad_norm": 2.642212280079548e-07, "learning_rate": 2.704172223188428e-05, "loss": 0.0, "num_input_tokens_seen": 5308192, "step": 18960 }, { "epoch": 210.72222222222223, "grad_norm": 1.3580377355992823e-07, "learning_rate": 2.7031937392780334e-05, "loss": 0.0, "num_input_tokens_seen": 5309600, "step": 18965 }, { "epoch": 210.77777777777777, "grad_norm": 1.6323205898061133e-07, "learning_rate": 2.702215224032611e-05, "loss": 0.0, "num_input_tokens_seen": 5310960, "step": 18970 }, { "epoch": 210.83333333333334, "grad_norm": 3.661681802213934e-08, "learning_rate": 2.70123667760306e-05, "loss": 0.0, "num_input_tokens_seen": 5312352, "step": 18975 }, { "epoch": 210.88888888888889, "grad_norm": 3.3348942451993935e-07, "learning_rate": 2.7002581001402845e-05, "loss": 0.0, "num_input_tokens_seen": 5313776, "step": 18980 }, { "epoch": 210.94444444444446, "grad_norm": 3.4302431117794185e-07, "learning_rate": 2.6992794917951923e-05, "loss": 0.0, "num_input_tokens_seen": 5315136, "step": 18985 }, { "epoch": 211.0, "grad_norm": 2.1736584585596574e-06, "learning_rate": 2.6983008527187e-05, "loss": 0.0, "num_input_tokens_seen": 5316560, "step": 18990 }, { "epoch": 211.05555555555554, "grad_norm": 5.9311724953659e-07, "learning_rate": 2.697322183061723e-05, "loss": 0.0, "num_input_tokens_seen": 5317952, "step": 18995 }, { "epoch": 211.11111111111111, "grad_norm": 3.5543013154892833e-07, "learning_rate": 2.696343482975186e-05, "loss": 0.0, "num_input_tokens_seen": 5319344, "step": 19000 }, { "epoch": 211.11111111111111, "eval_loss": 1.3527178764343262, "eval_runtime": 1.3881, "eval_samples_per_second": 28.816, "eval_steps_per_second": 14.408, "num_input_tokens_seen": 5319344, "step": 19000 }, { "epoch": 211.16666666666666, "grad_norm": 5.123660002936958e-07, "learning_rate": 2.695364752610016e-05, "loss": 0.0, "num_input_tokens_seen": 5320736, "step": 19005 }, { "epoch": 211.22222222222223, "grad_norm": 6.980450848459441e-07, "learning_rate": 2.6943859921171467e-05, "loss": 0.0, "num_input_tokens_seen": 5322176, "step": 19010 }, { "epoch": 211.27777777777777, "grad_norm": 2.9368032983256853e-07, "learning_rate": 2.6934072016475143e-05, "loss": 0.0, "num_input_tokens_seen": 5323552, "step": 19015 }, { "epoch": 211.33333333333334, "grad_norm": 3.442018510213529e-07, "learning_rate": 2.6924283813520606e-05, "loss": 0.0, "num_input_tokens_seen": 5324992, "step": 19020 }, { "epoch": 211.38888888888889, "grad_norm": 4.4288174194662133e-07, "learning_rate": 2.691449531381733e-05, "loss": 0.0, "num_input_tokens_seen": 5326416, "step": 19025 }, { "epoch": 211.44444444444446, "grad_norm": 3.049728718451661e-07, "learning_rate": 2.6904706518874816e-05, "loss": 0.0, "num_input_tokens_seen": 5327808, "step": 19030 }, { "epoch": 211.5, "grad_norm": 2.1097439457662404e-06, "learning_rate": 2.6894917430202615e-05, "loss": 0.0, "num_input_tokens_seen": 5329216, "step": 19035 }, { "epoch": 211.55555555555554, "grad_norm": 1.0691087481973227e-07, "learning_rate": 2.6885128049310343e-05, "loss": 0.0, "num_input_tokens_seen": 5330624, "step": 19040 }, { "epoch": 211.61111111111111, "grad_norm": 3.536116537361522e-07, "learning_rate": 2.687533837770762e-05, "loss": 0.0, "num_input_tokens_seen": 5332016, "step": 19045 }, { "epoch": 211.66666666666666, "grad_norm": 2.00925185822598e-07, "learning_rate": 2.6865548416904162e-05, "loss": 0.0, "num_input_tokens_seen": 5333424, "step": 19050 }, { "epoch": 211.72222222222223, "grad_norm": 8.953381325227383e-07, "learning_rate": 2.68557581684097e-05, "loss": 0.0, "num_input_tokens_seen": 5334832, "step": 19055 }, { "epoch": 211.77777777777777, "grad_norm": 8.87430360307917e-07, "learning_rate": 2.6845967633733998e-05, "loss": 0.0, "num_input_tokens_seen": 5336208, "step": 19060 }, { "epoch": 211.83333333333334, "grad_norm": 5.454376150737517e-07, "learning_rate": 2.683617681438689e-05, "loss": 0.0, "num_input_tokens_seen": 5337584, "step": 19065 }, { "epoch": 211.88888888888889, "grad_norm": 2.9518014343921095e-07, "learning_rate": 2.682638571187825e-05, "loss": 0.0, "num_input_tokens_seen": 5338944, "step": 19070 }, { "epoch": 211.94444444444446, "grad_norm": 7.345549306592147e-07, "learning_rate": 2.6816594327717976e-05, "loss": 0.0, "num_input_tokens_seen": 5340368, "step": 19075 }, { "epoch": 212.0, "grad_norm": 1.5012524556823337e-07, "learning_rate": 2.680680266341603e-05, "loss": 0.0, "num_input_tokens_seen": 5341776, "step": 19080 }, { "epoch": 212.05555555555554, "grad_norm": 2.9624254693771945e-07, "learning_rate": 2.67970107204824e-05, "loss": 0.0, "num_input_tokens_seen": 5343120, "step": 19085 }, { "epoch": 212.11111111111111, "grad_norm": 7.13305496446992e-07, "learning_rate": 2.6787218500427142e-05, "loss": 0.0, "num_input_tokens_seen": 5344496, "step": 19090 }, { "epoch": 212.16666666666666, "grad_norm": 2.763686097750906e-07, "learning_rate": 2.6777426004760332e-05, "loss": 0.0, "num_input_tokens_seen": 5345920, "step": 19095 }, { "epoch": 212.22222222222223, "grad_norm": 5.340245365914598e-07, "learning_rate": 2.6767633234992094e-05, "loss": 0.0, "num_input_tokens_seen": 5347312, "step": 19100 }, { "epoch": 212.27777777777777, "grad_norm": 3.719499943599658e-07, "learning_rate": 2.6757840192632598e-05, "loss": 0.0, "num_input_tokens_seen": 5348768, "step": 19105 }, { "epoch": 212.33333333333334, "grad_norm": 2.384907020314131e-06, "learning_rate": 2.6748046879192052e-05, "loss": 0.0, "num_input_tokens_seen": 5350192, "step": 19110 }, { "epoch": 212.38888888888889, "grad_norm": 1.4096136169428064e-07, "learning_rate": 2.673825329618071e-05, "loss": 0.0, "num_input_tokens_seen": 5351600, "step": 19115 }, { "epoch": 212.44444444444446, "grad_norm": 2.8828776521550026e-07, "learning_rate": 2.6728459445108866e-05, "loss": 0.0, "num_input_tokens_seen": 5353040, "step": 19120 }, { "epoch": 212.5, "grad_norm": 4.645971500849555e-07, "learning_rate": 2.6718665327486854e-05, "loss": 0.0, "num_input_tokens_seen": 5354416, "step": 19125 }, { "epoch": 212.55555555555554, "grad_norm": 4.638314692329004e-07, "learning_rate": 2.6708870944825048e-05, "loss": 0.0, "num_input_tokens_seen": 5355808, "step": 19130 }, { "epoch": 212.61111111111111, "grad_norm": 3.507271912894794e-07, "learning_rate": 2.6699076298633874e-05, "loss": 0.0, "num_input_tokens_seen": 5357248, "step": 19135 }, { "epoch": 212.66666666666666, "grad_norm": 1.5305084843930672e-07, "learning_rate": 2.6689281390423788e-05, "loss": 0.0, "num_input_tokens_seen": 5358656, "step": 19140 }, { "epoch": 212.72222222222223, "grad_norm": 3.580534269076452e-07, "learning_rate": 2.667948622170527e-05, "loss": 0.0, "num_input_tokens_seen": 5360016, "step": 19145 }, { "epoch": 212.77777777777777, "grad_norm": 1.5724157265140093e-07, "learning_rate": 2.6669690793988873e-05, "loss": 0.0, "num_input_tokens_seen": 5361376, "step": 19150 }, { "epoch": 212.83333333333334, "grad_norm": 6.233592557691736e-07, "learning_rate": 2.665989510878518e-05, "loss": 0.0, "num_input_tokens_seen": 5362784, "step": 19155 }, { "epoch": 212.88888888888889, "grad_norm": 1.0320396626184447e-07, "learning_rate": 2.6650099167604793e-05, "loss": 0.0, "num_input_tokens_seen": 5364176, "step": 19160 }, { "epoch": 212.94444444444446, "grad_norm": 3.583755869840388e-07, "learning_rate": 2.6640302971958376e-05, "loss": 0.0, "num_input_tokens_seen": 5365584, "step": 19165 }, { "epoch": 213.0, "grad_norm": 3.1892858487481135e-07, "learning_rate": 2.6630506523356635e-05, "loss": 0.0, "num_input_tokens_seen": 5366960, "step": 19170 }, { "epoch": 213.05555555555554, "grad_norm": 1.5588145174660895e-07, "learning_rate": 2.6620709823310297e-05, "loss": 0.0, "num_input_tokens_seen": 5368320, "step": 19175 }, { "epoch": 213.11111111111111, "grad_norm": 1.9972179643445997e-07, "learning_rate": 2.661091287333014e-05, "loss": 0.0, "num_input_tokens_seen": 5369760, "step": 19180 }, { "epoch": 213.16666666666666, "grad_norm": 7.530643415520899e-06, "learning_rate": 2.660111567492696e-05, "loss": 0.0, "num_input_tokens_seen": 5371136, "step": 19185 }, { "epoch": 213.22222222222223, "grad_norm": 1.756599345981158e-07, "learning_rate": 2.6591318229611635e-05, "loss": 0.0, "num_input_tokens_seen": 5372528, "step": 19190 }, { "epoch": 213.27777777777777, "grad_norm": 3.5613876292472924e-08, "learning_rate": 2.6581520538895037e-05, "loss": 0.0, "num_input_tokens_seen": 5373888, "step": 19195 }, { "epoch": 213.33333333333334, "grad_norm": 2.1710769715355127e-07, "learning_rate": 2.6571722604288102e-05, "loss": 0.0, "num_input_tokens_seen": 5375280, "step": 19200 }, { "epoch": 213.33333333333334, "eval_loss": 1.3433246612548828, "eval_runtime": 1.3858, "eval_samples_per_second": 28.865, "eval_steps_per_second": 14.432, "num_input_tokens_seen": 5375280, "step": 19200 }, { "epoch": 213.38888888888889, "grad_norm": 1.6593635621120484e-07, "learning_rate": 2.656192442730179e-05, "loss": 0.0, "num_input_tokens_seen": 5376704, "step": 19205 }, { "epoch": 213.44444444444446, "grad_norm": 7.893683573456656e-08, "learning_rate": 2.6552126009447098e-05, "loss": 0.0, "num_input_tokens_seen": 5378080, "step": 19210 }, { "epoch": 213.5, "grad_norm": 1.4507766366023134e-07, "learning_rate": 2.654232735223507e-05, "loss": 0.0, "num_input_tokens_seen": 5379504, "step": 19215 }, { "epoch": 213.55555555555554, "grad_norm": 4.1907966874532576e-08, "learning_rate": 2.6532528457176787e-05, "loss": 0.0, "num_input_tokens_seen": 5380896, "step": 19220 }, { "epoch": 213.61111111111111, "grad_norm": 2.736946100867499e-07, "learning_rate": 2.6522729325783348e-05, "loss": 0.0, "num_input_tokens_seen": 5382304, "step": 19225 }, { "epoch": 213.66666666666666, "grad_norm": 3.325803845655173e-07, "learning_rate": 2.6512929959565914e-05, "loss": 0.0, "num_input_tokens_seen": 5383664, "step": 19230 }, { "epoch": 213.72222222222223, "grad_norm": 2.6881397729994205e-07, "learning_rate": 2.6503130360035673e-05, "loss": 0.0, "num_input_tokens_seen": 5385088, "step": 19235 }, { "epoch": 213.77777777777777, "grad_norm": 1.0505081604605948e-07, "learning_rate": 2.6493330528703835e-05, "loss": 0.0, "num_input_tokens_seen": 5386512, "step": 19240 }, { "epoch": 213.83333333333334, "grad_norm": 2.1789928439375217e-07, "learning_rate": 2.648353046708167e-05, "loss": 0.0, "num_input_tokens_seen": 5387968, "step": 19245 }, { "epoch": 213.88888888888889, "grad_norm": 4.6415084398176987e-07, "learning_rate": 2.647373017668046e-05, "loss": 0.0, "num_input_tokens_seen": 5389392, "step": 19250 }, { "epoch": 213.94444444444446, "grad_norm": 4.2233745034536696e-07, "learning_rate": 2.6463929659011537e-05, "loss": 0.0, "num_input_tokens_seen": 5390800, "step": 19255 }, { "epoch": 214.0, "grad_norm": 1.0692299667880434e-07, "learning_rate": 2.6454128915586262e-05, "loss": 0.0, "num_input_tokens_seen": 5392192, "step": 19260 }, { "epoch": 214.05555555555554, "grad_norm": 1.0060682598123094e-07, "learning_rate": 2.6444327947916036e-05, "loss": 0.0, "num_input_tokens_seen": 5393568, "step": 19265 }, { "epoch": 214.11111111111111, "grad_norm": 5.153991651241085e-07, "learning_rate": 2.6434526757512292e-05, "loss": 0.0, "num_input_tokens_seen": 5395008, "step": 19270 }, { "epoch": 214.16666666666666, "grad_norm": 3.2453149856337404e-07, "learning_rate": 2.6424725345886486e-05, "loss": 0.0, "num_input_tokens_seen": 5396384, "step": 19275 }, { "epoch": 214.22222222222223, "grad_norm": 5.892503054383269e-07, "learning_rate": 2.641492371455014e-05, "loss": 0.0, "num_input_tokens_seen": 5397744, "step": 19280 }, { "epoch": 214.27777777777777, "grad_norm": 2.9057629191697742e-08, "learning_rate": 2.640512186501477e-05, "loss": 0.0, "num_input_tokens_seen": 5399168, "step": 19285 }, { "epoch": 214.33333333333334, "grad_norm": 5.545785484173393e-07, "learning_rate": 2.639531979879195e-05, "loss": 0.0, "num_input_tokens_seen": 5400560, "step": 19290 }, { "epoch": 214.38888888888889, "grad_norm": 9.464358896593694e-08, "learning_rate": 2.638551751739328e-05, "loss": 0.0, "num_input_tokens_seen": 5401984, "step": 19295 }, { "epoch": 214.44444444444446, "grad_norm": 1.4632404088388284e-07, "learning_rate": 2.6375715022330404e-05, "loss": 0.0, "num_input_tokens_seen": 5403392, "step": 19300 }, { "epoch": 214.5, "grad_norm": 5.09679580318334e-07, "learning_rate": 2.6365912315114976e-05, "loss": 0.0, "num_input_tokens_seen": 5404800, "step": 19305 }, { "epoch": 214.55555555555554, "grad_norm": 5.388327508626389e-07, "learning_rate": 2.6356109397258704e-05, "loss": 0.0, "num_input_tokens_seen": 5406208, "step": 19310 }, { "epoch": 214.61111111111111, "grad_norm": 1.4982256857365428e-07, "learning_rate": 2.6346306270273325e-05, "loss": 0.0, "num_input_tokens_seen": 5407616, "step": 19315 }, { "epoch": 214.66666666666666, "grad_norm": 1.1723859643097967e-06, "learning_rate": 2.6336502935670608e-05, "loss": 0.0, "num_input_tokens_seen": 5409040, "step": 19320 }, { "epoch": 214.72222222222223, "grad_norm": 3.3240414154533937e-07, "learning_rate": 2.6326699394962333e-05, "loss": 0.0, "num_input_tokens_seen": 5410448, "step": 19325 }, { "epoch": 214.77777777777777, "grad_norm": 3.612364594118844e-07, "learning_rate": 2.6316895649660334e-05, "loss": 0.0, "num_input_tokens_seen": 5411808, "step": 19330 }, { "epoch": 214.83333333333334, "grad_norm": 4.208760913115839e-07, "learning_rate": 2.6307091701276486e-05, "loss": 0.0, "num_input_tokens_seen": 5413232, "step": 19335 }, { "epoch": 214.88888888888889, "grad_norm": 8.563258546701036e-08, "learning_rate": 2.629728755132267e-05, "loss": 0.0, "num_input_tokens_seen": 5414656, "step": 19340 }, { "epoch": 214.94444444444446, "grad_norm": 5.279449624140398e-07, "learning_rate": 2.628748320131081e-05, "loss": 0.0, "num_input_tokens_seen": 5416080, "step": 19345 }, { "epoch": 215.0, "grad_norm": 1.340999631338491e-07, "learning_rate": 2.6277678652752856e-05, "loss": 0.0, "num_input_tokens_seen": 5417472, "step": 19350 }, { "epoch": 215.05555555555554, "grad_norm": 1.6564356997150753e-07, "learning_rate": 2.6267873907160807e-05, "loss": 0.0, "num_input_tokens_seen": 5418880, "step": 19355 }, { "epoch": 215.11111111111111, "grad_norm": 2.256891065144373e-07, "learning_rate": 2.6258068966046668e-05, "loss": 0.0, "num_input_tokens_seen": 5420272, "step": 19360 }, { "epoch": 215.16666666666666, "grad_norm": 6.373911105583829e-07, "learning_rate": 2.6248263830922475e-05, "loss": 0.0, "num_input_tokens_seen": 5421664, "step": 19365 }, { "epoch": 215.22222222222223, "grad_norm": 1.3824489997205092e-07, "learning_rate": 2.6238458503300318e-05, "loss": 0.0, "num_input_tokens_seen": 5423072, "step": 19370 }, { "epoch": 215.27777777777777, "grad_norm": 4.788275873579551e-07, "learning_rate": 2.6228652984692292e-05, "loss": 0.0, "num_input_tokens_seen": 5424496, "step": 19375 }, { "epoch": 215.33333333333334, "grad_norm": 6.558531140399282e-07, "learning_rate": 2.621884727661054e-05, "loss": 0.0, "num_input_tokens_seen": 5425904, "step": 19380 }, { "epoch": 215.38888888888889, "grad_norm": 7.200677032415115e-08, "learning_rate": 2.6209041380567222e-05, "loss": 0.0, "num_input_tokens_seen": 5427296, "step": 19385 }, { "epoch": 215.44444444444446, "grad_norm": 3.0399354500332265e-07, "learning_rate": 2.6199235298074527e-05, "loss": 0.0, "num_input_tokens_seen": 5428688, "step": 19390 }, { "epoch": 215.5, "grad_norm": 1.5754686444324761e-07, "learning_rate": 2.618942903064468e-05, "loss": 0.0, "num_input_tokens_seen": 5430080, "step": 19395 }, { "epoch": 215.55555555555554, "grad_norm": 1.1168813216499984e-06, "learning_rate": 2.6179622579789932e-05, "loss": 0.0, "num_input_tokens_seen": 5431520, "step": 19400 }, { "epoch": 215.55555555555554, "eval_loss": 1.2959202527999878, "eval_runtime": 1.3894, "eval_samples_per_second": 28.789, "eval_steps_per_second": 14.395, "num_input_tokens_seen": 5431520, "step": 19400 }, { "epoch": 215.61111111111111, "grad_norm": 2.1069298838938266e-07, "learning_rate": 2.6169815947022553e-05, "loss": 0.0, "num_input_tokens_seen": 5432960, "step": 19405 }, { "epoch": 215.66666666666666, "grad_norm": 1.06870720628649e-06, "learning_rate": 2.6160009133854853e-05, "loss": 0.0, "num_input_tokens_seen": 5434352, "step": 19410 }, { "epoch": 215.72222222222223, "grad_norm": 7.728884554580873e-08, "learning_rate": 2.6150202141799168e-05, "loss": 0.0, "num_input_tokens_seen": 5435760, "step": 19415 }, { "epoch": 215.77777777777777, "grad_norm": 1.1273810684997443e-07, "learning_rate": 2.614039497236786e-05, "loss": 0.0, "num_input_tokens_seen": 5437168, "step": 19420 }, { "epoch": 215.83333333333334, "grad_norm": 1.2356524337064911e-07, "learning_rate": 2.6130587627073315e-05, "loss": 0.0, "num_input_tokens_seen": 5438544, "step": 19425 }, { "epoch": 215.88888888888889, "grad_norm": 2.3656565417695674e-07, "learning_rate": 2.6120780107427956e-05, "loss": 0.0, "num_input_tokens_seen": 5439904, "step": 19430 }, { "epoch": 215.94444444444446, "grad_norm": 2.81353550235508e-07, "learning_rate": 2.6110972414944214e-05, "loss": 0.0, "num_input_tokens_seen": 5441360, "step": 19435 }, { "epoch": 216.0, "grad_norm": 3.526109253471077e-07, "learning_rate": 2.6101164551134565e-05, "loss": 0.0, "num_input_tokens_seen": 5442752, "step": 19440 }, { "epoch": 216.05555555555554, "grad_norm": 4.5847039586988103e-07, "learning_rate": 2.6091356517511505e-05, "loss": 0.0, "num_input_tokens_seen": 5444176, "step": 19445 }, { "epoch": 216.11111111111111, "grad_norm": 5.193561491978471e-07, "learning_rate": 2.608154831558755e-05, "loss": 0.0, "num_input_tokens_seen": 5445584, "step": 19450 }, { "epoch": 216.16666666666666, "grad_norm": 4.5727014708063507e-07, "learning_rate": 2.607173994687526e-05, "loss": 0.0, "num_input_tokens_seen": 5446976, "step": 19455 }, { "epoch": 216.22222222222223, "grad_norm": 1.5638633499293064e-07, "learning_rate": 2.6061931412887196e-05, "loss": 0.0, "num_input_tokens_seen": 5448384, "step": 19460 }, { "epoch": 216.27777777777777, "grad_norm": 2.119425772661998e-07, "learning_rate": 2.6052122715135973e-05, "loss": 0.0, "num_input_tokens_seen": 5449808, "step": 19465 }, { "epoch": 216.33333333333334, "grad_norm": 1.7396456541973748e-07, "learning_rate": 2.60423138551342e-05, "loss": 0.0, "num_input_tokens_seen": 5451232, "step": 19470 }, { "epoch": 216.38888888888889, "grad_norm": 3.757735100862192e-07, "learning_rate": 2.6032504834394527e-05, "loss": 0.0, "num_input_tokens_seen": 5452624, "step": 19475 }, { "epoch": 216.44444444444446, "grad_norm": 6.974596544750966e-08, "learning_rate": 2.602269565442964e-05, "loss": 0.0, "num_input_tokens_seen": 5453968, "step": 19480 }, { "epoch": 216.5, "grad_norm": 1.0018094798169841e-07, "learning_rate": 2.6012886316752227e-05, "loss": 0.0, "num_input_tokens_seen": 5455392, "step": 19485 }, { "epoch": 216.55555555555554, "grad_norm": 5.421743480837904e-07, "learning_rate": 2.6003076822875018e-05, "loss": 0.0, "num_input_tokens_seen": 5456800, "step": 19490 }, { "epoch": 216.61111111111111, "grad_norm": 1.9132848194658436e-07, "learning_rate": 2.5993267174310755e-05, "loss": 0.0, "num_input_tokens_seen": 5458192, "step": 19495 }, { "epoch": 216.66666666666666, "grad_norm": 4.052427584610996e-07, "learning_rate": 2.5983457372572218e-05, "loss": 0.0, "num_input_tokens_seen": 5459568, "step": 19500 }, { "epoch": 216.72222222222223, "grad_norm": 4.955715553478512e-07, "learning_rate": 2.597364741917219e-05, "loss": 0.0, "num_input_tokens_seen": 5460960, "step": 19505 }, { "epoch": 216.77777777777777, "grad_norm": 4.0740820850260206e-07, "learning_rate": 2.5963837315623492e-05, "loss": 0.0, "num_input_tokens_seen": 5462368, "step": 19510 }, { "epoch": 216.83333333333334, "grad_norm": 3.8680818192915467e-07, "learning_rate": 2.595402706343897e-05, "loss": 0.0, "num_input_tokens_seen": 5463792, "step": 19515 }, { "epoch": 216.88888888888889, "grad_norm": 9.213825791221097e-08, "learning_rate": 2.594421666413148e-05, "loss": 0.0, "num_input_tokens_seen": 5465232, "step": 19520 }, { "epoch": 216.94444444444446, "grad_norm": 7.390004839180619e-07, "learning_rate": 2.5934406119213928e-05, "loss": 0.0, "num_input_tokens_seen": 5466640, "step": 19525 }, { "epoch": 217.0, "grad_norm": 6.471713618338981e-07, "learning_rate": 2.5924595430199193e-05, "loss": 0.0, "num_input_tokens_seen": 5467984, "step": 19530 }, { "epoch": 217.05555555555554, "grad_norm": 4.214740272345807e-07, "learning_rate": 2.5914784598600238e-05, "loss": 0.0, "num_input_tokens_seen": 5469376, "step": 19535 }, { "epoch": 217.11111111111111, "grad_norm": 2.904037330608844e-07, "learning_rate": 2.5904973625930002e-05, "loss": 0.0, "num_input_tokens_seen": 5470720, "step": 19540 }, { "epoch": 217.16666666666666, "grad_norm": 1.8254392841754452e-07, "learning_rate": 2.5895162513701456e-05, "loss": 0.0, "num_input_tokens_seen": 5472112, "step": 19545 }, { "epoch": 217.22222222222223, "grad_norm": 1.6177249051452236e-07, "learning_rate": 2.5885351263427593e-05, "loss": 0.0, "num_input_tokens_seen": 5473520, "step": 19550 }, { "epoch": 217.27777777777777, "grad_norm": 8.382461658129614e-08, "learning_rate": 2.5875539876621448e-05, "loss": 0.0, "num_input_tokens_seen": 5474912, "step": 19555 }, { "epoch": 217.33333333333334, "grad_norm": 2.483960315657896e-06, "learning_rate": 2.586572835479605e-05, "loss": 0.0, "num_input_tokens_seen": 5476288, "step": 19560 }, { "epoch": 217.38888888888889, "grad_norm": 1.453630176229126e-07, "learning_rate": 2.585591669946446e-05, "loss": 0.0, "num_input_tokens_seen": 5477632, "step": 19565 }, { "epoch": 217.44444444444446, "grad_norm": 2.6038608780254435e-07, "learning_rate": 2.5846104912139756e-05, "loss": 0.0, "num_input_tokens_seen": 5479056, "step": 19570 }, { "epoch": 217.5, "grad_norm": 5.219082481744408e-07, "learning_rate": 2.583629299433505e-05, "loss": 0.0, "num_input_tokens_seen": 5480480, "step": 19575 }, { "epoch": 217.55555555555554, "grad_norm": 2.2606828053994832e-07, "learning_rate": 2.582648094756345e-05, "loss": 0.0, "num_input_tokens_seen": 5481904, "step": 19580 }, { "epoch": 217.61111111111111, "grad_norm": 4.774120725414832e-07, "learning_rate": 2.5816668773338098e-05, "loss": 0.0, "num_input_tokens_seen": 5483280, "step": 19585 }, { "epoch": 217.66666666666666, "grad_norm": 1.1305229463687283e-07, "learning_rate": 2.580685647317216e-05, "loss": 0.0, "num_input_tokens_seen": 5484688, "step": 19590 }, { "epoch": 217.72222222222223, "grad_norm": 6.92908088240074e-07, "learning_rate": 2.5797044048578818e-05, "loss": 0.0, "num_input_tokens_seen": 5486064, "step": 19595 }, { "epoch": 217.77777777777777, "grad_norm": 1.9278003549061395e-07, "learning_rate": 2.5787231501071262e-05, "loss": 0.0, "num_input_tokens_seen": 5487472, "step": 19600 }, { "epoch": 217.77777777777777, "eval_loss": 1.3495502471923828, "eval_runtime": 1.3869, "eval_samples_per_second": 28.841, "eval_steps_per_second": 14.42, "num_input_tokens_seen": 5487472, "step": 19600 }, { "epoch": 217.83333333333334, "grad_norm": 5.834133389726048e-07, "learning_rate": 2.577741883216272e-05, "loss": 0.0, "num_input_tokens_seen": 5488880, "step": 19605 }, { "epoch": 217.88888888888889, "grad_norm": 1.7415109709872922e-07, "learning_rate": 2.576760604336642e-05, "loss": 0.0, "num_input_tokens_seen": 5490304, "step": 19610 }, { "epoch": 217.94444444444446, "grad_norm": 1.9534122941422538e-07, "learning_rate": 2.575779313619563e-05, "loss": 0.0, "num_input_tokens_seen": 5491712, "step": 19615 }, { "epoch": 218.0, "grad_norm": 6.420194864631412e-08, "learning_rate": 2.5747980112163605e-05, "loss": 0.0, "num_input_tokens_seen": 5493136, "step": 19620 }, { "epoch": 218.05555555555554, "grad_norm": 1.6924292367548333e-07, "learning_rate": 2.5738166972783656e-05, "loss": 0.0, "num_input_tokens_seen": 5494528, "step": 19625 }, { "epoch": 218.11111111111111, "grad_norm": 4.770473083226534e-07, "learning_rate": 2.5728353719569075e-05, "loss": 0.0, "num_input_tokens_seen": 5495920, "step": 19630 }, { "epoch": 218.16666666666666, "grad_norm": 4.996651910005312e-07, "learning_rate": 2.57185403540332e-05, "loss": 0.0, "num_input_tokens_seen": 5497296, "step": 19635 }, { "epoch": 218.22222222222223, "grad_norm": 2.650729697961651e-07, "learning_rate": 2.5708726877689375e-05, "loss": 0.0, "num_input_tokens_seen": 5498688, "step": 19640 }, { "epoch": 218.27777777777777, "grad_norm": 3.5234157280683576e-07, "learning_rate": 2.5698913292050964e-05, "loss": 0.0, "num_input_tokens_seen": 5500096, "step": 19645 }, { "epoch": 218.33333333333334, "grad_norm": 5.060130092715553e-07, "learning_rate": 2.568909959863133e-05, "loss": 0.0, "num_input_tokens_seen": 5501520, "step": 19650 }, { "epoch": 218.38888888888889, "grad_norm": 1.6351270915038185e-07, "learning_rate": 2.5679285798943887e-05, "loss": 0.0, "num_input_tokens_seen": 5502928, "step": 19655 }, { "epoch": 218.44444444444446, "grad_norm": 2.134685530563729e-07, "learning_rate": 2.5669471894502035e-05, "loss": 0.0, "num_input_tokens_seen": 5504304, "step": 19660 }, { "epoch": 218.5, "grad_norm": 1.2538204430256883e-07, "learning_rate": 2.56596578868192e-05, "loss": 0.0, "num_input_tokens_seen": 5505664, "step": 19665 }, { "epoch": 218.55555555555554, "grad_norm": 9.595792107575107e-07, "learning_rate": 2.564984377740883e-05, "loss": 0.0, "num_input_tokens_seen": 5507088, "step": 19670 }, { "epoch": 218.61111111111111, "grad_norm": 3.999521709374676e-07, "learning_rate": 2.564002956778438e-05, "loss": 0.0, "num_input_tokens_seen": 5508512, "step": 19675 }, { "epoch": 218.66666666666666, "grad_norm": 5.950071795268741e-07, "learning_rate": 2.563021525945934e-05, "loss": 0.0, "num_input_tokens_seen": 5509952, "step": 19680 }, { "epoch": 218.72222222222223, "grad_norm": 1.5468592096112843e-07, "learning_rate": 2.562040085394718e-05, "loss": 0.0, "num_input_tokens_seen": 5511344, "step": 19685 }, { "epoch": 218.77777777777777, "grad_norm": 8.393578809773317e-08, "learning_rate": 2.56105863527614e-05, "loss": 0.0, "num_input_tokens_seen": 5512720, "step": 19690 }, { "epoch": 218.83333333333334, "grad_norm": 1.1320621240429318e-07, "learning_rate": 2.5600771757415548e-05, "loss": 0.0, "num_input_tokens_seen": 5514112, "step": 19695 }, { "epoch": 218.88888888888889, "grad_norm": 3.974610081058927e-06, "learning_rate": 2.5590957069423134e-05, "loss": 0.0, "num_input_tokens_seen": 5515568, "step": 19700 }, { "epoch": 218.94444444444446, "grad_norm": 5.839301593368873e-07, "learning_rate": 2.5581142290297716e-05, "loss": 0.0, "num_input_tokens_seen": 5516992, "step": 19705 }, { "epoch": 219.0, "grad_norm": 6.31684372365271e-07, "learning_rate": 2.557132742155285e-05, "loss": 0.0, "num_input_tokens_seen": 5518368, "step": 19710 }, { "epoch": 219.05555555555554, "grad_norm": 1.161074720812394e-07, "learning_rate": 2.556151246470212e-05, "loss": 0.0, "num_input_tokens_seen": 5519776, "step": 19715 }, { "epoch": 219.11111111111111, "grad_norm": 1.4522900926294824e-07, "learning_rate": 2.5551697421259114e-05, "loss": 0.0, "num_input_tokens_seen": 5521216, "step": 19720 }, { "epoch": 219.16666666666666, "grad_norm": 9.389082578081798e-08, "learning_rate": 2.554188229273743e-05, "loss": 0.0, "num_input_tokens_seen": 5522608, "step": 19725 }, { "epoch": 219.22222222222223, "grad_norm": 2.0556724678044702e-07, "learning_rate": 2.5532067080650678e-05, "loss": 0.0, "num_input_tokens_seen": 5524016, "step": 19730 }, { "epoch": 219.27777777777777, "grad_norm": 5.975713293082663e-07, "learning_rate": 2.55222517865125e-05, "loss": 0.0, "num_input_tokens_seen": 5525424, "step": 19735 }, { "epoch": 219.33333333333334, "grad_norm": 9.412924555363134e-07, "learning_rate": 2.5512436411836538e-05, "loss": 0.0, "num_input_tokens_seen": 5526816, "step": 19740 }, { "epoch": 219.38888888888889, "grad_norm": 4.640346560336184e-07, "learning_rate": 2.5502620958136443e-05, "loss": 0.0, "num_input_tokens_seen": 5528192, "step": 19745 }, { "epoch": 219.44444444444446, "grad_norm": 1.341929021236865e-07, "learning_rate": 2.5492805426925874e-05, "loss": 0.0, "num_input_tokens_seen": 5529616, "step": 19750 }, { "epoch": 219.5, "grad_norm": 4.2934743760270067e-07, "learning_rate": 2.5482989819718523e-05, "loss": 0.0, "num_input_tokens_seen": 5531008, "step": 19755 }, { "epoch": 219.55555555555554, "grad_norm": 3.815165712239832e-07, "learning_rate": 2.5473174138028065e-05, "loss": 0.0, "num_input_tokens_seen": 5532384, "step": 19760 }, { "epoch": 219.61111111111111, "grad_norm": 4.353912288479478e-07, "learning_rate": 2.5463358383368212e-05, "loss": 0.0, "num_input_tokens_seen": 5533728, "step": 19765 }, { "epoch": 219.66666666666666, "grad_norm": 1.7323128531643306e-07, "learning_rate": 2.545354255725267e-05, "loss": 0.0, "num_input_tokens_seen": 5535152, "step": 19770 }, { "epoch": 219.72222222222223, "grad_norm": 7.539583180005138e-07, "learning_rate": 2.5443726661195165e-05, "loss": 0.0, "num_input_tokens_seen": 5536560, "step": 19775 }, { "epoch": 219.77777777777777, "grad_norm": 1.8726753125974938e-07, "learning_rate": 2.543391069670944e-05, "loss": 0.0, "num_input_tokens_seen": 5537936, "step": 19780 }, { "epoch": 219.83333333333334, "grad_norm": 6.361494229167874e-07, "learning_rate": 2.5424094665309228e-05, "loss": 0.0, "num_input_tokens_seen": 5539360, "step": 19785 }, { "epoch": 219.88888888888889, "grad_norm": 6.642357561759127e-07, "learning_rate": 2.5414278568508292e-05, "loss": 0.0, "num_input_tokens_seen": 5540784, "step": 19790 }, { "epoch": 219.94444444444446, "grad_norm": 2.262953273657331e-07, "learning_rate": 2.540446240782039e-05, "loss": 0.0, "num_input_tokens_seen": 5542160, "step": 19795 }, { "epoch": 220.0, "grad_norm": 1.6678028202932182e-07, "learning_rate": 2.5394646184759307e-05, "loss": 0.0, "num_input_tokens_seen": 5543504, "step": 19800 }, { "epoch": 220.0, "eval_loss": 1.3567408323287964, "eval_runtime": 1.3841, "eval_samples_per_second": 28.901, "eval_steps_per_second": 14.45, "num_input_tokens_seen": 5543504, "step": 19800 }, { "epoch": 220.05555555555554, "grad_norm": 1.543659919889251e-07, "learning_rate": 2.538482990083882e-05, "loss": 0.0, "num_input_tokens_seen": 5544928, "step": 19805 }, { "epoch": 220.11111111111111, "grad_norm": 1.6422247028913262e-07, "learning_rate": 2.5375013557572725e-05, "loss": 0.0, "num_input_tokens_seen": 5546352, "step": 19810 }, { "epoch": 220.16666666666666, "grad_norm": 5.320854938872799e-07, "learning_rate": 2.536519715647483e-05, "loss": 0.0, "num_input_tokens_seen": 5547712, "step": 19815 }, { "epoch": 220.22222222222223, "grad_norm": 4.304707204028091e-07, "learning_rate": 2.535538069905894e-05, "loss": 0.0, "num_input_tokens_seen": 5549168, "step": 19820 }, { "epoch": 220.27777777777777, "grad_norm": 1.5439434264408192e-07, "learning_rate": 2.534556418683888e-05, "loss": 0.0, "num_input_tokens_seen": 5550576, "step": 19825 }, { "epoch": 220.33333333333334, "grad_norm": 2.890728580950963e-07, "learning_rate": 2.5335747621328486e-05, "loss": 0.0, "num_input_tokens_seen": 5551968, "step": 19830 }, { "epoch": 220.38888888888889, "grad_norm": 1.4687277882785565e-07, "learning_rate": 2.5325931004041586e-05, "loss": 0.0, "num_input_tokens_seen": 5553376, "step": 19835 }, { "epoch": 220.44444444444446, "grad_norm": 1.7885375314108387e-07, "learning_rate": 2.5316114336492032e-05, "loss": 0.0, "num_input_tokens_seen": 5554736, "step": 19840 }, { "epoch": 220.5, "grad_norm": 7.522105249790911e-08, "learning_rate": 2.530629762019367e-05, "loss": 0.0, "num_input_tokens_seen": 5556112, "step": 19845 }, { "epoch": 220.55555555555554, "grad_norm": 6.013701181473152e-08, "learning_rate": 2.5296480856660364e-05, "loss": 0.0, "num_input_tokens_seen": 5557520, "step": 19850 }, { "epoch": 220.61111111111111, "grad_norm": 1.5375289308394713e-07, "learning_rate": 2.528666404740599e-05, "loss": 0.0, "num_input_tokens_seen": 5558944, "step": 19855 }, { "epoch": 220.66666666666666, "grad_norm": 2.0333526151716796e-07, "learning_rate": 2.527684719394442e-05, "loss": 0.0, "num_input_tokens_seen": 5560352, "step": 19860 }, { "epoch": 220.72222222222223, "grad_norm": 3.387881974958873e-07, "learning_rate": 2.526703029778953e-05, "loss": 0.0, "num_input_tokens_seen": 5561776, "step": 19865 }, { "epoch": 220.77777777777777, "grad_norm": 3.4768218029057607e-06, "learning_rate": 2.5257213360455208e-05, "loss": 0.0, "num_input_tokens_seen": 5563136, "step": 19870 }, { "epoch": 220.83333333333334, "grad_norm": 1.9692627972744958e-07, "learning_rate": 2.5247396383455353e-05, "loss": 0.0, "num_input_tokens_seen": 5564528, "step": 19875 }, { "epoch": 220.88888888888889, "grad_norm": 4.917872331589024e-08, "learning_rate": 2.523757936830387e-05, "loss": 0.0, "num_input_tokens_seen": 5565888, "step": 19880 }, { "epoch": 220.94444444444446, "grad_norm": 2.988653236002392e-08, "learning_rate": 2.5227762316514662e-05, "loss": 0.0, "num_input_tokens_seen": 5567312, "step": 19885 }, { "epoch": 221.0, "grad_norm": 2.8231102078279946e-06, "learning_rate": 2.5217945229601648e-05, "loss": 0.0, "num_input_tokens_seen": 5568672, "step": 19890 }, { "epoch": 221.05555555555554, "grad_norm": 6.90942556502705e-07, "learning_rate": 2.5208128109078738e-05, "loss": 0.0, "num_input_tokens_seen": 5570096, "step": 19895 }, { "epoch": 221.11111111111111, "grad_norm": 1.8944008672860946e-07, "learning_rate": 2.5198310956459853e-05, "loss": 0.0, "num_input_tokens_seen": 5571456, "step": 19900 }, { "epoch": 221.16666666666666, "grad_norm": 1.0350603929509816e-07, "learning_rate": 2.518849377325893e-05, "loss": 0.0, "num_input_tokens_seen": 5572848, "step": 19905 }, { "epoch": 221.22222222222223, "grad_norm": 2.736061901487119e-07, "learning_rate": 2.51786765609899e-05, "loss": 0.0, "num_input_tokens_seen": 5574240, "step": 19910 }, { "epoch": 221.27777777777777, "grad_norm": 2.5599635478101845e-07, "learning_rate": 2.5168859321166694e-05, "loss": 0.0, "num_input_tokens_seen": 5575616, "step": 19915 }, { "epoch": 221.33333333333334, "grad_norm": 3.412961646631629e-08, "learning_rate": 2.515904205530326e-05, "loss": 0.0, "num_input_tokens_seen": 5577040, "step": 19920 }, { "epoch": 221.38888888888889, "grad_norm": 2.394466775967885e-07, "learning_rate": 2.514922476491355e-05, "loss": 0.0, "num_input_tokens_seen": 5578432, "step": 19925 }, { "epoch": 221.44444444444446, "grad_norm": 2.1277513440054463e-07, "learning_rate": 2.51394074515115e-05, "loss": 0.0, "num_input_tokens_seen": 5579856, "step": 19930 }, { "epoch": 221.5, "grad_norm": 1.4111476787093125e-07, "learning_rate": 2.5129590116611067e-05, "loss": 0.0, "num_input_tokens_seen": 5581216, "step": 19935 }, { "epoch": 221.55555555555554, "grad_norm": 3.2723448839533376e-08, "learning_rate": 2.5119772761726212e-05, "loss": 0.0, "num_input_tokens_seen": 5582656, "step": 19940 }, { "epoch": 221.61111111111111, "grad_norm": 6.220194848083338e-08, "learning_rate": 2.5109955388370893e-05, "loss": 0.0, "num_input_tokens_seen": 5584032, "step": 19945 }, { "epoch": 221.66666666666666, "grad_norm": 2.095714393135495e-07, "learning_rate": 2.510013799805907e-05, "loss": 0.0, "num_input_tokens_seen": 5585408, "step": 19950 }, { "epoch": 221.72222222222223, "grad_norm": 3.6508563994175347e-07, "learning_rate": 2.5090320592304706e-05, "loss": 0.0, "num_input_tokens_seen": 5586848, "step": 19955 }, { "epoch": 221.77777777777777, "grad_norm": 8.804835971432112e-08, "learning_rate": 2.5080503172621777e-05, "loss": 0.0, "num_input_tokens_seen": 5588224, "step": 19960 }, { "epoch": 221.83333333333334, "grad_norm": 4.1259156091655314e-07, "learning_rate": 2.5070685740524246e-05, "loss": 0.0, "num_input_tokens_seen": 5589664, "step": 19965 }, { "epoch": 221.88888888888889, "grad_norm": 1.8076728736104997e-07, "learning_rate": 2.5060868297526084e-05, "loss": 0.0, "num_input_tokens_seen": 5591056, "step": 19970 }, { "epoch": 221.94444444444446, "grad_norm": 2.4352169702979154e-07, "learning_rate": 2.5051050845141267e-05, "loss": 0.0, "num_input_tokens_seen": 5592432, "step": 19975 }, { "epoch": 222.0, "grad_norm": 4.766021177715629e-08, "learning_rate": 2.5041233384883765e-05, "loss": 0.0, "num_input_tokens_seen": 5593824, "step": 19980 }, { "epoch": 222.05555555555554, "grad_norm": 3.4607523957674857e-07, "learning_rate": 2.5031415918267564e-05, "loss": 0.0, "num_input_tokens_seen": 5595264, "step": 19985 }, { "epoch": 222.11111111111111, "grad_norm": 1.4702081330142391e-07, "learning_rate": 2.5021598446806626e-05, "loss": 0.0, "num_input_tokens_seen": 5596688, "step": 19990 }, { "epoch": 222.16666666666666, "grad_norm": 2.419461395675171e-07, "learning_rate": 2.5011780972014937e-05, "loss": 0.0, "num_input_tokens_seen": 5598080, "step": 19995 }, { "epoch": 222.22222222222223, "grad_norm": 5.6968765704823454e-08, "learning_rate": 2.5001963495406478e-05, "loss": 0.0, "num_input_tokens_seen": 5599440, "step": 20000 }, { "epoch": 222.22222222222223, "eval_loss": 1.3599047660827637, "eval_runtime": 1.3848, "eval_samples_per_second": 28.886, "eval_steps_per_second": 14.443, "num_input_tokens_seen": 5599440, "step": 20000 }, { "epoch": 222.27777777777777, "grad_norm": 9.830192482240818e-08, "learning_rate": 2.499214601849522e-05, "loss": 0.0, "num_input_tokens_seen": 5600848, "step": 20005 }, { "epoch": 222.33333333333334, "grad_norm": 3.1124079669098137e-07, "learning_rate": 2.4982328542795148e-05, "loss": 0.0, "num_input_tokens_seen": 5602240, "step": 20010 }, { "epoch": 222.38888888888889, "grad_norm": 4.0712589566282986e-07, "learning_rate": 2.497251106982024e-05, "loss": 0.0, "num_input_tokens_seen": 5603600, "step": 20015 }, { "epoch": 222.44444444444446, "grad_norm": 3.531148990987276e-07, "learning_rate": 2.4962693601084458e-05, "loss": 0.0, "num_input_tokens_seen": 5604992, "step": 20020 }, { "epoch": 222.5, "grad_norm": 1.0593993948759817e-07, "learning_rate": 2.4952876138101794e-05, "loss": 0.0, "num_input_tokens_seen": 5606384, "step": 20025 }, { "epoch": 222.55555555555554, "grad_norm": 4.198142846689734e-07, "learning_rate": 2.4943058682386233e-05, "loss": 0.0, "num_input_tokens_seen": 5607792, "step": 20030 }, { "epoch": 222.61111111111111, "grad_norm": 5.9567209831357104e-08, "learning_rate": 2.493324123545173e-05, "loss": 0.0, "num_input_tokens_seen": 5609216, "step": 20035 }, { "epoch": 222.66666666666666, "grad_norm": 9.065754511539126e-07, "learning_rate": 2.4923423798812272e-05, "loss": 0.0, "num_input_tokens_seen": 5610656, "step": 20040 }, { "epoch": 222.72222222222223, "grad_norm": 8.80482460274834e-08, "learning_rate": 2.4913606373981825e-05, "loss": 0.0, "num_input_tokens_seen": 5612064, "step": 20045 }, { "epoch": 222.77777777777777, "grad_norm": 1.851410758035854e-07, "learning_rate": 2.4903788962474357e-05, "loss": 0.0, "num_input_tokens_seen": 5613472, "step": 20050 }, { "epoch": 222.83333333333334, "grad_norm": 1.0850421716668279e-07, "learning_rate": 2.489397156580385e-05, "loss": 0.0, "num_input_tokens_seen": 5614816, "step": 20055 }, { "epoch": 222.88888888888889, "grad_norm": 5.653795938087569e-07, "learning_rate": 2.4884154185484246e-05, "loss": 0.0, "num_input_tokens_seen": 5616208, "step": 20060 }, { "epoch": 222.94444444444446, "grad_norm": 7.922640747892729e-07, "learning_rate": 2.4874336823029526e-05, "loss": 0.0, "num_input_tokens_seen": 5617632, "step": 20065 }, { "epoch": 223.0, "grad_norm": 1.5899338734470803e-07, "learning_rate": 2.4864519479953656e-05, "loss": 0.0, "num_input_tokens_seen": 5619024, "step": 20070 }, { "epoch": 223.05555555555554, "grad_norm": 3.3930895426692587e-08, "learning_rate": 2.485470215777058e-05, "loss": 0.0, "num_input_tokens_seen": 5620448, "step": 20075 }, { "epoch": 223.11111111111111, "grad_norm": 9.038627979407465e-08, "learning_rate": 2.4844884857994258e-05, "loss": 0.0, "num_input_tokens_seen": 5621840, "step": 20080 }, { "epoch": 223.16666666666666, "grad_norm": 5.732227492671882e-08, "learning_rate": 2.4835067582138638e-05, "loss": 0.0, "num_input_tokens_seen": 5623280, "step": 20085 }, { "epoch": 223.22222222222223, "grad_norm": 5.652997856486763e-07, "learning_rate": 2.4825250331717666e-05, "loss": 0.0, "num_input_tokens_seen": 5624672, "step": 20090 }, { "epoch": 223.27777777777777, "grad_norm": 2.980133899654902e-07, "learning_rate": 2.4815433108245298e-05, "loss": 0.0, "num_input_tokens_seen": 5626112, "step": 20095 }, { "epoch": 223.33333333333334, "grad_norm": 1.0087925517154872e-07, "learning_rate": 2.4805615913235456e-05, "loss": 0.0, "num_input_tokens_seen": 5627552, "step": 20100 }, { "epoch": 223.38888888888889, "grad_norm": 6.917895802871499e-07, "learning_rate": 2.479579874820208e-05, "loss": 0.0, "num_input_tokens_seen": 5628944, "step": 20105 }, { "epoch": 223.44444444444446, "grad_norm": 1.0287584473189781e-06, "learning_rate": 2.4785981614659115e-05, "loss": 0.0, "num_input_tokens_seen": 5630320, "step": 20110 }, { "epoch": 223.5, "grad_norm": 2.528476272800617e-07, "learning_rate": 2.477616451412047e-05, "loss": 0.0, "num_input_tokens_seen": 5631664, "step": 20115 }, { "epoch": 223.55555555555554, "grad_norm": 1.0106361258976904e-07, "learning_rate": 2.476634744810007e-05, "loss": 0.0, "num_input_tokens_seen": 5633056, "step": 20120 }, { "epoch": 223.61111111111111, "grad_norm": 6.680862441044155e-08, "learning_rate": 2.475653041811183e-05, "loss": 0.0, "num_input_tokens_seen": 5634448, "step": 20125 }, { "epoch": 223.66666666666666, "grad_norm": 5.408071501733502e-07, "learning_rate": 2.4746713425669652e-05, "loss": 0.0, "num_input_tokens_seen": 5635824, "step": 20130 }, { "epoch": 223.72222222222223, "grad_norm": 2.051319825113751e-07, "learning_rate": 2.4736896472287458e-05, "loss": 0.0, "num_input_tokens_seen": 5637232, "step": 20135 }, { "epoch": 223.77777777777777, "grad_norm": 5.676607273130685e-08, "learning_rate": 2.4727079559479124e-05, "loss": 0.0, "num_input_tokens_seen": 5638640, "step": 20140 }, { "epoch": 223.83333333333334, "grad_norm": 8.616704150199439e-08, "learning_rate": 2.4717262688758557e-05, "loss": 0.0, "num_input_tokens_seen": 5640048, "step": 20145 }, { "epoch": 223.88888888888889, "grad_norm": 1.3731855119658576e-07, "learning_rate": 2.4707445861639637e-05, "loss": 0.0, "num_input_tokens_seen": 5641472, "step": 20150 }, { "epoch": 223.94444444444446, "grad_norm": 2.375853114244819e-08, "learning_rate": 2.4697629079636244e-05, "loss": 0.0, "num_input_tokens_seen": 5642880, "step": 20155 }, { "epoch": 224.0, "grad_norm": 4.7765791322262885e-08, "learning_rate": 2.4687812344262244e-05, "loss": 0.0, "num_input_tokens_seen": 5644224, "step": 20160 }, { "epoch": 224.05555555555554, "grad_norm": 2.5878449605443166e-07, "learning_rate": 2.46779956570315e-05, "loss": 0.0, "num_input_tokens_seen": 5645600, "step": 20165 }, { "epoch": 224.11111111111111, "grad_norm": 1.5311987056065846e-07, "learning_rate": 2.466817901945787e-05, "loss": 0.0, "num_input_tokens_seen": 5647024, "step": 20170 }, { "epoch": 224.16666666666666, "grad_norm": 7.048343064752771e-08, "learning_rate": 2.4658362433055217e-05, "loss": 0.0, "num_input_tokens_seen": 5648448, "step": 20175 }, { "epoch": 224.22222222222223, "grad_norm": 3.0691805363858293e-07, "learning_rate": 2.4648545899337356e-05, "loss": 0.0, "num_input_tokens_seen": 5649856, "step": 20180 }, { "epoch": 224.27777777777777, "grad_norm": 6.258918006096792e-07, "learning_rate": 2.4638729419818143e-05, "loss": 0.0, "num_input_tokens_seen": 5651216, "step": 20185 }, { "epoch": 224.33333333333334, "grad_norm": 4.161626350196457e-07, "learning_rate": 2.46289129960114e-05, "loss": 0.0, "num_input_tokens_seen": 5652624, "step": 20190 }, { "epoch": 224.38888888888889, "grad_norm": 1.754815741605853e-07, "learning_rate": 2.4619096629430924e-05, "loss": 0.0, "num_input_tokens_seen": 5654016, "step": 20195 }, { "epoch": 224.44444444444446, "grad_norm": 7.028979354117837e-08, "learning_rate": 2.4609280321590543e-05, "loss": 0.0, "num_input_tokens_seen": 5655424, "step": 20200 }, { "epoch": 224.44444444444446, "eval_loss": 1.4161657094955444, "eval_runtime": 1.386, "eval_samples_per_second": 28.86, "eval_steps_per_second": 14.43, "num_input_tokens_seen": 5655424, "step": 20200 }, { "epoch": 224.5, "grad_norm": 1.2845679009387823e-07, "learning_rate": 2.4599464074004037e-05, "loss": 0.0, "num_input_tokens_seen": 5656832, "step": 20205 }, { "epoch": 224.55555555555554, "grad_norm": 1.480932922959255e-07, "learning_rate": 2.4589647888185204e-05, "loss": 0.0, "num_input_tokens_seen": 5658208, "step": 20210 }, { "epoch": 224.61111111111111, "grad_norm": 2.3782978075814754e-07, "learning_rate": 2.4579831765647836e-05, "loss": 0.0, "num_input_tokens_seen": 5659568, "step": 20215 }, { "epoch": 224.66666666666666, "grad_norm": 1.210723041822348e-07, "learning_rate": 2.4570015707905676e-05, "loss": 0.0, "num_input_tokens_seen": 5661024, "step": 20220 }, { "epoch": 224.72222222222223, "grad_norm": 1.0460385624355695e-07, "learning_rate": 2.4560199716472508e-05, "loss": 0.0, "num_input_tokens_seen": 5662416, "step": 20225 }, { "epoch": 224.77777777777777, "grad_norm": 3.112665183380159e-07, "learning_rate": 2.455038379286207e-05, "loss": 0.0, "num_input_tokens_seen": 5663840, "step": 20230 }, { "epoch": 224.83333333333334, "grad_norm": 2.664958032028153e-08, "learning_rate": 2.4540567938588095e-05, "loss": 0.0, "num_input_tokens_seen": 5665264, "step": 20235 }, { "epoch": 224.88888888888889, "grad_norm": 1.0665775818097245e-07, "learning_rate": 2.4530752155164328e-05, "loss": 0.0, "num_input_tokens_seen": 5666656, "step": 20240 }, { "epoch": 224.94444444444446, "grad_norm": 3.028233095392352e-07, "learning_rate": 2.4520936444104463e-05, "loss": 0.0, "num_input_tokens_seen": 5668064, "step": 20245 }, { "epoch": 225.0, "grad_norm": 1.2014005790206284e-07, "learning_rate": 2.4511120806922218e-05, "loss": 0.0, "num_input_tokens_seen": 5669456, "step": 20250 }, { "epoch": 225.05555555555554, "grad_norm": 2.4789591179796844e-07, "learning_rate": 2.45013052451313e-05, "loss": 0.0, "num_input_tokens_seen": 5670800, "step": 20255 }, { "epoch": 225.11111111111111, "grad_norm": 2.5557253024999227e-07, "learning_rate": 2.4491489760245376e-05, "loss": 0.0, "num_input_tokens_seen": 5672160, "step": 20260 }, { "epoch": 225.16666666666666, "grad_norm": 2.3039886798414955e-07, "learning_rate": 2.4481674353778115e-05, "loss": 0.0, "num_input_tokens_seen": 5673568, "step": 20265 }, { "epoch": 225.22222222222223, "grad_norm": 8.742235735326176e-08, "learning_rate": 2.447185902724319e-05, "loss": 0.0, "num_input_tokens_seen": 5674960, "step": 20270 }, { "epoch": 225.27777777777777, "grad_norm": 2.5063198449970514e-07, "learning_rate": 2.4462043782154233e-05, "loss": 0.0, "num_input_tokens_seen": 5676352, "step": 20275 }, { "epoch": 225.33333333333334, "grad_norm": 2.0598390904069674e-07, "learning_rate": 2.4452228620024895e-05, "loss": 0.0, "num_input_tokens_seen": 5677776, "step": 20280 }, { "epoch": 225.38888888888889, "grad_norm": 2.230034255035207e-07, "learning_rate": 2.4442413542368776e-05, "loss": 0.0, "num_input_tokens_seen": 5679216, "step": 20285 }, { "epoch": 225.44444444444446, "grad_norm": 1.66100662113422e-07, "learning_rate": 2.4432598550699502e-05, "loss": 0.0, "num_input_tokens_seen": 5680640, "step": 20290 }, { "epoch": 225.5, "grad_norm": 2.4325191816387814e-07, "learning_rate": 2.4422783646530663e-05, "loss": 0.0, "num_input_tokens_seen": 5682048, "step": 20295 }, { "epoch": 225.55555555555554, "grad_norm": 7.106003607759703e-08, "learning_rate": 2.441296883137584e-05, "loss": 0.0, "num_input_tokens_seen": 5683440, "step": 20300 }, { "epoch": 225.61111111111111, "grad_norm": 1.5744977588383335e-07, "learning_rate": 2.4403154106748592e-05, "loss": 0.0, "num_input_tokens_seen": 5684816, "step": 20305 }, { "epoch": 225.66666666666666, "grad_norm": 7.192149809043258e-08, "learning_rate": 2.4393339474162494e-05, "loss": 0.0, "num_input_tokens_seen": 5686208, "step": 20310 }, { "epoch": 225.72222222222223, "grad_norm": 2.0753445539867243e-07, "learning_rate": 2.4383524935131062e-05, "loss": 0.0, "num_input_tokens_seen": 5687600, "step": 20315 }, { "epoch": 225.77777777777777, "grad_norm": 1.1706164571023692e-07, "learning_rate": 2.437371049116784e-05, "loss": 0.0, "num_input_tokens_seen": 5689008, "step": 20320 }, { "epoch": 225.83333333333334, "grad_norm": 2.662842746303795e-07, "learning_rate": 2.436389614378632e-05, "loss": 0.0, "num_input_tokens_seen": 5690416, "step": 20325 }, { "epoch": 225.88888888888889, "grad_norm": 5.490372956273859e-08, "learning_rate": 2.435408189450002e-05, "loss": 0.0, "num_input_tokens_seen": 5691792, "step": 20330 }, { "epoch": 225.94444444444446, "grad_norm": 2.421726819079595e-08, "learning_rate": 2.4344267744822406e-05, "loss": 0.0, "num_input_tokens_seen": 5693200, "step": 20335 }, { "epoch": 226.0, "grad_norm": 1.931333457605433e-07, "learning_rate": 2.4334453696266944e-05, "loss": 0.0, "num_input_tokens_seen": 5694624, "step": 20340 }, { "epoch": 226.05555555555554, "grad_norm": 3.306008409253991e-07, "learning_rate": 2.432463975034708e-05, "loss": 0.0, "num_input_tokens_seen": 5695984, "step": 20345 }, { "epoch": 226.11111111111111, "grad_norm": 1.1233436225666082e-07, "learning_rate": 2.4314825908576265e-05, "loss": 0.0, "num_input_tokens_seen": 5697344, "step": 20350 }, { "epoch": 226.16666666666666, "grad_norm": 4.371192119378975e-07, "learning_rate": 2.4305012172467897e-05, "loss": 0.0, "num_input_tokens_seen": 5698704, "step": 20355 }, { "epoch": 226.22222222222223, "grad_norm": 3.0168476428116264e-07, "learning_rate": 2.4295198543535393e-05, "loss": 0.0, "num_input_tokens_seen": 5700096, "step": 20360 }, { "epoch": 226.27777777777777, "grad_norm": 5.1345434570748694e-08, "learning_rate": 2.4285385023292124e-05, "loss": 0.0, "num_input_tokens_seen": 5701504, "step": 20365 }, { "epoch": 226.33333333333334, "grad_norm": 2.266678507112374e-07, "learning_rate": 2.427557161325147e-05, "loss": 0.0, "num_input_tokens_seen": 5702864, "step": 20370 }, { "epoch": 226.38888888888889, "grad_norm": 3.682821443362627e-07, "learning_rate": 2.4265758314926778e-05, "loss": 0.0, "num_input_tokens_seen": 5704304, "step": 20375 }, { "epoch": 226.44444444444446, "grad_norm": 1.63583706580539e-07, "learning_rate": 2.4255945129831373e-05, "loss": 0.0, "num_input_tokens_seen": 5705760, "step": 20380 }, { "epoch": 226.5, "grad_norm": 3.7261611396388616e-07, "learning_rate": 2.4246132059478578e-05, "loss": 0.0, "num_input_tokens_seen": 5707168, "step": 20385 }, { "epoch": 226.55555555555554, "grad_norm": 5.298121550367796e-07, "learning_rate": 2.4236319105381706e-05, "loss": 0.0, "num_input_tokens_seen": 5708528, "step": 20390 }, { "epoch": 226.61111111111111, "grad_norm": 3.398354309069873e-08, "learning_rate": 2.422650626905401e-05, "loss": 0.0, "num_input_tokens_seen": 5709920, "step": 20395 }, { "epoch": 226.66666666666666, "grad_norm": 3.679754172480898e-07, "learning_rate": 2.4216693552008785e-05, "loss": 0.0, "num_input_tokens_seen": 5711344, "step": 20400 }, { "epoch": 226.66666666666666, "eval_loss": 1.4205596446990967, "eval_runtime": 1.4033, "eval_samples_per_second": 28.504, "eval_steps_per_second": 14.252, "num_input_tokens_seen": 5711344, "step": 20400 }, { "epoch": 226.72222222222223, "grad_norm": 2.864450152628706e-07, "learning_rate": 2.4206880955759247e-05, "loss": 0.0, "num_input_tokens_seen": 5712688, "step": 20405 }, { "epoch": 226.77777777777777, "grad_norm": 2.0476242923450627e-07, "learning_rate": 2.419706848181863e-05, "loss": 0.0, "num_input_tokens_seen": 5714128, "step": 20410 }, { "epoch": 226.83333333333334, "grad_norm": 1.561457736443117e-07, "learning_rate": 2.4187256131700153e-05, "loss": 0.0, "num_input_tokens_seen": 5715520, "step": 20415 }, { "epoch": 226.88888888888889, "grad_norm": 4.0012514546106104e-07, "learning_rate": 2.4177443906916985e-05, "loss": 0.0, "num_input_tokens_seen": 5716992, "step": 20420 }, { "epoch": 226.94444444444446, "grad_norm": 4.932132924295729e-07, "learning_rate": 2.4167631808982303e-05, "loss": 0.0, "num_input_tokens_seen": 5718416, "step": 20425 }, { "epoch": 227.0, "grad_norm": 2.470679021371325e-07, "learning_rate": 2.4157819839409264e-05, "loss": 0.0, "num_input_tokens_seen": 5719840, "step": 20430 }, { "epoch": 227.05555555555554, "grad_norm": 7.102024142113805e-07, "learning_rate": 2.414800799971098e-05, "loss": 0.0, "num_input_tokens_seen": 5721232, "step": 20435 }, { "epoch": 227.11111111111111, "grad_norm": 2.535939813697041e-07, "learning_rate": 2.4138196291400582e-05, "loss": 0.0, "num_input_tokens_seen": 5722688, "step": 20440 }, { "epoch": 227.16666666666666, "grad_norm": 6.42959776087082e-07, "learning_rate": 2.412838471599114e-05, "loss": 0.0, "num_input_tokens_seen": 5724080, "step": 20445 }, { "epoch": 227.22222222222223, "grad_norm": 1.622516094812454e-07, "learning_rate": 2.411857327499572e-05, "loss": 0.0, "num_input_tokens_seen": 5725504, "step": 20450 }, { "epoch": 227.27777777777777, "grad_norm": 4.820947765438177e-07, "learning_rate": 2.410876196992739e-05, "loss": 0.0, "num_input_tokens_seen": 5726896, "step": 20455 }, { "epoch": 227.33333333333334, "grad_norm": 1.349215636992085e-07, "learning_rate": 2.4098950802299156e-05, "loss": 0.0, "num_input_tokens_seen": 5728240, "step": 20460 }, { "epoch": 227.38888888888889, "grad_norm": 6.90414435666753e-06, "learning_rate": 2.4089139773624027e-05, "loss": 0.0, "num_input_tokens_seen": 5729632, "step": 20465 }, { "epoch": 227.44444444444446, "grad_norm": 4.992343605181304e-08, "learning_rate": 2.4079328885415007e-05, "loss": 0.0, "num_input_tokens_seen": 5731040, "step": 20470 }, { "epoch": 227.5, "grad_norm": 2.619995029817801e-07, "learning_rate": 2.4069518139185036e-05, "loss": 0.0, "num_input_tokens_seen": 5732464, "step": 20475 }, { "epoch": 227.55555555555554, "grad_norm": 3.1070206318872806e-07, "learning_rate": 2.405970753644706e-05, "loss": 0.0, "num_input_tokens_seen": 5733824, "step": 20480 }, { "epoch": 227.61111111111111, "grad_norm": 2.6008879672190233e-07, "learning_rate": 2.4049897078714e-05, "loss": 0.0, "num_input_tokens_seen": 5735232, "step": 20485 }, { "epoch": 227.66666666666666, "grad_norm": 3.5022341648982547e-07, "learning_rate": 2.404008676749874e-05, "loss": 0.0, "num_input_tokens_seen": 5736608, "step": 20490 }, { "epoch": 227.72222222222223, "grad_norm": 1.001601162897714e-06, "learning_rate": 2.403027660431418e-05, "loss": 0.0, "num_input_tokens_seen": 5738032, "step": 20495 }, { "epoch": 227.77777777777777, "grad_norm": 1.1962377755025955e-07, "learning_rate": 2.402046659067314e-05, "loss": 0.0, "num_input_tokens_seen": 5739424, "step": 20500 }, { "epoch": 227.83333333333334, "grad_norm": 3.522917424447769e-08, "learning_rate": 2.401065672808847e-05, "loss": 0.0, "num_input_tokens_seen": 5740800, "step": 20505 }, { "epoch": 227.88888888888889, "grad_norm": 4.640947963707731e-07, "learning_rate": 2.400084701807296e-05, "loss": 0.0, "num_input_tokens_seen": 5742240, "step": 20510 }, { "epoch": 227.94444444444446, "grad_norm": 6.222618367246469e-07, "learning_rate": 2.39910374621394e-05, "loss": 0.0, "num_input_tokens_seen": 5743616, "step": 20515 }, { "epoch": 228.0, "grad_norm": 2.3070388976975664e-07, "learning_rate": 2.3981228061800544e-05, "loss": 0.0, "num_input_tokens_seen": 5745024, "step": 20520 }, { "epoch": 228.05555555555554, "grad_norm": 3.7540073094533e-07, "learning_rate": 2.3971418818569115e-05, "loss": 0.0, "num_input_tokens_seen": 5746448, "step": 20525 }, { "epoch": 228.11111111111111, "grad_norm": 5.243593477644026e-06, "learning_rate": 2.3961609733957832e-05, "loss": 0.0, "num_input_tokens_seen": 5747872, "step": 20530 }, { "epoch": 228.16666666666666, "grad_norm": 1.6041900607888238e-07, "learning_rate": 2.395180080947939e-05, "loss": 0.0, "num_input_tokens_seen": 5749264, "step": 20535 }, { "epoch": 228.22222222222223, "grad_norm": 8.992592626100304e-08, "learning_rate": 2.394199204664642e-05, "loss": 0.0, "num_input_tokens_seen": 5750704, "step": 20540 }, { "epoch": 228.27777777777777, "grad_norm": 7.098017817952496e-08, "learning_rate": 2.3932183446971583e-05, "loss": 0.0, "num_input_tokens_seen": 5752080, "step": 20545 }, { "epoch": 228.33333333333334, "grad_norm": 9.906759146360855e-08, "learning_rate": 2.3922375011967473e-05, "loss": 0.0, "num_input_tokens_seen": 5753472, "step": 20550 }, { "epoch": 228.38888888888889, "grad_norm": 1.040856929535039e-07, "learning_rate": 2.3912566743146676e-05, "loss": 0.0, "num_input_tokens_seen": 5754880, "step": 20555 }, { "epoch": 228.44444444444446, "grad_norm": 4.764545451507729e-07, "learning_rate": 2.390275864202176e-05, "loss": 0.0, "num_input_tokens_seen": 5756272, "step": 20560 }, { "epoch": 228.5, "grad_norm": 7.020139491942246e-07, "learning_rate": 2.3892950710105243e-05, "loss": 0.0, "num_input_tokens_seen": 5757632, "step": 20565 }, { "epoch": 228.55555555555554, "grad_norm": 2.8598086032616266e-07, "learning_rate": 2.3883142948909635e-05, "loss": 0.0, "num_input_tokens_seen": 5758976, "step": 20570 }, { "epoch": 228.61111111111111, "grad_norm": 6.442377298299107e-07, "learning_rate": 2.3873335359947433e-05, "loss": 0.0, "num_input_tokens_seen": 5760384, "step": 20575 }, { "epoch": 228.66666666666666, "grad_norm": 3.395296062080888e-07, "learning_rate": 2.3863527944731066e-05, "loss": 0.0, "num_input_tokens_seen": 5761792, "step": 20580 }, { "epoch": 228.72222222222223, "grad_norm": 4.222133327402844e-07, "learning_rate": 2.385372070477298e-05, "loss": 0.0, "num_input_tokens_seen": 5763216, "step": 20585 }, { "epoch": 228.77777777777777, "grad_norm": 4.97989276482258e-07, "learning_rate": 2.384391364158556e-05, "loss": 0.0, "num_input_tokens_seen": 5764576, "step": 20590 }, { "epoch": 228.83333333333334, "grad_norm": 1.707109191784184e-07, "learning_rate": 2.3834106756681185e-05, "loss": 0.0, "num_input_tokens_seen": 5765968, "step": 20595 }, { "epoch": 228.88888888888889, "grad_norm": 8.419031871653715e-08, "learning_rate": 2.3824300051572206e-05, "loss": 0.0, "num_input_tokens_seen": 5767376, "step": 20600 }, { "epoch": 228.88888888888889, "eval_loss": 1.3969475030899048, "eval_runtime": 1.4035, "eval_samples_per_second": 28.501, "eval_steps_per_second": 14.25, "num_input_tokens_seen": 5767376, "step": 20600 }, { "epoch": 228.94444444444446, "grad_norm": 4.6205863668546954e-07, "learning_rate": 2.3814493527770923e-05, "loss": 0.0, "num_input_tokens_seen": 5768832, "step": 20605 }, { "epoch": 229.0, "grad_norm": 3.336305383072613e-07, "learning_rate": 2.3804687186789637e-05, "loss": 0.0, "num_input_tokens_seen": 5770256, "step": 20610 }, { "epoch": 229.05555555555554, "grad_norm": 8.1821767139445e-08, "learning_rate": 2.379488103014062e-05, "loss": 0.0, "num_input_tokens_seen": 5771632, "step": 20615 }, { "epoch": 229.11111111111111, "grad_norm": 4.932622346132121e-07, "learning_rate": 2.3785075059336086e-05, "loss": 0.0, "num_input_tokens_seen": 5773024, "step": 20620 }, { "epoch": 229.16666666666666, "grad_norm": 7.731234319408031e-08, "learning_rate": 2.3775269275888248e-05, "loss": 0.0, "num_input_tokens_seen": 5774464, "step": 20625 }, { "epoch": 229.22222222222223, "grad_norm": 1.0295289030182175e-05, "learning_rate": 2.3765463681309274e-05, "loss": 0.0, "num_input_tokens_seen": 5775824, "step": 20630 }, { "epoch": 229.27777777777777, "grad_norm": 4.8820503195656784e-08, "learning_rate": 2.3755658277111313e-05, "loss": 0.0, "num_input_tokens_seen": 5777232, "step": 20635 }, { "epoch": 229.33333333333334, "grad_norm": 4.542490898984397e-07, "learning_rate": 2.374585306480649e-05, "loss": 0.0, "num_input_tokens_seen": 5778592, "step": 20640 }, { "epoch": 229.38888888888889, "grad_norm": 5.0777288151948596e-08, "learning_rate": 2.3736048045906877e-05, "loss": 0.0, "num_input_tokens_seen": 5779984, "step": 20645 }, { "epoch": 229.44444444444446, "grad_norm": 5.588903917441712e-08, "learning_rate": 2.372624322192454e-05, "loss": 0.0, "num_input_tokens_seen": 5781376, "step": 20650 }, { "epoch": 229.5, "grad_norm": 3.387398805898556e-08, "learning_rate": 2.3716438594371516e-05, "loss": 0.0, "num_input_tokens_seen": 5782752, "step": 20655 }, { "epoch": 229.55555555555554, "grad_norm": 2.1267258887291973e-07, "learning_rate": 2.3706634164759784e-05, "loss": 0.0, "num_input_tokens_seen": 5784176, "step": 20660 }, { "epoch": 229.61111111111111, "grad_norm": 5.135307219461538e-07, "learning_rate": 2.3696829934601323e-05, "loss": 0.0, "num_input_tokens_seen": 5785552, "step": 20665 }, { "epoch": 229.66666666666666, "grad_norm": 3.7779710027052715e-08, "learning_rate": 2.3687025905408053e-05, "loss": 0.0, "num_input_tokens_seen": 5786960, "step": 20670 }, { "epoch": 229.72222222222223, "grad_norm": 1.5595570346249588e-07, "learning_rate": 2.3677222078691886e-05, "loss": 0.0, "num_input_tokens_seen": 5788352, "step": 20675 }, { "epoch": 229.77777777777777, "grad_norm": 2.857501897324255e-07, "learning_rate": 2.366741845596471e-05, "loss": 0.0, "num_input_tokens_seen": 5789696, "step": 20680 }, { "epoch": 229.83333333333334, "grad_norm": 4.287167243433032e-08, "learning_rate": 2.3657615038738343e-05, "loss": 0.0, "num_input_tokens_seen": 5791072, "step": 20685 }, { "epoch": 229.88888888888889, "grad_norm": 2.4442209678454674e-07, "learning_rate": 2.3647811828524614e-05, "loss": 0.0, "num_input_tokens_seen": 5792528, "step": 20690 }, { "epoch": 229.94444444444446, "grad_norm": 5.950982426838891e-07, "learning_rate": 2.363800882683529e-05, "loss": 0.0, "num_input_tokens_seen": 5793920, "step": 20695 }, { "epoch": 230.0, "grad_norm": 3.5428138289717026e-07, "learning_rate": 2.3628206035182125e-05, "loss": 0.0, "num_input_tokens_seen": 5795360, "step": 20700 }, { "epoch": 230.05555555555554, "grad_norm": 2.863853012513573e-07, "learning_rate": 2.361840345507683e-05, "loss": 0.0, "num_input_tokens_seen": 5796736, "step": 20705 }, { "epoch": 230.11111111111111, "grad_norm": 4.9494660743221175e-06, "learning_rate": 2.3608601088031073e-05, "loss": 0.0, "num_input_tokens_seen": 5798144, "step": 20710 }, { "epoch": 230.16666666666666, "grad_norm": 4.2128149857489916e-08, "learning_rate": 2.3598798935556516e-05, "loss": 0.0, "num_input_tokens_seen": 5799520, "step": 20715 }, { "epoch": 230.22222222222223, "grad_norm": 1.9897096592558228e-07, "learning_rate": 2.3588996999164784e-05, "loss": 0.0, "num_input_tokens_seen": 5800896, "step": 20720 }, { "epoch": 230.27777777777777, "grad_norm": 1.482339939684607e-07, "learning_rate": 2.3579195280367434e-05, "loss": 0.0, "num_input_tokens_seen": 5802256, "step": 20725 }, { "epoch": 230.33333333333334, "grad_norm": 9.941526286638691e-07, "learning_rate": 2.356939378067603e-05, "loss": 0.0, "num_input_tokens_seen": 5803616, "step": 20730 }, { "epoch": 230.38888888888889, "grad_norm": 3.3580440685909707e-06, "learning_rate": 2.3559592501602092e-05, "loss": 0.0, "num_input_tokens_seen": 5804960, "step": 20735 }, { "epoch": 230.44444444444446, "grad_norm": 1.0354721098337905e-06, "learning_rate": 2.3549791444657076e-05, "loss": 0.0, "num_input_tokens_seen": 5806368, "step": 20740 }, { "epoch": 230.5, "grad_norm": 1.2477471500460524e-06, "learning_rate": 2.353999061135246e-05, "loss": 0.0, "num_input_tokens_seen": 5807712, "step": 20745 }, { "epoch": 230.55555555555554, "grad_norm": 2.0819967971874576e-07, "learning_rate": 2.3530190003199626e-05, "loss": 0.0, "num_input_tokens_seen": 5809072, "step": 20750 }, { "epoch": 230.61111111111111, "grad_norm": 1.6372847255752276e-07, "learning_rate": 2.3520389621709965e-05, "loss": 0.0, "num_input_tokens_seen": 5810480, "step": 20755 }, { "epoch": 230.66666666666666, "grad_norm": 8.243955562647898e-08, "learning_rate": 2.351058946839483e-05, "loss": 0.0, "num_input_tokens_seen": 5811888, "step": 20760 }, { "epoch": 230.72222222222223, "grad_norm": 4.66766216788983e-08, "learning_rate": 2.350078954476551e-05, "loss": 0.0, "num_input_tokens_seen": 5813280, "step": 20765 }, { "epoch": 230.77777777777777, "grad_norm": 2.1331877064767468e-07, "learning_rate": 2.3490989852333272e-05, "loss": 0.0, "num_input_tokens_seen": 5814704, "step": 20770 }, { "epoch": 230.83333333333334, "grad_norm": 5.605052137980238e-07, "learning_rate": 2.3481190392609377e-05, "loss": 0.0, "num_input_tokens_seen": 5816144, "step": 20775 }, { "epoch": 230.88888888888889, "grad_norm": 4.7152082061074907e-07, "learning_rate": 2.3471391167105e-05, "loss": 0.0, "num_input_tokens_seen": 5817584, "step": 20780 }, { "epoch": 230.94444444444446, "grad_norm": 9.405201808476704e-07, "learning_rate": 2.3461592177331325e-05, "loss": 0.0, "num_input_tokens_seen": 5819008, "step": 20785 }, { "epoch": 231.0, "grad_norm": 2.6239431960561888e-08, "learning_rate": 2.345179342479946e-05, "loss": 0.0, "num_input_tokens_seen": 5820432, "step": 20790 }, { "epoch": 231.05555555555554, "grad_norm": 4.0161515357794997e-07, "learning_rate": 2.3441994911020503e-05, "loss": 0.0, "num_input_tokens_seen": 5821872, "step": 20795 }, { "epoch": 231.11111111111111, "grad_norm": 6.0268519064266e-07, "learning_rate": 2.3432196637505522e-05, "loss": 0.0, "num_input_tokens_seen": 5823264, "step": 20800 }, { "epoch": 231.11111111111111, "eval_loss": 1.4180264472961426, "eval_runtime": 1.3839, "eval_samples_per_second": 28.904, "eval_steps_per_second": 14.452, "num_input_tokens_seen": 5823264, "step": 20800 }, { "epoch": 231.16666666666666, "grad_norm": 4.091637890724087e-07, "learning_rate": 2.3422398605765515e-05, "loss": 0.0, "num_input_tokens_seen": 5824640, "step": 20805 }, { "epoch": 231.22222222222223, "grad_norm": 8.838638336783333e-07, "learning_rate": 2.3412600817311462e-05, "loss": 0.0, "num_input_tokens_seen": 5826000, "step": 20810 }, { "epoch": 231.27777777777777, "grad_norm": 1.754783056640008e-07, "learning_rate": 2.3402803273654326e-05, "loss": 0.0, "num_input_tokens_seen": 5827392, "step": 20815 }, { "epoch": 231.33333333333334, "grad_norm": 1.6743454978040972e-07, "learning_rate": 2.3393005976304983e-05, "loss": 0.0, "num_input_tokens_seen": 5828768, "step": 20820 }, { "epoch": 231.38888888888889, "grad_norm": 9.218297947199972e-08, "learning_rate": 2.338320892677432e-05, "loss": 0.0, "num_input_tokens_seen": 5830160, "step": 20825 }, { "epoch": 231.44444444444446, "grad_norm": 4.561562434446387e-07, "learning_rate": 2.3373412126573155e-05, "loss": 0.0, "num_input_tokens_seen": 5831600, "step": 20830 }, { "epoch": 231.5, "grad_norm": 1.81948919930619e-07, "learning_rate": 2.3363615577212285e-05, "loss": 0.0, "num_input_tokens_seen": 5832992, "step": 20835 }, { "epoch": 231.55555555555554, "grad_norm": 4.079780921983911e-07, "learning_rate": 2.3353819280202455e-05, "loss": 0.0, "num_input_tokens_seen": 5834384, "step": 20840 }, { "epoch": 231.61111111111111, "grad_norm": 8.805355946606142e-07, "learning_rate": 2.334402323705438e-05, "loss": 0.0, "num_input_tokens_seen": 5835776, "step": 20845 }, { "epoch": 231.66666666666666, "grad_norm": 8.662428854222526e-07, "learning_rate": 2.3334227449278725e-05, "loss": 0.0, "num_input_tokens_seen": 5837184, "step": 20850 }, { "epoch": 231.72222222222223, "grad_norm": 3.942719558835961e-06, "learning_rate": 2.3324431918386143e-05, "loss": 0.0, "num_input_tokens_seen": 5838592, "step": 20855 }, { "epoch": 231.77777777777777, "grad_norm": 7.610066887764333e-08, "learning_rate": 2.3314636645887207e-05, "loss": 0.0, "num_input_tokens_seen": 5840000, "step": 20860 }, { "epoch": 231.83333333333334, "grad_norm": 1.639757698512767e-07, "learning_rate": 2.3304841633292487e-05, "loss": 0.0, "num_input_tokens_seen": 5841424, "step": 20865 }, { "epoch": 231.88888888888889, "grad_norm": 3.625803230988822e-07, "learning_rate": 2.329504688211248e-05, "loss": 0.0, "num_input_tokens_seen": 5842816, "step": 20870 }, { "epoch": 231.94444444444446, "grad_norm": 4.7140396475242596e-08, "learning_rate": 2.3285252393857677e-05, "loss": 0.0, "num_input_tokens_seen": 5844224, "step": 20875 }, { "epoch": 232.0, "grad_norm": 4.686225167915836e-07, "learning_rate": 2.327545817003851e-05, "loss": 0.0, "num_input_tokens_seen": 5845632, "step": 20880 }, { "epoch": 232.05555555555554, "grad_norm": 3.882219971274026e-06, "learning_rate": 2.326566421216535e-05, "loss": 0.0, "num_input_tokens_seen": 5847040, "step": 20885 }, { "epoch": 232.11111111111111, "grad_norm": 1.1545427014425513e-07, "learning_rate": 2.3255870521748565e-05, "loss": 0.0, "num_input_tokens_seen": 5848448, "step": 20890 }, { "epoch": 232.16666666666666, "grad_norm": 2.1710825137688516e-07, "learning_rate": 2.3246077100298474e-05, "loss": 0.0, "num_input_tokens_seen": 5849824, "step": 20895 }, { "epoch": 232.22222222222223, "grad_norm": 3.1637310371479543e-07, "learning_rate": 2.3236283949325328e-05, "loss": 0.0, "num_input_tokens_seen": 5851200, "step": 20900 }, { "epoch": 232.27777777777777, "grad_norm": 6.525853279981675e-08, "learning_rate": 2.3226491070339368e-05, "loss": 0.0, "num_input_tokens_seen": 5852624, "step": 20905 }, { "epoch": 232.33333333333334, "grad_norm": 1.0439846676035813e-07, "learning_rate": 2.3216698464850762e-05, "loss": 0.0, "num_input_tokens_seen": 5854032, "step": 20910 }, { "epoch": 232.38888888888889, "grad_norm": 7.46610226087796e-07, "learning_rate": 2.320690613436967e-05, "loss": 0.0, "num_input_tokens_seen": 5855456, "step": 20915 }, { "epoch": 232.44444444444446, "grad_norm": 1.196197985109393e-07, "learning_rate": 2.3197114080406192e-05, "loss": 0.0, "num_input_tokens_seen": 5856880, "step": 20920 }, { "epoch": 232.5, "grad_norm": 1.2708075303180522e-07, "learning_rate": 2.3187322304470365e-05, "loss": 0.0, "num_input_tokens_seen": 5858256, "step": 20925 }, { "epoch": 232.55555555555554, "grad_norm": 9.940934120322709e-08, "learning_rate": 2.3177530808072222e-05, "loss": 0.0, "num_input_tokens_seen": 5859680, "step": 20930 }, { "epoch": 232.61111111111111, "grad_norm": 7.662212624381937e-07, "learning_rate": 2.316773959272174e-05, "loss": 0.0, "num_input_tokens_seen": 5861040, "step": 20935 }, { "epoch": 232.66666666666666, "grad_norm": 8.8778648432708e-07, "learning_rate": 2.3157948659928823e-05, "loss": 0.0, "num_input_tokens_seen": 5862496, "step": 20940 }, { "epoch": 232.72222222222223, "grad_norm": 1.653933878742464e-07, "learning_rate": 2.3148158011203388e-05, "loss": 0.0, "num_input_tokens_seen": 5863872, "step": 20945 }, { "epoch": 232.77777777777777, "grad_norm": 2.718191183248564e-07, "learning_rate": 2.3138367648055253e-05, "loss": 0.0, "num_input_tokens_seen": 5865328, "step": 20950 }, { "epoch": 232.83333333333334, "grad_norm": 1.5591531621339527e-07, "learning_rate": 2.312857757199422e-05, "loss": 0.0, "num_input_tokens_seen": 5866720, "step": 20955 }, { "epoch": 232.88888888888889, "grad_norm": 4.800179453923192e-07, "learning_rate": 2.3118787784530048e-05, "loss": 0.0, "num_input_tokens_seen": 5868096, "step": 20960 }, { "epoch": 232.94444444444446, "grad_norm": 2.8552784669955145e-07, "learning_rate": 2.310899828717243e-05, "loss": 0.0, "num_input_tokens_seen": 5869488, "step": 20965 }, { "epoch": 233.0, "grad_norm": 9.799547484590221e-08, "learning_rate": 2.309920908143104e-05, "loss": 0.0, "num_input_tokens_seen": 5870896, "step": 20970 }, { "epoch": 233.05555555555554, "grad_norm": 1.7274072661166429e-07, "learning_rate": 2.308942016881551e-05, "loss": 0.0, "num_input_tokens_seen": 5872272, "step": 20975 }, { "epoch": 233.11111111111111, "grad_norm": 3.9515459349104276e-08, "learning_rate": 2.307963155083539e-05, "loss": 0.0, "num_input_tokens_seen": 5873632, "step": 20980 }, { "epoch": 233.16666666666666, "grad_norm": 3.1908760433907446e-07, "learning_rate": 2.306984322900022e-05, "loss": 0.0, "num_input_tokens_seen": 5874992, "step": 20985 }, { "epoch": 233.22222222222223, "grad_norm": 8.801390549706412e-08, "learning_rate": 2.3060055204819482e-05, "loss": 0.0, "num_input_tokens_seen": 5876368, "step": 20990 }, { "epoch": 233.27777777777777, "grad_norm": 8.700380504933491e-08, "learning_rate": 2.3050267479802604e-05, "loss": 0.0, "num_input_tokens_seen": 5877824, "step": 20995 }, { "epoch": 233.33333333333334, "grad_norm": 3.664073133791135e-08, "learning_rate": 2.304048005545899e-05, "loss": 0.0, "num_input_tokens_seen": 5879248, "step": 21000 }, { "epoch": 233.33333333333334, "eval_loss": 1.417594313621521, "eval_runtime": 1.3864, "eval_samples_per_second": 28.852, "eval_steps_per_second": 14.426, "num_input_tokens_seen": 5879248, "step": 21000 }, { "epoch": 233.38888888888889, "grad_norm": 5.041276267547801e-07, "learning_rate": 2.3030692933297972e-05, "loss": 0.0, "num_input_tokens_seen": 5880688, "step": 21005 }, { "epoch": 233.44444444444446, "grad_norm": 1.5742949699415476e-07, "learning_rate": 2.3020906114828843e-05, "loss": 0.0, "num_input_tokens_seen": 5882080, "step": 21010 }, { "epoch": 233.5, "grad_norm": 5.494711103892769e-07, "learning_rate": 2.301111960156088e-05, "loss": 0.0, "num_input_tokens_seen": 5883472, "step": 21015 }, { "epoch": 233.55555555555554, "grad_norm": 4.927543386656907e-07, "learning_rate": 2.300133339500326e-05, "loss": 0.0, "num_input_tokens_seen": 5884912, "step": 21020 }, { "epoch": 233.61111111111111, "grad_norm": 2.407419401606603e-07, "learning_rate": 2.2991547496665148e-05, "loss": 0.0, "num_input_tokens_seen": 5886288, "step": 21025 }, { "epoch": 233.66666666666666, "grad_norm": 3.7973563848936465e-06, "learning_rate": 2.298176190805565e-05, "loss": 0.0, "num_input_tokens_seen": 5887712, "step": 21030 }, { "epoch": 233.72222222222223, "grad_norm": 1.0900453872864091e-07, "learning_rate": 2.2971976630683826e-05, "loss": 0.0, "num_input_tokens_seen": 5889088, "step": 21035 }, { "epoch": 233.77777777777777, "grad_norm": 7.602270102324837e-08, "learning_rate": 2.29621916660587e-05, "loss": 0.0, "num_input_tokens_seen": 5890512, "step": 21040 }, { "epoch": 233.83333333333334, "grad_norm": 4.1543906093011174e-08, "learning_rate": 2.295240701568922e-05, "loss": 0.0, "num_input_tokens_seen": 5891872, "step": 21045 }, { "epoch": 233.88888888888889, "grad_norm": 1.5101748829238204e-07, "learning_rate": 2.2942622681084312e-05, "loss": 0.0, "num_input_tokens_seen": 5893264, "step": 21050 }, { "epoch": 233.94444444444446, "grad_norm": 1.0532436789389976e-07, "learning_rate": 2.293283866375284e-05, "loss": 0.0, "num_input_tokens_seen": 5894672, "step": 21055 }, { "epoch": 234.0, "grad_norm": 1.1193449722668447e-07, "learning_rate": 2.2923054965203627e-05, "loss": 0.0, "num_input_tokens_seen": 5896096, "step": 21060 }, { "epoch": 234.05555555555554, "grad_norm": 4.6805604370092624e-07, "learning_rate": 2.2913271586945443e-05, "loss": 0.0, "num_input_tokens_seen": 5897440, "step": 21065 }, { "epoch": 234.11111111111111, "grad_norm": 6.256690312511637e-07, "learning_rate": 2.290348853048699e-05, "loss": 0.0, "num_input_tokens_seen": 5898848, "step": 21070 }, { "epoch": 234.16666666666666, "grad_norm": 4.474760771699948e-06, "learning_rate": 2.2893705797336956e-05, "loss": 0.0, "num_input_tokens_seen": 5900256, "step": 21075 }, { "epoch": 234.22222222222223, "grad_norm": 1.713686259563474e-07, "learning_rate": 2.288392338900397e-05, "loss": 0.0, "num_input_tokens_seen": 5901664, "step": 21080 }, { "epoch": 234.27777777777777, "grad_norm": 1.2723005227144313e-07, "learning_rate": 2.2874141306996576e-05, "loss": 0.0, "num_input_tokens_seen": 5903008, "step": 21085 }, { "epoch": 234.33333333333334, "grad_norm": 3.9745629010212724e-07, "learning_rate": 2.2864359552823312e-05, "loss": 0.0, "num_input_tokens_seen": 5904432, "step": 21090 }, { "epoch": 234.38888888888889, "grad_norm": 8.070517196756555e-07, "learning_rate": 2.2854578127992648e-05, "loss": 0.0, "num_input_tokens_seen": 5905840, "step": 21095 }, { "epoch": 234.44444444444446, "grad_norm": 4.021866288894671e-07, "learning_rate": 2.2844797034012988e-05, "loss": 0.0, "num_input_tokens_seen": 5907264, "step": 21100 }, { "epoch": 234.5, "grad_norm": 1.4081983579217194e-07, "learning_rate": 2.2835016272392722e-05, "loss": 0.0, "num_input_tokens_seen": 5908640, "step": 21105 }, { "epoch": 234.55555555555554, "grad_norm": 1.2533838855688373e-07, "learning_rate": 2.2825235844640142e-05, "loss": 0.0, "num_input_tokens_seen": 5910032, "step": 21110 }, { "epoch": 234.61111111111111, "grad_norm": 1.320958489259283e-07, "learning_rate": 2.2815455752263522e-05, "loss": 0.0, "num_input_tokens_seen": 5911376, "step": 21115 }, { "epoch": 234.66666666666666, "grad_norm": 2.2853626546748274e-07, "learning_rate": 2.2805675996771092e-05, "loss": 0.0, "num_input_tokens_seen": 5912800, "step": 21120 }, { "epoch": 234.72222222222223, "grad_norm": 3.9126530282374006e-07, "learning_rate": 2.2795896579670987e-05, "loss": 0.0, "num_input_tokens_seen": 5914192, "step": 21125 }, { "epoch": 234.77777777777777, "grad_norm": 3.1997146265894116e-07, "learning_rate": 2.2786117502471337e-05, "loss": 0.0, "num_input_tokens_seen": 5915600, "step": 21130 }, { "epoch": 234.83333333333334, "grad_norm": 5.036648644818342e-07, "learning_rate": 2.2776338766680185e-05, "loss": 0.0, "num_input_tokens_seen": 5916928, "step": 21135 }, { "epoch": 234.88888888888889, "grad_norm": 9.643832044048395e-08, "learning_rate": 2.2766560373805533e-05, "loss": 0.0, "num_input_tokens_seen": 5918368, "step": 21140 }, { "epoch": 234.94444444444446, "grad_norm": 1.3994365133385145e-07, "learning_rate": 2.2756782325355353e-05, "loss": 0.0, "num_input_tokens_seen": 5919840, "step": 21145 }, { "epoch": 235.0, "grad_norm": 2.0279884438423323e-07, "learning_rate": 2.2747004622837514e-05, "loss": 0.0, "num_input_tokens_seen": 5921216, "step": 21150 }, { "epoch": 235.05555555555554, "grad_norm": 1.6732545304876112e-07, "learning_rate": 2.2737227267759878e-05, "loss": 0.0, "num_input_tokens_seen": 5922608, "step": 21155 }, { "epoch": 235.11111111111111, "grad_norm": 3.5374105777918885e-07, "learning_rate": 2.272745026163024e-05, "loss": 0.0, "num_input_tokens_seen": 5924000, "step": 21160 }, { "epoch": 235.16666666666666, "grad_norm": 2.1572942898728797e-07, "learning_rate": 2.271767360595633e-05, "loss": 0.0, "num_input_tokens_seen": 5925376, "step": 21165 }, { "epoch": 235.22222222222223, "grad_norm": 3.1320320204031304e-07, "learning_rate": 2.270789730224583e-05, "loss": 0.0, "num_input_tokens_seen": 5926784, "step": 21170 }, { "epoch": 235.27777777777777, "grad_norm": 4.0956866342867215e-08, "learning_rate": 2.2698121352006367e-05, "loss": 0.0, "num_input_tokens_seen": 5928240, "step": 21175 }, { "epoch": 235.33333333333334, "grad_norm": 6.357621487040888e-07, "learning_rate": 2.2688345756745517e-05, "loss": 0.0, "num_input_tokens_seen": 5929600, "step": 21180 }, { "epoch": 235.38888888888889, "grad_norm": 3.308268787804991e-07, "learning_rate": 2.267857051797081e-05, "loss": 0.0, "num_input_tokens_seen": 5930992, "step": 21185 }, { "epoch": 235.44444444444446, "grad_norm": 7.636305667801935e-07, "learning_rate": 2.2668795637189695e-05, "loss": 0.0, "num_input_tokens_seen": 5932368, "step": 21190 }, { "epoch": 235.5, "grad_norm": 3.1963211313268403e-06, "learning_rate": 2.2659021115909586e-05, "loss": 0.0, "num_input_tokens_seen": 5933744, "step": 21195 }, { "epoch": 235.55555555555554, "grad_norm": 1.22063568142039e-06, "learning_rate": 2.2649246955637847e-05, "loss": 0.0, "num_input_tokens_seen": 5935168, "step": 21200 }, { "epoch": 235.55555555555554, "eval_loss": 1.4442393779754639, "eval_runtime": 1.3855, "eval_samples_per_second": 28.871, "eval_steps_per_second": 14.435, "num_input_tokens_seen": 5935168, "step": 21200 }, { "epoch": 235.61111111111111, "grad_norm": 1.3635370521569712e-07, "learning_rate": 2.2639473157881766e-05, "loss": 0.0, "num_input_tokens_seen": 5936544, "step": 21205 }, { "epoch": 235.66666666666666, "grad_norm": 1.9862937961079297e-07, "learning_rate": 2.2629699724148594e-05, "loss": 0.0, "num_input_tokens_seen": 5937904, "step": 21210 }, { "epoch": 235.72222222222223, "grad_norm": 1.3137760390691255e-07, "learning_rate": 2.26199266559455e-05, "loss": 0.0, "num_input_tokens_seen": 5939296, "step": 21215 }, { "epoch": 235.77777777777777, "grad_norm": 2.6881295411840256e-07, "learning_rate": 2.2610153954779625e-05, "loss": 0.0, "num_input_tokens_seen": 5940704, "step": 21220 }, { "epoch": 235.83333333333334, "grad_norm": 5.49983906239504e-06, "learning_rate": 2.2600381622158056e-05, "loss": 0.0, "num_input_tokens_seen": 5942080, "step": 21225 }, { "epoch": 235.88888888888889, "grad_norm": 7.336913085964625e-07, "learning_rate": 2.2590609659587783e-05, "loss": 0.0, "num_input_tokens_seen": 5943536, "step": 21230 }, { "epoch": 235.94444444444446, "grad_norm": 7.319222277146764e-07, "learning_rate": 2.2580838068575787e-05, "loss": 0.0, "num_input_tokens_seen": 5944912, "step": 21235 }, { "epoch": 236.0, "grad_norm": 2.92012515501483e-07, "learning_rate": 2.257106685062896e-05, "loss": 0.0, "num_input_tokens_seen": 5946352, "step": 21240 }, { "epoch": 236.05555555555554, "grad_norm": 2.9194694661782705e-07, "learning_rate": 2.256129600725415e-05, "loss": 0.0, "num_input_tokens_seen": 5947744, "step": 21245 }, { "epoch": 236.11111111111111, "grad_norm": 1.7812898533975385e-07, "learning_rate": 2.2551525539958145e-05, "loss": 0.0, "num_input_tokens_seen": 5949136, "step": 21250 }, { "epoch": 236.16666666666666, "grad_norm": 2.1336713018627051e-07, "learning_rate": 2.2541755450247663e-05, "loss": 0.0, "num_input_tokens_seen": 5950544, "step": 21255 }, { "epoch": 236.22222222222223, "grad_norm": 2.5093410727095034e-07, "learning_rate": 2.2531985739629382e-05, "loss": 0.0, "num_input_tokens_seen": 5951968, "step": 21260 }, { "epoch": 236.27777777777777, "grad_norm": 1.3710514394915663e-06, "learning_rate": 2.2522216409609924e-05, "loss": 0.0, "num_input_tokens_seen": 5953376, "step": 21265 }, { "epoch": 236.33333333333334, "grad_norm": 5.213702820583421e-07, "learning_rate": 2.2512447461695826e-05, "loss": 0.0, "num_input_tokens_seen": 5954816, "step": 21270 }, { "epoch": 236.38888888888889, "grad_norm": 2.0255814092706714e-07, "learning_rate": 2.2502678897393593e-05, "loss": 0.0, "num_input_tokens_seen": 5956208, "step": 21275 }, { "epoch": 236.44444444444446, "grad_norm": 7.925127931684983e-08, "learning_rate": 2.2492910718209665e-05, "loss": 0.0, "num_input_tokens_seen": 5957616, "step": 21280 }, { "epoch": 236.5, "grad_norm": 4.686998806846532e-07, "learning_rate": 2.2483142925650398e-05, "loss": 0.0, "num_input_tokens_seen": 5959024, "step": 21285 }, { "epoch": 236.55555555555554, "grad_norm": 4.039529528654384e-07, "learning_rate": 2.247337552122213e-05, "loss": 0.0, "num_input_tokens_seen": 5960400, "step": 21290 }, { "epoch": 236.61111111111111, "grad_norm": 3.1807519462745404e-07, "learning_rate": 2.24636085064311e-05, "loss": 0.0, "num_input_tokens_seen": 5961840, "step": 21295 }, { "epoch": 236.66666666666666, "grad_norm": 7.115988154282604e-08, "learning_rate": 2.245384188278351e-05, "loss": 0.0, "num_input_tokens_seen": 5963232, "step": 21300 }, { "epoch": 236.72222222222223, "grad_norm": 9.68554104474606e-07, "learning_rate": 2.2444075651785513e-05, "loss": 0.0, "num_input_tokens_seen": 5964656, "step": 21305 }, { "epoch": 236.77777777777777, "grad_norm": 9.083971690415638e-07, "learning_rate": 2.243430981494316e-05, "loss": 0.0, "num_input_tokens_seen": 5966080, "step": 21310 }, { "epoch": 236.83333333333334, "grad_norm": 3.4806109283636033e-07, "learning_rate": 2.2424544373762475e-05, "loss": 0.0, "num_input_tokens_seen": 5967504, "step": 21315 }, { "epoch": 236.88888888888889, "grad_norm": 9.557974181007012e-07, "learning_rate": 2.2414779329749418e-05, "loss": 0.0, "num_input_tokens_seen": 5968896, "step": 21320 }, { "epoch": 236.94444444444446, "grad_norm": 5.052976575825596e-06, "learning_rate": 2.2405014684409873e-05, "loss": 0.0, "num_input_tokens_seen": 5970272, "step": 21325 }, { "epoch": 237.0, "grad_norm": 1.1394912036166716e-07, "learning_rate": 2.239525043924968e-05, "loss": 0.0, "num_input_tokens_seen": 5971664, "step": 21330 }, { "epoch": 237.05555555555554, "grad_norm": 1.2259280879334256e-07, "learning_rate": 2.2385486595774592e-05, "loss": 0.0, "num_input_tokens_seen": 5972992, "step": 21335 }, { "epoch": 237.11111111111111, "grad_norm": 2.1385540094343014e-07, "learning_rate": 2.237572315549033e-05, "loss": 0.0, "num_input_tokens_seen": 5974368, "step": 21340 }, { "epoch": 237.16666666666666, "grad_norm": 4.0225043562713836e-07, "learning_rate": 2.2365960119902545e-05, "loss": 0.0, "num_input_tokens_seen": 5975728, "step": 21345 }, { "epoch": 237.22222222222223, "grad_norm": 1.0622452464303933e-06, "learning_rate": 2.2356197490516806e-05, "loss": 0.0, "num_input_tokens_seen": 5977152, "step": 21350 }, { "epoch": 237.27777777777777, "grad_norm": 1.9462066802589106e-07, "learning_rate": 2.234643526883863e-05, "loss": 0.0, "num_input_tokens_seen": 5978608, "step": 21355 }, { "epoch": 237.33333333333334, "grad_norm": 6.286054485826753e-07, "learning_rate": 2.2336673456373497e-05, "loss": 0.0, "num_input_tokens_seen": 5979968, "step": 21360 }, { "epoch": 237.38888888888889, "grad_norm": 3.711715805820859e-07, "learning_rate": 2.2326912054626772e-05, "loss": 0.0, "num_input_tokens_seen": 5981376, "step": 21365 }, { "epoch": 237.44444444444446, "grad_norm": 4.223570044814551e-07, "learning_rate": 2.2317151065103813e-05, "loss": 0.0, "num_input_tokens_seen": 5982816, "step": 21370 }, { "epoch": 237.5, "grad_norm": 1.1151912815421383e-07, "learning_rate": 2.2307390489309865e-05, "loss": 0.0, "num_input_tokens_seen": 5984208, "step": 21375 }, { "epoch": 237.55555555555554, "grad_norm": 2.9622759711855906e-07, "learning_rate": 2.2297630328750146e-05, "loss": 0.0, "num_input_tokens_seen": 5985616, "step": 21380 }, { "epoch": 237.61111111111111, "grad_norm": 1.0368736411692225e-06, "learning_rate": 2.228787058492979e-05, "loss": 0.0, "num_input_tokens_seen": 5987024, "step": 21385 }, { "epoch": 237.66666666666666, "grad_norm": 1.8288513103925652e-07, "learning_rate": 2.2278111259353875e-05, "loss": 0.0, "num_input_tokens_seen": 5988448, "step": 21390 }, { "epoch": 237.72222222222223, "grad_norm": 5.37750509010948e-07, "learning_rate": 2.2268352353527395e-05, "loss": 0.0, "num_input_tokens_seen": 5989856, "step": 21395 }, { "epoch": 237.77777777777777, "grad_norm": 3.688455052497375e-08, "learning_rate": 2.225859386895533e-05, "loss": 0.0, "num_input_tokens_seen": 5991232, "step": 21400 }, { "epoch": 237.77777777777777, "eval_loss": 1.4511137008666992, "eval_runtime": 1.3867, "eval_samples_per_second": 28.845, "eval_steps_per_second": 14.422, "num_input_tokens_seen": 5991232, "step": 21400 }, { "epoch": 237.83333333333334, "grad_norm": 4.875039394391933e-07, "learning_rate": 2.2248835807142525e-05, "loss": 0.0, "num_input_tokens_seen": 5992656, "step": 21405 }, { "epoch": 237.88888888888889, "grad_norm": 7.169869320478028e-08, "learning_rate": 2.2239078169593826e-05, "loss": 0.0, "num_input_tokens_seen": 5994032, "step": 21410 }, { "epoch": 237.94444444444446, "grad_norm": 5.648157639370766e-07, "learning_rate": 2.222932095781396e-05, "loss": 0.0, "num_input_tokens_seen": 5995456, "step": 21415 }, { "epoch": 238.0, "grad_norm": 4.207609194395445e-08, "learning_rate": 2.221956417330762e-05, "loss": 0.0, "num_input_tokens_seen": 5996896, "step": 21420 }, { "epoch": 238.05555555555554, "grad_norm": 8.288333447126206e-06, "learning_rate": 2.2209807817579438e-05, "loss": 0.0, "num_input_tokens_seen": 5998288, "step": 21425 }, { "epoch": 238.11111111111111, "grad_norm": 1.766805439729069e-07, "learning_rate": 2.220005189213394e-05, "loss": 0.0, "num_input_tokens_seen": 5999712, "step": 21430 }, { "epoch": 238.16666666666666, "grad_norm": 6.568137678186758e-07, "learning_rate": 2.2190296398475624e-05, "loss": 0.0, "num_input_tokens_seen": 6001136, "step": 21435 }, { "epoch": 238.22222222222223, "grad_norm": 2.0553218860186462e-07, "learning_rate": 2.2180541338108926e-05, "loss": 0.0, "num_input_tokens_seen": 6002512, "step": 21440 }, { "epoch": 238.27777777777777, "grad_norm": 3.318526182738424e-07, "learning_rate": 2.2170786712538176e-05, "loss": 0.0, "num_input_tokens_seen": 6003888, "step": 21445 }, { "epoch": 238.33333333333334, "grad_norm": 8.074315474004834e-07, "learning_rate": 2.216103252326768e-05, "loss": 0.0, "num_input_tokens_seen": 6005280, "step": 21450 }, { "epoch": 238.38888888888889, "grad_norm": 6.150445130970184e-08, "learning_rate": 2.2151278771801635e-05, "loss": 0.0, "num_input_tokens_seen": 6006688, "step": 21455 }, { "epoch": 238.44444444444446, "grad_norm": 4.792896106664557e-07, "learning_rate": 2.21415254596442e-05, "loss": 0.0, "num_input_tokens_seen": 6008096, "step": 21460 }, { "epoch": 238.5, "grad_norm": 4.119996077633914e-08, "learning_rate": 2.213177258829947e-05, "loss": 0.0, "num_input_tokens_seen": 6009568, "step": 21465 }, { "epoch": 238.55555555555554, "grad_norm": 2.1038009379026334e-07, "learning_rate": 2.2122020159271445e-05, "loss": 0.0, "num_input_tokens_seen": 6010992, "step": 21470 }, { "epoch": 238.61111111111111, "grad_norm": 7.471787739632418e-07, "learning_rate": 2.2112268174064075e-05, "loss": 0.0, "num_input_tokens_seen": 6012384, "step": 21475 }, { "epoch": 238.66666666666666, "grad_norm": 9.035125003720168e-07, "learning_rate": 2.2102516634181253e-05, "loss": 0.0, "num_input_tokens_seen": 6013776, "step": 21480 }, { "epoch": 238.72222222222223, "grad_norm": 2.0043887616338907e-07, "learning_rate": 2.209276554112677e-05, "loss": 0.0, "num_input_tokens_seen": 6015200, "step": 21485 }, { "epoch": 238.77777777777777, "grad_norm": 3.919692517229123e-06, "learning_rate": 2.2083014896404384e-05, "loss": 0.0, "num_input_tokens_seen": 6016624, "step": 21490 }, { "epoch": 238.83333333333334, "grad_norm": 5.165816219232511e-07, "learning_rate": 2.207326470151775e-05, "loss": 0.0, "num_input_tokens_seen": 6018032, "step": 21495 }, { "epoch": 238.88888888888889, "grad_norm": 1.4131645912129898e-05, "learning_rate": 2.2063514957970477e-05, "loss": 0.0, "num_input_tokens_seen": 6019424, "step": 21500 }, { "epoch": 238.94444444444446, "grad_norm": 1.381536520739246e-07, "learning_rate": 2.205376566726611e-05, "loss": 0.0, "num_input_tokens_seen": 6020832, "step": 21505 }, { "epoch": 239.0, "grad_norm": 8.19332939272499e-08, "learning_rate": 2.204401683090809e-05, "loss": 0.0, "num_input_tokens_seen": 6022192, "step": 21510 }, { "epoch": 239.05555555555554, "grad_norm": 6.136347963092703e-08, "learning_rate": 2.203426845039982e-05, "loss": 0.0, "num_input_tokens_seen": 6023616, "step": 21515 }, { "epoch": 239.11111111111111, "grad_norm": 8.183211264167767e-08, "learning_rate": 2.202452052724464e-05, "loss": 0.0, "num_input_tokens_seen": 6025008, "step": 21520 }, { "epoch": 239.16666666666666, "grad_norm": 7.744543495391554e-08, "learning_rate": 2.2014773062945777e-05, "loss": 0.0, "num_input_tokens_seen": 6026368, "step": 21525 }, { "epoch": 239.22222222222223, "grad_norm": 5.098243960333093e-08, "learning_rate": 2.2005026059006427e-05, "loss": 0.0, "num_input_tokens_seen": 6027760, "step": 21530 }, { "epoch": 239.27777777777777, "grad_norm": 1.6015931691981677e-07, "learning_rate": 2.1995279516929695e-05, "loss": 0.0, "num_input_tokens_seen": 6029120, "step": 21535 }, { "epoch": 239.33333333333334, "grad_norm": 1.3554466704590595e-07, "learning_rate": 2.1985533438218613e-05, "loss": 0.0, "num_input_tokens_seen": 6030560, "step": 21540 }, { "epoch": 239.38888888888889, "grad_norm": 7.966964687966538e-08, "learning_rate": 2.197578782437617e-05, "loss": 0.0, "num_input_tokens_seen": 6031952, "step": 21545 }, { "epoch": 239.44444444444446, "grad_norm": 5.007461467698704e-08, "learning_rate": 2.196604267690524e-05, "loss": 0.0, "num_input_tokens_seen": 6033376, "step": 21550 }, { "epoch": 239.5, "grad_norm": 1.4965625894092227e-07, "learning_rate": 2.195629799730865e-05, "loss": 0.0, "num_input_tokens_seen": 6034768, "step": 21555 }, { "epoch": 239.55555555555554, "grad_norm": 1.4880527032801183e-07, "learning_rate": 2.1946553787089173e-05, "loss": 0.0, "num_input_tokens_seen": 6036176, "step": 21560 }, { "epoch": 239.61111111111111, "grad_norm": 1.1309962388850181e-07, "learning_rate": 2.193681004774947e-05, "loss": 0.0, "num_input_tokens_seen": 6037568, "step": 21565 }, { "epoch": 239.66666666666666, "grad_norm": 1.7291962706167396e-07, "learning_rate": 2.1927066780792154e-05, "loss": 0.0, "num_input_tokens_seen": 6038992, "step": 21570 }, { "epoch": 239.72222222222223, "grad_norm": 1.3786522856662486e-07, "learning_rate": 2.191732398771975e-05, "loss": 0.0, "num_input_tokens_seen": 6040384, "step": 21575 }, { "epoch": 239.77777777777777, "grad_norm": 8.668610718132186e-08, "learning_rate": 2.1907581670034725e-05, "loss": 0.0, "num_input_tokens_seen": 6041792, "step": 21580 }, { "epoch": 239.83333333333334, "grad_norm": 8.144236574025854e-08, "learning_rate": 2.189783982923948e-05, "loss": 0.0, "num_input_tokens_seen": 6043168, "step": 21585 }, { "epoch": 239.88888888888889, "grad_norm": 2.7096345434074465e-07, "learning_rate": 2.1888098466836303e-05, "loss": 0.0, "num_input_tokens_seen": 6044608, "step": 21590 }, { "epoch": 239.94444444444446, "grad_norm": 8.550693308961854e-08, "learning_rate": 2.1878357584327457e-05, "loss": 0.0, "num_input_tokens_seen": 6045984, "step": 21595 }, { "epoch": 240.0, "grad_norm": 1.1949586564696801e-07, "learning_rate": 2.1868617183215103e-05, "loss": 0.0, "num_input_tokens_seen": 6047376, "step": 21600 }, { "epoch": 240.0, "eval_loss": 1.4868799448013306, "eval_runtime": 1.3972, "eval_samples_per_second": 28.629, "eval_steps_per_second": 14.315, "num_input_tokens_seen": 6047376, "step": 21600 }, { "epoch": 240.05555555555554, "grad_norm": 2.4193676040340506e-07, "learning_rate": 2.1858877265001327e-05, "loss": 0.0, "num_input_tokens_seen": 6048816, "step": 21605 }, { "epoch": 240.11111111111111, "grad_norm": 1.2512568048350659e-07, "learning_rate": 2.184913783118816e-05, "loss": 0.0, "num_input_tokens_seen": 6050192, "step": 21610 }, { "epoch": 240.16666666666666, "grad_norm": 1.0043883946764254e-07, "learning_rate": 2.1839398883277522e-05, "loss": 0.0, "num_input_tokens_seen": 6051600, "step": 21615 }, { "epoch": 240.22222222222223, "grad_norm": 4.120336072332975e-08, "learning_rate": 2.182966042277129e-05, "loss": 0.0, "num_input_tokens_seen": 6052992, "step": 21620 }, { "epoch": 240.27777777777777, "grad_norm": 1.0717267429072308e-07, "learning_rate": 2.181992245117128e-05, "loss": 0.0, "num_input_tokens_seen": 6054384, "step": 21625 }, { "epoch": 240.33333333333334, "grad_norm": 1.8287957459506288e-07, "learning_rate": 2.181018496997918e-05, "loss": 0.0, "num_input_tokens_seen": 6055744, "step": 21630 }, { "epoch": 240.38888888888889, "grad_norm": 3.524711758018384e-07, "learning_rate": 2.1800447980696648e-05, "loss": 0.0, "num_input_tokens_seen": 6057184, "step": 21635 }, { "epoch": 240.44444444444446, "grad_norm": 1.1147587741788811e-07, "learning_rate": 2.1790711484825248e-05, "loss": 0.0, "num_input_tokens_seen": 6058576, "step": 21640 }, { "epoch": 240.5, "grad_norm": 1.7026322893798351e-06, "learning_rate": 2.178097548386646e-05, "loss": 0.0, "num_input_tokens_seen": 6059968, "step": 21645 }, { "epoch": 240.55555555555554, "grad_norm": 2.95141120432163e-07, "learning_rate": 2.1771239979321712e-05, "loss": 0.0, "num_input_tokens_seen": 6061424, "step": 21650 }, { "epoch": 240.61111111111111, "grad_norm": 2.9427067715914745e-07, "learning_rate": 2.1761504972692327e-05, "loss": 0.0, "num_input_tokens_seen": 6062800, "step": 21655 }, { "epoch": 240.66666666666666, "grad_norm": 6.020581508892064e-07, "learning_rate": 2.1751770465479572e-05, "loss": 0.0, "num_input_tokens_seen": 6064208, "step": 21660 }, { "epoch": 240.72222222222223, "grad_norm": 5.29731551068835e-06, "learning_rate": 2.174203645918464e-05, "loss": 0.0, "num_input_tokens_seen": 6065616, "step": 21665 }, { "epoch": 240.77777777777777, "grad_norm": 5.557010922530026e-07, "learning_rate": 2.1732302955308624e-05, "loss": 0.0, "num_input_tokens_seen": 6067008, "step": 21670 }, { "epoch": 240.83333333333334, "grad_norm": 7.086847517712158e-07, "learning_rate": 2.172256995535255e-05, "loss": 0.0, "num_input_tokens_seen": 6068400, "step": 21675 }, { "epoch": 240.88888888888889, "grad_norm": 3.2913652603383525e-07, "learning_rate": 2.171283746081739e-05, "loss": 0.0, "num_input_tokens_seen": 6069760, "step": 21680 }, { "epoch": 240.94444444444446, "grad_norm": 1.2351907230367942e-07, "learning_rate": 2.1703105473203988e-05, "loss": 0.0, "num_input_tokens_seen": 6071120, "step": 21685 }, { "epoch": 241.0, "grad_norm": 7.60244063258142e-08, "learning_rate": 2.1693373994013168e-05, "loss": 0.0, "num_input_tokens_seen": 6072512, "step": 21690 }, { "epoch": 241.05555555555554, "grad_norm": 1.1080349082703833e-07, "learning_rate": 2.168364302474562e-05, "loss": 0.0, "num_input_tokens_seen": 6073872, "step": 21695 }, { "epoch": 241.11111111111111, "grad_norm": 2.3761580791870074e-07, "learning_rate": 2.167391256690199e-05, "loss": 0.0, "num_input_tokens_seen": 6075280, "step": 21700 }, { "epoch": 241.16666666666666, "grad_norm": 3.459527988525224e-07, "learning_rate": 2.1664182621982855e-05, "loss": 0.0, "num_input_tokens_seen": 6076704, "step": 21705 }, { "epoch": 241.22222222222223, "grad_norm": 2.893385442348517e-07, "learning_rate": 2.1654453191488673e-05, "loss": 0.0, "num_input_tokens_seen": 6078080, "step": 21710 }, { "epoch": 241.27777777777777, "grad_norm": 2.3567646678657184e-07, "learning_rate": 2.1644724276919846e-05, "loss": 0.0, "num_input_tokens_seen": 6079504, "step": 21715 }, { "epoch": 241.33333333333334, "grad_norm": 3.012499689702963e-07, "learning_rate": 2.1634995879776715e-05, "loss": 0.0, "num_input_tokens_seen": 6080896, "step": 21720 }, { "epoch": 241.38888888888889, "grad_norm": 2.2385584941275738e-07, "learning_rate": 2.162526800155949e-05, "loss": 0.0, "num_input_tokens_seen": 6082304, "step": 21725 }, { "epoch": 241.44444444444446, "grad_norm": 3.4950753047269245e-07, "learning_rate": 2.1615540643768363e-05, "loss": 0.0, "num_input_tokens_seen": 6083696, "step": 21730 }, { "epoch": 241.5, "grad_norm": 3.91056033777204e-07, "learning_rate": 2.160581380790339e-05, "loss": 0.0, "num_input_tokens_seen": 6085072, "step": 21735 }, { "epoch": 241.55555555555554, "grad_norm": 2.465643262894446e-07, "learning_rate": 2.1596087495464586e-05, "loss": 0.0, "num_input_tokens_seen": 6086464, "step": 21740 }, { "epoch": 241.61111111111111, "grad_norm": 2.928816229541553e-07, "learning_rate": 2.1586361707951866e-05, "loss": 0.0, "num_input_tokens_seen": 6087808, "step": 21745 }, { "epoch": 241.66666666666666, "grad_norm": 1.3073678246655618e-06, "learning_rate": 2.157663644686507e-05, "loss": 0.0, "num_input_tokens_seen": 6089216, "step": 21750 }, { "epoch": 241.72222222222223, "grad_norm": 2.630443418638606e-07, "learning_rate": 2.156691171370396e-05, "loss": 0.0, "num_input_tokens_seen": 6090640, "step": 21755 }, { "epoch": 241.77777777777777, "grad_norm": 3.274414837051154e-07, "learning_rate": 2.1557187509968195e-05, "loss": 0.0, "num_input_tokens_seen": 6092016, "step": 21760 }, { "epoch": 241.83333333333334, "grad_norm": 5.66183643968543e-07, "learning_rate": 2.1547463837157382e-05, "loss": 0.0, "num_input_tokens_seen": 6093456, "step": 21765 }, { "epoch": 241.88888888888889, "grad_norm": 4.5496079792428645e-07, "learning_rate": 2.1537740696771045e-05, "loss": 0.0, "num_input_tokens_seen": 6094880, "step": 21770 }, { "epoch": 241.94444444444446, "grad_norm": 3.567613475752296e-07, "learning_rate": 2.1528018090308587e-05, "loss": 0.0, "num_input_tokens_seen": 6096352, "step": 21775 }, { "epoch": 242.0, "grad_norm": 2.8935585305589484e-07, "learning_rate": 2.151829601926938e-05, "loss": 0.0, "num_input_tokens_seen": 6097712, "step": 21780 }, { "epoch": 242.05555555555554, "grad_norm": 1.0490395396800523e-07, "learning_rate": 2.1508574485152684e-05, "loss": 0.0, "num_input_tokens_seen": 6099072, "step": 21785 }, { "epoch": 242.11111111111111, "grad_norm": 7.706439220100947e-08, "learning_rate": 2.1498853489457667e-05, "loss": 0.0, "num_input_tokens_seen": 6100496, "step": 21790 }, { "epoch": 242.16666666666666, "grad_norm": 2.793308908621839e-07, "learning_rate": 2.1489133033683455e-05, "loss": 0.0, "num_input_tokens_seen": 6101888, "step": 21795 }, { "epoch": 242.22222222222223, "grad_norm": 1.3486157968145562e-07, "learning_rate": 2.1479413119329038e-05, "loss": 0.0, "num_input_tokens_seen": 6103328, "step": 21800 }, { "epoch": 242.22222222222223, "eval_loss": 1.5167688131332397, "eval_runtime": 1.3865, "eval_samples_per_second": 28.849, "eval_steps_per_second": 14.424, "num_input_tokens_seen": 6103328, "step": 21800 }, { "epoch": 242.27777777777777, "grad_norm": 8.567105425072441e-08, "learning_rate": 2.1469693747893355e-05, "loss": 0.0, "num_input_tokens_seen": 6104704, "step": 21805 }, { "epoch": 242.33333333333334, "grad_norm": 7.513096988986945e-08, "learning_rate": 2.1459974920875274e-05, "loss": 0.0, "num_input_tokens_seen": 6106144, "step": 21810 }, { "epoch": 242.38888888888889, "grad_norm": 1.5440495815255417e-07, "learning_rate": 2.145025663977354e-05, "loss": 0.0, "num_input_tokens_seen": 6107552, "step": 21815 }, { "epoch": 242.44444444444446, "grad_norm": 2.675568282484164e-07, "learning_rate": 2.1440538906086844e-05, "loss": 0.0, "num_input_tokens_seen": 6108912, "step": 21820 }, { "epoch": 242.5, "grad_norm": 6.180564469104866e-07, "learning_rate": 2.1430821721313782e-05, "loss": 0.0, "num_input_tokens_seen": 6110320, "step": 21825 }, { "epoch": 242.55555555555554, "grad_norm": 2.1086803769776452e-07, "learning_rate": 2.142110508695286e-05, "loss": 0.0, "num_input_tokens_seen": 6111760, "step": 21830 }, { "epoch": 242.61111111111111, "grad_norm": 3.16955436119315e-07, "learning_rate": 2.1411389004502515e-05, "loss": 0.0, "num_input_tokens_seen": 6113152, "step": 21835 }, { "epoch": 242.66666666666666, "grad_norm": 2.5226987077076046e-07, "learning_rate": 2.140167347546107e-05, "loss": 0.0, "num_input_tokens_seen": 6114512, "step": 21840 }, { "epoch": 242.72222222222223, "grad_norm": 2.7199305918657046e-07, "learning_rate": 2.1391958501326793e-05, "loss": 0.0, "num_input_tokens_seen": 6115888, "step": 21845 }, { "epoch": 242.77777777777777, "grad_norm": 3.2067550392866906e-08, "learning_rate": 2.1382244083597873e-05, "loss": 0.0, "num_input_tokens_seen": 6117296, "step": 21850 }, { "epoch": 242.83333333333334, "grad_norm": 7.375405743914598e-07, "learning_rate": 2.137253022377237e-05, "loss": 0.0, "num_input_tokens_seen": 6118688, "step": 21855 }, { "epoch": 242.88888888888889, "grad_norm": 5.905239390813222e-07, "learning_rate": 2.136281692334829e-05, "loss": 0.0, "num_input_tokens_seen": 6120096, "step": 21860 }, { "epoch": 242.94444444444446, "grad_norm": 7.775596344572477e-08, "learning_rate": 2.135310418382356e-05, "loss": 0.0, "num_input_tokens_seen": 6121504, "step": 21865 }, { "epoch": 243.0, "grad_norm": 6.23291327883635e-07, "learning_rate": 2.134339200669598e-05, "loss": 0.0, "num_input_tokens_seen": 6122896, "step": 21870 }, { "epoch": 243.05555555555554, "grad_norm": 2.4165893819372286e-07, "learning_rate": 2.133368039346331e-05, "loss": 0.0, "num_input_tokens_seen": 6124304, "step": 21875 }, { "epoch": 243.11111111111111, "grad_norm": 3.628089757512498e-07, "learning_rate": 2.1323969345623195e-05, "loss": 0.0, "num_input_tokens_seen": 6125744, "step": 21880 }, { "epoch": 243.16666666666666, "grad_norm": 1.7321220013855054e-07, "learning_rate": 2.1314258864673207e-05, "loss": 0.0, "num_input_tokens_seen": 6127168, "step": 21885 }, { "epoch": 243.22222222222223, "grad_norm": 1.4728830421972816e-07, "learning_rate": 2.130454895211082e-05, "loss": 0.0, "num_input_tokens_seen": 6128544, "step": 21890 }, { "epoch": 243.27777777777777, "grad_norm": 3.7350865511598386e-08, "learning_rate": 2.129483960943342e-05, "loss": 0.0, "num_input_tokens_seen": 6129936, "step": 21895 }, { "epoch": 243.33333333333334, "grad_norm": 1.6679408645359217e-06, "learning_rate": 2.128513083813831e-05, "loss": 0.0, "num_input_tokens_seen": 6131376, "step": 21900 }, { "epoch": 243.38888888888889, "grad_norm": 8.428358455603302e-07, "learning_rate": 2.1275422639722724e-05, "loss": 0.0, "num_input_tokens_seen": 6132784, "step": 21905 }, { "epoch": 243.44444444444446, "grad_norm": 4.206742687529186e-06, "learning_rate": 2.126571501568376e-05, "loss": 0.0, "num_input_tokens_seen": 6134176, "step": 21910 }, { "epoch": 243.5, "grad_norm": 1.2430600691004656e-06, "learning_rate": 2.1256007967518478e-05, "loss": 0.0, "num_input_tokens_seen": 6135568, "step": 21915 }, { "epoch": 243.55555555555554, "grad_norm": 3.129262609036232e-07, "learning_rate": 2.124630149672381e-05, "loss": 0.0, "num_input_tokens_seen": 6136960, "step": 21920 }, { "epoch": 243.61111111111111, "grad_norm": 1.3014177113745973e-07, "learning_rate": 2.1236595604796624e-05, "loss": 0.0, "num_input_tokens_seen": 6138336, "step": 21925 }, { "epoch": 243.66666666666666, "grad_norm": 9.188894978251483e-08, "learning_rate": 2.1226890293233693e-05, "loss": 0.0, "num_input_tokens_seen": 6139744, "step": 21930 }, { "epoch": 243.72222222222223, "grad_norm": 2.158668621632387e-07, "learning_rate": 2.1217185563531694e-05, "loss": 0.0, "num_input_tokens_seen": 6141168, "step": 21935 }, { "epoch": 243.77777777777777, "grad_norm": 3.2002120065044437e-07, "learning_rate": 2.120748141718721e-05, "loss": 0.0, "num_input_tokens_seen": 6142592, "step": 21940 }, { "epoch": 243.83333333333334, "grad_norm": 4.117764973443627e-08, "learning_rate": 2.1197777855696765e-05, "loss": 0.0, "num_input_tokens_seen": 6143984, "step": 21945 }, { "epoch": 243.88888888888889, "grad_norm": 2.6710105771599046e-07, "learning_rate": 2.1188074880556746e-05, "loss": 0.0, "num_input_tokens_seen": 6145392, "step": 21950 }, { "epoch": 243.94444444444446, "grad_norm": 3.9899188664094254e-07, "learning_rate": 2.1178372493263495e-05, "loss": 0.0, "num_input_tokens_seen": 6146800, "step": 21955 }, { "epoch": 244.0, "grad_norm": 2.0093339969662338e-07, "learning_rate": 2.116867069531322e-05, "loss": 0.0, "num_input_tokens_seen": 6148176, "step": 21960 }, { "epoch": 244.05555555555554, "grad_norm": 3.29223809103496e-07, "learning_rate": 2.1158969488202073e-05, "loss": 0.0, "num_input_tokens_seen": 6149616, "step": 21965 }, { "epoch": 244.11111111111111, "grad_norm": 9.249824017842911e-08, "learning_rate": 2.114926887342611e-05, "loss": 0.0, "num_input_tokens_seen": 6150992, "step": 21970 }, { "epoch": 244.16666666666666, "grad_norm": 1.1453884241063861e-07, "learning_rate": 2.113956885248127e-05, "loss": 0.0, "num_input_tokens_seen": 6152384, "step": 21975 }, { "epoch": 244.22222222222223, "grad_norm": 1.5835310023248894e-06, "learning_rate": 2.112986942686342e-05, "loss": 0.0, "num_input_tokens_seen": 6153792, "step": 21980 }, { "epoch": 244.27777777777777, "grad_norm": 7.265074941642524e-07, "learning_rate": 2.112017059806835e-05, "loss": 0.0, "num_input_tokens_seen": 6155200, "step": 21985 }, { "epoch": 244.33333333333334, "grad_norm": 2.434791497307742e-07, "learning_rate": 2.1110472367591724e-05, "loss": 0.0, "num_input_tokens_seen": 6156576, "step": 21990 }, { "epoch": 244.38888888888889, "grad_norm": 4.228396619510022e-07, "learning_rate": 2.1100774736929145e-05, "loss": 0.0, "num_input_tokens_seen": 6158000, "step": 21995 }, { "epoch": 244.44444444444446, "grad_norm": 1.6345325093425345e-07, "learning_rate": 2.10910777075761e-05, "loss": 0.0, "num_input_tokens_seen": 6159376, "step": 22000 }, { "epoch": 244.44444444444446, "eval_loss": 1.4850764274597168, "eval_runtime": 1.3848, "eval_samples_per_second": 28.885, "eval_steps_per_second": 14.442, "num_input_tokens_seen": 6159376, "step": 22000 }, { "epoch": 244.5, "grad_norm": 6.942794783526551e-08, "learning_rate": 2.108138128102799e-05, "loss": 0.0, "num_input_tokens_seen": 6160800, "step": 22005 }, { "epoch": 244.55555555555554, "grad_norm": 5.434667116333003e-08, "learning_rate": 2.107168545878014e-05, "loss": 0.0, "num_input_tokens_seen": 6162192, "step": 22010 }, { "epoch": 244.61111111111111, "grad_norm": 4.3646929270835244e-07, "learning_rate": 2.106199024232775e-05, "loss": 0.0, "num_input_tokens_seen": 6163600, "step": 22015 }, { "epoch": 244.66666666666666, "grad_norm": 3.478369023923733e-07, "learning_rate": 2.105229563316595e-05, "loss": 0.0, "num_input_tokens_seen": 6165024, "step": 22020 }, { "epoch": 244.72222222222223, "grad_norm": 1.2673997673573467e-07, "learning_rate": 2.1042601632789784e-05, "loss": 0.0, "num_input_tokens_seen": 6166432, "step": 22025 }, { "epoch": 244.77777777777777, "grad_norm": 1.2770560431363265e-07, "learning_rate": 2.103290824269417e-05, "loss": 0.0, "num_input_tokens_seen": 6167824, "step": 22030 }, { "epoch": 244.83333333333334, "grad_norm": 1.4097538780788454e-07, "learning_rate": 2.1023215464373965e-05, "loss": 0.0, "num_input_tokens_seen": 6169184, "step": 22035 }, { "epoch": 244.88888888888889, "grad_norm": 3.280187570453563e-07, "learning_rate": 2.1013523299323908e-05, "loss": 0.0, "num_input_tokens_seen": 6170592, "step": 22040 }, { "epoch": 244.94444444444446, "grad_norm": 4.049790902627137e-07, "learning_rate": 2.1003831749038654e-05, "loss": 0.0, "num_input_tokens_seen": 6171984, "step": 22045 }, { "epoch": 245.0, "grad_norm": 2.408887951332872e-07, "learning_rate": 2.099414081501277e-05, "loss": 0.0, "num_input_tokens_seen": 6173376, "step": 22050 }, { "epoch": 245.05555555555554, "grad_norm": 8.472218837596301e-08, "learning_rate": 2.09844504987407e-05, "loss": 0.0, "num_input_tokens_seen": 6174752, "step": 22055 }, { "epoch": 245.11111111111111, "grad_norm": 4.443285561706034e-08, "learning_rate": 2.097476080171683e-05, "loss": 0.0, "num_input_tokens_seen": 6176128, "step": 22060 }, { "epoch": 245.16666666666666, "grad_norm": 7.787613753862388e-08, "learning_rate": 2.0965071725435436e-05, "loss": 0.0, "num_input_tokens_seen": 6177504, "step": 22065 }, { "epoch": 245.22222222222223, "grad_norm": 1.2527281967322779e-07, "learning_rate": 2.0955383271390684e-05, "loss": 0.0, "num_input_tokens_seen": 6178944, "step": 22070 }, { "epoch": 245.27777777777777, "grad_norm": 2.7452514927972516e-07, "learning_rate": 2.094569544107666e-05, "loss": 0.0, "num_input_tokens_seen": 6180352, "step": 22075 }, { "epoch": 245.33333333333334, "grad_norm": 1.9789835903338826e-07, "learning_rate": 2.093600823598735e-05, "loss": 0.0, "num_input_tokens_seen": 6181680, "step": 22080 }, { "epoch": 245.38888888888889, "grad_norm": 3.5251835583949287e-07, "learning_rate": 2.092632165761663e-05, "loss": 0.0, "num_input_tokens_seen": 6183088, "step": 22085 }, { "epoch": 245.44444444444446, "grad_norm": 5.168258212506771e-07, "learning_rate": 2.091663570745832e-05, "loss": 0.0, "num_input_tokens_seen": 6184464, "step": 22090 }, { "epoch": 245.5, "grad_norm": 3.1426039640791714e-05, "learning_rate": 2.0906950387006086e-05, "loss": 0.0, "num_input_tokens_seen": 6185808, "step": 22095 }, { "epoch": 245.55555555555554, "grad_norm": 1.120863203141198e-06, "learning_rate": 2.0897265697753543e-05, "loss": 0.0, "num_input_tokens_seen": 6187216, "step": 22100 }, { "epoch": 245.61111111111111, "grad_norm": 7.698521358179278e-07, "learning_rate": 2.088758164119419e-05, "loss": 0.0, "num_input_tokens_seen": 6188592, "step": 22105 }, { "epoch": 245.66666666666666, "grad_norm": 8.493918812746415e-08, "learning_rate": 2.0877898218821428e-05, "loss": 0.0, "num_input_tokens_seen": 6190016, "step": 22110 }, { "epoch": 245.72222222222223, "grad_norm": 7.46813313412531e-08, "learning_rate": 2.0868215432128565e-05, "loss": 0.0, "num_input_tokens_seen": 6191408, "step": 22115 }, { "epoch": 245.77777777777777, "grad_norm": 5.245479997029179e-07, "learning_rate": 2.0858533282608796e-05, "loss": 0.0, "num_input_tokens_seen": 6192832, "step": 22120 }, { "epoch": 245.83333333333334, "grad_norm": 8.030235676415032e-08, "learning_rate": 2.084885177175524e-05, "loss": 0.0, "num_input_tokens_seen": 6194272, "step": 22125 }, { "epoch": 245.88888888888889, "grad_norm": 5.299131444758132e-08, "learning_rate": 2.0839170901060917e-05, "loss": 0.0, "num_input_tokens_seen": 6195696, "step": 22130 }, { "epoch": 245.94444444444446, "grad_norm": 2.6751752812970153e-08, "learning_rate": 2.082949067201872e-05, "loss": 0.0, "num_input_tokens_seen": 6197104, "step": 22135 }, { "epoch": 246.0, "grad_norm": 6.49815561359901e-08, "learning_rate": 2.0819811086121475e-05, "loss": 0.0, "num_input_tokens_seen": 6198512, "step": 22140 }, { "epoch": 246.05555555555554, "grad_norm": 4.035315015471497e-08, "learning_rate": 2.08101321448619e-05, "loss": 0.0, "num_input_tokens_seen": 6199888, "step": 22145 }, { "epoch": 246.11111111111111, "grad_norm": 3.778173507384963e-08, "learning_rate": 2.080045384973259e-05, "loss": 0.0, "num_input_tokens_seen": 6201296, "step": 22150 }, { "epoch": 246.16666666666666, "grad_norm": 1.1834242741315393e-07, "learning_rate": 2.0790776202226082e-05, "loss": 0.0, "num_input_tokens_seen": 6202688, "step": 22155 }, { "epoch": 246.22222222222223, "grad_norm": 2.9515118171730137e-07, "learning_rate": 2.078109920383477e-05, "loss": 0.0, "num_input_tokens_seen": 6204048, "step": 22160 }, { "epoch": 246.27777777777777, "grad_norm": 1.2020981898785976e-07, "learning_rate": 2.0771422856050978e-05, "loss": 0.0, "num_input_tokens_seen": 6205488, "step": 22165 }, { "epoch": 246.33333333333334, "grad_norm": 8.21570367293134e-08, "learning_rate": 2.076174716036693e-05, "loss": 0.0, "num_input_tokens_seen": 6206912, "step": 22170 }, { "epoch": 246.38888888888889, "grad_norm": 8.518431116044667e-08, "learning_rate": 2.075207211827472e-05, "loss": 0.0, "num_input_tokens_seen": 6208304, "step": 22175 }, { "epoch": 246.44444444444446, "grad_norm": 1.7697279020012502e-07, "learning_rate": 2.074239773126638e-05, "loss": 0.0, "num_input_tokens_seen": 6209728, "step": 22180 }, { "epoch": 246.5, "grad_norm": 1.9140617268931237e-07, "learning_rate": 2.073272400083382e-05, "loss": 0.0, "num_input_tokens_seen": 6211168, "step": 22185 }, { "epoch": 246.55555555555554, "grad_norm": 4.508193995889087e-08, "learning_rate": 2.072305092846883e-05, "loss": 0.0, "num_input_tokens_seen": 6212560, "step": 22190 }, { "epoch": 246.61111111111111, "grad_norm": 5.645712519708468e-08, "learning_rate": 2.0713378515663152e-05, "loss": 0.0, "num_input_tokens_seen": 6213968, "step": 22195 }, { "epoch": 246.66666666666666, "grad_norm": 4.6178186607903626e-08, "learning_rate": 2.070370676390836e-05, "loss": 0.0, "num_input_tokens_seen": 6215360, "step": 22200 }, { "epoch": 246.66666666666666, "eval_loss": 1.5619878768920898, "eval_runtime": 1.387, "eval_samples_per_second": 28.839, "eval_steps_per_second": 14.419, "num_input_tokens_seen": 6215360, "step": 22200 }, { "epoch": 246.72222222222223, "grad_norm": 7.106504540388414e-08, "learning_rate": 2.0694035674695974e-05, "loss": 0.0, "num_input_tokens_seen": 6216784, "step": 22205 }, { "epoch": 246.77777777777777, "grad_norm": 6.993318635295509e-08, "learning_rate": 2.0684365249517416e-05, "loss": 0.0, "num_input_tokens_seen": 6218192, "step": 22210 }, { "epoch": 246.83333333333334, "grad_norm": 5.928243851371917e-08, "learning_rate": 2.067469548986396e-05, "loss": 0.0, "num_input_tokens_seen": 6219616, "step": 22215 }, { "epoch": 246.88888888888889, "grad_norm": 2.3698075892752968e-06, "learning_rate": 2.066502639722681e-05, "loss": 0.0, "num_input_tokens_seen": 6220976, "step": 22220 }, { "epoch": 246.94444444444446, "grad_norm": 4.312887824653444e-08, "learning_rate": 2.065535797309708e-05, "loss": 0.0, "num_input_tokens_seen": 6222352, "step": 22225 }, { "epoch": 247.0, "grad_norm": 2.3080775690687005e-07, "learning_rate": 2.0645690218965736e-05, "loss": 0.0, "num_input_tokens_seen": 6223728, "step": 22230 }, { "epoch": 247.05555555555554, "grad_norm": 1.3109276153500105e-07, "learning_rate": 2.063602313632369e-05, "loss": 0.0, "num_input_tokens_seen": 6225104, "step": 22235 }, { "epoch": 247.11111111111111, "grad_norm": 1.671404419312239e-07, "learning_rate": 2.0626356726661704e-05, "loss": 0.0, "num_input_tokens_seen": 6226464, "step": 22240 }, { "epoch": 247.16666666666666, "grad_norm": 4.2396049337867225e-08, "learning_rate": 2.0616690991470477e-05, "loss": 0.0, "num_input_tokens_seen": 6227840, "step": 22245 }, { "epoch": 247.22222222222223, "grad_norm": 8.193173073323123e-08, "learning_rate": 2.0607025932240595e-05, "loss": 0.0, "num_input_tokens_seen": 6229232, "step": 22250 }, { "epoch": 247.27777777777777, "grad_norm": 6.676923476334196e-07, "learning_rate": 2.059736155046251e-05, "loss": 0.0, "num_input_tokens_seen": 6230624, "step": 22255 }, { "epoch": 247.33333333333334, "grad_norm": 5.83471937432023e-08, "learning_rate": 2.0587697847626603e-05, "loss": 0.0, "num_input_tokens_seen": 6232032, "step": 22260 }, { "epoch": 247.38888888888889, "grad_norm": 9.155074565114774e-08, "learning_rate": 2.057803482522314e-05, "loss": 0.0, "num_input_tokens_seen": 6233424, "step": 22265 }, { "epoch": 247.44444444444446, "grad_norm": 2.2264205767896783e-07, "learning_rate": 2.056837248474227e-05, "loss": 0.0, "num_input_tokens_seen": 6234800, "step": 22270 }, { "epoch": 247.5, "grad_norm": 1.993205955841404e-07, "learning_rate": 2.0558710827674064e-05, "loss": 0.0, "num_input_tokens_seen": 6236208, "step": 22275 }, { "epoch": 247.55555555555554, "grad_norm": 3.6792229707316437e-07, "learning_rate": 2.054904985550845e-05, "loss": 0.0, "num_input_tokens_seen": 6237632, "step": 22280 }, { "epoch": 247.61111111111111, "grad_norm": 1.2030139373564452e-07, "learning_rate": 2.0539389569735287e-05, "loss": 0.0, "num_input_tokens_seen": 6239072, "step": 22285 }, { "epoch": 247.66666666666666, "grad_norm": 7.341849141084822e-06, "learning_rate": 2.052972997184431e-05, "loss": 0.0, "num_input_tokens_seen": 6240464, "step": 22290 }, { "epoch": 247.72222222222223, "grad_norm": 3.020994938651711e-07, "learning_rate": 2.0520071063325146e-05, "loss": 0.0, "num_input_tokens_seen": 6241840, "step": 22295 }, { "epoch": 247.77777777777777, "grad_norm": 3.385311586612261e-08, "learning_rate": 2.051041284566732e-05, "loss": 0.0, "num_input_tokens_seen": 6243280, "step": 22300 }, { "epoch": 247.83333333333334, "grad_norm": 2.0641053311010182e-08, "learning_rate": 2.050075532036026e-05, "loss": 0.0, "num_input_tokens_seen": 6244704, "step": 22305 }, { "epoch": 247.88888888888889, "grad_norm": 2.331307314307196e-06, "learning_rate": 2.0491098488893264e-05, "loss": 0.0, "num_input_tokens_seen": 6246128, "step": 22310 }, { "epoch": 247.94444444444446, "grad_norm": 4.3302343044615554e-08, "learning_rate": 2.0481442352755546e-05, "loss": 0.0, "num_input_tokens_seen": 6247504, "step": 22315 }, { "epoch": 248.0, "grad_norm": 1.9372428994302027e-08, "learning_rate": 2.0471786913436198e-05, "loss": 0.0, "num_input_tokens_seen": 6248896, "step": 22320 }, { "epoch": 248.05555555555554, "grad_norm": 1.2904936852464743e-07, "learning_rate": 2.0462132172424218e-05, "loss": 0.0, "num_input_tokens_seen": 6250288, "step": 22325 }, { "epoch": 248.11111111111111, "grad_norm": 5.544166370441417e-08, "learning_rate": 2.0452478131208484e-05, "loss": 0.0, "num_input_tokens_seen": 6251696, "step": 22330 }, { "epoch": 248.16666666666666, "grad_norm": 2.695985870104778e-07, "learning_rate": 2.0442824791277765e-05, "loss": 0.0, "num_input_tokens_seen": 6253056, "step": 22335 }, { "epoch": 248.22222222222223, "grad_norm": 4.056696312204622e-08, "learning_rate": 2.0433172154120727e-05, "loss": 0.0, "num_input_tokens_seen": 6254464, "step": 22340 }, { "epoch": 248.27777777777777, "grad_norm": 4.690870625267962e-08, "learning_rate": 2.0423520221225947e-05, "loss": 0.0, "num_input_tokens_seen": 6255872, "step": 22345 }, { "epoch": 248.33333333333334, "grad_norm": 1.2172137076049694e-07, "learning_rate": 2.0413868994081848e-05, "loss": 0.0, "num_input_tokens_seen": 6257248, "step": 22350 }, { "epoch": 248.38888888888889, "grad_norm": 6.649617745324576e-08, "learning_rate": 2.0404218474176795e-05, "loss": 0.0, "num_input_tokens_seen": 6258608, "step": 22355 }, { "epoch": 248.44444444444446, "grad_norm": 1.7532713059154048e-07, "learning_rate": 2.0394568662999002e-05, "loss": 0.0, "num_input_tokens_seen": 6260000, "step": 22360 }, { "epoch": 248.5, "grad_norm": 7.533758861200113e-08, "learning_rate": 2.0384919562036593e-05, "loss": 0.0, "num_input_tokens_seen": 6261440, "step": 22365 }, { "epoch": 248.55555555555554, "grad_norm": 7.62720802072181e-08, "learning_rate": 2.0375271172777593e-05, "loss": 0.0, "num_input_tokens_seen": 6262896, "step": 22370 }, { "epoch": 248.61111111111111, "grad_norm": 5.434147709593162e-08, "learning_rate": 2.0365623496709885e-05, "loss": 0.0, "num_input_tokens_seen": 6264272, "step": 22375 }, { "epoch": 248.66666666666666, "grad_norm": 2.764825808299065e-07, "learning_rate": 2.0355976535321283e-05, "loss": 0.0, "num_input_tokens_seen": 6265648, "step": 22380 }, { "epoch": 248.72222222222223, "grad_norm": 3.665007852760027e-07, "learning_rate": 2.034633029009945e-05, "loss": 0.0, "num_input_tokens_seen": 6267040, "step": 22385 }, { "epoch": 248.77777777777777, "grad_norm": 3.713531171456452e-08, "learning_rate": 2.0336684762531972e-05, "loss": 0.0, "num_input_tokens_seen": 6268432, "step": 22390 }, { "epoch": 248.83333333333334, "grad_norm": 2.3288363593110262e-07, "learning_rate": 2.032703995410631e-05, "loss": 0.0, "num_input_tokens_seen": 6269824, "step": 22395 }, { "epoch": 248.88888888888889, "grad_norm": 5.114656005389406e-07, "learning_rate": 2.031739586630981e-05, "loss": 0.0, "num_input_tokens_seen": 6271232, "step": 22400 }, { "epoch": 248.88888888888889, "eval_loss": 1.526442289352417, "eval_runtime": 1.4041, "eval_samples_per_second": 28.487, "eval_steps_per_second": 14.244, "num_input_tokens_seen": 6271232, "step": 22400 }, { "epoch": 248.94444444444446, "grad_norm": 5.999203835926892e-07, "learning_rate": 2.0307752500629707e-05, "loss": 0.0, "num_input_tokens_seen": 6272656, "step": 22405 }, { "epoch": 249.0, "grad_norm": 2.928622393483238e-07, "learning_rate": 2.0298109858553144e-05, "loss": 0.0, "num_input_tokens_seen": 6274048, "step": 22410 }, { "epoch": 249.05555555555554, "grad_norm": 4.367420842754655e-06, "learning_rate": 2.028846794156712e-05, "loss": 0.0, "num_input_tokens_seen": 6275504, "step": 22415 }, { "epoch": 249.11111111111111, "grad_norm": 1.3288060074501118e-07, "learning_rate": 2.027882675115856e-05, "loss": 0.0, "num_input_tokens_seen": 6276880, "step": 22420 }, { "epoch": 249.16666666666666, "grad_norm": 1.0426148122633094e-07, "learning_rate": 2.026918628881423e-05, "loss": 0.0, "num_input_tokens_seen": 6278288, "step": 22425 }, { "epoch": 249.22222222222223, "grad_norm": 1.199488650627245e-07, "learning_rate": 2.0259546556020833e-05, "loss": 0.0, "num_input_tokens_seen": 6279680, "step": 22430 }, { "epoch": 249.27777777777777, "grad_norm": 5.670083851327945e-07, "learning_rate": 2.024990755426493e-05, "loss": 0.0, "num_input_tokens_seen": 6281088, "step": 22435 }, { "epoch": 249.33333333333334, "grad_norm": 7.453060391071631e-08, "learning_rate": 2.0240269285032975e-05, "loss": 0.0, "num_input_tokens_seen": 6282480, "step": 22440 }, { "epoch": 249.38888888888889, "grad_norm": 3.557137233656249e-07, "learning_rate": 2.0230631749811306e-05, "loss": 0.0, "num_input_tokens_seen": 6283872, "step": 22445 }, { "epoch": 249.44444444444446, "grad_norm": 1.2160275275618915e-07, "learning_rate": 2.0220994950086162e-05, "loss": 0.0, "num_input_tokens_seen": 6285232, "step": 22450 }, { "epoch": 249.5, "grad_norm": 8.358618686088448e-08, "learning_rate": 2.021135888734365e-05, "loss": 0.0, "num_input_tokens_seen": 6286592, "step": 22455 }, { "epoch": 249.55555555555554, "grad_norm": 6.122441931211142e-08, "learning_rate": 2.0201723563069783e-05, "loss": 0.0, "num_input_tokens_seen": 6288000, "step": 22460 }, { "epoch": 249.61111111111111, "grad_norm": 1.15203738459968e-06, "learning_rate": 2.0192088978750433e-05, "loss": 0.0, "num_input_tokens_seen": 6289392, "step": 22465 }, { "epoch": 249.66666666666666, "grad_norm": 6.185898655530764e-07, "learning_rate": 2.0182455135871385e-05, "loss": 0.0, "num_input_tokens_seen": 6290800, "step": 22470 }, { "epoch": 249.72222222222223, "grad_norm": 2.6706177891355765e-07, "learning_rate": 2.0172822035918305e-05, "loss": 0.0, "num_input_tokens_seen": 6292176, "step": 22475 }, { "epoch": 249.77777777777777, "grad_norm": 2.2325961879232636e-07, "learning_rate": 2.016318968037671e-05, "loss": 0.0, "num_input_tokens_seen": 6293552, "step": 22480 }, { "epoch": 249.83333333333334, "grad_norm": 1.6355673437828955e-07, "learning_rate": 2.015355807073206e-05, "loss": 0.0, "num_input_tokens_seen": 6294944, "step": 22485 }, { "epoch": 249.88888888888889, "grad_norm": 1.0236310998834597e-07, "learning_rate": 2.0143927208469664e-05, "loss": 0.0, "num_input_tokens_seen": 6296384, "step": 22490 }, { "epoch": 249.94444444444446, "grad_norm": 1.4690735383737774e-07, "learning_rate": 2.0134297095074708e-05, "loss": 0.0, "num_input_tokens_seen": 6297808, "step": 22495 }, { "epoch": 250.0, "grad_norm": 1.7438934207802959e-07, "learning_rate": 2.0124667732032297e-05, "loss": 0.0, "num_input_tokens_seen": 6299216, "step": 22500 }, { "epoch": 250.05555555555554, "grad_norm": 1.2028858975554613e-07, "learning_rate": 2.011503912082738e-05, "loss": 0.0, "num_input_tokens_seen": 6300656, "step": 22505 }, { "epoch": 250.11111111111111, "grad_norm": 8.312940735777374e-08, "learning_rate": 2.0105411262944823e-05, "loss": 0.0, "num_input_tokens_seen": 6302080, "step": 22510 }, { "epoch": 250.16666666666666, "grad_norm": 2.9549991609201243e-07, "learning_rate": 2.0095784159869366e-05, "loss": 0.0, "num_input_tokens_seen": 6303456, "step": 22515 }, { "epoch": 250.22222222222223, "grad_norm": 1.522654201835394e-06, "learning_rate": 2.0086157813085608e-05, "loss": 0.0, "num_input_tokens_seen": 6304816, "step": 22520 }, { "epoch": 250.27777777777777, "grad_norm": 2.0407114220688527e-07, "learning_rate": 2.0076532224078068e-05, "loss": 0.0, "num_input_tokens_seen": 6306192, "step": 22525 }, { "epoch": 250.33333333333334, "grad_norm": 3.174419020979258e-07, "learning_rate": 2.0066907394331142e-05, "loss": 0.0, "num_input_tokens_seen": 6307568, "step": 22530 }, { "epoch": 250.38888888888889, "grad_norm": 1.0428910854898277e-06, "learning_rate": 2.0057283325329077e-05, "loss": 0.0, "num_input_tokens_seen": 6308944, "step": 22535 }, { "epoch": 250.44444444444446, "grad_norm": 1.375027181893529e-06, "learning_rate": 2.0047660018556047e-05, "loss": 0.0, "num_input_tokens_seen": 6310352, "step": 22540 }, { "epoch": 250.5, "grad_norm": 4.508478014031425e-06, "learning_rate": 2.0038037475496075e-05, "loss": 0.0, "num_input_tokens_seen": 6311712, "step": 22545 }, { "epoch": 250.55555555555554, "grad_norm": 8.070511512414669e-07, "learning_rate": 2.0028415697633073e-05, "loss": 0.0, "num_input_tokens_seen": 6313184, "step": 22550 }, { "epoch": 250.61111111111111, "grad_norm": 3.552577609866603e-08, "learning_rate": 2.0018794686450858e-05, "loss": 0.0, "num_input_tokens_seen": 6314560, "step": 22555 }, { "epoch": 250.66666666666666, "grad_norm": 1.8048656613700587e-07, "learning_rate": 2.0009174443433088e-05, "loss": 0.0, "num_input_tokens_seen": 6315984, "step": 22560 }, { "epoch": 250.72222222222223, "grad_norm": 1.3038014401445253e-07, "learning_rate": 1.999955497006334e-05, "loss": 0.0, "num_input_tokens_seen": 6317360, "step": 22565 }, { "epoch": 250.77777777777777, "grad_norm": 9.781147269904977e-08, "learning_rate": 1.9989936267825067e-05, "loss": 0.0, "num_input_tokens_seen": 6318768, "step": 22570 }, { "epoch": 250.83333333333334, "grad_norm": 1.2445967456642393e-07, "learning_rate": 1.9980318338201572e-05, "loss": 0.0, "num_input_tokens_seen": 6320144, "step": 22575 }, { "epoch": 250.88888888888889, "grad_norm": 1.1781520470321993e-07, "learning_rate": 1.997070118267607e-05, "loss": 0.0, "num_input_tokens_seen": 6321568, "step": 22580 }, { "epoch": 250.94444444444446, "grad_norm": 1.3287262845551595e-07, "learning_rate": 1.9961084802731654e-05, "loss": 0.0, "num_input_tokens_seen": 6322960, "step": 22585 }, { "epoch": 251.0, "grad_norm": 3.8395450019379496e-07, "learning_rate": 1.9951469199851273e-05, "loss": 0.0, "num_input_tokens_seen": 6324320, "step": 22590 }, { "epoch": 251.05555555555554, "grad_norm": 1.2526118098321604e-07, "learning_rate": 1.99418543755178e-05, "loss": 0.0, "num_input_tokens_seen": 6325760, "step": 22595 }, { "epoch": 251.11111111111111, "grad_norm": 3.112051274456462e-07, "learning_rate": 1.9932240331213936e-05, "loss": 0.0, "num_input_tokens_seen": 6327136, "step": 22600 }, { "epoch": 251.11111111111111, "eval_loss": 1.6043952703475952, "eval_runtime": 1.3855, "eval_samples_per_second": 28.87, "eval_steps_per_second": 14.435, "num_input_tokens_seen": 6327136, "step": 22600 }, { "epoch": 251.16666666666666, "grad_norm": 1.2868433429957804e-07, "learning_rate": 1.9922627068422297e-05, "loss": 0.0, "num_input_tokens_seen": 6328512, "step": 22605 }, { "epoch": 251.22222222222223, "grad_norm": 4.10170315490177e-08, "learning_rate": 1.991301458862538e-05, "loss": 0.0, "num_input_tokens_seen": 6329904, "step": 22610 }, { "epoch": 251.27777777777777, "grad_norm": 1.4395817515833187e-07, "learning_rate": 1.9903402893305536e-05, "loss": 0.0, "num_input_tokens_seen": 6331344, "step": 22615 }, { "epoch": 251.33333333333334, "grad_norm": 5.458698737470513e-08, "learning_rate": 1.9893791983945016e-05, "loss": 0.0, "num_input_tokens_seen": 6332768, "step": 22620 }, { "epoch": 251.38888888888889, "grad_norm": 1.1144057765477555e-07, "learning_rate": 1.988418186202594e-05, "loss": 0.0, "num_input_tokens_seen": 6334192, "step": 22625 }, { "epoch": 251.44444444444446, "grad_norm": 7.342641197283228e-07, "learning_rate": 1.98745725290303e-05, "loss": 0.0, "num_input_tokens_seen": 6335568, "step": 22630 }, { "epoch": 251.5, "grad_norm": 2.330127273353355e-07, "learning_rate": 1.986496398644e-05, "loss": 0.0, "num_input_tokens_seen": 6336976, "step": 22635 }, { "epoch": 251.55555555555554, "grad_norm": 2.970699028992385e-07, "learning_rate": 1.9855356235736777e-05, "loss": 0.0, "num_input_tokens_seen": 6338384, "step": 22640 }, { "epoch": 251.61111111111111, "grad_norm": 4.834582867374593e-08, "learning_rate": 1.9845749278402277e-05, "loss": 0.0, "num_input_tokens_seen": 6339792, "step": 22645 }, { "epoch": 251.66666666666666, "grad_norm": 1.942795080367432e-07, "learning_rate": 1.9836143115918006e-05, "loss": 0.0, "num_input_tokens_seen": 6341184, "step": 22650 }, { "epoch": 251.72222222222223, "grad_norm": 2.1709804798319965e-07, "learning_rate": 1.9826537749765367e-05, "loss": 0.0, "num_input_tokens_seen": 6342608, "step": 22655 }, { "epoch": 251.77777777777777, "grad_norm": 3.983476304370015e-08, "learning_rate": 1.9816933181425625e-05, "loss": 0.0, "num_input_tokens_seen": 6344048, "step": 22660 }, { "epoch": 251.83333333333334, "grad_norm": 4.878911113337381e-06, "learning_rate": 1.9807329412379903e-05, "loss": 0.0, "num_input_tokens_seen": 6345472, "step": 22665 }, { "epoch": 251.88888888888889, "grad_norm": 2.152020357470974e-07, "learning_rate": 1.9797726444109247e-05, "loss": 0.0, "num_input_tokens_seen": 6346896, "step": 22670 }, { "epoch": 251.94444444444446, "grad_norm": 6.624069470717586e-08, "learning_rate": 1.9788124278094557e-05, "loss": 0.0, "num_input_tokens_seen": 6348272, "step": 22675 }, { "epoch": 252.0, "grad_norm": 1.0302550208507455e-06, "learning_rate": 1.9778522915816594e-05, "loss": 0.0, "num_input_tokens_seen": 6349664, "step": 22680 }, { "epoch": 252.05555555555554, "grad_norm": 5.549155730477651e-07, "learning_rate": 1.9768922358756014e-05, "loss": 0.0, "num_input_tokens_seen": 6351056, "step": 22685 }, { "epoch": 252.11111111111111, "grad_norm": 5.26827534486074e-07, "learning_rate": 1.9759322608393353e-05, "loss": 0.0, "num_input_tokens_seen": 6352464, "step": 22690 }, { "epoch": 252.16666666666666, "grad_norm": 3.814414242242492e-07, "learning_rate": 1.9749723666208992e-05, "loss": 0.0, "num_input_tokens_seen": 6353840, "step": 22695 }, { "epoch": 252.22222222222223, "grad_norm": 6.461273756031005e-07, "learning_rate": 1.9740125533683235e-05, "loss": 0.0, "num_input_tokens_seen": 6355232, "step": 22700 }, { "epoch": 252.27777777777777, "grad_norm": 2.746986069723789e-07, "learning_rate": 1.9730528212296208e-05, "loss": 0.0, "num_input_tokens_seen": 6356672, "step": 22705 }, { "epoch": 252.33333333333334, "grad_norm": 4.7064901309568086e-07, "learning_rate": 1.9720931703527945e-05, "loss": 0.0, "num_input_tokens_seen": 6358096, "step": 22710 }, { "epoch": 252.38888888888889, "grad_norm": 4.0811482904246077e-07, "learning_rate": 1.9711336008858373e-05, "loss": 0.0, "num_input_tokens_seen": 6359520, "step": 22715 }, { "epoch": 252.44444444444446, "grad_norm": 3.227976037578628e-07, "learning_rate": 1.9701741129767233e-05, "loss": 0.0, "num_input_tokens_seen": 6360928, "step": 22720 }, { "epoch": 252.5, "grad_norm": 1.3644579155425163e-07, "learning_rate": 1.9692147067734202e-05, "loss": 0.0, "num_input_tokens_seen": 6362352, "step": 22725 }, { "epoch": 252.55555555555554, "grad_norm": 8.97483062090032e-07, "learning_rate": 1.96825538242388e-05, "loss": 0.0, "num_input_tokens_seen": 6363712, "step": 22730 }, { "epoch": 252.61111111111111, "grad_norm": 1.9583862354011217e-07, "learning_rate": 1.967296140076041e-05, "loss": 0.0, "num_input_tokens_seen": 6365136, "step": 22735 }, { "epoch": 252.66666666666666, "grad_norm": 1.7089645609758009e-07, "learning_rate": 1.966336979877833e-05, "loss": 0.0, "num_input_tokens_seen": 6366528, "step": 22740 }, { "epoch": 252.72222222222223, "grad_norm": 3.8049216755098314e-07, "learning_rate": 1.9653779019771678e-05, "loss": 0.0, "num_input_tokens_seen": 6367968, "step": 22745 }, { "epoch": 252.77777777777777, "grad_norm": 3.180215912834683e-07, "learning_rate": 1.9644189065219488e-05, "loss": 0.0, "num_input_tokens_seen": 6369360, "step": 22750 }, { "epoch": 252.83333333333334, "grad_norm": 3.0027050001990574e-07, "learning_rate": 1.9634599936600655e-05, "loss": 0.0, "num_input_tokens_seen": 6370752, "step": 22755 }, { "epoch": 252.88888888888889, "grad_norm": 4.3745541233874974e-07, "learning_rate": 1.9625011635393935e-05, "loss": 0.0, "num_input_tokens_seen": 6372128, "step": 22760 }, { "epoch": 252.94444444444446, "grad_norm": 1.730253188725328e-06, "learning_rate": 1.9615424163077963e-05, "loss": 0.0, "num_input_tokens_seen": 6373488, "step": 22765 }, { "epoch": 253.0, "grad_norm": 1.0640731318289909e-07, "learning_rate": 1.9605837521131263e-05, "loss": 0.0, "num_input_tokens_seen": 6374912, "step": 22770 }, { "epoch": 253.05555555555554, "grad_norm": 4.997141900275892e-07, "learning_rate": 1.9596251711032192e-05, "loss": 0.0, "num_input_tokens_seen": 6376352, "step": 22775 }, { "epoch": 253.11111111111111, "grad_norm": 3.9616236335859867e-07, "learning_rate": 1.958666673425903e-05, "loss": 0.0, "num_input_tokens_seen": 6377776, "step": 22780 }, { "epoch": 253.16666666666666, "grad_norm": 9.412752888238174e-08, "learning_rate": 1.957708259228987e-05, "loss": 0.0, "num_input_tokens_seen": 6379152, "step": 22785 }, { "epoch": 253.22222222222223, "grad_norm": 4.353949861979345e-06, "learning_rate": 1.956749928660273e-05, "loss": 0.0, "num_input_tokens_seen": 6380528, "step": 22790 }, { "epoch": 253.27777777777777, "grad_norm": 4.981002348358743e-07, "learning_rate": 1.955791681867547e-05, "loss": 0.0, "num_input_tokens_seen": 6381840, "step": 22795 }, { "epoch": 253.33333333333334, "grad_norm": 6.773690301997704e-07, "learning_rate": 1.9548335189985824e-05, "loss": 0.0, "num_input_tokens_seen": 6383248, "step": 22800 }, { "epoch": 253.33333333333334, "eval_loss": 1.5395723581314087, "eval_runtime": 1.4002, "eval_samples_per_second": 28.567, "eval_steps_per_second": 14.283, "num_input_tokens_seen": 6383248, "step": 22800 }, { "epoch": 253.38888888888889, "grad_norm": 3.5219116512053006e-07, "learning_rate": 1.9538754402011396e-05, "loss": 0.0, "num_input_tokens_seen": 6384608, "step": 22805 }, { "epoch": 253.44444444444446, "grad_norm": 4.552474308638921e-07, "learning_rate": 1.952917445622968e-05, "loss": 0.0, "num_input_tokens_seen": 6385968, "step": 22810 }, { "epoch": 253.5, "grad_norm": 9.601532156011672e-07, "learning_rate": 1.9519595354118005e-05, "loss": 0.0, "num_input_tokens_seen": 6387376, "step": 22815 }, { "epoch": 253.55555555555554, "grad_norm": 5.711308403988369e-07, "learning_rate": 1.951001709715361e-05, "loss": 0.0, "num_input_tokens_seen": 6388800, "step": 22820 }, { "epoch": 253.61111111111111, "grad_norm": 1.1083918138865556e-07, "learning_rate": 1.9500439686813556e-05, "loss": 0.0, "num_input_tokens_seen": 6390176, "step": 22825 }, { "epoch": 253.66666666666666, "grad_norm": 2.595232047042373e-07, "learning_rate": 1.949086312457482e-05, "loss": 0.0, "num_input_tokens_seen": 6391568, "step": 22830 }, { "epoch": 253.72222222222223, "grad_norm": 1.0080467518491787e-06, "learning_rate": 1.9481287411914223e-05, "loss": 0.0, "num_input_tokens_seen": 6392976, "step": 22835 }, { "epoch": 253.77777777777777, "grad_norm": 4.0118533206623397e-07, "learning_rate": 1.9471712550308457e-05, "loss": 0.0, "num_input_tokens_seen": 6394400, "step": 22840 }, { "epoch": 253.83333333333334, "grad_norm": 4.807995992450742e-07, "learning_rate": 1.946213854123409e-05, "loss": 0.0, "num_input_tokens_seen": 6395840, "step": 22845 }, { "epoch": 253.88888888888889, "grad_norm": 3.7233155580906896e-07, "learning_rate": 1.9452565386167554e-05, "loss": 0.0, "num_input_tokens_seen": 6397232, "step": 22850 }, { "epoch": 253.94444444444446, "grad_norm": 8.00945326773217e-06, "learning_rate": 1.9442993086585142e-05, "loss": 0.0, "num_input_tokens_seen": 6398624, "step": 22855 }, { "epoch": 254.0, "grad_norm": 1.1268760857774396e-07, "learning_rate": 1.9433421643963043e-05, "loss": 0.0, "num_input_tokens_seen": 6400048, "step": 22860 }, { "epoch": 254.05555555555554, "grad_norm": 1.09009349102962e-07, "learning_rate": 1.942385105977727e-05, "loss": 0.0, "num_input_tokens_seen": 6401504, "step": 22865 }, { "epoch": 254.11111111111111, "grad_norm": 2.683912896372931e-07, "learning_rate": 1.9414281335503743e-05, "loss": 0.0, "num_input_tokens_seen": 6402896, "step": 22870 }, { "epoch": 254.16666666666666, "grad_norm": 4.463191771719721e-07, "learning_rate": 1.9404712472618232e-05, "loss": 0.0, "num_input_tokens_seen": 6404288, "step": 22875 }, { "epoch": 254.22222222222223, "grad_norm": 8.749854600864637e-07, "learning_rate": 1.939514447259636e-05, "loss": 0.0, "num_input_tokens_seen": 6405664, "step": 22880 }, { "epoch": 254.27777777777777, "grad_norm": 1.3658075658895541e-05, "learning_rate": 1.938557733691365e-05, "loss": 0.0, "num_input_tokens_seen": 6407072, "step": 22885 }, { "epoch": 254.33333333333334, "grad_norm": 4.0259445199808397e-07, "learning_rate": 1.9376011067045476e-05, "loss": 0.0, "num_input_tokens_seen": 6408496, "step": 22890 }, { "epoch": 254.38888888888889, "grad_norm": 2.925151676436144e-08, "learning_rate": 1.9366445664467065e-05, "loss": 0.0, "num_input_tokens_seen": 6409888, "step": 22895 }, { "epoch": 254.44444444444446, "grad_norm": 9.497902908606193e-08, "learning_rate": 1.9356881130653533e-05, "loss": 0.0, "num_input_tokens_seen": 6411280, "step": 22900 }, { "epoch": 254.5, "grad_norm": 9.059259298282996e-08, "learning_rate": 1.9347317467079846e-05, "loss": 0.0, "num_input_tokens_seen": 6412672, "step": 22905 }, { "epoch": 254.55555555555554, "grad_norm": 2.322775003449351e-07, "learning_rate": 1.9337754675220836e-05, "loss": 0.0, "num_input_tokens_seen": 6414048, "step": 22910 }, { "epoch": 254.61111111111111, "grad_norm": 4.059003089196267e-08, "learning_rate": 1.9328192756551218e-05, "loss": 0.0, "num_input_tokens_seen": 6415456, "step": 22915 }, { "epoch": 254.66666666666666, "grad_norm": 5.012292092487769e-08, "learning_rate": 1.931863171254555e-05, "loss": 0.0, "num_input_tokens_seen": 6416832, "step": 22920 }, { "epoch": 254.72222222222223, "grad_norm": 3.5406927167969116e-07, "learning_rate": 1.930907154467826e-05, "loss": 0.0, "num_input_tokens_seen": 6418224, "step": 22925 }, { "epoch": 254.77777777777777, "grad_norm": 3.187849273444954e-08, "learning_rate": 1.9299512254423673e-05, "loss": 0.0, "num_input_tokens_seen": 6419632, "step": 22930 }, { "epoch": 254.83333333333334, "grad_norm": 3.649091979696095e-07, "learning_rate": 1.9289953843255914e-05, "loss": 0.0, "num_input_tokens_seen": 6421088, "step": 22935 }, { "epoch": 254.88888888888889, "grad_norm": 1.584615461069916e-07, "learning_rate": 1.9280396312649048e-05, "loss": 0.0, "num_input_tokens_seen": 6422496, "step": 22940 }, { "epoch": 254.94444444444446, "grad_norm": 1.063031902504008e-07, "learning_rate": 1.9270839664076936e-05, "loss": 0.0, "num_input_tokens_seen": 6423824, "step": 22945 }, { "epoch": 255.0, "grad_norm": 1.7484013881130522e-07, "learning_rate": 1.9261283899013345e-05, "loss": 0.0, "num_input_tokens_seen": 6425200, "step": 22950 }, { "epoch": 255.05555555555554, "grad_norm": 4.055480573583736e-08, "learning_rate": 1.92517290189319e-05, "loss": 0.0, "num_input_tokens_seen": 6426640, "step": 22955 }, { "epoch": 255.11111111111111, "grad_norm": 1.0178797538173967e-07, "learning_rate": 1.924217502530607e-05, "loss": 0.0, "num_input_tokens_seen": 6428032, "step": 22960 }, { "epoch": 255.16666666666666, "grad_norm": 4.0753153740524795e-08, "learning_rate": 1.9232621919609207e-05, "loss": 0.0, "num_input_tokens_seen": 6429440, "step": 22965 }, { "epoch": 255.22222222222223, "grad_norm": 8.253531547097737e-08, "learning_rate": 1.9223069703314534e-05, "loss": 0.0, "num_input_tokens_seen": 6430816, "step": 22970 }, { "epoch": 255.27777777777777, "grad_norm": 1.4912689039192628e-07, "learning_rate": 1.92135183778951e-05, "loss": 0.0, "num_input_tokens_seen": 6432144, "step": 22975 }, { "epoch": 255.33333333333334, "grad_norm": 2.5898546596181404e-07, "learning_rate": 1.9203967944823857e-05, "loss": 0.0, "num_input_tokens_seen": 6433520, "step": 22980 }, { "epoch": 255.38888888888889, "grad_norm": 5.0913151028453285e-08, "learning_rate": 1.9194418405573588e-05, "loss": 0.0, "num_input_tokens_seen": 6434976, "step": 22985 }, { "epoch": 255.44444444444446, "grad_norm": 1.1597102655969138e-07, "learning_rate": 1.9184869761616954e-05, "loss": 0.0, "num_input_tokens_seen": 6436384, "step": 22990 }, { "epoch": 255.5, "grad_norm": 1.92743058846645e-07, "learning_rate": 1.9175322014426495e-05, "loss": 0.0, "num_input_tokens_seen": 6437792, "step": 22995 }, { "epoch": 255.55555555555554, "grad_norm": 5.2976567133100616e-08, "learning_rate": 1.9165775165474565e-05, "loss": 0.0, "num_input_tokens_seen": 6439168, "step": 23000 }, { "epoch": 255.55555555555554, "eval_loss": 1.5609865188598633, "eval_runtime": 1.3914, "eval_samples_per_second": 28.747, "eval_steps_per_second": 14.374, "num_input_tokens_seen": 6439168, "step": 23000 }, { "epoch": 255.61111111111111, "grad_norm": 1.2462618315112195e-07, "learning_rate": 1.9156229216233434e-05, "loss": 0.0, "num_input_tokens_seen": 6440576, "step": 23005 }, { "epoch": 255.66666666666666, "grad_norm": 1.3878199922601198e-07, "learning_rate": 1.9146684168175184e-05, "loss": 0.0, "num_input_tokens_seen": 6441968, "step": 23010 }, { "epoch": 255.72222222222223, "grad_norm": 1.8823234881892859e-07, "learning_rate": 1.9137140022771796e-05, "loss": 0.0, "num_input_tokens_seen": 6443408, "step": 23015 }, { "epoch": 255.77777777777777, "grad_norm": 2.8104361149416945e-07, "learning_rate": 1.9127596781495103e-05, "loss": 0.0, "num_input_tokens_seen": 6444768, "step": 23020 }, { "epoch": 255.83333333333334, "grad_norm": 9.874083417571455e-08, "learning_rate": 1.9118054445816767e-05, "loss": 0.0, "num_input_tokens_seen": 6446160, "step": 23025 }, { "epoch": 255.88888888888889, "grad_norm": 1.6819475945339946e-07, "learning_rate": 1.9108513017208356e-05, "loss": 0.0, "num_input_tokens_seen": 6447600, "step": 23030 }, { "epoch": 255.94444444444446, "grad_norm": 7.961057235661428e-08, "learning_rate": 1.9098972497141287e-05, "loss": 0.0, "num_input_tokens_seen": 6449024, "step": 23035 }, { "epoch": 256.0, "grad_norm": 1.2741971033847221e-07, "learning_rate": 1.9089432887086806e-05, "loss": 0.0, "num_input_tokens_seen": 6450432, "step": 23040 }, { "epoch": 256.05555555555554, "grad_norm": 8.625013236951418e-08, "learning_rate": 1.9079894188516056e-05, "loss": 0.0, "num_input_tokens_seen": 6451856, "step": 23045 }, { "epoch": 256.1111111111111, "grad_norm": 3.314350394134635e-08, "learning_rate": 1.907035640290002e-05, "loss": 0.0, "num_input_tokens_seen": 6453280, "step": 23050 }, { "epoch": 256.1666666666667, "grad_norm": 3.4358080824858916e-07, "learning_rate": 1.9060819531709534e-05, "loss": 0.0, "num_input_tokens_seen": 6454688, "step": 23055 }, { "epoch": 256.22222222222223, "grad_norm": 2.543748678363045e-07, "learning_rate": 1.9051283576415325e-05, "loss": 0.0, "num_input_tokens_seen": 6456080, "step": 23060 }, { "epoch": 256.27777777777777, "grad_norm": 1.9446368071385223e-07, "learning_rate": 1.904174853848793e-05, "loss": 0.0, "num_input_tokens_seen": 6457536, "step": 23065 }, { "epoch": 256.3333333333333, "grad_norm": 2.2556312728738703e-07, "learning_rate": 1.903221441939779e-05, "loss": 0.0, "num_input_tokens_seen": 6458928, "step": 23070 }, { "epoch": 256.3888888888889, "grad_norm": 1.0532518501804589e-07, "learning_rate": 1.9022681220615194e-05, "loss": 0.0, "num_input_tokens_seen": 6460336, "step": 23075 }, { "epoch": 256.44444444444446, "grad_norm": 4.249893379437708e-07, "learning_rate": 1.9013148943610255e-05, "loss": 0.0, "num_input_tokens_seen": 6461744, "step": 23080 }, { "epoch": 256.5, "grad_norm": 3.9515583694083034e-07, "learning_rate": 1.9003617589852998e-05, "loss": 0.0, "num_input_tokens_seen": 6463184, "step": 23085 }, { "epoch": 256.55555555555554, "grad_norm": 6.794565479140147e-07, "learning_rate": 1.899408716081326e-05, "loss": 0.0, "num_input_tokens_seen": 6464544, "step": 23090 }, { "epoch": 256.6111111111111, "grad_norm": 3.387339120308752e-07, "learning_rate": 1.898455765796075e-05, "loss": 0.0, "num_input_tokens_seen": 6465936, "step": 23095 }, { "epoch": 256.6666666666667, "grad_norm": 1.2040032970617176e-07, "learning_rate": 1.8975029082765053e-05, "loss": 0.0, "num_input_tokens_seen": 6467328, "step": 23100 }, { "epoch": 256.72222222222223, "grad_norm": 2.163420163014962e-07, "learning_rate": 1.8965501436695577e-05, "loss": 0.0, "num_input_tokens_seen": 6468704, "step": 23105 }, { "epoch": 256.77777777777777, "grad_norm": 3.874768026435049e-06, "learning_rate": 1.895597472122161e-05, "loss": 0.0, "num_input_tokens_seen": 6470096, "step": 23110 }, { "epoch": 256.8333333333333, "grad_norm": 7.888770028330327e-07, "learning_rate": 1.894644893781231e-05, "loss": 0.0, "num_input_tokens_seen": 6471488, "step": 23115 }, { "epoch": 256.8888888888889, "grad_norm": 5.102298814563255e-07, "learning_rate": 1.893692408793665e-05, "loss": 0.0, "num_input_tokens_seen": 6472864, "step": 23120 }, { "epoch": 256.94444444444446, "grad_norm": 8.159059206036545e-08, "learning_rate": 1.8927400173063493e-05, "loss": 0.0, "num_input_tokens_seen": 6474272, "step": 23125 }, { "epoch": 257.0, "grad_norm": 1.393321440446016e-07, "learning_rate": 1.891787719466154e-05, "loss": 0.0, "num_input_tokens_seen": 6475680, "step": 23130 }, { "epoch": 257.05555555555554, "grad_norm": 6.622580599469075e-07, "learning_rate": 1.8908355154199346e-05, "loss": 0.0, "num_input_tokens_seen": 6477088, "step": 23135 }, { "epoch": 257.1111111111111, "grad_norm": 5.730739758291747e-07, "learning_rate": 1.8898834053145357e-05, "loss": 0.0, "num_input_tokens_seen": 6478496, "step": 23140 }, { "epoch": 257.1666666666667, "grad_norm": 1.7922823758453887e-07, "learning_rate": 1.8889313892967813e-05, "loss": 0.0, "num_input_tokens_seen": 6479904, "step": 23145 }, { "epoch": 257.22222222222223, "grad_norm": 1.0836235730948829e-07, "learning_rate": 1.8879794675134863e-05, "loss": 0.0, "num_input_tokens_seen": 6481280, "step": 23150 }, { "epoch": 257.27777777777777, "grad_norm": 2.4149804289663734e-07, "learning_rate": 1.8870276401114494e-05, "loss": 0.0, "num_input_tokens_seen": 6482688, "step": 23155 }, { "epoch": 257.3333333333333, "grad_norm": 6.140327855064243e-07, "learning_rate": 1.886075907237453e-05, "loss": 0.0, "num_input_tokens_seen": 6484096, "step": 23160 }, { "epoch": 257.3888888888889, "grad_norm": 1.8807087087679975e-07, "learning_rate": 1.8851242690382672e-05, "loss": 0.0, "num_input_tokens_seen": 6485504, "step": 23165 }, { "epoch": 257.44444444444446, "grad_norm": 1.4658660347777186e-06, "learning_rate": 1.884172725660645e-05, "loss": 0.0, "num_input_tokens_seen": 6486928, "step": 23170 }, { "epoch": 257.5, "grad_norm": 2.3778602553647943e-07, "learning_rate": 1.8832212772513277e-05, "loss": 0.0, "num_input_tokens_seen": 6488288, "step": 23175 }, { "epoch": 257.55555555555554, "grad_norm": 4.2453058313185466e-07, "learning_rate": 1.8822699239570414e-05, "loss": 0.0, "num_input_tokens_seen": 6489664, "step": 23180 }, { "epoch": 257.6111111111111, "grad_norm": 1.5470303083020553e-07, "learning_rate": 1.8813186659244943e-05, "loss": 0.0, "num_input_tokens_seen": 6491088, "step": 23185 }, { "epoch": 257.6666666666667, "grad_norm": 1.613256017662934e-07, "learning_rate": 1.880367503300385e-05, "loss": 0.0, "num_input_tokens_seen": 6492464, "step": 23190 }, { "epoch": 257.72222222222223, "grad_norm": 3.961706909194618e-07, "learning_rate": 1.8794164362313927e-05, "loss": 0.0, "num_input_tokens_seen": 6493840, "step": 23195 }, { "epoch": 257.77777777777777, "grad_norm": 1.9040853658225387e-06, "learning_rate": 1.878465464864185e-05, "loss": 0.0, "num_input_tokens_seen": 6495280, "step": 23200 }, { "epoch": 257.77777777777777, "eval_loss": 1.5008690357208252, "eval_runtime": 1.3846, "eval_samples_per_second": 28.889, "eval_steps_per_second": 14.445, "num_input_tokens_seen": 6495280, "step": 23200 }, { "epoch": 257.8333333333333, "grad_norm": 6.205777935974766e-07, "learning_rate": 1.877514589345414e-05, "loss": 0.0, "num_input_tokens_seen": 6496720, "step": 23205 }, { "epoch": 257.8888888888889, "grad_norm": 9.571222108206712e-07, "learning_rate": 1.876563809821715e-05, "loss": 0.0, "num_input_tokens_seen": 6498144, "step": 23210 }, { "epoch": 257.94444444444446, "grad_norm": 4.351912821221049e-07, "learning_rate": 1.8756131264397106e-05, "loss": 0.0, "num_input_tokens_seen": 6499504, "step": 23215 }, { "epoch": 258.0, "grad_norm": 3.926036242773989e-07, "learning_rate": 1.87466253934601e-05, "loss": 0.0, "num_input_tokens_seen": 6500912, "step": 23220 }, { "epoch": 258.05555555555554, "grad_norm": 7.455400350409036e-07, "learning_rate": 1.8737120486872033e-05, "loss": 0.0, "num_input_tokens_seen": 6502272, "step": 23225 }, { "epoch": 258.1111111111111, "grad_norm": 7.1140084401122294e-06, "learning_rate": 1.8727616546098696e-05, "loss": 0.0, "num_input_tokens_seen": 6503632, "step": 23230 }, { "epoch": 258.1666666666667, "grad_norm": 5.1767283792969465e-08, "learning_rate": 1.8718113572605716e-05, "loss": 0.0, "num_input_tokens_seen": 6505040, "step": 23235 }, { "epoch": 258.22222222222223, "grad_norm": 7.014224223667043e-08, "learning_rate": 1.8708611567858554e-05, "loss": 0.0, "num_input_tokens_seen": 6506448, "step": 23240 }, { "epoch": 258.27777777777777, "grad_norm": 6.770438432113224e-08, "learning_rate": 1.8699110533322565e-05, "loss": 0.0, "num_input_tokens_seen": 6507888, "step": 23245 }, { "epoch": 258.3333333333333, "grad_norm": 7.948607816388176e-08, "learning_rate": 1.8689610470462897e-05, "loss": 0.0, "num_input_tokens_seen": 6509328, "step": 23250 }, { "epoch": 258.3888888888889, "grad_norm": 2.2618890227477095e-07, "learning_rate": 1.8680111380744604e-05, "loss": 0.0, "num_input_tokens_seen": 6510704, "step": 23255 }, { "epoch": 258.44444444444446, "grad_norm": 4.5707938767236556e-08, "learning_rate": 1.8670613265632564e-05, "loss": 0.0, "num_input_tokens_seen": 6512096, "step": 23260 }, { "epoch": 258.5, "grad_norm": 5.4419860617827e-08, "learning_rate": 1.866111612659149e-05, "loss": 0.0, "num_input_tokens_seen": 6513456, "step": 23265 }, { "epoch": 258.55555555555554, "grad_norm": 3.036814746337768e-07, "learning_rate": 1.8651619965085967e-05, "loss": 0.0, "num_input_tokens_seen": 6514896, "step": 23270 }, { "epoch": 258.6111111111111, "grad_norm": 3.7218473636357885e-08, "learning_rate": 1.8642124782580433e-05, "loss": 0.0, "num_input_tokens_seen": 6516336, "step": 23275 }, { "epoch": 258.6666666666667, "grad_norm": 5.778768752406904e-08, "learning_rate": 1.8632630580539144e-05, "loss": 0.0, "num_input_tokens_seen": 6517744, "step": 23280 }, { "epoch": 258.72222222222223, "grad_norm": 2.6475629510969156e-07, "learning_rate": 1.862313736042625e-05, "loss": 0.0, "num_input_tokens_seen": 6519120, "step": 23285 }, { "epoch": 258.77777777777777, "grad_norm": 6.98948738886429e-08, "learning_rate": 1.8613645123705703e-05, "loss": 0.0, "num_input_tokens_seen": 6520512, "step": 23290 }, { "epoch": 258.8333333333333, "grad_norm": 5.5844392221615635e-08, "learning_rate": 1.8604153871841328e-05, "loss": 0.0, "num_input_tokens_seen": 6521904, "step": 23295 }, { "epoch": 258.8888888888889, "grad_norm": 4.6631615902015255e-08, "learning_rate": 1.859466360629682e-05, "loss": 0.0, "num_input_tokens_seen": 6523296, "step": 23300 }, { "epoch": 258.94444444444446, "grad_norm": 3.72177844099042e-08, "learning_rate": 1.8585174328535666e-05, "loss": 0.0, "num_input_tokens_seen": 6524672, "step": 23305 }, { "epoch": 259.0, "grad_norm": 1.1027859869727763e-07, "learning_rate": 1.857568604002124e-05, "loss": 0.0, "num_input_tokens_seen": 6526112, "step": 23310 }, { "epoch": 259.05555555555554, "grad_norm": 1.2256847981007013e-07, "learning_rate": 1.8566198742216774e-05, "loss": 0.0, "num_input_tokens_seen": 6527456, "step": 23315 }, { "epoch": 259.1111111111111, "grad_norm": 4.2665107002903824e-08, "learning_rate": 1.85567124365853e-05, "loss": 0.0, "num_input_tokens_seen": 6528848, "step": 23320 }, { "epoch": 259.1666666666667, "grad_norm": 4.4830080980773346e-08, "learning_rate": 1.854722712458975e-05, "loss": 0.0, "num_input_tokens_seen": 6530288, "step": 23325 }, { "epoch": 259.22222222222223, "grad_norm": 1.672678564546004e-07, "learning_rate": 1.853774280769286e-05, "loss": 0.0, "num_input_tokens_seen": 6531712, "step": 23330 }, { "epoch": 259.27777777777777, "grad_norm": 6.069170126465906e-07, "learning_rate": 1.852825948735724e-05, "loss": 0.0, "num_input_tokens_seen": 6533104, "step": 23335 }, { "epoch": 259.3333333333333, "grad_norm": 1.1585159853666482e-07, "learning_rate": 1.851877716504534e-05, "loss": 0.0, "num_input_tokens_seen": 6534480, "step": 23340 }, { "epoch": 259.3888888888889, "grad_norm": 3.151857583816309e-07, "learning_rate": 1.8509295842219448e-05, "loss": 0.0, "num_input_tokens_seen": 6535872, "step": 23345 }, { "epoch": 259.44444444444446, "grad_norm": 1.7240877525637188e-07, "learning_rate": 1.8499815520341697e-05, "loss": 0.0, "num_input_tokens_seen": 6537264, "step": 23350 }, { "epoch": 259.5, "grad_norm": 8.770201702645863e-07, "learning_rate": 1.8490336200874094e-05, "loss": 0.0, "num_input_tokens_seen": 6538688, "step": 23355 }, { "epoch": 259.55555555555554, "grad_norm": 3.4597564990690444e-07, "learning_rate": 1.848085788527844e-05, "loss": 0.0, "num_input_tokens_seen": 6540096, "step": 23360 }, { "epoch": 259.6111111111111, "grad_norm": 5.5871343107583016e-08, "learning_rate": 1.847138057501644e-05, "loss": 0.0, "num_input_tokens_seen": 6541504, "step": 23365 }, { "epoch": 259.6666666666667, "grad_norm": 4.575805121476151e-07, "learning_rate": 1.8461904271549582e-05, "loss": 0.0, "num_input_tokens_seen": 6542912, "step": 23370 }, { "epoch": 259.72222222222223, "grad_norm": 2.0305184023072798e-07, "learning_rate": 1.845242897633926e-05, "loss": 0.0, "num_input_tokens_seen": 6544336, "step": 23375 }, { "epoch": 259.77777777777777, "grad_norm": 2.812664092743944e-07, "learning_rate": 1.844295469084667e-05, "loss": 0.0, "num_input_tokens_seen": 6545696, "step": 23380 }, { "epoch": 259.8333333333333, "grad_norm": 5.713122277484217e-07, "learning_rate": 1.843348141653286e-05, "loss": 0.0, "num_input_tokens_seen": 6547024, "step": 23385 }, { "epoch": 259.8888888888889, "grad_norm": 4.760164813433221e-07, "learning_rate": 1.842400915485874e-05, "loss": 0.0, "num_input_tokens_seen": 6548400, "step": 23390 }, { "epoch": 259.94444444444446, "grad_norm": 1.9653568017474754e-07, "learning_rate": 1.8414537907285053e-05, "loss": 0.0, "num_input_tokens_seen": 6549824, "step": 23395 }, { "epoch": 260.0, "grad_norm": 1.2732590448649717e-07, "learning_rate": 1.840506767527237e-05, "loss": 0.0, "num_input_tokens_seen": 6551264, "step": 23400 }, { "epoch": 260.0, "eval_loss": 1.5016049146652222, "eval_runtime": 1.3874, "eval_samples_per_second": 28.832, "eval_steps_per_second": 14.416, "num_input_tokens_seen": 6551264, "step": 23400 }, { "epoch": 260.05555555555554, "grad_norm": 6.626696631428786e-07, "learning_rate": 1.8395598460281137e-05, "loss": 0.0, "num_input_tokens_seen": 6552720, "step": 23405 }, { "epoch": 260.1111111111111, "grad_norm": 4.586446422649715e-08, "learning_rate": 1.838613026377161e-05, "loss": 0.0, "num_input_tokens_seen": 6554144, "step": 23410 }, { "epoch": 260.1666666666667, "grad_norm": 1.5110516926597484e-07, "learning_rate": 1.8376663087203917e-05, "loss": 0.0, "num_input_tokens_seen": 6555504, "step": 23415 }, { "epoch": 260.22222222222223, "grad_norm": 6.666416538791964e-07, "learning_rate": 1.8367196932038014e-05, "loss": 0.0, "num_input_tokens_seen": 6556960, "step": 23420 }, { "epoch": 260.27777777777777, "grad_norm": 4.075537844983046e-07, "learning_rate": 1.8357731799733686e-05, "loss": 0.0, "num_input_tokens_seen": 6558336, "step": 23425 }, { "epoch": 260.3333333333333, "grad_norm": 1.2862693665738334e-07, "learning_rate": 1.8348267691750586e-05, "loss": 0.0, "num_input_tokens_seen": 6559744, "step": 23430 }, { "epoch": 260.3888888888889, "grad_norm": 5.632465445160051e-07, "learning_rate": 1.833880460954821e-05, "loss": 0.0, "num_input_tokens_seen": 6561152, "step": 23435 }, { "epoch": 260.44444444444446, "grad_norm": 1.9948431884131423e-07, "learning_rate": 1.8329342554585866e-05, "loss": 0.0, "num_input_tokens_seen": 6562544, "step": 23440 }, { "epoch": 260.5, "grad_norm": 2.458813810335414e-07, "learning_rate": 1.8319881528322735e-05, "loss": 0.0, "num_input_tokens_seen": 6563888, "step": 23445 }, { "epoch": 260.55555555555554, "grad_norm": 1.6426950821823993e-07, "learning_rate": 1.8310421532217815e-05, "loss": 0.0, "num_input_tokens_seen": 6565264, "step": 23450 }, { "epoch": 260.6111111111111, "grad_norm": 1.3063809944924287e-07, "learning_rate": 1.8300962567729958e-05, "loss": 0.0, "num_input_tokens_seen": 6566656, "step": 23455 }, { "epoch": 260.6666666666667, "grad_norm": 7.588369044242427e-07, "learning_rate": 1.8291504636317866e-05, "loss": 0.0, "num_input_tokens_seen": 6568048, "step": 23460 }, { "epoch": 260.72222222222223, "grad_norm": 4.7487878873653244e-07, "learning_rate": 1.8282047739440055e-05, "loss": 0.0, "num_input_tokens_seen": 6569472, "step": 23465 }, { "epoch": 260.77777777777777, "grad_norm": 1.3456099168251967e-07, "learning_rate": 1.8272591878554903e-05, "loss": 0.0, "num_input_tokens_seen": 6570880, "step": 23470 }, { "epoch": 260.8333333333333, "grad_norm": 1.3648507035668445e-07, "learning_rate": 1.8263137055120638e-05, "loss": 0.0, "num_input_tokens_seen": 6572304, "step": 23475 }, { "epoch": 260.8888888888889, "grad_norm": 1.1483525241828829e-07, "learning_rate": 1.8253683270595295e-05, "loss": 0.0, "num_input_tokens_seen": 6573728, "step": 23480 }, { "epoch": 260.94444444444446, "grad_norm": 7.57361107162069e-08, "learning_rate": 1.824423052643677e-05, "loss": 0.0, "num_input_tokens_seen": 6575120, "step": 23485 }, { "epoch": 261.0, "grad_norm": 7.996843720547986e-08, "learning_rate": 1.82347788241028e-05, "loss": 0.0, "num_input_tokens_seen": 6576512, "step": 23490 }, { "epoch": 261.05555555555554, "grad_norm": 3.246942625878546e-08, "learning_rate": 1.8225328165050942e-05, "loss": 0.0, "num_input_tokens_seen": 6577920, "step": 23495 }, { "epoch": 261.1111111111111, "grad_norm": 8.072095170064131e-08, "learning_rate": 1.821587855073863e-05, "loss": 0.0, "num_input_tokens_seen": 6579312, "step": 23500 }, { "epoch": 261.1666666666667, "grad_norm": 1.3069802662357688e-07, "learning_rate": 1.8206429982623086e-05, "loss": 0.0, "num_input_tokens_seen": 6580688, "step": 23505 }, { "epoch": 261.22222222222223, "grad_norm": 4.77130356557609e-07, "learning_rate": 1.8196982462161416e-05, "loss": 0.0, "num_input_tokens_seen": 6582080, "step": 23510 }, { "epoch": 261.27777777777777, "grad_norm": 3.2644102532231045e-08, "learning_rate": 1.818753599081055e-05, "loss": 0.0, "num_input_tokens_seen": 6583504, "step": 23515 }, { "epoch": 261.3333333333333, "grad_norm": 1.8164759296723787e-07, "learning_rate": 1.817809057002724e-05, "loss": 0.0, "num_input_tokens_seen": 6584928, "step": 23520 }, { "epoch": 261.3888888888889, "grad_norm": 4.2088240093107743e-07, "learning_rate": 1.8168646201268096e-05, "loss": 0.0, "num_input_tokens_seen": 6586336, "step": 23525 }, { "epoch": 261.44444444444446, "grad_norm": 4.480728321709648e-08, "learning_rate": 1.8159202885989557e-05, "loss": 0.0, "num_input_tokens_seen": 6587760, "step": 23530 }, { "epoch": 261.5, "grad_norm": 9.40005051575099e-08, "learning_rate": 1.814976062564789e-05, "loss": 0.0, "num_input_tokens_seen": 6589168, "step": 23535 }, { "epoch": 261.55555555555554, "grad_norm": 3.3962148648924995e-08, "learning_rate": 1.8140319421699234e-05, "loss": 0.0, "num_input_tokens_seen": 6590544, "step": 23540 }, { "epoch": 261.6111111111111, "grad_norm": 4.074330206549348e-08, "learning_rate": 1.8130879275599515e-05, "loss": 0.0, "num_input_tokens_seen": 6591968, "step": 23545 }, { "epoch": 261.6666666666667, "grad_norm": 3.4395728931713165e-08, "learning_rate": 1.8121440188804544e-05, "loss": 0.0, "num_input_tokens_seen": 6593440, "step": 23550 }, { "epoch": 261.72222222222223, "grad_norm": 9.959887137256374e-08, "learning_rate": 1.811200216276993e-05, "loss": 0.0, "num_input_tokens_seen": 6594816, "step": 23555 }, { "epoch": 261.77777777777777, "grad_norm": 1.6373959965676477e-07, "learning_rate": 1.810256519895115e-05, "loss": 0.0, "num_input_tokens_seen": 6596192, "step": 23560 }, { "epoch": 261.8333333333333, "grad_norm": 4.71640255739203e-08, "learning_rate": 1.8093129298803494e-05, "loss": 0.0, "num_input_tokens_seen": 6597600, "step": 23565 }, { "epoch": 261.8888888888889, "grad_norm": 4.289590549433342e-08, "learning_rate": 1.808369446378209e-05, "loss": 0.0, "num_input_tokens_seen": 6598976, "step": 23570 }, { "epoch": 261.94444444444446, "grad_norm": 2.4717570212828832e-08, "learning_rate": 1.8074260695341914e-05, "loss": 0.0, "num_input_tokens_seen": 6600368, "step": 23575 }, { "epoch": 262.0, "grad_norm": 4.137582720886712e-07, "learning_rate": 1.8064827994937782e-05, "loss": 0.0, "num_input_tokens_seen": 6601760, "step": 23580 }, { "epoch": 262.05555555555554, "grad_norm": 3.106440260580712e-08, "learning_rate": 1.8055396364024317e-05, "loss": 0.0, "num_input_tokens_seen": 6603152, "step": 23585 }, { "epoch": 262.1111111111111, "grad_norm": 5.521402002273135e-08, "learning_rate": 1.804596580405601e-05, "loss": 0.0, "num_input_tokens_seen": 6604592, "step": 23590 }, { "epoch": 262.1666666666667, "grad_norm": 5.362243271633815e-08, "learning_rate": 1.8036536316487174e-05, "loss": 0.0, "num_input_tokens_seen": 6606032, "step": 23595 }, { "epoch": 262.22222222222223, "grad_norm": 1.288908038077352e-07, "learning_rate": 1.802710790277193e-05, "loss": 0.0, "num_input_tokens_seen": 6607424, "step": 23600 }, { "epoch": 262.22222222222223, "eval_loss": 1.5534162521362305, "eval_runtime": 1.3851, "eval_samples_per_second": 28.878, "eval_steps_per_second": 14.439, "num_input_tokens_seen": 6607424, "step": 23600 }, { "epoch": 262.27777777777777, "grad_norm": 8.421515929057932e-08, "learning_rate": 1.801768056436429e-05, "loss": 0.0, "num_input_tokens_seen": 6608768, "step": 23605 }, { "epoch": 262.3333333333333, "grad_norm": 3.74793671653606e-07, "learning_rate": 1.8008254302718035e-05, "loss": 0.0, "num_input_tokens_seen": 6610176, "step": 23610 }, { "epoch": 262.3888888888889, "grad_norm": 8.499146986196138e-08, "learning_rate": 1.7998829119286837e-05, "loss": 0.0, "num_input_tokens_seen": 6611552, "step": 23615 }, { "epoch": 262.44444444444446, "grad_norm": 1.0691076113289455e-07, "learning_rate": 1.798940501552418e-05, "loss": 0.0, "num_input_tokens_seen": 6612944, "step": 23620 }, { "epoch": 262.5, "grad_norm": 6.213380743247399e-08, "learning_rate": 1.797998199288336e-05, "loss": 0.0, "num_input_tokens_seen": 6614272, "step": 23625 }, { "epoch": 262.55555555555554, "grad_norm": 1.3168489942927408e-07, "learning_rate": 1.7970560052817543e-05, "loss": 0.0, "num_input_tokens_seen": 6615680, "step": 23630 }, { "epoch": 262.6111111111111, "grad_norm": 1.7119519668540306e-07, "learning_rate": 1.7961139196779702e-05, "loss": 0.0, "num_input_tokens_seen": 6617072, "step": 23635 }, { "epoch": 262.6666666666667, "grad_norm": 4.152671451151946e-08, "learning_rate": 1.7951719426222647e-05, "loss": 0.0, "num_input_tokens_seen": 6618464, "step": 23640 }, { "epoch": 262.72222222222223, "grad_norm": 1.1358535800809477e-07, "learning_rate": 1.794230074259904e-05, "loss": 0.0, "num_input_tokens_seen": 6619888, "step": 23645 }, { "epoch": 262.77777777777777, "grad_norm": 8.866717848832195e-08, "learning_rate": 1.7932883147361336e-05, "loss": 0.0, "num_input_tokens_seen": 6621296, "step": 23650 }, { "epoch": 262.8333333333333, "grad_norm": 2.8211366043251473e-07, "learning_rate": 1.7923466641961865e-05, "loss": 0.0, "num_input_tokens_seen": 6622704, "step": 23655 }, { "epoch": 262.8888888888889, "grad_norm": 4.3458040721588986e-08, "learning_rate": 1.791405122785278e-05, "loss": 0.0, "num_input_tokens_seen": 6624112, "step": 23660 }, { "epoch": 262.94444444444446, "grad_norm": 1.45222983860549e-07, "learning_rate": 1.7904636906486037e-05, "loss": 0.0, "num_input_tokens_seen": 6625504, "step": 23665 }, { "epoch": 263.0, "grad_norm": 8.156384865287691e-07, "learning_rate": 1.7895223679313448e-05, "loss": 0.0, "num_input_tokens_seen": 6626880, "step": 23670 }, { "epoch": 263.05555555555554, "grad_norm": 1.526653079508833e-07, "learning_rate": 1.7885811547786653e-05, "loss": 0.0, "num_input_tokens_seen": 6628288, "step": 23675 }, { "epoch": 263.1111111111111, "grad_norm": 9.315699855960702e-08, "learning_rate": 1.7876400513357115e-05, "loss": 0.0, "num_input_tokens_seen": 6629648, "step": 23680 }, { "epoch": 263.1666666666667, "grad_norm": 2.918025643339206e-07, "learning_rate": 1.7866990577476146e-05, "loss": 0.0, "num_input_tokens_seen": 6631040, "step": 23685 }, { "epoch": 263.22222222222223, "grad_norm": 4.205198820272926e-07, "learning_rate": 1.7857581741594863e-05, "loss": 0.0, "num_input_tokens_seen": 6632448, "step": 23690 }, { "epoch": 263.27777777777777, "grad_norm": 7.965216468619474e-07, "learning_rate": 1.7848174007164237e-05, "loss": 0.0, "num_input_tokens_seen": 6633824, "step": 23695 }, { "epoch": 263.3333333333333, "grad_norm": 5.551260642278066e-07, "learning_rate": 1.7838767375635052e-05, "loss": 0.0, "num_input_tokens_seen": 6635280, "step": 23700 }, { "epoch": 263.3888888888889, "grad_norm": 4.767748293943441e-07, "learning_rate": 1.782936184845793e-05, "loss": 0.0, "num_input_tokens_seen": 6636656, "step": 23705 }, { "epoch": 263.44444444444446, "grad_norm": 5.986120868328726e-06, "learning_rate": 1.7819957427083334e-05, "loss": 0.0, "num_input_tokens_seen": 6638048, "step": 23710 }, { "epoch": 263.5, "grad_norm": 3.528167837885121e-07, "learning_rate": 1.7810554112961516e-05, "loss": 0.0, "num_input_tokens_seen": 6639488, "step": 23715 }, { "epoch": 263.55555555555554, "grad_norm": 3.634130507634836e-07, "learning_rate": 1.7801151907542607e-05, "loss": 0.0, "num_input_tokens_seen": 6640880, "step": 23720 }, { "epoch": 263.6111111111111, "grad_norm": 2.4822256250445207e-07, "learning_rate": 1.7791750812276547e-05, "loss": 0.0, "num_input_tokens_seen": 6642288, "step": 23725 }, { "epoch": 263.6666666666667, "grad_norm": 4.820464027943672e-07, "learning_rate": 1.778235082861309e-05, "loss": 0.0, "num_input_tokens_seen": 6643664, "step": 23730 }, { "epoch": 263.72222222222223, "grad_norm": 8.829057662751438e-08, "learning_rate": 1.777295195800184e-05, "loss": 0.0, "num_input_tokens_seen": 6645056, "step": 23735 }, { "epoch": 263.77777777777777, "grad_norm": 6.762948601135577e-08, "learning_rate": 1.7763554201892215e-05, "loss": 0.0, "num_input_tokens_seen": 6646448, "step": 23740 }, { "epoch": 263.8333333333333, "grad_norm": 4.2750112783096483e-08, "learning_rate": 1.7754157561733476e-05, "loss": 0.0, "num_input_tokens_seen": 6647840, "step": 23745 }, { "epoch": 263.8888888888889, "grad_norm": 1.2865257303928956e-07, "learning_rate": 1.7744762038974702e-05, "loss": 0.0, "num_input_tokens_seen": 6649248, "step": 23750 }, { "epoch": 263.94444444444446, "grad_norm": 1.7433005439215776e-07, "learning_rate": 1.7735367635064788e-05, "loss": 0.0, "num_input_tokens_seen": 6650688, "step": 23755 }, { "epoch": 264.0, "grad_norm": 1.2209508781779732e-07, "learning_rate": 1.7725974351452474e-05, "loss": 0.0, "num_input_tokens_seen": 6652080, "step": 23760 }, { "epoch": 264.05555555555554, "grad_norm": 4.5542100224338355e-07, "learning_rate": 1.771658218958634e-05, "loss": 0.0, "num_input_tokens_seen": 6653472, "step": 23765 }, { "epoch": 264.1111111111111, "grad_norm": 7.290971382190037e-08, "learning_rate": 1.770719115091475e-05, "loss": 0.0, "num_input_tokens_seen": 6654880, "step": 23770 }, { "epoch": 264.1666666666667, "grad_norm": 1.6045822803789633e-07, "learning_rate": 1.7697801236885935e-05, "loss": 0.0, "num_input_tokens_seen": 6656256, "step": 23775 }, { "epoch": 264.22222222222223, "grad_norm": 1.664060533812517e-07, "learning_rate": 1.7688412448947944e-05, "loss": 0.0, "num_input_tokens_seen": 6657664, "step": 23780 }, { "epoch": 264.27777777777777, "grad_norm": 1.593867295923701e-07, "learning_rate": 1.767902478854862e-05, "loss": 0.0, "num_input_tokens_seen": 6659040, "step": 23785 }, { "epoch": 264.3333333333333, "grad_norm": 1.3391159825459908e-07, "learning_rate": 1.766963825713569e-05, "loss": 0.0, "num_input_tokens_seen": 6660384, "step": 23790 }, { "epoch": 264.3888888888889, "grad_norm": 1.0416567164384105e-07, "learning_rate": 1.766025285615665e-05, "loss": 0.0, "num_input_tokens_seen": 6661792, "step": 23795 }, { "epoch": 264.44444444444446, "grad_norm": 1.0916463821786238e-07, "learning_rate": 1.7650868587058854e-05, "loss": 0.0, "num_input_tokens_seen": 6663168, "step": 23800 }, { "epoch": 264.44444444444446, "eval_loss": 1.5226383209228516, "eval_runtime": 1.3865, "eval_samples_per_second": 28.849, "eval_steps_per_second": 14.424, "num_input_tokens_seen": 6663168, "step": 23800 }, { "epoch": 264.5, "grad_norm": 4.15483157212293e-07, "learning_rate": 1.7641485451289484e-05, "loss": 0.0, "num_input_tokens_seen": 6664608, "step": 23805 }, { "epoch": 264.55555555555554, "grad_norm": 2.527742424263124e-07, "learning_rate": 1.7632103450295534e-05, "loss": 0.0, "num_input_tokens_seen": 6665952, "step": 23810 }, { "epoch": 264.6111111111111, "grad_norm": 5.6586145547044e-07, "learning_rate": 1.762272258552381e-05, "loss": 0.0, "num_input_tokens_seen": 6667376, "step": 23815 }, { "epoch": 264.6666666666667, "grad_norm": 1.765029367106763e-07, "learning_rate": 1.7613342858420988e-05, "loss": 0.0, "num_input_tokens_seen": 6668784, "step": 23820 }, { "epoch": 264.72222222222223, "grad_norm": 1.1896146645540284e-07, "learning_rate": 1.760396427043351e-05, "loss": 0.0, "num_input_tokens_seen": 6670176, "step": 23825 }, { "epoch": 264.77777777777777, "grad_norm": 3.369304977240972e-07, "learning_rate": 1.7594586823007696e-05, "loss": 0.0, "num_input_tokens_seen": 6671584, "step": 23830 }, { "epoch": 264.8333333333333, "grad_norm": 9.682738664196222e-07, "learning_rate": 1.7585210517589646e-05, "loss": 0.0, "num_input_tokens_seen": 6673024, "step": 23835 }, { "epoch": 264.8888888888889, "grad_norm": 3.232688925436378e-07, "learning_rate": 1.7575835355625314e-05, "loss": 0.0, "num_input_tokens_seen": 6674448, "step": 23840 }, { "epoch": 264.94444444444446, "grad_norm": 4.498718055856443e-07, "learning_rate": 1.756646133856048e-05, "loss": 0.0, "num_input_tokens_seen": 6675840, "step": 23845 }, { "epoch": 265.0, "grad_norm": 3.536963504302548e-07, "learning_rate": 1.7557088467840714e-05, "loss": 0.0, "num_input_tokens_seen": 6677248, "step": 23850 }, { "epoch": 265.05555555555554, "grad_norm": 5.14869225298753e-06, "learning_rate": 1.7547716744911438e-05, "loss": 0.0, "num_input_tokens_seen": 6678608, "step": 23855 }, { "epoch": 265.1111111111111, "grad_norm": 7.461277107267961e-08, "learning_rate": 1.7538346171217902e-05, "loss": 0.0, "num_input_tokens_seen": 6680000, "step": 23860 }, { "epoch": 265.1666666666667, "grad_norm": 2.9611342711177713e-07, "learning_rate": 1.7528976748205146e-05, "loss": 0.0, "num_input_tokens_seen": 6681376, "step": 23865 }, { "epoch": 265.22222222222223, "grad_norm": 6.737286639690865e-08, "learning_rate": 1.751960847731807e-05, "loss": 0.0, "num_input_tokens_seen": 6682768, "step": 23870 }, { "epoch": 265.27777777777777, "grad_norm": 1.3921206232225813e-07, "learning_rate": 1.7510241360001362e-05, "loss": 0.0, "num_input_tokens_seen": 6684208, "step": 23875 }, { "epoch": 265.3333333333333, "grad_norm": 2.0784607102086738e-07, "learning_rate": 1.7500875397699562e-05, "loss": 0.0, "num_input_tokens_seen": 6685632, "step": 23880 }, { "epoch": 265.3888888888889, "grad_norm": 1.7098663818160276e-07, "learning_rate": 1.7491510591857015e-05, "loss": 0.0, "num_input_tokens_seen": 6687040, "step": 23885 }, { "epoch": 265.44444444444446, "grad_norm": 1.387144266118412e-07, "learning_rate": 1.7482146943917896e-05, "loss": 0.0, "num_input_tokens_seen": 6688480, "step": 23890 }, { "epoch": 265.5, "grad_norm": 1.2688664696725027e-07, "learning_rate": 1.7472784455326185e-05, "loss": 0.0, "num_input_tokens_seen": 6689872, "step": 23895 }, { "epoch": 265.55555555555554, "grad_norm": 1.3682711141882464e-07, "learning_rate": 1.746342312752572e-05, "loss": 0.0, "num_input_tokens_seen": 6691280, "step": 23900 }, { "epoch": 265.6111111111111, "grad_norm": 5.71638771873495e-08, "learning_rate": 1.74540629619601e-05, "loss": 0.0, "num_input_tokens_seen": 6692640, "step": 23905 }, { "epoch": 265.6666666666667, "grad_norm": 1.0511998738138573e-07, "learning_rate": 1.7444703960072815e-05, "loss": 0.0, "num_input_tokens_seen": 6694048, "step": 23910 }, { "epoch": 265.72222222222223, "grad_norm": 3.174487233081891e-07, "learning_rate": 1.7435346123307118e-05, "loss": 0.0, "num_input_tokens_seen": 6695440, "step": 23915 }, { "epoch": 265.77777777777777, "grad_norm": 6.208121590134397e-07, "learning_rate": 1.742598945310611e-05, "loss": 0.0, "num_input_tokens_seen": 6696832, "step": 23920 }, { "epoch": 265.8333333333333, "grad_norm": 6.092006969993236e-07, "learning_rate": 1.741663395091272e-05, "loss": 0.0, "num_input_tokens_seen": 6698288, "step": 23925 }, { "epoch": 265.8888888888889, "grad_norm": 4.4159875756122347e-08, "learning_rate": 1.7407279618169657e-05, "loss": 0.0, "num_input_tokens_seen": 6699712, "step": 23930 }, { "epoch": 265.94444444444446, "grad_norm": 9.070681272760339e-08, "learning_rate": 1.73979264563195e-05, "loss": 0.0, "num_input_tokens_seen": 6701120, "step": 23935 }, { "epoch": 266.0, "grad_norm": 1.0153274843105464e-06, "learning_rate": 1.7388574466804625e-05, "loss": 0.0, "num_input_tokens_seen": 6702544, "step": 23940 }, { "epoch": 266.05555555555554, "grad_norm": 4.210994575259974e-08, "learning_rate": 1.7379223651067207e-05, "loss": 0.0, "num_input_tokens_seen": 6703952, "step": 23945 }, { "epoch": 266.1111111111111, "grad_norm": 6.497846527508955e-08, "learning_rate": 1.736987401054928e-05, "loss": 0.0, "num_input_tokens_seen": 6705360, "step": 23950 }, { "epoch": 266.1666666666667, "grad_norm": 1.2742313515445858e-07, "learning_rate": 1.736052554669266e-05, "loss": 0.0, "num_input_tokens_seen": 6706752, "step": 23955 }, { "epoch": 266.22222222222223, "grad_norm": 1.3653929897827766e-07, "learning_rate": 1.7351178260939007e-05, "loss": 0.0, "num_input_tokens_seen": 6708176, "step": 23960 }, { "epoch": 266.27777777777777, "grad_norm": 5.041356416768394e-07, "learning_rate": 1.7341832154729794e-05, "loss": 0.0, "num_input_tokens_seen": 6709584, "step": 23965 }, { "epoch": 266.3333333333333, "grad_norm": 1.380542897777559e-07, "learning_rate": 1.7332487229506286e-05, "loss": 0.0, "num_input_tokens_seen": 6710912, "step": 23970 }, { "epoch": 266.3888888888889, "grad_norm": 2.5002083248182316e-07, "learning_rate": 1.732314348670961e-05, "loss": 0.0, "num_input_tokens_seen": 6712272, "step": 23975 }, { "epoch": 266.44444444444446, "grad_norm": 4.0550597191213456e-07, "learning_rate": 1.7313800927780686e-05, "loss": 0.0, "num_input_tokens_seen": 6713648, "step": 23980 }, { "epoch": 266.5, "grad_norm": 1.1566820745656514e-07, "learning_rate": 1.7304459554160245e-05, "loss": 0.0, "num_input_tokens_seen": 6715040, "step": 23985 }, { "epoch": 266.55555555555554, "grad_norm": 3.330430686787622e-08, "learning_rate": 1.7295119367288853e-05, "loss": 0.0, "num_input_tokens_seen": 6716432, "step": 23990 }, { "epoch": 266.6111111111111, "grad_norm": 3.837524502614542e-07, "learning_rate": 1.728578036860688e-05, "loss": 0.0, "num_input_tokens_seen": 6717808, "step": 23995 }, { "epoch": 266.6666666666667, "grad_norm": 4.341410431152326e-08, "learning_rate": 1.7276442559554513e-05, "loss": 0.0, "num_input_tokens_seen": 6719216, "step": 24000 }, { "epoch": 266.6666666666667, "eval_loss": 1.6124918460845947, "eval_runtime": 1.3881, "eval_samples_per_second": 28.816, "eval_steps_per_second": 14.408, "num_input_tokens_seen": 6719216, "step": 24000 }, { "epoch": 266.72222222222223, "grad_norm": 8.917046443457366e-07, "learning_rate": 1.726710594157177e-05, "loss": 0.0, "num_input_tokens_seen": 6720592, "step": 24005 }, { "epoch": 266.77777777777777, "grad_norm": 3.5918674257118255e-07, "learning_rate": 1.725777051609846e-05, "loss": 0.0, "num_input_tokens_seen": 6722032, "step": 24010 }, { "epoch": 266.8333333333333, "grad_norm": 5.328053589437332e-07, "learning_rate": 1.7248436284574228e-05, "loss": 0.0, "num_input_tokens_seen": 6723456, "step": 24015 }, { "epoch": 266.8888888888889, "grad_norm": 3.220264659375971e-07, "learning_rate": 1.723910324843855e-05, "loss": 0.0, "num_input_tokens_seen": 6724880, "step": 24020 }, { "epoch": 266.94444444444446, "grad_norm": 2.1975537833895942e-07, "learning_rate": 1.722977140913067e-05, "loss": 0.0, "num_input_tokens_seen": 6726304, "step": 24025 }, { "epoch": 267.0, "grad_norm": 3.353712259013264e-07, "learning_rate": 1.7220440768089688e-05, "loss": 0.0, "num_input_tokens_seen": 6727728, "step": 24030 }, { "epoch": 267.05555555555554, "grad_norm": 2.3555232075977983e-07, "learning_rate": 1.7211111326754505e-05, "loss": 0.0, "num_input_tokens_seen": 6729120, "step": 24035 }, { "epoch": 267.1111111111111, "grad_norm": 3.016892833329621e-07, "learning_rate": 1.720178308656383e-05, "loss": 0.0, "num_input_tokens_seen": 6730560, "step": 24040 }, { "epoch": 267.1666666666667, "grad_norm": 2.125926243934373e-07, "learning_rate": 1.719245604895621e-05, "loss": 0.0, "num_input_tokens_seen": 6731952, "step": 24045 }, { "epoch": 267.22222222222223, "grad_norm": 1.007849093070945e-07, "learning_rate": 1.7183130215369972e-05, "loss": 0.0, "num_input_tokens_seen": 6733344, "step": 24050 }, { "epoch": 267.27777777777777, "grad_norm": 1.0773415226594807e-07, "learning_rate": 1.7173805587243292e-05, "loss": 0.0, "num_input_tokens_seen": 6734752, "step": 24055 }, { "epoch": 267.3333333333333, "grad_norm": 1.9144174245866452e-07, "learning_rate": 1.7164482166014147e-05, "loss": 0.0, "num_input_tokens_seen": 6736176, "step": 24060 }, { "epoch": 267.3888888888889, "grad_norm": 1.0195667243806383e-07, "learning_rate": 1.7155159953120313e-05, "loss": 0.0, "num_input_tokens_seen": 6737552, "step": 24065 }, { "epoch": 267.44444444444446, "grad_norm": 3.233733423257945e-07, "learning_rate": 1.714583894999941e-05, "loss": 0.0, "num_input_tokens_seen": 6738944, "step": 24070 }, { "epoch": 267.5, "grad_norm": 4.0647060473020247e-07, "learning_rate": 1.7136519158088826e-05, "loss": 0.0, "num_input_tokens_seen": 6740352, "step": 24075 }, { "epoch": 267.55555555555554, "grad_norm": 2.787896562495007e-07, "learning_rate": 1.712720057882581e-05, "loss": 0.0, "num_input_tokens_seen": 6741760, "step": 24080 }, { "epoch": 267.6111111111111, "grad_norm": 6.575787097062857e-07, "learning_rate": 1.7117883213647413e-05, "loss": 0.0, "num_input_tokens_seen": 6743152, "step": 24085 }, { "epoch": 267.6666666666667, "grad_norm": 5.423156039796595e-07, "learning_rate": 1.710856706399046e-05, "loss": 0.0, "num_input_tokens_seen": 6744608, "step": 24090 }, { "epoch": 267.72222222222223, "grad_norm": 3.218503707103082e-06, "learning_rate": 1.7099252131291648e-05, "loss": 0.0, "num_input_tokens_seen": 6746016, "step": 24095 }, { "epoch": 267.77777777777777, "grad_norm": 8.104390758489899e-07, "learning_rate": 1.708993841698744e-05, "loss": 0.0, "num_input_tokens_seen": 6747344, "step": 24100 }, { "epoch": 267.8333333333333, "grad_norm": 1.2417604011716321e-06, "learning_rate": 1.7080625922514132e-05, "loss": 0.0, "num_input_tokens_seen": 6748768, "step": 24105 }, { "epoch": 267.8888888888889, "grad_norm": 6.553266757691745e-06, "learning_rate": 1.7071314649307836e-05, "loss": 0.0, "num_input_tokens_seen": 6750176, "step": 24110 }, { "epoch": 267.94444444444446, "grad_norm": 1.3739652615640807e-07, "learning_rate": 1.7062004598804448e-05, "loss": 0.0, "num_input_tokens_seen": 6751600, "step": 24115 }, { "epoch": 268.0, "grad_norm": 4.64571627389887e-07, "learning_rate": 1.7052695772439702e-05, "loss": 0.0, "num_input_tokens_seen": 6753008, "step": 24120 }, { "epoch": 268.05555555555554, "grad_norm": 9.43555050980649e-07, "learning_rate": 1.7043388171649154e-05, "loss": 0.0, "num_input_tokens_seen": 6754416, "step": 24125 }, { "epoch": 268.1111111111111, "grad_norm": 9.63078946369933e-07, "learning_rate": 1.7034081797868127e-05, "loss": 0.0, "num_input_tokens_seen": 6755792, "step": 24130 }, { "epoch": 268.1666666666667, "grad_norm": 4.324828921653534e-07, "learning_rate": 1.70247766525318e-05, "loss": 0.0, "num_input_tokens_seen": 6757216, "step": 24135 }, { "epoch": 268.22222222222223, "grad_norm": 2.8068741812603548e-05, "learning_rate": 1.701547273707514e-05, "loss": 0.0, "num_input_tokens_seen": 6758624, "step": 24140 }, { "epoch": 268.27777777777777, "grad_norm": 8.347446822654092e-08, "learning_rate": 1.7006170052932916e-05, "loss": 0.0, "num_input_tokens_seen": 6760032, "step": 24145 }, { "epoch": 268.3333333333333, "grad_norm": 1.1290498491689505e-07, "learning_rate": 1.6996868601539735e-05, "loss": 0.0, "num_input_tokens_seen": 6761424, "step": 24150 }, { "epoch": 268.3888888888889, "grad_norm": 1.1847773606632472e-07, "learning_rate": 1.6987568384329977e-05, "loss": 0.0, "num_input_tokens_seen": 6762800, "step": 24155 }, { "epoch": 268.44444444444446, "grad_norm": 3.004807922479813e-07, "learning_rate": 1.6978269402737866e-05, "loss": 0.0, "num_input_tokens_seen": 6764192, "step": 24160 }, { "epoch": 268.5, "grad_norm": 7.177284544468421e-08, "learning_rate": 1.696897165819743e-05, "loss": 0.0, "num_input_tokens_seen": 6765616, "step": 24165 }, { "epoch": 268.55555555555554, "grad_norm": 9.431340686205658e-08, "learning_rate": 1.6959675152142487e-05, "loss": 0.0, "num_input_tokens_seen": 6766992, "step": 24170 }, { "epoch": 268.6111111111111, "grad_norm": 7.812364799519855e-08, "learning_rate": 1.6950379886006667e-05, "loss": 0.0, "num_input_tokens_seen": 6768352, "step": 24175 }, { "epoch": 268.6666666666667, "grad_norm": 2.565916759067477e-07, "learning_rate": 1.6941085861223438e-05, "loss": 0.0, "num_input_tokens_seen": 6769744, "step": 24180 }, { "epoch": 268.72222222222223, "grad_norm": 1.0909020176086415e-07, "learning_rate": 1.6931793079226034e-05, "loss": 0.0, "num_input_tokens_seen": 6771168, "step": 24185 }, { "epoch": 268.77777777777777, "grad_norm": 2.9428863967950747e-07, "learning_rate": 1.692250154144754e-05, "loss": 0.0, "num_input_tokens_seen": 6772512, "step": 24190 }, { "epoch": 268.8333333333333, "grad_norm": 2.8302920895839634e-07, "learning_rate": 1.6913211249320807e-05, "loss": 0.0, "num_input_tokens_seen": 6773920, "step": 24195 }, { "epoch": 268.8888888888889, "grad_norm": 7.103360388782676e-08, "learning_rate": 1.6903922204278522e-05, "loss": 0.0, "num_input_tokens_seen": 6775344, "step": 24200 }, { "epoch": 268.8888888888889, "eval_loss": 1.5949466228485107, "eval_runtime": 1.3886, "eval_samples_per_second": 28.807, "eval_steps_per_second": 14.403, "num_input_tokens_seen": 6775344, "step": 24200 }, { "epoch": 268.94444444444446, "grad_norm": 7.704967686095188e-08, "learning_rate": 1.6894634407753186e-05, "loss": 0.0, "num_input_tokens_seen": 6776736, "step": 24205 }, { "epoch": 269.0, "grad_norm": 2.0970891512206435e-07, "learning_rate": 1.6885347861177077e-05, "loss": 0.0, "num_input_tokens_seen": 6778144, "step": 24210 }, { "epoch": 269.05555555555554, "grad_norm": 6.330805035759113e-08, "learning_rate": 1.6876062565982298e-05, "loss": 0.0, "num_input_tokens_seen": 6779552, "step": 24215 }, { "epoch": 269.1111111111111, "grad_norm": 5.85882332870824e-07, "learning_rate": 1.6866778523600774e-05, "loss": 0.0, "num_input_tokens_seen": 6780960, "step": 24220 }, { "epoch": 269.1666666666667, "grad_norm": 1.9612670598689874e-07, "learning_rate": 1.6857495735464195e-05, "loss": 0.0, "num_input_tokens_seen": 6782352, "step": 24225 }, { "epoch": 269.22222222222223, "grad_norm": 7.795811285404852e-08, "learning_rate": 1.6848214203004115e-05, "loss": 0.0, "num_input_tokens_seen": 6783696, "step": 24230 }, { "epoch": 269.27777777777777, "grad_norm": 8.729240335014765e-07, "learning_rate": 1.6838933927651835e-05, "loss": 0.0, "num_input_tokens_seen": 6785104, "step": 24235 }, { "epoch": 269.3333333333333, "grad_norm": 5.228996613482195e-08, "learning_rate": 1.6829654910838506e-05, "loss": 0.0, "num_input_tokens_seen": 6786496, "step": 24240 }, { "epoch": 269.3888888888889, "grad_norm": 5.871746679986245e-07, "learning_rate": 1.6820377153995065e-05, "loss": 0.0, "num_input_tokens_seen": 6787888, "step": 24245 }, { "epoch": 269.44444444444446, "grad_norm": 1.0853995746629153e-07, "learning_rate": 1.681110065855226e-05, "loss": 0.0, "num_input_tokens_seen": 6789280, "step": 24250 }, { "epoch": 269.5, "grad_norm": 6.219305248578166e-08, "learning_rate": 1.6801825425940642e-05, "loss": 0.0, "num_input_tokens_seen": 6790656, "step": 24255 }, { "epoch": 269.55555555555554, "grad_norm": 2.166615331589128e-07, "learning_rate": 1.679255145759056e-05, "loss": 0.0, "num_input_tokens_seen": 6792064, "step": 24260 }, { "epoch": 269.6111111111111, "grad_norm": 6.585385676771693e-07, "learning_rate": 1.6783278754932187e-05, "loss": 0.0, "num_input_tokens_seen": 6793472, "step": 24265 }, { "epoch": 269.6666666666667, "grad_norm": 1.2651541680952505e-07, "learning_rate": 1.6774007319395496e-05, "loss": 0.0, "num_input_tokens_seen": 6794880, "step": 24270 }, { "epoch": 269.72222222222223, "grad_norm": 9.511752097068893e-08, "learning_rate": 1.6764737152410243e-05, "loss": 0.0, "num_input_tokens_seen": 6796304, "step": 24275 }, { "epoch": 269.77777777777777, "grad_norm": 8.613580604333038e-08, "learning_rate": 1.6755468255406016e-05, "loss": 0.0, "num_input_tokens_seen": 6797696, "step": 24280 }, { "epoch": 269.8333333333333, "grad_norm": 3.9374262428282236e-07, "learning_rate": 1.674620062981219e-05, "loss": 0.0, "num_input_tokens_seen": 6799088, "step": 24285 }, { "epoch": 269.8888888888889, "grad_norm": 9.479212792484759e-08, "learning_rate": 1.6736934277057947e-05, "loss": 0.0, "num_input_tokens_seen": 6800480, "step": 24290 }, { "epoch": 269.94444444444446, "grad_norm": 1.289490256795034e-07, "learning_rate": 1.6727669198572286e-05, "loss": 0.0, "num_input_tokens_seen": 6801888, "step": 24295 }, { "epoch": 270.0, "grad_norm": 8.453285431642144e-07, "learning_rate": 1.6718405395783984e-05, "loss": 0.0, "num_input_tokens_seen": 6803344, "step": 24300 }, { "epoch": 270.05555555555554, "grad_norm": 4.5221912614579196e-07, "learning_rate": 1.6709142870121643e-05, "loss": 0.0, "num_input_tokens_seen": 6804736, "step": 24305 }, { "epoch": 270.1111111111111, "grad_norm": 7.431415127712171e-08, "learning_rate": 1.669988162301367e-05, "loss": 0.0, "num_input_tokens_seen": 6806112, "step": 24310 }, { "epoch": 270.1666666666667, "grad_norm": 2.51022242991894e-07, "learning_rate": 1.6690621655888243e-05, "loss": 0.0, "num_input_tokens_seen": 6807488, "step": 24315 }, { "epoch": 270.22222222222223, "grad_norm": 2.1912700987058997e-08, "learning_rate": 1.6681362970173386e-05, "loss": 0.0, "num_input_tokens_seen": 6808896, "step": 24320 }, { "epoch": 270.27777777777777, "grad_norm": 4.3280810047008345e-08, "learning_rate": 1.6672105567296904e-05, "loss": 0.0, "num_input_tokens_seen": 6810320, "step": 24325 }, { "epoch": 270.3333333333333, "grad_norm": 4.0507501353204134e-08, "learning_rate": 1.666284944868639e-05, "loss": 0.0, "num_input_tokens_seen": 6811712, "step": 24330 }, { "epoch": 270.3888888888889, "grad_norm": 4.82079144603631e-07, "learning_rate": 1.665359461576927e-05, "loss": 0.0, "num_input_tokens_seen": 6813120, "step": 24335 }, { "epoch": 270.44444444444446, "grad_norm": 2.813455921568675e-07, "learning_rate": 1.6644341069972736e-05, "loss": 0.0, "num_input_tokens_seen": 6814528, "step": 24340 }, { "epoch": 270.5, "grad_norm": 4.221512028834695e-07, "learning_rate": 1.6635088812723813e-05, "loss": 0.0, "num_input_tokens_seen": 6815904, "step": 24345 }, { "epoch": 270.55555555555554, "grad_norm": 7.142650702007813e-07, "learning_rate": 1.6625837845449328e-05, "loss": 0.0, "num_input_tokens_seen": 6817344, "step": 24350 }, { "epoch": 270.6111111111111, "grad_norm": 4.820273034056299e-07, "learning_rate": 1.6616588169575874e-05, "loss": 0.0, "num_input_tokens_seen": 6818736, "step": 24355 }, { "epoch": 270.6666666666667, "grad_norm": 1.3916860552853905e-07, "learning_rate": 1.6607339786529878e-05, "loss": 0.0, "num_input_tokens_seen": 6820144, "step": 24360 }, { "epoch": 270.72222222222223, "grad_norm": 2.274289379045058e-08, "learning_rate": 1.659809269773756e-05, "loss": 0.0, "num_input_tokens_seen": 6821504, "step": 24365 }, { "epoch": 270.77777777777777, "grad_norm": 3.2195032417803304e-07, "learning_rate": 1.658884690462493e-05, "loss": 0.0, "num_input_tokens_seen": 6822880, "step": 24370 }, { "epoch": 270.8333333333333, "grad_norm": 5.073442110870019e-08, "learning_rate": 1.6579602408617813e-05, "loss": 0.0, "num_input_tokens_seen": 6824272, "step": 24375 }, { "epoch": 270.8888888888889, "grad_norm": 9.477042084427012e-08, "learning_rate": 1.657035921114181e-05, "loss": 0.0, "num_input_tokens_seen": 6825680, "step": 24380 }, { "epoch": 270.94444444444446, "grad_norm": 1.731834515794617e-07, "learning_rate": 1.656111731362236e-05, "loss": 0.0, "num_input_tokens_seen": 6827088, "step": 24385 }, { "epoch": 271.0, "grad_norm": 1.1640631925047273e-07, "learning_rate": 1.6551876717484666e-05, "loss": 0.0, "num_input_tokens_seen": 6828512, "step": 24390 }, { "epoch": 271.05555555555554, "grad_norm": 7.027787773949967e-08, "learning_rate": 1.6542637424153752e-05, "loss": 0.0, "num_input_tokens_seen": 6829904, "step": 24395 }, { "epoch": 271.1111111111111, "grad_norm": 3.3422017509110447e-07, "learning_rate": 1.6533399435054418e-05, "loss": 0.0, "num_input_tokens_seen": 6831344, "step": 24400 }, { "epoch": 271.1111111111111, "eval_loss": 1.6079126596450806, "eval_runtime": 1.4025, "eval_samples_per_second": 28.521, "eval_steps_per_second": 14.26, "num_input_tokens_seen": 6831344, "step": 24400 }, { "epoch": 271.1666666666667, "grad_norm": 3.5702015566130285e-07, "learning_rate": 1.6524162751611304e-05, "loss": 0.0, "num_input_tokens_seen": 6832704, "step": 24405 }, { "epoch": 271.22222222222223, "grad_norm": 5.91684226947109e-07, "learning_rate": 1.6514927375248796e-05, "loss": 0.0, "num_input_tokens_seen": 6834096, "step": 24410 }, { "epoch": 271.27777777777777, "grad_norm": 3.511571264880331e-07, "learning_rate": 1.6505693307391127e-05, "loss": 0.0, "num_input_tokens_seen": 6835472, "step": 24415 }, { "epoch": 271.3333333333333, "grad_norm": 1.8788578870498895e-07, "learning_rate": 1.6496460549462288e-05, "loss": 0.0, "num_input_tokens_seen": 6836848, "step": 24420 }, { "epoch": 271.3888888888889, "grad_norm": 2.3408350102727127e-07, "learning_rate": 1.6487229102886097e-05, "loss": 0.0, "num_input_tokens_seen": 6838240, "step": 24425 }, { "epoch": 271.44444444444446, "grad_norm": 1.8938513335342577e-07, "learning_rate": 1.6477998969086155e-05, "loss": 0.0, "num_input_tokens_seen": 6839584, "step": 24430 }, { "epoch": 271.5, "grad_norm": 1.0369168421675568e-06, "learning_rate": 1.646877014948587e-05, "loss": 0.0, "num_input_tokens_seen": 6841008, "step": 24435 }, { "epoch": 271.55555555555554, "grad_norm": 8.829334063875649e-08, "learning_rate": 1.6459542645508433e-05, "loss": 0.0, "num_input_tokens_seen": 6842464, "step": 24440 }, { "epoch": 271.6111111111111, "grad_norm": 1.0788340887302184e-07, "learning_rate": 1.6450316458576852e-05, "loss": 0.0, "num_input_tokens_seen": 6843824, "step": 24445 }, { "epoch": 271.6666666666667, "grad_norm": 2.3439655549850613e-08, "learning_rate": 1.6441091590113912e-05, "loss": 0.0, "num_input_tokens_seen": 6845248, "step": 24450 }, { "epoch": 271.72222222222223, "grad_norm": 4.86884971451218e-07, "learning_rate": 1.6431868041542213e-05, "loss": 0.0, "num_input_tokens_seen": 6846656, "step": 24455 }, { "epoch": 271.77777777777777, "grad_norm": 2.8150235920065825e-08, "learning_rate": 1.6422645814284123e-05, "loss": 0.0, "num_input_tokens_seen": 6848048, "step": 24460 }, { "epoch": 271.8333333333333, "grad_norm": 1.939148290830417e-07, "learning_rate": 1.6413424909761846e-05, "loss": 0.0, "num_input_tokens_seen": 6849456, "step": 24465 }, { "epoch": 271.8888888888889, "grad_norm": 5.389368595842825e-08, "learning_rate": 1.640420532939736e-05, "loss": 0.0, "num_input_tokens_seen": 6850928, "step": 24470 }, { "epoch": 271.94444444444446, "grad_norm": 2.980339672831178e-07, "learning_rate": 1.639498707461242e-05, "loss": 0.0, "num_input_tokens_seen": 6852304, "step": 24475 }, { "epoch": 272.0, "grad_norm": 9.664517364171843e-08, "learning_rate": 1.6385770146828614e-05, "loss": 0.0, "num_input_tokens_seen": 6853712, "step": 24480 }, { "epoch": 272.05555555555554, "grad_norm": 1.885629927755872e-07, "learning_rate": 1.637655454746731e-05, "loss": 0.0, "num_input_tokens_seen": 6855120, "step": 24485 }, { "epoch": 272.1111111111111, "grad_norm": 1.3654293695708475e-07, "learning_rate": 1.6367340277949658e-05, "loss": 0.0, "num_input_tokens_seen": 6856496, "step": 24490 }, { "epoch": 272.1666666666667, "grad_norm": 3.315489038868691e-07, "learning_rate": 1.635812733969663e-05, "loss": 0.0, "num_input_tokens_seen": 6857872, "step": 24495 }, { "epoch": 272.22222222222223, "grad_norm": 4.413406884395954e-07, "learning_rate": 1.634891573412896e-05, "loss": 0.0, "num_input_tokens_seen": 6859296, "step": 24500 }, { "epoch": 272.27777777777777, "grad_norm": 1.957526478690852e-07, "learning_rate": 1.6339705462667196e-05, "loss": 0.0, "num_input_tokens_seen": 6860688, "step": 24505 }, { "epoch": 272.3333333333333, "grad_norm": 7.863364999138867e-07, "learning_rate": 1.633049652673169e-05, "loss": 0.0, "num_input_tokens_seen": 6862048, "step": 24510 }, { "epoch": 272.3888888888889, "grad_norm": 1.7611870362088666e-07, "learning_rate": 1.632128892774256e-05, "loss": 0.0, "num_input_tokens_seen": 6863520, "step": 24515 }, { "epoch": 272.44444444444446, "grad_norm": 1.8471186535862216e-07, "learning_rate": 1.6312082667119737e-05, "loss": 0.0, "num_input_tokens_seen": 6864928, "step": 24520 }, { "epoch": 272.5, "grad_norm": 4.2264986177542596e-08, "learning_rate": 1.630287774628296e-05, "loss": 0.0, "num_input_tokens_seen": 6866352, "step": 24525 }, { "epoch": 272.55555555555554, "grad_norm": 3.679015492252802e-08, "learning_rate": 1.6293674166651718e-05, "loss": 0.0, "num_input_tokens_seen": 6867744, "step": 24530 }, { "epoch": 272.6111111111111, "grad_norm": 3.940023418635974e-08, "learning_rate": 1.6284471929645338e-05, "loss": 0.0, "num_input_tokens_seen": 6869152, "step": 24535 }, { "epoch": 272.6666666666667, "grad_norm": 8.00685953095126e-08, "learning_rate": 1.627527103668291e-05, "loss": 0.0, "num_input_tokens_seen": 6870560, "step": 24540 }, { "epoch": 272.72222222222223, "grad_norm": 3.956311189767803e-08, "learning_rate": 1.6266071489183327e-05, "loss": 0.0, "num_input_tokens_seen": 6872000, "step": 24545 }, { "epoch": 272.77777777777777, "grad_norm": 1.4922331104116893e-07, "learning_rate": 1.6256873288565283e-05, "loss": 0.0, "num_input_tokens_seen": 6873360, "step": 24550 }, { "epoch": 272.8333333333333, "grad_norm": 1.81984248115441e-07, "learning_rate": 1.6247676436247245e-05, "loss": 0.0, "num_input_tokens_seen": 6874736, "step": 24555 }, { "epoch": 272.8888888888889, "grad_norm": 1.8049671268727252e-07, "learning_rate": 1.6238480933647486e-05, "loss": 0.0, "num_input_tokens_seen": 6876128, "step": 24560 }, { "epoch": 272.94444444444446, "grad_norm": 3.7525051510556295e-08, "learning_rate": 1.6229286782184083e-05, "loss": 0.0, "num_input_tokens_seen": 6877504, "step": 24565 }, { "epoch": 273.0, "grad_norm": 8.41468050793992e-08, "learning_rate": 1.622009398327487e-05, "loss": 0.0, "num_input_tokens_seen": 6878928, "step": 24570 }, { "epoch": 273.05555555555554, "grad_norm": 6.66168986640514e-08, "learning_rate": 1.6210902538337502e-05, "loss": 0.0, "num_input_tokens_seen": 6880336, "step": 24575 }, { "epoch": 273.1111111111111, "grad_norm": 1.0117901183548383e-07, "learning_rate": 1.6201712448789413e-05, "loss": 0.0, "num_input_tokens_seen": 6881760, "step": 24580 }, { "epoch": 273.1666666666667, "grad_norm": 2.8396345896908315e-07, "learning_rate": 1.6192523716047827e-05, "loss": 0.0, "num_input_tokens_seen": 6883184, "step": 24585 }, { "epoch": 273.22222222222223, "grad_norm": 6.768884759367211e-07, "learning_rate": 1.6183336341529776e-05, "loss": 0.0, "num_input_tokens_seen": 6884560, "step": 24590 }, { "epoch": 273.27777777777777, "grad_norm": 1.0217100765430587e-07, "learning_rate": 1.6174150326652047e-05, "loss": 0.0, "num_input_tokens_seen": 6885904, "step": 24595 }, { "epoch": 273.3333333333333, "grad_norm": 3.2964102558707964e-08, "learning_rate": 1.6164965672831256e-05, "loss": 0.0, "num_input_tokens_seen": 6887344, "step": 24600 }, { "epoch": 273.3333333333333, "eval_loss": 1.6198837757110596, "eval_runtime": 1.3927, "eval_samples_per_second": 28.721, "eval_steps_per_second": 14.36, "num_input_tokens_seen": 6887344, "step": 24600 }, { "epoch": 273.3888888888889, "grad_norm": 2.5963620942093257e-07, "learning_rate": 1.6155782381483784e-05, "loss": 0.0, "num_input_tokens_seen": 6888752, "step": 24605 }, { "epoch": 273.44444444444446, "grad_norm": 2.9513702770600503e-07, "learning_rate": 1.6146600454025813e-05, "loss": 0.0, "num_input_tokens_seen": 6890160, "step": 24610 }, { "epoch": 273.5, "grad_norm": 7.068997831538582e-08, "learning_rate": 1.6137419891873317e-05, "loss": 0.0, "num_input_tokens_seen": 6891568, "step": 24615 }, { "epoch": 273.55555555555554, "grad_norm": 2.694322915886005e-07, "learning_rate": 1.6128240696442038e-05, "loss": 0.0, "num_input_tokens_seen": 6893008, "step": 24620 }, { "epoch": 273.6111111111111, "grad_norm": 3.7844667843955904e-08, "learning_rate": 1.611906286914753e-05, "loss": 0.0, "num_input_tokens_seen": 6894384, "step": 24625 }, { "epoch": 273.6666666666667, "grad_norm": 9.788224986095884e-08, "learning_rate": 1.6109886411405144e-05, "loss": 0.0, "num_input_tokens_seen": 6895792, "step": 24630 }, { "epoch": 273.72222222222223, "grad_norm": 2.16104410810658e-07, "learning_rate": 1.6100711324629985e-05, "loss": 0.0, "num_input_tokens_seen": 6897200, "step": 24635 }, { "epoch": 273.77777777777777, "grad_norm": 4.7501508504410594e-08, "learning_rate": 1.609153761023698e-05, "loss": 0.0, "num_input_tokens_seen": 6898624, "step": 24640 }, { "epoch": 273.8333333333333, "grad_norm": 1.663213566871491e-07, "learning_rate": 1.608236526964083e-05, "loss": 0.0, "num_input_tokens_seen": 6900032, "step": 24645 }, { "epoch": 273.8888888888889, "grad_norm": 3.9198445733745757e-07, "learning_rate": 1.607319430425601e-05, "loss": 0.0, "num_input_tokens_seen": 6901440, "step": 24650 }, { "epoch": 273.94444444444446, "grad_norm": 4.935926085636311e-07, "learning_rate": 1.606402471549682e-05, "loss": 0.0, "num_input_tokens_seen": 6902848, "step": 24655 }, { "epoch": 274.0, "grad_norm": 1.1162661195385226e-07, "learning_rate": 1.6054856504777312e-05, "loss": 0.0, "num_input_tokens_seen": 6904192, "step": 24660 }, { "epoch": 274.05555555555554, "grad_norm": 7.84088030059138e-08, "learning_rate": 1.6045689673511334e-05, "loss": 0.0, "num_input_tokens_seen": 6905600, "step": 24665 }, { "epoch": 274.1111111111111, "grad_norm": 3.0154734531606664e-07, "learning_rate": 1.6036524223112548e-05, "loss": 0.0, "num_input_tokens_seen": 6907024, "step": 24670 }, { "epoch": 274.1666666666667, "grad_norm": 5.972012218080636e-07, "learning_rate": 1.602736015499436e-05, "loss": 0.0, "num_input_tokens_seen": 6908448, "step": 24675 }, { "epoch": 274.22222222222223, "grad_norm": 2.534133614062739e-07, "learning_rate": 1.601819747057e-05, "loss": 0.0, "num_input_tokens_seen": 6909856, "step": 24680 }, { "epoch": 274.27777777777777, "grad_norm": 5.612561722045939e-07, "learning_rate": 1.6009036171252465e-05, "loss": 0.0, "num_input_tokens_seen": 6911248, "step": 24685 }, { "epoch": 274.3333333333333, "grad_norm": 2.2014859268892906e-07, "learning_rate": 1.599987625845453e-05, "loss": 0.0, "num_input_tokens_seen": 6912656, "step": 24690 }, { "epoch": 274.3888888888889, "grad_norm": 1.2434465190835908e-07, "learning_rate": 1.599071773358879e-05, "loss": 0.0, "num_input_tokens_seen": 6914096, "step": 24695 }, { "epoch": 274.44444444444446, "grad_norm": 2.8915974326082505e-07, "learning_rate": 1.598156059806758e-05, "loss": 0.0, "num_input_tokens_seen": 6915456, "step": 24700 }, { "epoch": 274.5, "grad_norm": 3.703819473344083e-08, "learning_rate": 1.5972404853303062e-05, "loss": 0.0, "num_input_tokens_seen": 6916848, "step": 24705 }, { "epoch": 274.55555555555554, "grad_norm": 1.3886057104173233e-07, "learning_rate": 1.5963250500707172e-05, "loss": 0.0, "num_input_tokens_seen": 6918272, "step": 24710 }, { "epoch": 274.6111111111111, "grad_norm": 7.354543640758493e-08, "learning_rate": 1.5954097541691612e-05, "loss": 0.0, "num_input_tokens_seen": 6919728, "step": 24715 }, { "epoch": 274.6666666666667, "grad_norm": 7.642840671451268e-08, "learning_rate": 1.5944945977667884e-05, "loss": 0.0, "num_input_tokens_seen": 6921120, "step": 24720 }, { "epoch": 274.72222222222223, "grad_norm": 5.058709007244033e-07, "learning_rate": 1.593579581004729e-05, "loss": 0.0, "num_input_tokens_seen": 6922448, "step": 24725 }, { "epoch": 274.77777777777777, "grad_norm": 1.2836210316891083e-07, "learning_rate": 1.592664704024088e-05, "loss": 0.0, "num_input_tokens_seen": 6923840, "step": 24730 }, { "epoch": 274.8333333333333, "grad_norm": 4.45152465999854e-08, "learning_rate": 1.591749966965953e-05, "loss": 0.0, "num_input_tokens_seen": 6925264, "step": 24735 }, { "epoch": 274.8888888888889, "grad_norm": 3.878952625768761e-08, "learning_rate": 1.5908353699713856e-05, "loss": 0.0, "num_input_tokens_seen": 6926640, "step": 24740 }, { "epoch": 274.94444444444446, "grad_norm": 6.988616974012984e-07, "learning_rate": 1.5899209131814298e-05, "loss": 0.0, "num_input_tokens_seen": 6928032, "step": 24745 }, { "epoch": 275.0, "grad_norm": 5.662762987412862e-07, "learning_rate": 1.5890065967371067e-05, "loss": 0.0, "num_input_tokens_seen": 6929456, "step": 24750 }, { "epoch": 275.05555555555554, "grad_norm": 2.1411361217360536e-07, "learning_rate": 1.5880924207794144e-05, "loss": 0.0, "num_input_tokens_seen": 6930880, "step": 24755 }, { "epoch": 275.1111111111111, "grad_norm": 4.339796788599415e-08, "learning_rate": 1.5871783854493298e-05, "loss": 0.0, "num_input_tokens_seen": 6932336, "step": 24760 }, { "epoch": 275.1666666666667, "grad_norm": 5.052090727986069e-07, "learning_rate": 1.5862644908878106e-05, "loss": 0.0, "num_input_tokens_seen": 6933728, "step": 24765 }, { "epoch": 275.22222222222223, "grad_norm": 1.473789126293923e-07, "learning_rate": 1.5853507372357885e-05, "loss": 0.0, "num_input_tokens_seen": 6935200, "step": 24770 }, { "epoch": 275.27777777777777, "grad_norm": 1.0290144700775272e-06, "learning_rate": 1.5844371246341776e-05, "loss": 0.0, "num_input_tokens_seen": 6936608, "step": 24775 }, { "epoch": 275.3333333333333, "grad_norm": 1.0829236885001592e-07, "learning_rate": 1.5835236532238674e-05, "loss": 0.0, "num_input_tokens_seen": 6938048, "step": 24780 }, { "epoch": 275.3888888888889, "grad_norm": 1.971747707329996e-07, "learning_rate": 1.582610323145727e-05, "loss": 0.0, "num_input_tokens_seen": 6939440, "step": 24785 }, { "epoch": 275.44444444444446, "grad_norm": 2.2349433947965736e-07, "learning_rate": 1.5816971345406035e-05, "loss": 0.0, "num_input_tokens_seen": 6940784, "step": 24790 }, { "epoch": 275.5, "grad_norm": 1.9987884058991767e-07, "learning_rate": 1.5807840875493225e-05, "loss": 0.0, "num_input_tokens_seen": 6942208, "step": 24795 }, { "epoch": 275.55555555555554, "grad_norm": 1.980130122092305e-07, "learning_rate": 1.5798711823126854e-05, "loss": 0.0, "num_input_tokens_seen": 6943632, "step": 24800 }, { "epoch": 275.55555555555554, "eval_loss": 1.5672613382339478, "eval_runtime": 1.3954, "eval_samples_per_second": 28.665, "eval_steps_per_second": 14.333, "num_input_tokens_seen": 6943632, "step": 24800 }, { "epoch": 275.6111111111111, "grad_norm": 1.6388810308853863e-07, "learning_rate": 1.578958418971477e-05, "loss": 0.0, "num_input_tokens_seen": 6945024, "step": 24805 }, { "epoch": 275.6666666666667, "grad_norm": 8.392150618874439e-08, "learning_rate": 1.578045797666453e-05, "loss": 0.0, "num_input_tokens_seen": 6946400, "step": 24810 }, { "epoch": 275.72222222222223, "grad_norm": 2.1863171184577368e-07, "learning_rate": 1.5771333185383548e-05, "loss": 0.0, "num_input_tokens_seen": 6947760, "step": 24815 }, { "epoch": 275.77777777777777, "grad_norm": 3.622930364599597e-07, "learning_rate": 1.576220981727895e-05, "loss": 0.0, "num_input_tokens_seen": 6949152, "step": 24820 }, { "epoch": 275.8333333333333, "grad_norm": 2.973958714846958e-07, "learning_rate": 1.575308787375769e-05, "loss": 0.0, "num_input_tokens_seen": 6950560, "step": 24825 }, { "epoch": 275.8888888888889, "grad_norm": 4.679182438849239e-06, "learning_rate": 1.5743967356226492e-05, "loss": 0.0, "num_input_tokens_seen": 6951952, "step": 24830 }, { "epoch": 275.94444444444446, "grad_norm": 1.50965297507355e-06, "learning_rate": 1.5734848266091835e-05, "loss": 0.0, "num_input_tokens_seen": 6953344, "step": 24835 }, { "epoch": 276.0, "grad_norm": 1.0314025757907075e-06, "learning_rate": 1.572573060476001e-05, "loss": 0.0, "num_input_tokens_seen": 6954752, "step": 24840 }, { "epoch": 276.05555555555554, "grad_norm": 1.7960558125196258e-06, "learning_rate": 1.5716614373637085e-05, "loss": 0.0, "num_input_tokens_seen": 6956160, "step": 24845 }, { "epoch": 276.1111111111111, "grad_norm": 4.923551273350313e-07, "learning_rate": 1.570749957412887e-05, "loss": 0.0, "num_input_tokens_seen": 6957536, "step": 24850 }, { "epoch": 276.1666666666667, "grad_norm": 4.752878339786548e-07, "learning_rate": 1.5698386207641013e-05, "loss": 0.0, "num_input_tokens_seen": 6958960, "step": 24855 }, { "epoch": 276.22222222222223, "grad_norm": 4.6470825054711895e-07, "learning_rate": 1.5689274275578884e-05, "loss": 0.0, "num_input_tokens_seen": 6960320, "step": 24860 }, { "epoch": 276.27777777777777, "grad_norm": 1.441779744482119e-07, "learning_rate": 1.5680163779347667e-05, "loss": 0.0, "num_input_tokens_seen": 6961696, "step": 24865 }, { "epoch": 276.3333333333333, "grad_norm": 3.7470285860763397e-06, "learning_rate": 1.5671054720352327e-05, "loss": 0.0, "num_input_tokens_seen": 6963136, "step": 24870 }, { "epoch": 276.3888888888889, "grad_norm": 4.007769405234285e-07, "learning_rate": 1.566194709999757e-05, "loss": 0.0, "num_input_tokens_seen": 6964560, "step": 24875 }, { "epoch": 276.44444444444446, "grad_norm": 3.192368467352935e-07, "learning_rate": 1.5652840919687933e-05, "loss": 0.0, "num_input_tokens_seen": 6965952, "step": 24880 }, { "epoch": 276.5, "grad_norm": 7.84253941787938e-08, "learning_rate": 1.5643736180827676e-05, "loss": 0.0, "num_input_tokens_seen": 6967344, "step": 24885 }, { "epoch": 276.55555555555554, "grad_norm": 8.576990495612336e-08, "learning_rate": 1.5634632884820878e-05, "loss": 0.0, "num_input_tokens_seen": 6968768, "step": 24890 }, { "epoch": 276.6111111111111, "grad_norm": 2.7740876262782876e-08, "learning_rate": 1.5625531033071395e-05, "loss": 0.0, "num_input_tokens_seen": 6970128, "step": 24895 }, { "epoch": 276.6666666666667, "grad_norm": 1.969989398276084e-07, "learning_rate": 1.5616430626982828e-05, "loss": 0.0, "num_input_tokens_seen": 6971568, "step": 24900 }, { "epoch": 276.72222222222223, "grad_norm": 9.371651543688131e-08, "learning_rate": 1.5607331667958575e-05, "loss": 0.0, "num_input_tokens_seen": 6972992, "step": 24905 }, { "epoch": 276.77777777777777, "grad_norm": 7.05572205106364e-08, "learning_rate": 1.5598234157401824e-05, "loss": 0.0, "num_input_tokens_seen": 6974368, "step": 24910 }, { "epoch": 276.8333333333333, "grad_norm": 8.373044835252585e-08, "learning_rate": 1.5589138096715503e-05, "loss": 0.0, "num_input_tokens_seen": 6975792, "step": 24915 }, { "epoch": 276.8888888888889, "grad_norm": 4.245909579481122e-08, "learning_rate": 1.5580043487302365e-05, "loss": 0.0, "num_input_tokens_seen": 6977184, "step": 24920 }, { "epoch": 276.94444444444446, "grad_norm": 8.86435245206485e-08, "learning_rate": 1.5570950330564888e-05, "loss": 0.0, "num_input_tokens_seen": 6978592, "step": 24925 }, { "epoch": 277.0, "grad_norm": 1.3127143461133528e-07, "learning_rate": 1.5561858627905367e-05, "loss": 0.0, "num_input_tokens_seen": 6980000, "step": 24930 }, { "epoch": 277.05555555555554, "grad_norm": 3.273714810347883e-08, "learning_rate": 1.5552768380725857e-05, "loss": 0.0, "num_input_tokens_seen": 6981392, "step": 24935 }, { "epoch": 277.1111111111111, "grad_norm": 3.1158765523287e-07, "learning_rate": 1.5543679590428183e-05, "loss": 0.0, "num_input_tokens_seen": 6982768, "step": 24940 }, { "epoch": 277.1666666666667, "grad_norm": 6.47224212002584e-08, "learning_rate": 1.5534592258413943e-05, "loss": 0.0, "num_input_tokens_seen": 6984160, "step": 24945 }, { "epoch": 277.22222222222223, "grad_norm": 7.361722964560613e-08, "learning_rate": 1.5525506386084538e-05, "loss": 0.0, "num_input_tokens_seen": 6985568, "step": 24950 }, { "epoch": 277.27777777777777, "grad_norm": 7.047372463375723e-08, "learning_rate": 1.55164219748411e-05, "loss": 0.0, "num_input_tokens_seen": 6986976, "step": 24955 }, { "epoch": 277.3333333333333, "grad_norm": 5.53084120724634e-08, "learning_rate": 1.550733902608459e-05, "loss": 0.0, "num_input_tokens_seen": 6988400, "step": 24960 }, { "epoch": 277.3888888888889, "grad_norm": 2.425335026146058e-07, "learning_rate": 1.549825754121568e-05, "loss": 0.0, "num_input_tokens_seen": 6989792, "step": 24965 }, { "epoch": 277.44444444444446, "grad_norm": 1.8330460704873985e-07, "learning_rate": 1.5489177521634864e-05, "loss": 0.0, "num_input_tokens_seen": 6991200, "step": 24970 }, { "epoch": 277.5, "grad_norm": 2.1333251254418428e-07, "learning_rate": 1.5480098968742402e-05, "loss": 0.0, "num_input_tokens_seen": 6992592, "step": 24975 }, { "epoch": 277.55555555555554, "grad_norm": 2.474939151397848e-07, "learning_rate": 1.5471021883938304e-05, "loss": 0.0, "num_input_tokens_seen": 6994016, "step": 24980 }, { "epoch": 277.6111111111111, "grad_norm": 6.639150456067e-07, "learning_rate": 1.546194626862238e-05, "loss": 0.0, "num_input_tokens_seen": 6995408, "step": 24985 }, { "epoch": 277.6666666666667, "grad_norm": 4.346954483480658e-06, "learning_rate": 1.5452872124194216e-05, "loss": 0.0, "num_input_tokens_seen": 6996864, "step": 24990 }, { "epoch": 277.72222222222223, "grad_norm": 1.7922266692949052e-07, "learning_rate": 1.5443799452053136e-05, "loss": 0.0, "num_input_tokens_seen": 6998256, "step": 24995 }, { "epoch": 277.77777777777777, "grad_norm": 1.6745025277487002e-07, "learning_rate": 1.543472825359828e-05, "loss": 0.0, "num_input_tokens_seen": 6999632, "step": 25000 }, { "epoch": 277.77777777777777, "eval_loss": 1.5776519775390625, "eval_runtime": 1.3912, "eval_samples_per_second": 28.751, "eval_steps_per_second": 14.376, "num_input_tokens_seen": 6999632, "step": 25000 }, { "epoch": 277.8333333333333, "grad_norm": 1.8020753600467287e-07, "learning_rate": 1.5425658530228522e-05, "loss": 0.0, "num_input_tokens_seen": 7001024, "step": 25005 }, { "epoch": 277.8888888888889, "grad_norm": 4.341025316989544e-08, "learning_rate": 1.5416590283342546e-05, "loss": 0.0, "num_input_tokens_seen": 7002400, "step": 25010 }, { "epoch": 277.94444444444446, "grad_norm": 8.76782522141184e-08, "learning_rate": 1.5407523514338783e-05, "loss": 0.0, "num_input_tokens_seen": 7003824, "step": 25015 }, { "epoch": 278.0, "grad_norm": 1.1654618248257975e-07, "learning_rate": 1.539845822461543e-05, "loss": 0.0, "num_input_tokens_seen": 7005184, "step": 25020 }, { "epoch": 278.05555555555554, "grad_norm": 2.6344767434238747e-07, "learning_rate": 1.538939441557048e-05, "loss": 0.0, "num_input_tokens_seen": 7006608, "step": 25025 }, { "epoch": 278.1111111111111, "grad_norm": 5.302546313146195e-08, "learning_rate": 1.5380332088601696e-05, "loss": 0.0, "num_input_tokens_seen": 7008000, "step": 25030 }, { "epoch": 278.1666666666667, "grad_norm": 1.46266501133141e-07, "learning_rate": 1.537127124510658e-05, "loss": 0.0, "num_input_tokens_seen": 7009424, "step": 25035 }, { "epoch": 278.22222222222223, "grad_norm": 6.970271471118394e-08, "learning_rate": 1.5362211886482457e-05, "loss": 0.0, "num_input_tokens_seen": 7010816, "step": 25040 }, { "epoch": 278.27777777777777, "grad_norm": 7.419141212494651e-08, "learning_rate": 1.5353154014126363e-05, "loss": 0.0, "num_input_tokens_seen": 7012192, "step": 25045 }, { "epoch": 278.3333333333333, "grad_norm": 2.2553696510385635e-07, "learning_rate": 1.534409762943515e-05, "loss": 0.0, "num_input_tokens_seen": 7013568, "step": 25050 }, { "epoch": 278.3888888888889, "grad_norm": 9.279985135890456e-08, "learning_rate": 1.5335042733805438e-05, "loss": 0.0, "num_input_tokens_seen": 7014976, "step": 25055 }, { "epoch": 278.44444444444446, "grad_norm": 7.242746846713999e-08, "learning_rate": 1.532598932863358e-05, "loss": 0.0, "num_input_tokens_seen": 7016384, "step": 25060 }, { "epoch": 278.5, "grad_norm": 1.2273650895622268e-07, "learning_rate": 1.531693741531574e-05, "loss": 0.0, "num_input_tokens_seen": 7017824, "step": 25065 }, { "epoch": 278.55555555555554, "grad_norm": 2.096210920399244e-07, "learning_rate": 1.5307886995247844e-05, "loss": 0.0, "num_input_tokens_seen": 7019184, "step": 25070 }, { "epoch": 278.6111111111111, "grad_norm": 4.289460662221245e-07, "learning_rate": 1.529883806982557e-05, "loss": 0.0, "num_input_tokens_seen": 7020608, "step": 25075 }, { "epoch": 278.6666666666667, "grad_norm": 6.616398309233773e-07, "learning_rate": 1.5289790640444376e-05, "loss": 0.0, "num_input_tokens_seen": 7022016, "step": 25080 }, { "epoch": 278.72222222222223, "grad_norm": 1.9022235164811718e-07, "learning_rate": 1.5280744708499494e-05, "loss": 0.0, "num_input_tokens_seen": 7023408, "step": 25085 }, { "epoch": 278.77777777777777, "grad_norm": 5.030452143728326e-07, "learning_rate": 1.527170027538591e-05, "loss": 0.0, "num_input_tokens_seen": 7024800, "step": 25090 }, { "epoch": 278.8333333333333, "grad_norm": 5.791468424831692e-07, "learning_rate": 1.5262657342498407e-05, "loss": 0.0, "num_input_tokens_seen": 7026208, "step": 25095 }, { "epoch": 278.8888888888889, "grad_norm": 5.371250608732225e-07, "learning_rate": 1.52536159112315e-05, "loss": 0.0, "num_input_tokens_seen": 7027568, "step": 25100 }, { "epoch": 278.94444444444446, "grad_norm": 1.9310469667743746e-07, "learning_rate": 1.5244575982979497e-05, "loss": 0.0, "num_input_tokens_seen": 7028992, "step": 25105 }, { "epoch": 279.0, "grad_norm": 5.100552584735851e-07, "learning_rate": 1.5235537559136487e-05, "loss": 0.0, "num_input_tokens_seen": 7030400, "step": 25110 }, { "epoch": 279.05555555555554, "grad_norm": 6.401914447451418e-07, "learning_rate": 1.5226500641096286e-05, "loss": 0.0, "num_input_tokens_seen": 7031792, "step": 25115 }, { "epoch": 279.1111111111111, "grad_norm": 2.2293791346328362e-07, "learning_rate": 1.5217465230252509e-05, "loss": 0.0, "num_input_tokens_seen": 7033136, "step": 25120 }, { "epoch": 279.1666666666667, "grad_norm": 5.398262032940693e-07, "learning_rate": 1.5208431327998523e-05, "loss": 0.0, "num_input_tokens_seen": 7034560, "step": 25125 }, { "epoch": 279.22222222222223, "grad_norm": 6.165342938402318e-07, "learning_rate": 1.5199398935727477e-05, "loss": 0.0, "num_input_tokens_seen": 7036016, "step": 25130 }, { "epoch": 279.27777777777777, "grad_norm": 3.433333972679975e-07, "learning_rate": 1.5190368054832282e-05, "loss": 0.0, "num_input_tokens_seen": 7037392, "step": 25135 }, { "epoch": 279.3333333333333, "grad_norm": 1.5867686670389958e-05, "learning_rate": 1.5181338686705601e-05, "loss": 0.0, "num_input_tokens_seen": 7038816, "step": 25140 }, { "epoch": 279.3888888888889, "grad_norm": 1.354267453734792e-07, "learning_rate": 1.5172310832739889e-05, "loss": 0.0, "num_input_tokens_seen": 7040192, "step": 25145 }, { "epoch": 279.44444444444446, "grad_norm": 1.0947768913638356e-07, "learning_rate": 1.5163284494327346e-05, "loss": 0.0, "num_input_tokens_seen": 7041552, "step": 25150 }, { "epoch": 279.5, "grad_norm": 1.7594493328942917e-07, "learning_rate": 1.5154259672859952e-05, "loss": 0.0, "num_input_tokens_seen": 7042960, "step": 25155 }, { "epoch": 279.55555555555554, "grad_norm": 1.0409338813133218e-07, "learning_rate": 1.5145236369729452e-05, "loss": 0.0, "num_input_tokens_seen": 7044400, "step": 25160 }, { "epoch": 279.6111111111111, "grad_norm": 9.050221478901221e-07, "learning_rate": 1.5136214586327335e-05, "loss": 0.0, "num_input_tokens_seen": 7045824, "step": 25165 }, { "epoch": 279.6666666666667, "grad_norm": 1.5974592315615155e-07, "learning_rate": 1.5127194324044885e-05, "loss": 0.0, "num_input_tokens_seen": 7047200, "step": 25170 }, { "epoch": 279.72222222222223, "grad_norm": 4.307622987198556e-07, "learning_rate": 1.5118175584273148e-05, "loss": 0.0, "num_input_tokens_seen": 7048592, "step": 25175 }, { "epoch": 279.77777777777777, "grad_norm": 4.191444133994082e-07, "learning_rate": 1.5109158368402909e-05, "loss": 0.0, "num_input_tokens_seen": 7050032, "step": 25180 }, { "epoch": 279.8333333333333, "grad_norm": 3.6158414786768844e-06, "learning_rate": 1.5100142677824753e-05, "loss": 0.0, "num_input_tokens_seen": 7051424, "step": 25185 }, { "epoch": 279.8888888888889, "grad_norm": 6.553040066137328e-07, "learning_rate": 1.509112851392901e-05, "loss": 0.0, "num_input_tokens_seen": 7052864, "step": 25190 }, { "epoch": 279.94444444444446, "grad_norm": 1.5214618542813696e-07, "learning_rate": 1.5082115878105763e-05, "loss": 0.0, "num_input_tokens_seen": 7054272, "step": 25195 }, { "epoch": 280.0, "grad_norm": 1.1837339997100571e-07, "learning_rate": 1.5073104771744892e-05, "loss": 0.0, "num_input_tokens_seen": 7055664, "step": 25200 }, { "epoch": 280.0, "eval_loss": 1.612060546875, "eval_runtime": 1.3963, "eval_samples_per_second": 28.646, "eval_steps_per_second": 14.323, "num_input_tokens_seen": 7055664, "step": 25200 }, { "epoch": 280.05555555555554, "grad_norm": 1.1914940500901139e-07, "learning_rate": 1.5064095196236006e-05, "loss": 0.0, "num_input_tokens_seen": 7057072, "step": 25205 }, { "epoch": 280.1111111111111, "grad_norm": 6.890341808230005e-08, "learning_rate": 1.50550871529685e-05, "loss": 0.0, "num_input_tokens_seen": 7058496, "step": 25210 }, { "epoch": 280.1666666666667, "grad_norm": 2.9225776643215795e-07, "learning_rate": 1.5046080643331546e-05, "loss": 0.0, "num_input_tokens_seen": 7059920, "step": 25215 }, { "epoch": 280.22222222222223, "grad_norm": 3.885303101469617e-07, "learning_rate": 1.5037075668714028e-05, "loss": 0.0, "num_input_tokens_seen": 7061360, "step": 25220 }, { "epoch": 280.27777777777777, "grad_norm": 5.418311843641277e-07, "learning_rate": 1.5028072230504656e-05, "loss": 0.0, "num_input_tokens_seen": 7062720, "step": 25225 }, { "epoch": 280.3333333333333, "grad_norm": 6.682407160951698e-07, "learning_rate": 1.5019070330091861e-05, "loss": 0.0, "num_input_tokens_seen": 7064160, "step": 25230 }, { "epoch": 280.3888888888889, "grad_norm": 2.0232175756973447e-07, "learning_rate": 1.5010069968863843e-05, "loss": 0.0, "num_input_tokens_seen": 7065536, "step": 25235 }, { "epoch": 280.44444444444446, "grad_norm": 5.178290507501515e-07, "learning_rate": 1.5001071148208584e-05, "loss": 0.0, "num_input_tokens_seen": 7066912, "step": 25240 }, { "epoch": 280.5, "grad_norm": 4.3671423100022366e-07, "learning_rate": 1.49920738695138e-05, "loss": 0.0, "num_input_tokens_seen": 7068304, "step": 25245 }, { "epoch": 280.55555555555554, "grad_norm": 2.6729370006250974e-07, "learning_rate": 1.4983078134166995e-05, "loss": 0.0, "num_input_tokens_seen": 7069680, "step": 25250 }, { "epoch": 280.6111111111111, "grad_norm": 4.026994417927199e-07, "learning_rate": 1.4974083943555428e-05, "loss": 0.0, "num_input_tokens_seen": 7071104, "step": 25255 }, { "epoch": 280.6666666666667, "grad_norm": 5.549940738092118e-07, "learning_rate": 1.496509129906611e-05, "loss": 0.0, "num_input_tokens_seen": 7072544, "step": 25260 }, { "epoch": 280.72222222222223, "grad_norm": 1.943700880246979e-07, "learning_rate": 1.4956100202085809e-05, "loss": 0.0, "num_input_tokens_seen": 7073920, "step": 25265 }, { "epoch": 280.77777777777777, "grad_norm": 3.562963968306576e-07, "learning_rate": 1.4947110654001093e-05, "loss": 0.0, "num_input_tokens_seen": 7075296, "step": 25270 }, { "epoch": 280.8333333333333, "grad_norm": 2.744339724358724e-07, "learning_rate": 1.4938122656198234e-05, "loss": 0.0, "num_input_tokens_seen": 7076656, "step": 25275 }, { "epoch": 280.8888888888889, "grad_norm": 6.979522737537991e-08, "learning_rate": 1.4929136210063316e-05, "loss": 0.0, "num_input_tokens_seen": 7078032, "step": 25280 }, { "epoch": 280.94444444444446, "grad_norm": 3.500821890156658e-07, "learning_rate": 1.4920151316982146e-05, "loss": 0.0, "num_input_tokens_seen": 7079440, "step": 25285 }, { "epoch": 281.0, "grad_norm": 5.559328997151169e-07, "learning_rate": 1.4911167978340312e-05, "loss": 0.0, "num_input_tokens_seen": 7080832, "step": 25290 }, { "epoch": 281.05555555555554, "grad_norm": 9.98689756670501e-07, "learning_rate": 1.4902186195523166e-05, "loss": 0.0, "num_input_tokens_seen": 7082224, "step": 25295 }, { "epoch": 281.1111111111111, "grad_norm": 1.738594392008963e-06, "learning_rate": 1.4893205969915805e-05, "loss": 0.0, "num_input_tokens_seen": 7083664, "step": 25300 }, { "epoch": 281.1666666666667, "grad_norm": 9.032885827764403e-06, "learning_rate": 1.4884227302903086e-05, "loss": 0.0, "num_input_tokens_seen": 7085120, "step": 25305 }, { "epoch": 281.22222222222223, "grad_norm": 2.850209455118602e-07, "learning_rate": 1.4875250195869653e-05, "loss": 0.0, "num_input_tokens_seen": 7086528, "step": 25310 }, { "epoch": 281.27777777777777, "grad_norm": 2.527716844724637e-07, "learning_rate": 1.4866274650199862e-05, "loss": 0.0, "num_input_tokens_seen": 7087920, "step": 25315 }, { "epoch": 281.3333333333333, "grad_norm": 5.996651566420041e-07, "learning_rate": 1.485730066727788e-05, "loss": 0.0, "num_input_tokens_seen": 7089312, "step": 25320 }, { "epoch": 281.3888888888889, "grad_norm": 2.1950084772015543e-07, "learning_rate": 1.4848328248487586e-05, "loss": 0.0, "num_input_tokens_seen": 7090672, "step": 25325 }, { "epoch": 281.44444444444446, "grad_norm": 4.372172952571418e-07, "learning_rate": 1.4839357395212656e-05, "loss": 0.0, "num_input_tokens_seen": 7092096, "step": 25330 }, { "epoch": 281.5, "grad_norm": 3.9976788457352086e-07, "learning_rate": 1.4830388108836502e-05, "loss": 0.0, "num_input_tokens_seen": 7093472, "step": 25335 }, { "epoch": 281.55555555555554, "grad_norm": 6.014622044858697e-07, "learning_rate": 1.4821420390742299e-05, "loss": 0.0, "num_input_tokens_seen": 7094848, "step": 25340 }, { "epoch": 281.6111111111111, "grad_norm": 1.4727807240433322e-07, "learning_rate": 1.4812454242312979e-05, "loss": 0.0, "num_input_tokens_seen": 7096272, "step": 25345 }, { "epoch": 281.6666666666667, "grad_norm": 6.190794579197245e-07, "learning_rate": 1.4803489664931253e-05, "loss": 0.0, "num_input_tokens_seen": 7097648, "step": 25350 }, { "epoch": 281.72222222222223, "grad_norm": 6.508836918328598e-07, "learning_rate": 1.4794526659979544e-05, "loss": 0.0, "num_input_tokens_seen": 7099088, "step": 25355 }, { "epoch": 281.77777777777777, "grad_norm": 1.1528029517648974e-06, "learning_rate": 1.4785565228840086e-05, "loss": 0.0, "num_input_tokens_seen": 7100496, "step": 25360 }, { "epoch": 281.8333333333333, "grad_norm": 8.091242875707394e-07, "learning_rate": 1.4776605372894819e-05, "loss": 0.0, "num_input_tokens_seen": 7101904, "step": 25365 }, { "epoch": 281.8888888888889, "grad_norm": 5.31609111931175e-06, "learning_rate": 1.4767647093525488e-05, "loss": 0.0, "num_input_tokens_seen": 7103296, "step": 25370 }, { "epoch": 281.94444444444446, "grad_norm": 2.863244503714668e-07, "learning_rate": 1.4758690392113566e-05, "loss": 0.0, "num_input_tokens_seen": 7104656, "step": 25375 }, { "epoch": 282.0, "grad_norm": 2.415005155853578e-07, "learning_rate": 1.4749735270040276e-05, "loss": 0.0, "num_input_tokens_seen": 7106048, "step": 25380 }, { "epoch": 282.05555555555554, "grad_norm": 1.7261255891298788e-07, "learning_rate": 1.4740781728686623e-05, "loss": 0.0, "num_input_tokens_seen": 7107472, "step": 25385 }, { "epoch": 282.1111111111111, "grad_norm": 8.646454716654262e-07, "learning_rate": 1.4731829769433358e-05, "loss": 0.0, "num_input_tokens_seen": 7108832, "step": 25390 }, { "epoch": 282.1666666666667, "grad_norm": 6.582995411008596e-07, "learning_rate": 1.4722879393660976e-05, "loss": 0.0, "num_input_tokens_seen": 7110256, "step": 25395 }, { "epoch": 282.22222222222223, "grad_norm": 5.830704594700364e-07, "learning_rate": 1.4713930602749748e-05, "loss": 0.0, "num_input_tokens_seen": 7111664, "step": 25400 }, { "epoch": 282.22222222222223, "eval_loss": 1.636621117591858, "eval_runtime": 1.3979, "eval_samples_per_second": 28.613, "eval_steps_per_second": 14.307, "num_input_tokens_seen": 7111664, "step": 25400 }, { "epoch": 282.27777777777777, "grad_norm": 6.663665885753289e-07, "learning_rate": 1.470498339807968e-05, "loss": 0.0, "num_input_tokens_seen": 7113072, "step": 25405 }, { "epoch": 282.3333333333333, "grad_norm": 6.178123612698982e-07, "learning_rate": 1.4696037781030542e-05, "loss": 0.0, "num_input_tokens_seen": 7114480, "step": 25410 }, { "epoch": 282.3888888888889, "grad_norm": 2.1310169984189997e-07, "learning_rate": 1.4687093752981876e-05, "loss": 0.0, "num_input_tokens_seen": 7115888, "step": 25415 }, { "epoch": 282.44444444444446, "grad_norm": 2.8880907620987273e-07, "learning_rate": 1.4678151315312943e-05, "loss": 0.0, "num_input_tokens_seen": 7117280, "step": 25420 }, { "epoch": 282.5, "grad_norm": 4.15383709650996e-07, "learning_rate": 1.4669210469402789e-05, "loss": 0.0, "num_input_tokens_seen": 7118656, "step": 25425 }, { "epoch": 282.55555555555554, "grad_norm": 4.7609049147467886e-07, "learning_rate": 1.4660271216630218e-05, "loss": 0.0, "num_input_tokens_seen": 7120048, "step": 25430 }, { "epoch": 282.6111111111111, "grad_norm": 7.394618251055363e-07, "learning_rate": 1.4651333558373748e-05, "loss": 0.0, "num_input_tokens_seen": 7121440, "step": 25435 }, { "epoch": 282.6666666666667, "grad_norm": 4.503382626808161e-07, "learning_rate": 1.4642397496011707e-05, "loss": 0.0, "num_input_tokens_seen": 7122880, "step": 25440 }, { "epoch": 282.72222222222223, "grad_norm": 4.330685214881669e-07, "learning_rate": 1.4633463030922129e-05, "loss": 0.0, "num_input_tokens_seen": 7124304, "step": 25445 }, { "epoch": 282.77777777777777, "grad_norm": 1.8269996360231744e-07, "learning_rate": 1.462453016448282e-05, "loss": 0.0, "num_input_tokens_seen": 7125696, "step": 25450 }, { "epoch": 282.8333333333333, "grad_norm": 3.5173582091374556e-07, "learning_rate": 1.4615598898071354e-05, "loss": 0.0, "num_input_tokens_seen": 7127120, "step": 25455 }, { "epoch": 282.8888888888889, "grad_norm": 4.4667845600088185e-07, "learning_rate": 1.4606669233065026e-05, "loss": 0.0, "num_input_tokens_seen": 7128496, "step": 25460 }, { "epoch": 282.94444444444446, "grad_norm": 3.5959197930424125e-07, "learning_rate": 1.4597741170840914e-05, "loss": 0.0, "num_input_tokens_seen": 7129936, "step": 25465 }, { "epoch": 283.0, "grad_norm": 3.781231328048307e-07, "learning_rate": 1.4588814712775853e-05, "loss": 0.0, "num_input_tokens_seen": 7131328, "step": 25470 }, { "epoch": 283.05555555555554, "grad_norm": 7.039905653982714e-07, "learning_rate": 1.4579889860246382e-05, "loss": 0.0, "num_input_tokens_seen": 7132720, "step": 25475 }, { "epoch": 283.1111111111111, "grad_norm": 9.950934327207506e-06, "learning_rate": 1.457096661462885e-05, "loss": 0.0, "num_input_tokens_seen": 7134096, "step": 25480 }, { "epoch": 283.1666666666667, "grad_norm": 7.638611236870929e-07, "learning_rate": 1.4562044977299322e-05, "loss": 0.0, "num_input_tokens_seen": 7135520, "step": 25485 }, { "epoch": 283.22222222222223, "grad_norm": 6.030363692843821e-07, "learning_rate": 1.4553124949633623e-05, "loss": 0.0, "num_input_tokens_seen": 7136944, "step": 25490 }, { "epoch": 283.27777777777777, "grad_norm": 3.5340005410944286e-07, "learning_rate": 1.4544206533007354e-05, "loss": 0.0, "num_input_tokens_seen": 7138352, "step": 25495 }, { "epoch": 283.3333333333333, "grad_norm": 1.7714154409986804e-07, "learning_rate": 1.4535289728795821e-05, "loss": 0.0, "num_input_tokens_seen": 7139744, "step": 25500 }, { "epoch": 283.3888888888889, "grad_norm": 3.9823089537094347e-07, "learning_rate": 1.4526374538374132e-05, "loss": 0.0, "num_input_tokens_seen": 7141136, "step": 25505 }, { "epoch": 283.44444444444446, "grad_norm": 6.924083209014498e-07, "learning_rate": 1.4517460963117097e-05, "loss": 0.0, "num_input_tokens_seen": 7142544, "step": 25510 }, { "epoch": 283.5, "grad_norm": 4.76595801046642e-07, "learning_rate": 1.4508549004399314e-05, "loss": 0.0, "num_input_tokens_seen": 7143968, "step": 25515 }, { "epoch": 283.55555555555554, "grad_norm": 4.708520464191679e-06, "learning_rate": 1.449963866359513e-05, "loss": 0.0, "num_input_tokens_seen": 7145328, "step": 25520 }, { "epoch": 283.6111111111111, "grad_norm": 1.5789045448855177e-07, "learning_rate": 1.4490729942078607e-05, "loss": 0.0, "num_input_tokens_seen": 7146704, "step": 25525 }, { "epoch": 283.6666666666667, "grad_norm": 2.0825447677452757e-07, "learning_rate": 1.4481822841223608e-05, "loss": 0.0, "num_input_tokens_seen": 7148112, "step": 25530 }, { "epoch": 283.72222222222223, "grad_norm": 4.02848081648699e-06, "learning_rate": 1.4472917362403704e-05, "loss": 0.0, "num_input_tokens_seen": 7149520, "step": 25535 }, { "epoch": 283.77777777777777, "grad_norm": 1.0136187711395905e-07, "learning_rate": 1.4464013506992224e-05, "loss": 0.0, "num_input_tokens_seen": 7150896, "step": 25540 }, { "epoch": 283.8333333333333, "grad_norm": 7.392218890345248e-07, "learning_rate": 1.4455111276362277e-05, "loss": 0.0, "num_input_tokens_seen": 7152336, "step": 25545 }, { "epoch": 283.8888888888889, "grad_norm": 3.6895914945489494e-07, "learning_rate": 1.4446210671886676e-05, "loss": 0.0, "num_input_tokens_seen": 7153728, "step": 25550 }, { "epoch": 283.94444444444446, "grad_norm": 6.655321271864523e-07, "learning_rate": 1.4437311694938015e-05, "loss": 0.0, "num_input_tokens_seen": 7155136, "step": 25555 }, { "epoch": 284.0, "grad_norm": 9.403281069353397e-07, "learning_rate": 1.442841434688864e-05, "loss": 0.0, "num_input_tokens_seen": 7156512, "step": 25560 }, { "epoch": 284.05555555555554, "grad_norm": 9.070353712559154e-07, "learning_rate": 1.4419518629110615e-05, "loss": 0.0, "num_input_tokens_seen": 7157936, "step": 25565 }, { "epoch": 284.1111111111111, "grad_norm": 1.140457356996194e-06, "learning_rate": 1.4410624542975778e-05, "loss": 0.0, "num_input_tokens_seen": 7159360, "step": 25570 }, { "epoch": 284.1666666666667, "grad_norm": 6.879417924210429e-07, "learning_rate": 1.4401732089855724e-05, "loss": 0.0, "num_input_tokens_seen": 7160704, "step": 25575 }, { "epoch": 284.22222222222223, "grad_norm": 5.029953626944916e-07, "learning_rate": 1.4392841271121754e-05, "loss": 0.0, "num_input_tokens_seen": 7162144, "step": 25580 }, { "epoch": 284.27777777777777, "grad_norm": 4.268594864242914e-07, "learning_rate": 1.438395208814497e-05, "loss": 0.0, "num_input_tokens_seen": 7163536, "step": 25585 }, { "epoch": 284.3333333333333, "grad_norm": 4.767059351706848e-07, "learning_rate": 1.4375064542296174e-05, "loss": 0.0, "num_input_tokens_seen": 7164960, "step": 25590 }, { "epoch": 284.3888888888889, "grad_norm": 3.0614359047831385e-07, "learning_rate": 1.4366178634945946e-05, "loss": 0.0, "num_input_tokens_seen": 7166320, "step": 25595 }, { "epoch": 284.44444444444446, "grad_norm": 7.454935371242755e-07, "learning_rate": 1.4357294367464616e-05, "loss": 0.0, "num_input_tokens_seen": 7167744, "step": 25600 }, { "epoch": 284.44444444444446, "eval_loss": 1.630126953125, "eval_runtime": 1.3926, "eval_samples_per_second": 28.722, "eval_steps_per_second": 14.361, "num_input_tokens_seen": 7167744, "step": 25600 }, { "epoch": 284.5, "grad_norm": 2.587413838739394e-08, "learning_rate": 1.434841174122224e-05, "loss": 0.0, "num_input_tokens_seen": 7169152, "step": 25605 }, { "epoch": 284.55555555555554, "grad_norm": 3.31883325088711e-06, "learning_rate": 1.4339530757588615e-05, "loss": 0.0, "num_input_tokens_seen": 7170560, "step": 25610 }, { "epoch": 284.6111111111111, "grad_norm": 5.2779405734781903e-08, "learning_rate": 1.433065141793333e-05, "loss": 0.0, "num_input_tokens_seen": 7171984, "step": 25615 }, { "epoch": 284.6666666666667, "grad_norm": 5.527460871235235e-07, "learning_rate": 1.4321773723625665e-05, "loss": 0.0, "num_input_tokens_seen": 7173376, "step": 25620 }, { "epoch": 284.72222222222223, "grad_norm": 3.169420779158827e-07, "learning_rate": 1.4312897676034693e-05, "loss": 0.0, "num_input_tokens_seen": 7174784, "step": 25625 }, { "epoch": 284.77777777777777, "grad_norm": 8.435166591880261e-07, "learning_rate": 1.4304023276529188e-05, "loss": 0.0, "num_input_tokens_seen": 7176176, "step": 25630 }, { "epoch": 284.8333333333333, "grad_norm": 3.923616532119922e-06, "learning_rate": 1.4295150526477712e-05, "loss": 0.0, "num_input_tokens_seen": 7177584, "step": 25635 }, { "epoch": 284.8888888888889, "grad_norm": 7.273053626022374e-08, "learning_rate": 1.4286279427248562e-05, "loss": 0.0, "num_input_tokens_seen": 7178944, "step": 25640 }, { "epoch": 284.94444444444446, "grad_norm": 5.803568683404592e-07, "learning_rate": 1.4277409980209747e-05, "loss": 0.0, "num_input_tokens_seen": 7180368, "step": 25645 }, { "epoch": 285.0, "grad_norm": 1.8501883403132524e-07, "learning_rate": 1.4268542186729061e-05, "loss": 0.0, "num_input_tokens_seen": 7181728, "step": 25650 }, { "epoch": 285.05555555555554, "grad_norm": 1.4172281908031437e-06, "learning_rate": 1.4259676048174043e-05, "loss": 0.0, "num_input_tokens_seen": 7183120, "step": 25655 }, { "epoch": 285.1111111111111, "grad_norm": 7.392528118543851e-07, "learning_rate": 1.4250811565911937e-05, "loss": 0.0, "num_input_tokens_seen": 7184528, "step": 25660 }, { "epoch": 285.1666666666667, "grad_norm": 7.560022936559108e-07, "learning_rate": 1.4241948741309782e-05, "loss": 0.0, "num_input_tokens_seen": 7185904, "step": 25665 }, { "epoch": 285.22222222222223, "grad_norm": 1.1485757340778946e-06, "learning_rate": 1.4233087575734317e-05, "loss": 0.0, "num_input_tokens_seen": 7187344, "step": 25670 }, { "epoch": 285.27777777777777, "grad_norm": 3.341945955526171e-07, "learning_rate": 1.422422807055206e-05, "loss": 0.0, "num_input_tokens_seen": 7188720, "step": 25675 }, { "epoch": 285.3333333333333, "grad_norm": 7.176869303293643e-07, "learning_rate": 1.4215370227129243e-05, "loss": 0.0, "num_input_tokens_seen": 7190144, "step": 25680 }, { "epoch": 285.3888888888889, "grad_norm": 3.6723022844853404e-07, "learning_rate": 1.4206514046831876e-05, "loss": 0.0, "num_input_tokens_seen": 7191536, "step": 25685 }, { "epoch": 285.44444444444446, "grad_norm": 2.7970580163128034e-07, "learning_rate": 1.419765953102567e-05, "loss": 0.0, "num_input_tokens_seen": 7192912, "step": 25690 }, { "epoch": 285.5, "grad_norm": 5.730850602958526e-07, "learning_rate": 1.4188806681076125e-05, "loss": 0.0, "num_input_tokens_seen": 7194304, "step": 25695 }, { "epoch": 285.55555555555554, "grad_norm": 3.317836672067642e-07, "learning_rate": 1.4179955498348443e-05, "loss": 0.0, "num_input_tokens_seen": 7195744, "step": 25700 }, { "epoch": 285.6111111111111, "grad_norm": 7.485871265089372e-07, "learning_rate": 1.4171105984207605e-05, "loss": 0.0, "num_input_tokens_seen": 7197120, "step": 25705 }, { "epoch": 285.6666666666667, "grad_norm": 5.42482439414016e-07, "learning_rate": 1.4162258140018304e-05, "loss": 0.0, "num_input_tokens_seen": 7198560, "step": 25710 }, { "epoch": 285.72222222222223, "grad_norm": 8.926170949052903e-07, "learning_rate": 1.4153411967144986e-05, "loss": 0.0, "num_input_tokens_seen": 7199968, "step": 25715 }, { "epoch": 285.77777777777777, "grad_norm": 4.291907202969014e-07, "learning_rate": 1.4144567466951864e-05, "loss": 0.0, "num_input_tokens_seen": 7201408, "step": 25720 }, { "epoch": 285.8333333333333, "grad_norm": 6.619293344556354e-07, "learning_rate": 1.4135724640802844e-05, "loss": 0.0, "num_input_tokens_seen": 7202768, "step": 25725 }, { "epoch": 285.8888888888889, "grad_norm": 2.4956841571111e-07, "learning_rate": 1.4126883490061615e-05, "loss": 0.0, "num_input_tokens_seen": 7204128, "step": 25730 }, { "epoch": 285.94444444444446, "grad_norm": 1.0061392003990477e-06, "learning_rate": 1.4118044016091603e-05, "loss": 0.0, "num_input_tokens_seen": 7205536, "step": 25735 }, { "epoch": 286.0, "grad_norm": 1.4990958163707546e-07, "learning_rate": 1.410920622025594e-05, "loss": 0.0, "num_input_tokens_seen": 7206960, "step": 25740 }, { "epoch": 286.05555555555554, "grad_norm": 1.5045580425976368e-07, "learning_rate": 1.4100370103917554e-05, "loss": 0.0, "num_input_tokens_seen": 7208352, "step": 25745 }, { "epoch": 286.1111111111111, "grad_norm": 5.206200270890804e-08, "learning_rate": 1.409153566843907e-05, "loss": 0.0, "num_input_tokens_seen": 7209696, "step": 25750 }, { "epoch": 286.1666666666667, "grad_norm": 3.888429773724056e-07, "learning_rate": 1.408270291518286e-05, "loss": 0.0, "num_input_tokens_seen": 7211120, "step": 25755 }, { "epoch": 286.22222222222223, "grad_norm": 2.903285007960221e-07, "learning_rate": 1.407387184551107e-05, "loss": 0.0, "num_input_tokens_seen": 7212528, "step": 25760 }, { "epoch": 286.27777777777777, "grad_norm": 8.34854120057571e-07, "learning_rate": 1.4065042460785532e-05, "loss": 0.0, "num_input_tokens_seen": 7213920, "step": 25765 }, { "epoch": 286.3333333333333, "grad_norm": 1.399820348524372e-07, "learning_rate": 1.405621476236787e-05, "loss": 0.0, "num_input_tokens_seen": 7215344, "step": 25770 }, { "epoch": 286.3888888888889, "grad_norm": 2.866556201297499e-07, "learning_rate": 1.4047388751619423e-05, "loss": 0.0, "num_input_tokens_seen": 7216752, "step": 25775 }, { "epoch": 286.44444444444446, "grad_norm": 4.1970969277826953e-07, "learning_rate": 1.4038564429901264e-05, "loss": 0.0, "num_input_tokens_seen": 7218096, "step": 25780 }, { "epoch": 286.5, "grad_norm": 1.0186890904151369e-06, "learning_rate": 1.4029741798574227e-05, "loss": 0.0, "num_input_tokens_seen": 7219520, "step": 25785 }, { "epoch": 286.55555555555554, "grad_norm": 9.886247198664933e-08, "learning_rate": 1.402092085899886e-05, "loss": 0.0, "num_input_tokens_seen": 7220928, "step": 25790 }, { "epoch": 286.6111111111111, "grad_norm": 2.1152993667783448e-07, "learning_rate": 1.4012101612535464e-05, "loss": 0.0, "num_input_tokens_seen": 7222336, "step": 25795 }, { "epoch": 286.6666666666667, "grad_norm": 3.0049849897295644e-07, "learning_rate": 1.4003284060544092e-05, "loss": 0.0, "num_input_tokens_seen": 7223696, "step": 25800 }, { "epoch": 286.6666666666667, "eval_loss": 1.5831866264343262, "eval_runtime": 1.3906, "eval_samples_per_second": 28.765, "eval_steps_per_second": 14.383, "num_input_tokens_seen": 7223696, "step": 25800 }, { "epoch": 286.72222222222223, "grad_norm": 4.867358711635461e-07, "learning_rate": 1.3994468204384504e-05, "loss": 0.0, "num_input_tokens_seen": 7225088, "step": 25805 }, { "epoch": 286.77777777777777, "grad_norm": 5.116495458423742e-07, "learning_rate": 1.398565404541622e-05, "loss": 0.0, "num_input_tokens_seen": 7226448, "step": 25810 }, { "epoch": 286.8333333333333, "grad_norm": 3.4620242672644963e-07, "learning_rate": 1.3976841584998513e-05, "loss": 0.0, "num_input_tokens_seen": 7227872, "step": 25815 }, { "epoch": 286.8888888888889, "grad_norm": 6.067318167879421e-07, "learning_rate": 1.3968030824490352e-05, "loss": 0.0, "num_input_tokens_seen": 7229296, "step": 25820 }, { "epoch": 286.94444444444446, "grad_norm": 3.7301988413673826e-06, "learning_rate": 1.3959221765250469e-05, "loss": 0.0, "num_input_tokens_seen": 7230704, "step": 25825 }, { "epoch": 287.0, "grad_norm": 1.0830460723809665e-06, "learning_rate": 1.3950414408637343e-05, "loss": 0.0, "num_input_tokens_seen": 7232128, "step": 25830 }, { "epoch": 287.05555555555554, "grad_norm": 3.200734965957963e-07, "learning_rate": 1.3941608756009166e-05, "loss": 0.0, "num_input_tokens_seen": 7233552, "step": 25835 }, { "epoch": 287.1111111111111, "grad_norm": 5.205150728215813e-07, "learning_rate": 1.3932804808723898e-05, "loss": 0.0, "num_input_tokens_seen": 7234976, "step": 25840 }, { "epoch": 287.1666666666667, "grad_norm": 2.0315395943271142e-07, "learning_rate": 1.3924002568139194e-05, "loss": 0.0, "num_input_tokens_seen": 7236352, "step": 25845 }, { "epoch": 287.22222222222223, "grad_norm": 1.0404780823591864e-06, "learning_rate": 1.3915202035612485e-05, "loss": 0.0, "num_input_tokens_seen": 7237744, "step": 25850 }, { "epoch": 287.27777777777777, "grad_norm": 4.5748504362563835e-07, "learning_rate": 1.3906403212500935e-05, "loss": 0.0, "num_input_tokens_seen": 7239152, "step": 25855 }, { "epoch": 287.3333333333333, "grad_norm": 7.784261413235072e-08, "learning_rate": 1.3897606100161409e-05, "loss": 0.0, "num_input_tokens_seen": 7240560, "step": 25860 }, { "epoch": 287.3888888888889, "grad_norm": 3.230773586437863e-07, "learning_rate": 1.388881069995055e-05, "loss": 0.0, "num_input_tokens_seen": 7241984, "step": 25865 }, { "epoch": 287.44444444444446, "grad_norm": 2.2831788726307423e-07, "learning_rate": 1.3880017013224708e-05, "loss": 0.0, "num_input_tokens_seen": 7243392, "step": 25870 }, { "epoch": 287.5, "grad_norm": 6.444644782277464e-07, "learning_rate": 1.3871225041339984e-05, "loss": 0.0, "num_input_tokens_seen": 7244784, "step": 25875 }, { "epoch": 287.55555555555554, "grad_norm": 2.5280607474087446e-07, "learning_rate": 1.386243478565222e-05, "loss": 0.0, "num_input_tokens_seen": 7246176, "step": 25880 }, { "epoch": 287.6111111111111, "grad_norm": 5.566172944782011e-07, "learning_rate": 1.3853646247516966e-05, "loss": 0.0, "num_input_tokens_seen": 7247552, "step": 25885 }, { "epoch": 287.6666666666667, "grad_norm": 5.497509505403286e-07, "learning_rate": 1.3844859428289545e-05, "loss": 0.0, "num_input_tokens_seen": 7248960, "step": 25890 }, { "epoch": 287.72222222222223, "grad_norm": 2.673481560577784e-07, "learning_rate": 1.3836074329324984e-05, "loss": 0.0, "num_input_tokens_seen": 7250368, "step": 25895 }, { "epoch": 287.77777777777777, "grad_norm": 1.5116910390133853e-06, "learning_rate": 1.3827290951978044e-05, "loss": 0.0, "num_input_tokens_seen": 7251760, "step": 25900 }, { "epoch": 287.8333333333333, "grad_norm": 3.18900760021279e-07, "learning_rate": 1.381850929760326e-05, "loss": 0.0, "num_input_tokens_seen": 7253120, "step": 25905 }, { "epoch": 287.8888888888889, "grad_norm": 4.6330680447681516e-07, "learning_rate": 1.3809729367554842e-05, "loss": 0.0, "num_input_tokens_seen": 7254496, "step": 25910 }, { "epoch": 287.94444444444446, "grad_norm": 9.417390174348839e-07, "learning_rate": 1.3800951163186784e-05, "loss": 0.0, "num_input_tokens_seen": 7255904, "step": 25915 }, { "epoch": 288.0, "grad_norm": 9.690950264484854e-07, "learning_rate": 1.3792174685852801e-05, "loss": 0.0, "num_input_tokens_seen": 7257296, "step": 25920 }, { "epoch": 288.05555555555554, "grad_norm": 7.93697438439267e-07, "learning_rate": 1.378339993690632e-05, "loss": 0.0, "num_input_tokens_seen": 7258688, "step": 25925 }, { "epoch": 288.1111111111111, "grad_norm": 4.896720383840147e-06, "learning_rate": 1.3774626917700523e-05, "loss": 0.0, "num_input_tokens_seen": 7260064, "step": 25930 }, { "epoch": 288.1666666666667, "grad_norm": 4.708052472324198e-07, "learning_rate": 1.3765855629588334e-05, "loss": 0.0, "num_input_tokens_seen": 7261472, "step": 25935 }, { "epoch": 288.22222222222223, "grad_norm": 9.655047250589632e-08, "learning_rate": 1.3757086073922374e-05, "loss": 0.0, "num_input_tokens_seen": 7262896, "step": 25940 }, { "epoch": 288.27777777777777, "grad_norm": 5.806852598766454e-08, "learning_rate": 1.3748318252055038e-05, "loss": 0.0, "num_input_tokens_seen": 7264288, "step": 25945 }, { "epoch": 288.3333333333333, "grad_norm": 3.3802180610109644e-07, "learning_rate": 1.3739552165338416e-05, "loss": 0.0, "num_input_tokens_seen": 7265696, "step": 25950 }, { "epoch": 288.3888888888889, "grad_norm": 5.133756886266383e-08, "learning_rate": 1.3730787815124354e-05, "loss": 0.0, "num_input_tokens_seen": 7267088, "step": 25955 }, { "epoch": 288.44444444444446, "grad_norm": 1.252086150316245e-07, "learning_rate": 1.3722025202764443e-05, "loss": 0.0, "num_input_tokens_seen": 7268496, "step": 25960 }, { "epoch": 288.5, "grad_norm": 1.596076089072085e-07, "learning_rate": 1.371326432960997e-05, "loss": 0.0, "num_input_tokens_seen": 7269888, "step": 25965 }, { "epoch": 288.55555555555554, "grad_norm": 1.1957538958995428e-07, "learning_rate": 1.3704505197011969e-05, "loss": 0.0, "num_input_tokens_seen": 7271296, "step": 25970 }, { "epoch": 288.6111111111111, "grad_norm": 1.3259965214729164e-07, "learning_rate": 1.3695747806321224e-05, "loss": 0.0, "num_input_tokens_seen": 7272720, "step": 25975 }, { "epoch": 288.6666666666667, "grad_norm": 8.10736864309547e-08, "learning_rate": 1.3686992158888212e-05, "loss": 0.0, "num_input_tokens_seen": 7274160, "step": 25980 }, { "epoch": 288.72222222222223, "grad_norm": 3.490858091481641e-07, "learning_rate": 1.367823825606319e-05, "loss": 0.0, "num_input_tokens_seen": 7275552, "step": 25985 }, { "epoch": 288.77777777777777, "grad_norm": 3.9136531881922565e-07, "learning_rate": 1.36694860991961e-05, "loss": 0.0, "num_input_tokens_seen": 7276960, "step": 25990 }, { "epoch": 288.8333333333333, "grad_norm": 5.42056056929141e-07, "learning_rate": 1.3660735689636636e-05, "loss": 0.0, "num_input_tokens_seen": 7278352, "step": 25995 }, { "epoch": 288.8888888888889, "grad_norm": 3.793850567035406e-07, "learning_rate": 1.365198702873424e-05, "loss": 0.0, "num_input_tokens_seen": 7279760, "step": 26000 }, { "epoch": 288.8888888888889, "eval_loss": 1.5041314363479614, "eval_runtime": 1.3872, "eval_samples_per_second": 28.836, "eval_steps_per_second": 14.418, "num_input_tokens_seen": 7279760, "step": 26000 }, { "epoch": 288.94444444444446, "grad_norm": 5.209158757679688e-07, "learning_rate": 1.364324011783804e-05, "loss": 0.0, "num_input_tokens_seen": 7281200, "step": 26005 }, { "epoch": 289.0, "grad_norm": 3.7833353871974396e-07, "learning_rate": 1.3634494958296934e-05, "loss": 0.0, "num_input_tokens_seen": 7282608, "step": 26010 }, { "epoch": 289.05555555555554, "grad_norm": 6.244829933166329e-07, "learning_rate": 1.3625751551459542e-05, "loss": 0.0, "num_input_tokens_seen": 7283984, "step": 26015 }, { "epoch": 289.1111111111111, "grad_norm": 3.5655273222801043e-07, "learning_rate": 1.3617009898674188e-05, "loss": 0.0, "num_input_tokens_seen": 7285408, "step": 26020 }, { "epoch": 289.1666666666667, "grad_norm": 5.645780447594007e-07, "learning_rate": 1.3608270001288967e-05, "loss": 0.0, "num_input_tokens_seen": 7286816, "step": 26025 }, { "epoch": 289.22222222222223, "grad_norm": 4.6047799173720705e-07, "learning_rate": 1.359953186065166e-05, "loss": 0.0, "num_input_tokens_seen": 7288224, "step": 26030 }, { "epoch": 289.27777777777777, "grad_norm": 5.659267117152922e-07, "learning_rate": 1.3590795478109814e-05, "loss": 0.0, "num_input_tokens_seen": 7289600, "step": 26035 }, { "epoch": 289.3333333333333, "grad_norm": 1.1048258556911605e-06, "learning_rate": 1.3582060855010675e-05, "loss": 0.0, "num_input_tokens_seen": 7290976, "step": 26040 }, { "epoch": 289.3888888888889, "grad_norm": 5.64607717024046e-07, "learning_rate": 1.3573327992701245e-05, "loss": 0.0, "num_input_tokens_seen": 7292368, "step": 26045 }, { "epoch": 289.44444444444446, "grad_norm": 3.3543025779181335e-07, "learning_rate": 1.356459689252823e-05, "loss": 0.0, "num_input_tokens_seen": 7293808, "step": 26050 }, { "epoch": 289.5, "grad_norm": 8.636051234134356e-07, "learning_rate": 1.3555867555838087e-05, "loss": 0.0, "num_input_tokens_seen": 7295184, "step": 26055 }, { "epoch": 289.55555555555554, "grad_norm": 5.156996962796256e-07, "learning_rate": 1.3547139983976975e-05, "loss": 0.0, "num_input_tokens_seen": 7296576, "step": 26060 }, { "epoch": 289.6111111111111, "grad_norm": 4.0264669110001705e-07, "learning_rate": 1.3538414178290815e-05, "loss": 0.0, "num_input_tokens_seen": 7297968, "step": 26065 }, { "epoch": 289.6666666666667, "grad_norm": 5.438304810923e-07, "learning_rate": 1.3529690140125209e-05, "loss": 0.0, "num_input_tokens_seen": 7299344, "step": 26070 }, { "epoch": 289.72222222222223, "grad_norm": 5.246818091109162e-07, "learning_rate": 1.352096787082553e-05, "loss": 0.0, "num_input_tokens_seen": 7300736, "step": 26075 }, { "epoch": 289.77777777777777, "grad_norm": 2.0744625999213895e-06, "learning_rate": 1.3512247371736871e-05, "loss": 0.0, "num_input_tokens_seen": 7302128, "step": 26080 }, { "epoch": 289.8333333333333, "grad_norm": 9.987485327656032e-07, "learning_rate": 1.3503528644204022e-05, "loss": 0.0, "num_input_tokens_seen": 7303536, "step": 26085 }, { "epoch": 289.8888888888889, "grad_norm": 1.7022523479681695e-07, "learning_rate": 1.349481168957153e-05, "loss": 0.0, "num_input_tokens_seen": 7304960, "step": 26090 }, { "epoch": 289.94444444444446, "grad_norm": 3.304283779925754e-07, "learning_rate": 1.3486096509183665e-05, "loss": 0.0, "num_input_tokens_seen": 7306368, "step": 26095 }, { "epoch": 290.0, "grad_norm": 5.662406579176604e-07, "learning_rate": 1.3477383104384406e-05, "loss": 0.0, "num_input_tokens_seen": 7307760, "step": 26100 }, { "epoch": 290.05555555555554, "grad_norm": 2.5598868091947224e-07, "learning_rate": 1.3468671476517481e-05, "loss": 0.0, "num_input_tokens_seen": 7309168, "step": 26105 }, { "epoch": 290.1111111111111, "grad_norm": 3.001815400693886e-07, "learning_rate": 1.3459961626926326e-05, "loss": 0.0, "num_input_tokens_seen": 7310624, "step": 26110 }, { "epoch": 290.1666666666667, "grad_norm": 1.545670755831452e-07, "learning_rate": 1.3451253556954101e-05, "loss": 0.0, "num_input_tokens_seen": 7312048, "step": 26115 }, { "epoch": 290.22222222222223, "grad_norm": 5.384200107982906e-07, "learning_rate": 1.3442547267943717e-05, "loss": 0.0, "num_input_tokens_seen": 7313456, "step": 26120 }, { "epoch": 290.27777777777777, "grad_norm": 9.786908776732162e-07, "learning_rate": 1.3433842761237774e-05, "loss": 0.0, "num_input_tokens_seen": 7314816, "step": 26125 }, { "epoch": 290.3333333333333, "grad_norm": 1.2208275848024641e-06, "learning_rate": 1.3425140038178639e-05, "loss": 0.0, "num_input_tokens_seen": 7316208, "step": 26130 }, { "epoch": 290.3888888888889, "grad_norm": 5.029726253269473e-07, "learning_rate": 1.3416439100108358e-05, "loss": 0.0, "num_input_tokens_seen": 7317600, "step": 26135 }, { "epoch": 290.44444444444446, "grad_norm": 3.247690756325028e-07, "learning_rate": 1.3407739948368734e-05, "loss": 0.0, "num_input_tokens_seen": 7318976, "step": 26140 }, { "epoch": 290.5, "grad_norm": 3.6011445558870037e-07, "learning_rate": 1.3399042584301298e-05, "loss": 0.0, "num_input_tokens_seen": 7320416, "step": 26145 }, { "epoch": 290.55555555555554, "grad_norm": 8.902025001589209e-07, "learning_rate": 1.3390347009247272e-05, "loss": 0.0, "num_input_tokens_seen": 7321824, "step": 26150 }, { "epoch": 290.6111111111111, "grad_norm": 4.289169055482489e-07, "learning_rate": 1.3381653224547635e-05, "loss": 0.0, "num_input_tokens_seen": 7323200, "step": 26155 }, { "epoch": 290.6666666666667, "grad_norm": 6.942746608729067e-07, "learning_rate": 1.3372961231543086e-05, "loss": 0.0, "num_input_tokens_seen": 7324576, "step": 26160 }, { "epoch": 290.72222222222223, "grad_norm": 6.422904334613122e-06, "learning_rate": 1.3364271031574016e-05, "loss": 0.0, "num_input_tokens_seen": 7326048, "step": 26165 }, { "epoch": 290.77777777777777, "grad_norm": 9.226275778928539e-07, "learning_rate": 1.335558262598059e-05, "loss": 0.0, "num_input_tokens_seen": 7327392, "step": 26170 }, { "epoch": 290.8333333333333, "grad_norm": 7.461728159796621e-07, "learning_rate": 1.3346896016102645e-05, "loss": 0.0, "num_input_tokens_seen": 7328800, "step": 26175 }, { "epoch": 290.8888888888889, "grad_norm": 3.918074185094156e-07, "learning_rate": 1.3338211203279788e-05, "loss": 0.0, "num_input_tokens_seen": 7330240, "step": 26180 }, { "epoch": 290.94444444444446, "grad_norm": 4.954728751727089e-07, "learning_rate": 1.3329528188851303e-05, "loss": 0.0, "num_input_tokens_seen": 7331632, "step": 26185 }, { "epoch": 291.0, "grad_norm": 5.3369444685813505e-06, "learning_rate": 1.3320846974156242e-05, "loss": 0.0, "num_input_tokens_seen": 7333008, "step": 26190 }, { "epoch": 291.05555555555554, "grad_norm": 2.574096242824453e-07, "learning_rate": 1.3312167560533337e-05, "loss": 0.0, "num_input_tokens_seen": 7334384, "step": 26195 }, { "epoch": 291.1111111111111, "grad_norm": 5.044757926953025e-07, "learning_rate": 1.3303489949321082e-05, "loss": 0.0, "num_input_tokens_seen": 7335792, "step": 26200 }, { "epoch": 291.1111111111111, "eval_loss": 1.570312738418579, "eval_runtime": 1.4073, "eval_samples_per_second": 28.423, "eval_steps_per_second": 14.211, "num_input_tokens_seen": 7335792, "step": 26200 }, { "epoch": 291.1666666666667, "grad_norm": 3.315798835501482e-07, "learning_rate": 1.3294814141857653e-05, "loss": 0.0, "num_input_tokens_seen": 7337120, "step": 26205 }, { "epoch": 291.22222222222223, "grad_norm": 1.2827628097511479e-06, "learning_rate": 1.3286140139480992e-05, "loss": 0.0, "num_input_tokens_seen": 7338496, "step": 26210 }, { "epoch": 291.27777777777777, "grad_norm": 1.1509812338772463e-06, "learning_rate": 1.3277467943528719e-05, "loss": 0.0, "num_input_tokens_seen": 7339888, "step": 26215 }, { "epoch": 291.3333333333333, "grad_norm": 9.559078080201289e-07, "learning_rate": 1.3268797555338203e-05, "loss": 0.0, "num_input_tokens_seen": 7341328, "step": 26220 }, { "epoch": 291.3888888888889, "grad_norm": 6.258503049139108e-07, "learning_rate": 1.3260128976246533e-05, "loss": 0.0, "num_input_tokens_seen": 7342768, "step": 26225 }, { "epoch": 291.44444444444446, "grad_norm": 6.33234503766289e-06, "learning_rate": 1.32514622075905e-05, "loss": 0.0, "num_input_tokens_seen": 7344160, "step": 26230 }, { "epoch": 291.5, "grad_norm": 8.367444479517872e-07, "learning_rate": 1.3242797250706638e-05, "loss": 0.0, "num_input_tokens_seen": 7345536, "step": 26235 }, { "epoch": 291.55555555555554, "grad_norm": 1.1417329233154305e-06, "learning_rate": 1.3234134106931195e-05, "loss": 0.0, "num_input_tokens_seen": 7346960, "step": 26240 }, { "epoch": 291.6111111111111, "grad_norm": 1.1539896860313092e-07, "learning_rate": 1.322547277760013e-05, "loss": 0.0, "num_input_tokens_seen": 7348320, "step": 26245 }, { "epoch": 291.6666666666667, "grad_norm": 3.523206828504044e-07, "learning_rate": 1.3216813264049132e-05, "loss": 0.0, "num_input_tokens_seen": 7349696, "step": 26250 }, { "epoch": 291.72222222222223, "grad_norm": 1.595562082457036e-07, "learning_rate": 1.32081555676136e-05, "loss": 0.0, "num_input_tokens_seen": 7351120, "step": 26255 }, { "epoch": 291.77777777777777, "grad_norm": 4.539683402526862e-07, "learning_rate": 1.3199499689628674e-05, "loss": 0.0, "num_input_tokens_seen": 7352560, "step": 26260 }, { "epoch": 291.8333333333333, "grad_norm": 5.149401204107562e-06, "learning_rate": 1.3190845631429192e-05, "loss": 0.0, "num_input_tokens_seen": 7354000, "step": 26265 }, { "epoch": 291.8888888888889, "grad_norm": 1.5586191466354649e-06, "learning_rate": 1.3182193394349704e-05, "loss": 0.0, "num_input_tokens_seen": 7355440, "step": 26270 }, { "epoch": 291.94444444444446, "grad_norm": 3.349282167164347e-07, "learning_rate": 1.3173542979724507e-05, "loss": 0.0, "num_input_tokens_seen": 7356864, "step": 26275 }, { "epoch": 292.0, "grad_norm": 8.698858522393493e-08, "learning_rate": 1.3164894388887617e-05, "loss": 0.0, "num_input_tokens_seen": 7358272, "step": 26280 }, { "epoch": 292.05555555555554, "grad_norm": 6.962600451743128e-08, "learning_rate": 1.3156247623172727e-05, "loss": 0.0, "num_input_tokens_seen": 7359632, "step": 26285 }, { "epoch": 292.1111111111111, "grad_norm": 8.986403088329098e-08, "learning_rate": 1.3147602683913302e-05, "loss": 0.0, "num_input_tokens_seen": 7361008, "step": 26290 }, { "epoch": 292.1666666666667, "grad_norm": 5.873219066643287e-08, "learning_rate": 1.3138959572442481e-05, "loss": 0.0, "num_input_tokens_seen": 7362400, "step": 26295 }, { "epoch": 292.22222222222223, "grad_norm": 2.1644407866006077e-07, "learning_rate": 1.3130318290093146e-05, "loss": 0.0, "num_input_tokens_seen": 7363808, "step": 26300 }, { "epoch": 292.27777777777777, "grad_norm": 1.2442126262612874e-06, "learning_rate": 1.3121678838197909e-05, "loss": 0.0, "num_input_tokens_seen": 7365232, "step": 26305 }, { "epoch": 292.3333333333333, "grad_norm": 4.163193239037355e-07, "learning_rate": 1.3113041218089056e-05, "loss": 0.0, "num_input_tokens_seen": 7366672, "step": 26310 }, { "epoch": 292.3888888888889, "grad_norm": 9.578801751786159e-08, "learning_rate": 1.3104405431098626e-05, "loss": 0.0, "num_input_tokens_seen": 7368096, "step": 26315 }, { "epoch": 292.44444444444446, "grad_norm": 8.361542995771742e-07, "learning_rate": 1.3095771478558377e-05, "loss": 0.0, "num_input_tokens_seen": 7369472, "step": 26320 }, { "epoch": 292.5, "grad_norm": 1.0770877452159766e-06, "learning_rate": 1.3087139361799766e-05, "loss": 0.0, "num_input_tokens_seen": 7370896, "step": 26325 }, { "epoch": 292.55555555555554, "grad_norm": 2.0057477456703054e-07, "learning_rate": 1.3078509082153964e-05, "loss": 0.0, "num_input_tokens_seen": 7372256, "step": 26330 }, { "epoch": 292.6111111111111, "grad_norm": 6.988075256231241e-06, "learning_rate": 1.3069880640951885e-05, "loss": 0.0, "num_input_tokens_seen": 7373680, "step": 26335 }, { "epoch": 292.6666666666667, "grad_norm": 1.1524521141836885e-06, "learning_rate": 1.3061254039524123e-05, "loss": 0.0, "num_input_tokens_seen": 7375120, "step": 26340 }, { "epoch": 292.72222222222223, "grad_norm": 9.625822485759272e-07, "learning_rate": 1.3052629279201028e-05, "loss": 0.0, "num_input_tokens_seen": 7376560, "step": 26345 }, { "epoch": 292.77777777777777, "grad_norm": 1.4002854186401237e-05, "learning_rate": 1.3044006361312633e-05, "loss": 0.0, "num_input_tokens_seen": 7377936, "step": 26350 }, { "epoch": 292.8333333333333, "grad_norm": 7.635952670170809e-08, "learning_rate": 1.30353852871887e-05, "loss": 0.0, "num_input_tokens_seen": 7379312, "step": 26355 }, { "epoch": 292.8888888888889, "grad_norm": 1.0912450676414664e-07, "learning_rate": 1.302676605815873e-05, "loss": 0.0, "num_input_tokens_seen": 7380688, "step": 26360 }, { "epoch": 292.94444444444446, "grad_norm": 3.6057724628335563e-07, "learning_rate": 1.3018148675551884e-05, "loss": 0.0, "num_input_tokens_seen": 7382096, "step": 26365 }, { "epoch": 293.0, "grad_norm": 1.3148218158676173e-07, "learning_rate": 1.3009533140697094e-05, "loss": 0.0, "num_input_tokens_seen": 7383440, "step": 26370 }, { "epoch": 293.05555555555554, "grad_norm": 3.0285036700661294e-07, "learning_rate": 1.3000919454922966e-05, "loss": 0.0, "num_input_tokens_seen": 7384784, "step": 26375 }, { "epoch": 293.1111111111111, "grad_norm": 9.867562056342649e-08, "learning_rate": 1.299230761955785e-05, "loss": 0.0, "num_input_tokens_seen": 7386208, "step": 26380 }, { "epoch": 293.1666666666667, "grad_norm": 2.995801082761318e-07, "learning_rate": 1.2983697635929807e-05, "loss": 0.0, "num_input_tokens_seen": 7387600, "step": 26385 }, { "epoch": 293.22222222222223, "grad_norm": 1.9256025041158864e-07, "learning_rate": 1.2975089505366584e-05, "loss": 0.0, "num_input_tokens_seen": 7389040, "step": 26390 }, { "epoch": 293.27777777777777, "grad_norm": 1.2697407328232657e-05, "learning_rate": 1.2966483229195683e-05, "loss": 0.0, "num_input_tokens_seen": 7390432, "step": 26395 }, { "epoch": 293.3333333333333, "grad_norm": 6.422542355721816e-07, "learning_rate": 1.2957878808744283e-05, "loss": 0.0, "num_input_tokens_seen": 7391808, "step": 26400 }, { "epoch": 293.3333333333333, "eval_loss": 1.5176831483840942, "eval_runtime": 1.3871, "eval_samples_per_second": 28.837, "eval_steps_per_second": 14.419, "num_input_tokens_seen": 7391808, "step": 26400 }, { "epoch": 293.3888888888889, "grad_norm": 7.54004716441159e-08, "learning_rate": 1.294927624533931e-05, "loss": 0.0, "num_input_tokens_seen": 7393200, "step": 26405 }, { "epoch": 293.44444444444446, "grad_norm": 1.5767597005833522e-07, "learning_rate": 1.2940675540307378e-05, "loss": 0.0, "num_input_tokens_seen": 7394640, "step": 26410 }, { "epoch": 293.5, "grad_norm": 2.405850523246045e-07, "learning_rate": 1.2932076694974814e-05, "loss": 0.0, "num_input_tokens_seen": 7396016, "step": 26415 }, { "epoch": 293.55555555555554, "grad_norm": 1.4842662210412527e-07, "learning_rate": 1.2923479710667682e-05, "loss": 0.0, "num_input_tokens_seen": 7397424, "step": 26420 }, { "epoch": 293.6111111111111, "grad_norm": 1.9164725983955577e-07, "learning_rate": 1.2914884588711751e-05, "loss": 0.0, "num_input_tokens_seen": 7398816, "step": 26425 }, { "epoch": 293.6666666666667, "grad_norm": 9.611262612452265e-07, "learning_rate": 1.2906291330432475e-05, "loss": 0.0, "num_input_tokens_seen": 7400224, "step": 26430 }, { "epoch": 293.72222222222223, "grad_norm": 8.740445878174796e-07, "learning_rate": 1.2897699937155055e-05, "loss": 0.0, "num_input_tokens_seen": 7401632, "step": 26435 }, { "epoch": 293.77777777777777, "grad_norm": 1.558300937176682e-07, "learning_rate": 1.2889110410204403e-05, "loss": 0.0, "num_input_tokens_seen": 7403040, "step": 26440 }, { "epoch": 293.8333333333333, "grad_norm": 6.997678525522133e-08, "learning_rate": 1.2880522750905111e-05, "loss": 0.0, "num_input_tokens_seen": 7404448, "step": 26445 }, { "epoch": 293.8888888888889, "grad_norm": 8.382193072975497e-08, "learning_rate": 1.2871936960581523e-05, "loss": 0.0, "num_input_tokens_seen": 7405808, "step": 26450 }, { "epoch": 293.94444444444446, "grad_norm": 6.872156887993697e-08, "learning_rate": 1.2863353040557658e-05, "loss": 0.0, "num_input_tokens_seen": 7407232, "step": 26455 }, { "epoch": 294.0, "grad_norm": 1.3151949929124385e-07, "learning_rate": 1.2854770992157273e-05, "loss": 0.0, "num_input_tokens_seen": 7408640, "step": 26460 }, { "epoch": 294.05555555555554, "grad_norm": 1.3080898497719318e-07, "learning_rate": 1.2846190816703835e-05, "loss": 0.0, "num_input_tokens_seen": 7410048, "step": 26465 }, { "epoch": 294.1111111111111, "grad_norm": 1.3870261739157286e-07, "learning_rate": 1.2837612515520498e-05, "loss": 0.0, "num_input_tokens_seen": 7411456, "step": 26470 }, { "epoch": 294.1666666666667, "grad_norm": 3.69757714224761e-07, "learning_rate": 1.2829036089930163e-05, "loss": 0.0, "num_input_tokens_seen": 7412912, "step": 26475 }, { "epoch": 294.22222222222223, "grad_norm": 1.2974695096090727e-07, "learning_rate": 1.2820461541255412e-05, "loss": 0.0, "num_input_tokens_seen": 7414320, "step": 26480 }, { "epoch": 294.27777777777777, "grad_norm": 1.660021382576815e-07, "learning_rate": 1.2811888870818543e-05, "loss": 0.0, "num_input_tokens_seen": 7415792, "step": 26485 }, { "epoch": 294.3333333333333, "grad_norm": 2.1245850234663521e-07, "learning_rate": 1.2803318079941581e-05, "loss": 0.0, "num_input_tokens_seen": 7417216, "step": 26490 }, { "epoch": 294.3888888888889, "grad_norm": 5.441271468953346e-07, "learning_rate": 1.2794749169946235e-05, "loss": 0.0, "num_input_tokens_seen": 7418608, "step": 26495 }, { "epoch": 294.44444444444446, "grad_norm": 6.664777885134754e-08, "learning_rate": 1.2786182142153952e-05, "loss": 0.0, "num_input_tokens_seen": 7420000, "step": 26500 }, { "epoch": 294.5, "grad_norm": 1.1437810343295496e-07, "learning_rate": 1.2777616997885878e-05, "loss": 0.0, "num_input_tokens_seen": 7421376, "step": 26505 }, { "epoch": 294.55555555555554, "grad_norm": 4.768904773300164e-07, "learning_rate": 1.2769053738462847e-05, "loss": 0.0, "num_input_tokens_seen": 7422768, "step": 26510 }, { "epoch": 294.6111111111111, "grad_norm": 3.674091999528173e-07, "learning_rate": 1.2760492365205434e-05, "loss": 0.0, "num_input_tokens_seen": 7424128, "step": 26515 }, { "epoch": 294.6666666666667, "grad_norm": 1.9858823918639246e-07, "learning_rate": 1.2751932879433919e-05, "loss": 0.0, "num_input_tokens_seen": 7425504, "step": 26520 }, { "epoch": 294.72222222222223, "grad_norm": 3.200182163709542e-07, "learning_rate": 1.2743375282468267e-05, "loss": 0.0, "num_input_tokens_seen": 7426848, "step": 26525 }, { "epoch": 294.77777777777777, "grad_norm": 2.4855179958649387e-07, "learning_rate": 1.2734819575628182e-05, "loss": 0.0, "num_input_tokens_seen": 7428240, "step": 26530 }, { "epoch": 294.8333333333333, "grad_norm": 3.011827516274934e-07, "learning_rate": 1.2726265760233039e-05, "loss": 0.0, "num_input_tokens_seen": 7429632, "step": 26535 }, { "epoch": 294.8888888888889, "grad_norm": 5.059062573309347e-07, "learning_rate": 1.271771383760197e-05, "loss": 0.0, "num_input_tokens_seen": 7431008, "step": 26540 }, { "epoch": 294.94444444444446, "grad_norm": 6.319471026472456e-07, "learning_rate": 1.2709163809053764e-05, "loss": 0.0, "num_input_tokens_seen": 7432416, "step": 26545 }, { "epoch": 295.0, "grad_norm": 1.7189375967063825e-07, "learning_rate": 1.2700615675906963e-05, "loss": 0.0, "num_input_tokens_seen": 7433808, "step": 26550 }, { "epoch": 295.05555555555554, "grad_norm": 2.9500972686946625e-07, "learning_rate": 1.269206943947978e-05, "loss": 0.0, "num_input_tokens_seen": 7435200, "step": 26555 }, { "epoch": 295.1111111111111, "grad_norm": 1.2221767065057065e-06, "learning_rate": 1.2683525101090177e-05, "loss": 0.0, "num_input_tokens_seen": 7436608, "step": 26560 }, { "epoch": 295.1666666666667, "grad_norm": 2.5284994080720935e-06, "learning_rate": 1.2674982662055765e-05, "loss": 0.0, "num_input_tokens_seen": 7438016, "step": 26565 }, { "epoch": 295.22222222222223, "grad_norm": 3.725028250300966e-07, "learning_rate": 1.2666442123693922e-05, "loss": 0.0, "num_input_tokens_seen": 7439376, "step": 26570 }, { "epoch": 295.27777777777777, "grad_norm": 3.738014981991e-07, "learning_rate": 1.265790348732169e-05, "loss": 0.0, "num_input_tokens_seen": 7440784, "step": 26575 }, { "epoch": 295.3333333333333, "grad_norm": 4.926952215100755e-07, "learning_rate": 1.264936675425584e-05, "loss": 0.0, "num_input_tokens_seen": 7442176, "step": 26580 }, { "epoch": 295.3888888888889, "grad_norm": 4.528395493252901e-06, "learning_rate": 1.2640831925812852e-05, "loss": 0.0, "num_input_tokens_seen": 7443568, "step": 26585 }, { "epoch": 295.44444444444446, "grad_norm": 3.198728563802433e-06, "learning_rate": 1.263229900330889e-05, "loss": 0.0, "num_input_tokens_seen": 7444992, "step": 26590 }, { "epoch": 295.5, "grad_norm": 3.9690999642516545e-07, "learning_rate": 1.2623767988059843e-05, "loss": 0.0, "num_input_tokens_seen": 7446368, "step": 26595 }, { "epoch": 295.55555555555554, "grad_norm": 1.6937973157382658e-07, "learning_rate": 1.2615238881381309e-05, "loss": 0.0, "num_input_tokens_seen": 7447808, "step": 26600 }, { "epoch": 295.55555555555554, "eval_loss": 1.5250283479690552, "eval_runtime": 1.3886, "eval_samples_per_second": 28.806, "eval_steps_per_second": 14.403, "num_input_tokens_seen": 7447808, "step": 26600 }, { "epoch": 295.6111111111111, "grad_norm": 1.0889863943930322e-07, "learning_rate": 1.2606711684588568e-05, "loss": 0.0, "num_input_tokens_seen": 7449200, "step": 26605 }, { "epoch": 295.6666666666667, "grad_norm": 4.4521942754727206e-07, "learning_rate": 1.2598186398996636e-05, "loss": 0.0, "num_input_tokens_seen": 7450560, "step": 26610 }, { "epoch": 295.72222222222223, "grad_norm": 2.748770668858924e-07, "learning_rate": 1.2589663025920207e-05, "loss": 0.0, "num_input_tokens_seen": 7451952, "step": 26615 }, { "epoch": 295.77777777777777, "grad_norm": 5.0195990297652315e-06, "learning_rate": 1.2581141566673705e-05, "loss": 0.0, "num_input_tokens_seen": 7453360, "step": 26620 }, { "epoch": 295.8333333333333, "grad_norm": 9.686948487797054e-07, "learning_rate": 1.257262202257124e-05, "loss": 0.0, "num_input_tokens_seen": 7454768, "step": 26625 }, { "epoch": 295.8888888888889, "grad_norm": 1.3625713108922355e-06, "learning_rate": 1.2564104394926618e-05, "loss": 0.0, "num_input_tokens_seen": 7456176, "step": 26630 }, { "epoch": 295.94444444444446, "grad_norm": 4.0318317928722536e-07, "learning_rate": 1.2555588685053383e-05, "loss": 0.0, "num_input_tokens_seen": 7457584, "step": 26635 }, { "epoch": 296.0, "grad_norm": 1.624153895818381e-07, "learning_rate": 1.2547074894264762e-05, "loss": 0.0, "num_input_tokens_seen": 7458944, "step": 26640 }, { "epoch": 296.05555555555554, "grad_norm": 5.942144980508601e-07, "learning_rate": 1.2538563023873679e-05, "loss": 0.0, "num_input_tokens_seen": 7460352, "step": 26645 }, { "epoch": 296.1111111111111, "grad_norm": 6.419480769181973e-07, "learning_rate": 1.2530053075192789e-05, "loss": 0.0, "num_input_tokens_seen": 7461776, "step": 26650 }, { "epoch": 296.1666666666667, "grad_norm": 3.3369990433129715e-06, "learning_rate": 1.252154504953441e-05, "loss": 0.0, "num_input_tokens_seen": 7463184, "step": 26655 }, { "epoch": 296.22222222222223, "grad_norm": 1.4812970050570584e-07, "learning_rate": 1.25130389482106e-05, "loss": 0.0, "num_input_tokens_seen": 7464528, "step": 26660 }, { "epoch": 296.27777777777777, "grad_norm": 1.330874823679551e-07, "learning_rate": 1.2504534772533116e-05, "loss": 0.0, "num_input_tokens_seen": 7465904, "step": 26665 }, { "epoch": 296.3333333333333, "grad_norm": 7.523934186792758e-08, "learning_rate": 1.2496032523813387e-05, "loss": 0.0, "num_input_tokens_seen": 7467312, "step": 26670 }, { "epoch": 296.3888888888889, "grad_norm": 1.7326256340766122e-07, "learning_rate": 1.2487532203362576e-05, "loss": 0.0, "num_input_tokens_seen": 7468736, "step": 26675 }, { "epoch": 296.44444444444446, "grad_norm": 6.455957191064954e-07, "learning_rate": 1.247903381249155e-05, "loss": 0.0, "num_input_tokens_seen": 7470176, "step": 26680 }, { "epoch": 296.5, "grad_norm": 6.249241550904117e-07, "learning_rate": 1.2470537352510853e-05, "loss": 0.0, "num_input_tokens_seen": 7471616, "step": 26685 }, { "epoch": 296.55555555555554, "grad_norm": 3.4113461424567504e-06, "learning_rate": 1.2462042824730758e-05, "loss": 0.0, "num_input_tokens_seen": 7472992, "step": 26690 }, { "epoch": 296.6111111111111, "grad_norm": 3.6338253721623914e-06, "learning_rate": 1.245355023046122e-05, "loss": 0.0, "num_input_tokens_seen": 7474368, "step": 26695 }, { "epoch": 296.6666666666667, "grad_norm": 3.3461178645666223e-06, "learning_rate": 1.2445059571011896e-05, "loss": 0.0, "num_input_tokens_seen": 7475776, "step": 26700 }, { "epoch": 296.72222222222223, "grad_norm": 9.365227526814124e-08, "learning_rate": 1.2436570847692173e-05, "loss": 0.0, "num_input_tokens_seen": 7477120, "step": 26705 }, { "epoch": 296.77777777777777, "grad_norm": 3.8556825643354387e-07, "learning_rate": 1.2428084061811096e-05, "loss": 0.0, "num_input_tokens_seen": 7478544, "step": 26710 }, { "epoch": 296.8333333333333, "grad_norm": 2.931525955318648e-07, "learning_rate": 1.2419599214677447e-05, "loss": 0.0, "num_input_tokens_seen": 7479952, "step": 26715 }, { "epoch": 296.8888888888889, "grad_norm": 5.291883553582011e-07, "learning_rate": 1.2411116307599702e-05, "loss": 0.0, "num_input_tokens_seen": 7481328, "step": 26720 }, { "epoch": 296.94444444444446, "grad_norm": 1.0437966011522803e-06, "learning_rate": 1.2402635341886016e-05, "loss": 0.0, "num_input_tokens_seen": 7482784, "step": 26725 }, { "epoch": 297.0, "grad_norm": 4.186495516478317e-06, "learning_rate": 1.2394156318844278e-05, "loss": 0.0, "num_input_tokens_seen": 7484208, "step": 26730 }, { "epoch": 297.05555555555554, "grad_norm": 6.001603765071195e-07, "learning_rate": 1.2385679239782039e-05, "loss": 0.0, "num_input_tokens_seen": 7485600, "step": 26735 }, { "epoch": 297.1111111111111, "grad_norm": 7.342106869145937e-07, "learning_rate": 1.2377204106006585e-05, "loss": 0.0, "num_input_tokens_seen": 7486960, "step": 26740 }, { "epoch": 297.1666666666667, "grad_norm": 6.478545628851862e-07, "learning_rate": 1.2368730918824891e-05, "loss": 0.0, "num_input_tokens_seen": 7488304, "step": 26745 }, { "epoch": 297.22222222222223, "grad_norm": 6.006139869896288e-07, "learning_rate": 1.236025967954362e-05, "loss": 0.0, "num_input_tokens_seen": 7489760, "step": 26750 }, { "epoch": 297.27777777777777, "grad_norm": 8.252287102550326e-07, "learning_rate": 1.2351790389469153e-05, "loss": 0.0, "num_input_tokens_seen": 7491136, "step": 26755 }, { "epoch": 297.3333333333333, "grad_norm": 1.2944160232564172e-07, "learning_rate": 1.234332304990755e-05, "loss": 0.0, "num_input_tokens_seen": 7492544, "step": 26760 }, { "epoch": 297.3888888888889, "grad_norm": 3.3638900731602916e-07, "learning_rate": 1.2334857662164593e-05, "loss": 0.0, "num_input_tokens_seen": 7493952, "step": 26765 }, { "epoch": 297.44444444444446, "grad_norm": 4.98609210808354e-07, "learning_rate": 1.2326394227545743e-05, "loss": 0.0, "num_input_tokens_seen": 7495376, "step": 26770 }, { "epoch": 297.5, "grad_norm": 9.372482736580423e-07, "learning_rate": 1.2317932747356162e-05, "loss": 0.0, "num_input_tokens_seen": 7496752, "step": 26775 }, { "epoch": 297.55555555555554, "grad_norm": 1.023978825287486e-06, "learning_rate": 1.2309473222900726e-05, "loss": 0.0, "num_input_tokens_seen": 7498176, "step": 26780 }, { "epoch": 297.6111111111111, "grad_norm": 3.1589132731824066e-07, "learning_rate": 1.2301015655484006e-05, "loss": 0.0, "num_input_tokens_seen": 7499552, "step": 26785 }, { "epoch": 297.6666666666667, "grad_norm": 6.804295367146551e-07, "learning_rate": 1.2292560046410245e-05, "loss": 0.0, "num_input_tokens_seen": 7500960, "step": 26790 }, { "epoch": 297.72222222222223, "grad_norm": 1.0910932815022534e-06, "learning_rate": 1.228410639698343e-05, "loss": 0.0, "num_input_tokens_seen": 7502384, "step": 26795 }, { "epoch": 297.77777777777777, "grad_norm": 8.697359135112492e-07, "learning_rate": 1.2275654708507195e-05, "loss": 0.0, "num_input_tokens_seen": 7503824, "step": 26800 }, { "epoch": 297.77777777777777, "eval_loss": 1.490308403968811, "eval_runtime": 1.408, "eval_samples_per_second": 28.409, "eval_steps_per_second": 14.204, "num_input_tokens_seen": 7503824, "step": 26800 }, { "epoch": 297.8333333333333, "grad_norm": 2.791995939332992e-06, "learning_rate": 1.2267204982284908e-05, "loss": 0.0, "num_input_tokens_seen": 7505216, "step": 26805 }, { "epoch": 297.8888888888889, "grad_norm": 1.0273179213982075e-06, "learning_rate": 1.2258757219619635e-05, "loss": 0.0, "num_input_tokens_seen": 7506608, "step": 26810 }, { "epoch": 297.94444444444446, "grad_norm": 4.469789587346895e-07, "learning_rate": 1.2250311421814104e-05, "loss": 0.0, "num_input_tokens_seen": 7508000, "step": 26815 }, { "epoch": 298.0, "grad_norm": 1.0191322417085757e-06, "learning_rate": 1.2241867590170772e-05, "loss": 0.0, "num_input_tokens_seen": 7509456, "step": 26820 }, { "epoch": 298.05555555555554, "grad_norm": 4.321857431932585e-06, "learning_rate": 1.2233425725991799e-05, "loss": 0.0, "num_input_tokens_seen": 7510784, "step": 26825 }, { "epoch": 298.1111111111111, "grad_norm": 1.2587491937665618e-06, "learning_rate": 1.2224985830579003e-05, "loss": 0.0, "num_input_tokens_seen": 7512176, "step": 26830 }, { "epoch": 298.1666666666667, "grad_norm": 1.3162180039216764e-06, "learning_rate": 1.2216547905233944e-05, "loss": 0.0, "num_input_tokens_seen": 7513568, "step": 26835 }, { "epoch": 298.22222222222223, "grad_norm": 7.3424729407634e-07, "learning_rate": 1.2208111951257842e-05, "loss": 0.0, "num_input_tokens_seen": 7514992, "step": 26840 }, { "epoch": 298.27777777777777, "grad_norm": 1.1649167390714865e-06, "learning_rate": 1.2199677969951622e-05, "loss": 0.0, "num_input_tokens_seen": 7516400, "step": 26845 }, { "epoch": 298.3333333333333, "grad_norm": 1.8637128960108384e-06, "learning_rate": 1.2191245962615927e-05, "loss": 0.0, "num_input_tokens_seen": 7517856, "step": 26850 }, { "epoch": 298.3888888888889, "grad_norm": 4.804584250450716e-07, "learning_rate": 1.218281593055106e-05, "loss": 0.0, "num_input_tokens_seen": 7519232, "step": 26855 }, { "epoch": 298.44444444444446, "grad_norm": 2.1849288600606087e-07, "learning_rate": 1.217438787505705e-05, "loss": 0.0, "num_input_tokens_seen": 7520672, "step": 26860 }, { "epoch": 298.5, "grad_norm": 5.302038061927306e-07, "learning_rate": 1.2165961797433615e-05, "loss": 0.0, "num_input_tokens_seen": 7522080, "step": 26865 }, { "epoch": 298.55555555555554, "grad_norm": 1.0765257911771187e-06, "learning_rate": 1.215753769898014e-05, "loss": 0.0, "num_input_tokens_seen": 7523488, "step": 26870 }, { "epoch": 298.6111111111111, "grad_norm": 1.332379042651155e-06, "learning_rate": 1.2149115580995755e-05, "loss": 0.0, "num_input_tokens_seen": 7524848, "step": 26875 }, { "epoch": 298.6666666666667, "grad_norm": 3.721070527262782e-07, "learning_rate": 1.2140695444779227e-05, "loss": 0.0, "num_input_tokens_seen": 7526224, "step": 26880 }, { "epoch": 298.72222222222223, "grad_norm": 5.230972419667523e-07, "learning_rate": 1.2132277291629066e-05, "loss": 0.0, "num_input_tokens_seen": 7527600, "step": 26885 }, { "epoch": 298.77777777777777, "grad_norm": 7.787019171701104e-07, "learning_rate": 1.2123861122843458e-05, "loss": 0.0, "num_input_tokens_seen": 7528992, "step": 26890 }, { "epoch": 298.8333333333333, "grad_norm": 7.731135156063829e-06, "learning_rate": 1.2115446939720271e-05, "loss": 0.0, "num_input_tokens_seen": 7530416, "step": 26895 }, { "epoch": 298.8888888888889, "grad_norm": 7.263722636707826e-07, "learning_rate": 1.210703474355708e-05, "loss": 0.0, "num_input_tokens_seen": 7531792, "step": 26900 }, { "epoch": 298.94444444444446, "grad_norm": 6.891668249409122e-07, "learning_rate": 1.2098624535651164e-05, "loss": 0.0, "num_input_tokens_seen": 7533184, "step": 26905 }, { "epoch": 299.0, "grad_norm": 9.846196462603984e-07, "learning_rate": 1.2090216317299477e-05, "loss": 0.0, "num_input_tokens_seen": 7534608, "step": 26910 }, { "epoch": 299.05555555555554, "grad_norm": 0.0016679210821166635, "learning_rate": 1.2081810089798668e-05, "loss": 0.0, "num_input_tokens_seen": 7536000, "step": 26915 }, { "epoch": 299.1111111111111, "grad_norm": 2.0825868887186516e-06, "learning_rate": 1.2073405854445072e-05, "loss": 0.0, "num_input_tokens_seen": 7537376, "step": 26920 }, { "epoch": 299.1666666666667, "grad_norm": 1.3015671811444918e-06, "learning_rate": 1.206500361253474e-05, "loss": 0.0, "num_input_tokens_seen": 7538784, "step": 26925 }, { "epoch": 299.22222222222223, "grad_norm": 1.5092946341610514e-06, "learning_rate": 1.2056603365363409e-05, "loss": 0.0, "num_input_tokens_seen": 7540176, "step": 26930 }, { "epoch": 299.27777777777777, "grad_norm": 6.370943310685107e-07, "learning_rate": 1.2048205114226487e-05, "loss": 0.0, "num_input_tokens_seen": 7541568, "step": 26935 }, { "epoch": 299.3333333333333, "grad_norm": 8.388674928028195e-07, "learning_rate": 1.2039808860419102e-05, "loss": 0.2574, "num_input_tokens_seen": 7542992, "step": 26940 }, { "epoch": 299.3888888888889, "grad_norm": 4.661611819756217e-05, "learning_rate": 1.2031414605236066e-05, "loss": 0.0, "num_input_tokens_seen": 7544416, "step": 26945 }, { "epoch": 299.44444444444446, "grad_norm": 0.1732649803161621, "learning_rate": 1.2023022349971862e-05, "loss": 0.3839, "num_input_tokens_seen": 7545840, "step": 26950 }, { "epoch": 299.5, "grad_norm": 0.0010531890438869596, "learning_rate": 1.20146320959207e-05, "loss": 0.0, "num_input_tokens_seen": 7547264, "step": 26955 }, { "epoch": 299.55555555555554, "grad_norm": 77.03644561767578, "learning_rate": 1.2006243844376445e-05, "loss": 0.1227, "num_input_tokens_seen": 7548608, "step": 26960 }, { "epoch": 299.6111111111111, "grad_norm": 0.0038903276436030865, "learning_rate": 1.1997857596632678e-05, "loss": 0.0001, "num_input_tokens_seen": 7549984, "step": 26965 }, { "epoch": 299.6666666666667, "grad_norm": 0.001155118108727038, "learning_rate": 1.1989473353982672e-05, "loss": 0.0, "num_input_tokens_seen": 7551408, "step": 26970 }, { "epoch": 299.72222222222223, "grad_norm": 0.003287806175649166, "learning_rate": 1.198109111771937e-05, "loss": 0.0, "num_input_tokens_seen": 7552800, "step": 26975 }, { "epoch": 299.77777777777777, "grad_norm": 0.0005199109436944127, "learning_rate": 1.197271088913543e-05, "loss": 0.0, "num_input_tokens_seen": 7554208, "step": 26980 }, { "epoch": 299.8333333333333, "grad_norm": 0.0008122653234750032, "learning_rate": 1.1964332669523182e-05, "loss": 0.0, "num_input_tokens_seen": 7555648, "step": 26985 }, { "epoch": 299.8888888888889, "grad_norm": 0.03117823787033558, "learning_rate": 1.1955956460174645e-05, "loss": 0.0778, "num_input_tokens_seen": 7557040, "step": 26990 }, { "epoch": 299.94444444444446, "grad_norm": 2.5390079021453857, "learning_rate": 1.1947582262381552e-05, "loss": 0.0059, "num_input_tokens_seen": 7558448, "step": 26995 }, { "epoch": 300.0, "grad_norm": 4.42853597633075e-05, "learning_rate": 1.1939210077435293e-05, "loss": 0.0001, "num_input_tokens_seen": 7559856, "step": 27000 }, { "epoch": 300.0, "eval_loss": 0.8738605380058289, "eval_runtime": 1.3844, "eval_samples_per_second": 28.894, "eval_steps_per_second": 14.447, "num_input_tokens_seen": 7559856, "step": 27000 }, { "epoch": 300.05555555555554, "grad_norm": 0.0035666227340698242, "learning_rate": 1.193083990662697e-05, "loss": 0.0, "num_input_tokens_seen": 7561312, "step": 27005 }, { "epoch": 300.1111111111111, "grad_norm": 7.802491745678708e-05, "learning_rate": 1.192247175124738e-05, "loss": 0.0, "num_input_tokens_seen": 7562688, "step": 27010 }, { "epoch": 300.1666666666667, "grad_norm": 12.933464050292969, "learning_rate": 1.191410561258698e-05, "loss": 0.0018, "num_input_tokens_seen": 7564112, "step": 27015 }, { "epoch": 300.22222222222223, "grad_norm": 5.873112058907282e-06, "learning_rate": 1.1905741491935944e-05, "loss": 0.0, "num_input_tokens_seen": 7565504, "step": 27020 }, { "epoch": 300.27777777777777, "grad_norm": 1.4034518471817137e-06, "learning_rate": 1.1897379390584129e-05, "loss": 0.0, "num_input_tokens_seen": 7566928, "step": 27025 }, { "epoch": 300.3333333333333, "grad_norm": 2.7989273803541437e-05, "learning_rate": 1.1889019309821062e-05, "loss": 0.0, "num_input_tokens_seen": 7568352, "step": 27030 }, { "epoch": 300.3888888888889, "grad_norm": 1.2628539707293385e-06, "learning_rate": 1.188066125093599e-05, "loss": 0.0, "num_input_tokens_seen": 7569792, "step": 27035 }, { "epoch": 300.44444444444446, "grad_norm": 5.440564109449042e-06, "learning_rate": 1.1872305215217811e-05, "loss": 0.0, "num_input_tokens_seen": 7571232, "step": 27040 }, { "epoch": 300.5, "grad_norm": 1.0176735258937697e-06, "learning_rate": 1.186395120395514e-05, "loss": 0.0, "num_input_tokens_seen": 7572624, "step": 27045 }, { "epoch": 300.55555555555554, "grad_norm": 1.8061717128148302e-05, "learning_rate": 1.1855599218436283e-05, "loss": 0.0, "num_input_tokens_seen": 7574000, "step": 27050 }, { "epoch": 300.6111111111111, "grad_norm": 8.759829484006332e-07, "learning_rate": 1.1847249259949209e-05, "loss": 0.2273, "num_input_tokens_seen": 7575392, "step": 27055 }, { "epoch": 300.6666666666667, "grad_norm": 2.029967618000228e-06, "learning_rate": 1.1838901329781574e-05, "loss": 0.0, "num_input_tokens_seen": 7576784, "step": 27060 }, { "epoch": 300.72222222222223, "grad_norm": 2.069667516479967e-06, "learning_rate": 1.1830555429220758e-05, "loss": 0.0, "num_input_tokens_seen": 7578176, "step": 27065 }, { "epoch": 300.77777777777777, "grad_norm": 2.0645009044528706e-06, "learning_rate": 1.1822211559553784e-05, "loss": 0.2313, "num_input_tokens_seen": 7579568, "step": 27070 }, { "epoch": 300.8333333333333, "grad_norm": 0.00010377292346674949, "learning_rate": 1.18138697220674e-05, "loss": 0.0001, "num_input_tokens_seen": 7581008, "step": 27075 }, { "epoch": 300.8888888888889, "grad_norm": 0.0013194341445341706, "learning_rate": 1.1805529918048e-05, "loss": 0.0, "num_input_tokens_seen": 7582368, "step": 27080 }, { "epoch": 300.94444444444446, "grad_norm": 0.0021449499763548374, "learning_rate": 1.1797192148781702e-05, "loss": 0.0122, "num_input_tokens_seen": 7583744, "step": 27085 }, { "epoch": 301.0, "grad_norm": 42.75700378417969, "learning_rate": 1.1788856415554297e-05, "loss": 0.1221, "num_input_tokens_seen": 7585184, "step": 27090 }, { "epoch": 301.05555555555554, "grad_norm": 0.0022186487913131714, "learning_rate": 1.1780522719651249e-05, "loss": 0.0, "num_input_tokens_seen": 7586608, "step": 27095 }, { "epoch": 301.1111111111111, "grad_norm": 0.0035103403497487307, "learning_rate": 1.1772191062357721e-05, "loss": 0.0005, "num_input_tokens_seen": 7588032, "step": 27100 }, { "epoch": 301.1666666666667, "grad_norm": 0.0033982787281274796, "learning_rate": 1.1763861444958573e-05, "loss": 0.0001, "num_input_tokens_seen": 7589408, "step": 27105 }, { "epoch": 301.22222222222223, "grad_norm": 0.02388514205813408, "learning_rate": 1.1755533868738317e-05, "loss": 0.0002, "num_input_tokens_seen": 7590816, "step": 27110 }, { "epoch": 301.27777777777777, "grad_norm": 0.0019368481589481235, "learning_rate": 1.1747208334981185e-05, "loss": 0.0001, "num_input_tokens_seen": 7592176, "step": 27115 }, { "epoch": 301.3333333333333, "grad_norm": 0.38929662108421326, "learning_rate": 1.1738884844971067e-05, "loss": 0.0002, "num_input_tokens_seen": 7593600, "step": 27120 }, { "epoch": 301.3888888888889, "grad_norm": 0.0005925617297179997, "learning_rate": 1.1730563399991563e-05, "loss": 0.0, "num_input_tokens_seen": 7595008, "step": 27125 }, { "epoch": 301.44444444444446, "grad_norm": 0.0010848018573597074, "learning_rate": 1.1722244001325938e-05, "loss": 0.0, "num_input_tokens_seen": 7596448, "step": 27130 }, { "epoch": 301.5, "grad_norm": 0.004433032125234604, "learning_rate": 1.1713926650257137e-05, "loss": 0.0, "num_input_tokens_seen": 7597872, "step": 27135 }, { "epoch": 301.55555555555554, "grad_norm": 0.0008170646033249795, "learning_rate": 1.170561134806781e-05, "loss": 0.0, "num_input_tokens_seen": 7599232, "step": 27140 }, { "epoch": 301.6111111111111, "grad_norm": 0.0031124399974942207, "learning_rate": 1.1697298096040287e-05, "loss": 0.0, "num_input_tokens_seen": 7600576, "step": 27145 }, { "epoch": 301.6666666666667, "grad_norm": 0.0009384946315549314, "learning_rate": 1.1688986895456567e-05, "loss": 0.0, "num_input_tokens_seen": 7601968, "step": 27150 }, { "epoch": 301.72222222222223, "grad_norm": 0.0007556932396255434, "learning_rate": 1.1680677747598349e-05, "loss": 0.0, "num_input_tokens_seen": 7603360, "step": 27155 }, { "epoch": 301.77777777777777, "grad_norm": 0.003943213727325201, "learning_rate": 1.1672370653746995e-05, "loss": 0.0, "num_input_tokens_seen": 7604768, "step": 27160 }, { "epoch": 301.8333333333333, "grad_norm": 0.00015823809371795505, "learning_rate": 1.166406561518357e-05, "loss": 0.0, "num_input_tokens_seen": 7606176, "step": 27165 }, { "epoch": 301.8888888888889, "grad_norm": 0.014708645641803741, "learning_rate": 1.1655762633188826e-05, "loss": 0.0, "num_input_tokens_seen": 7607584, "step": 27170 }, { "epoch": 301.94444444444446, "grad_norm": 0.02296844869852066, "learning_rate": 1.1647461709043172e-05, "loss": 0.0, "num_input_tokens_seen": 7608976, "step": 27175 }, { "epoch": 302.0, "grad_norm": 0.0011293638963252306, "learning_rate": 1.1639162844026722e-05, "loss": 0.0, "num_input_tokens_seen": 7610368, "step": 27180 }, { "epoch": 302.05555555555554, "grad_norm": 0.0010438176104798913, "learning_rate": 1.163086603941927e-05, "loss": 0.0, "num_input_tokens_seen": 7611776, "step": 27185 }, { "epoch": 302.1111111111111, "grad_norm": 0.0005148957134224474, "learning_rate": 1.1622571296500273e-05, "loss": 0.0, "num_input_tokens_seen": 7613136, "step": 27190 }, { "epoch": 302.1666666666667, "grad_norm": 0.003624177537858486, "learning_rate": 1.1614278616548904e-05, "loss": 0.0, "num_input_tokens_seen": 7614496, "step": 27195 }, { "epoch": 302.22222222222223, "grad_norm": 0.0002270576951559633, "learning_rate": 1.1605988000843986e-05, "loss": 0.0, "num_input_tokens_seen": 7615904, "step": 27200 }, { "epoch": 302.22222222222223, "eval_loss": 0.6719339489936829, "eval_runtime": 1.3993, "eval_samples_per_second": 28.586, "eval_steps_per_second": 14.293, "num_input_tokens_seen": 7615904, "step": 27200 }, { "epoch": 302.27777777777777, "grad_norm": 0.0002894201024901122, "learning_rate": 1.1597699450664028e-05, "loss": 0.0, "num_input_tokens_seen": 7617328, "step": 27205 }, { "epoch": 302.3333333333333, "grad_norm": 0.0007625645957887173, "learning_rate": 1.1589412967287252e-05, "loss": 0.0, "num_input_tokens_seen": 7618784, "step": 27210 }, { "epoch": 302.3888888888889, "grad_norm": 0.000670638692099601, "learning_rate": 1.1581128551991514e-05, "loss": 0.0, "num_input_tokens_seen": 7620144, "step": 27215 }, { "epoch": 302.44444444444446, "grad_norm": 0.0015695035690441728, "learning_rate": 1.1572846206054383e-05, "loss": 0.0, "num_input_tokens_seen": 7621520, "step": 27220 }, { "epoch": 302.5, "grad_norm": 0.0012007232289761305, "learning_rate": 1.1564565930753113e-05, "loss": 0.0, "num_input_tokens_seen": 7622928, "step": 27225 }, { "epoch": 302.55555555555554, "grad_norm": 9.066532220458612e-05, "learning_rate": 1.1556287727364606e-05, "loss": 0.0, "num_input_tokens_seen": 7624304, "step": 27230 }, { "epoch": 302.6111111111111, "grad_norm": 0.0017218360444530845, "learning_rate": 1.1548011597165489e-05, "loss": 0.0, "num_input_tokens_seen": 7625728, "step": 27235 }, { "epoch": 302.6666666666667, "grad_norm": 0.00011525802983669564, "learning_rate": 1.1539737541432019e-05, "loss": 0.0, "num_input_tokens_seen": 7627120, "step": 27240 }, { "epoch": 302.72222222222223, "grad_norm": 0.00043190098949708045, "learning_rate": 1.1531465561440174e-05, "loss": 0.0, "num_input_tokens_seen": 7628544, "step": 27245 }, { "epoch": 302.77777777777777, "grad_norm": 0.006121811456978321, "learning_rate": 1.1523195658465605e-05, "loss": 0.0, "num_input_tokens_seen": 7629936, "step": 27250 }, { "epoch": 302.8333333333333, "grad_norm": 0.00017632472736295313, "learning_rate": 1.1514927833783618e-05, "loss": 0.0, "num_input_tokens_seen": 7631344, "step": 27255 }, { "epoch": 302.8888888888889, "grad_norm": 0.0003455379046499729, "learning_rate": 1.150666208866922e-05, "loss": 0.0, "num_input_tokens_seen": 7632704, "step": 27260 }, { "epoch": 302.94444444444446, "grad_norm": 0.0007743372698314488, "learning_rate": 1.1498398424397106e-05, "loss": 0.0, "num_input_tokens_seen": 7634128, "step": 27265 }, { "epoch": 303.0, "grad_norm": 0.0006271468591876328, "learning_rate": 1.1490136842241628e-05, "loss": 0.0, "num_input_tokens_seen": 7635536, "step": 27270 }, { "epoch": 303.05555555555554, "grad_norm": 0.0005353181040845811, "learning_rate": 1.1481877343476813e-05, "loss": 0.0, "num_input_tokens_seen": 7636928, "step": 27275 }, { "epoch": 303.1111111111111, "grad_norm": 0.0001232068898389116, "learning_rate": 1.14736199293764e-05, "loss": 0.0, "num_input_tokens_seen": 7638304, "step": 27280 }, { "epoch": 303.1666666666667, "grad_norm": 0.0004032693977933377, "learning_rate": 1.1465364601213771e-05, "loss": 0.0, "num_input_tokens_seen": 7639696, "step": 27285 }, { "epoch": 303.22222222222223, "grad_norm": 0.0001788782828953117, "learning_rate": 1.1457111360262012e-05, "loss": 0.0, "num_input_tokens_seen": 7641072, "step": 27290 }, { "epoch": 303.27777777777777, "grad_norm": 9.886035695672035e-05, "learning_rate": 1.1448860207793869e-05, "loss": 0.0, "num_input_tokens_seen": 7642448, "step": 27295 }, { "epoch": 303.3333333333333, "grad_norm": 0.0011784002417698503, "learning_rate": 1.144061114508177e-05, "loss": 0.0, "num_input_tokens_seen": 7643856, "step": 27300 }, { "epoch": 303.3888888888889, "grad_norm": 0.00029991610790602863, "learning_rate": 1.1432364173397842e-05, "loss": 0.0, "num_input_tokens_seen": 7645264, "step": 27305 }, { "epoch": 303.44444444444446, "grad_norm": 9.979232709156349e-05, "learning_rate": 1.1424119294013852e-05, "loss": 0.0, "num_input_tokens_seen": 7646672, "step": 27310 }, { "epoch": 303.5, "grad_norm": 0.0018093913095071912, "learning_rate": 1.1415876508201279e-05, "loss": 0.0, "num_input_tokens_seen": 7648080, "step": 27315 }, { "epoch": 303.55555555555554, "grad_norm": 6.256101187318563e-05, "learning_rate": 1.140763581723125e-05, "loss": 0.0, "num_input_tokens_seen": 7649504, "step": 27320 }, { "epoch": 303.6111111111111, "grad_norm": 5.318809053278528e-05, "learning_rate": 1.1399397222374588e-05, "loss": 0.0, "num_input_tokens_seen": 7650912, "step": 27325 }, { "epoch": 303.6666666666667, "grad_norm": 0.00011212049139430746, "learning_rate": 1.1391160724901804e-05, "loss": 0.0, "num_input_tokens_seen": 7652336, "step": 27330 }, { "epoch": 303.72222222222223, "grad_norm": 0.00042860579560510814, "learning_rate": 1.138292632608304e-05, "loss": 0.0, "num_input_tokens_seen": 7653744, "step": 27335 }, { "epoch": 303.77777777777777, "grad_norm": 0.0004163425473961979, "learning_rate": 1.1374694027188174e-05, "loss": 0.0, "num_input_tokens_seen": 7655184, "step": 27340 }, { "epoch": 303.8333333333333, "grad_norm": 0.0024434744846075773, "learning_rate": 1.1366463829486711e-05, "loss": 0.0, "num_input_tokens_seen": 7656608, "step": 27345 }, { "epoch": 303.8888888888889, "grad_norm": 3.197714249836281e-05, "learning_rate": 1.1358235734247849e-05, "loss": 0.0, "num_input_tokens_seen": 7657984, "step": 27350 }, { "epoch": 303.94444444444446, "grad_norm": 0.00019759126007556915, "learning_rate": 1.1350009742740478e-05, "loss": 0.0, "num_input_tokens_seen": 7659376, "step": 27355 }, { "epoch": 304.0, "grad_norm": 0.0013757984852418303, "learning_rate": 1.134178585623313e-05, "loss": 0.0, "num_input_tokens_seen": 7660800, "step": 27360 }, { "epoch": 304.05555555555554, "grad_norm": 0.0001831062836572528, "learning_rate": 1.1333564075994047e-05, "loss": 0.0, "num_input_tokens_seen": 7662224, "step": 27365 }, { "epoch": 304.1111111111111, "grad_norm": 0.0002766374673228711, "learning_rate": 1.1325344403291133e-05, "loss": 0.0, "num_input_tokens_seen": 7663648, "step": 27370 }, { "epoch": 304.1666666666667, "grad_norm": 6.386523455148563e-05, "learning_rate": 1.1317126839391951e-05, "loss": 0.0, "num_input_tokens_seen": 7665024, "step": 27375 }, { "epoch": 304.22222222222223, "grad_norm": 0.00034662839607335627, "learning_rate": 1.1308911385563766e-05, "loss": 0.0, "num_input_tokens_seen": 7666448, "step": 27380 }, { "epoch": 304.27777777777777, "grad_norm": 3.6819921660935506e-05, "learning_rate": 1.1300698043073494e-05, "loss": 0.0, "num_input_tokens_seen": 7667840, "step": 27385 }, { "epoch": 304.3333333333333, "grad_norm": 7.731190271442756e-05, "learning_rate": 1.1292486813187736e-05, "loss": 0.0, "num_input_tokens_seen": 7669232, "step": 27390 }, { "epoch": 304.3888888888889, "grad_norm": 0.0003375771048013121, "learning_rate": 1.1284277697172782e-05, "loss": 0.0, "num_input_tokens_seen": 7670608, "step": 27395 }, { "epoch": 304.44444444444446, "grad_norm": 0.0009354552021250129, "learning_rate": 1.127607069629456e-05, "loss": 0.0, "num_input_tokens_seen": 7672000, "step": 27400 }, { "epoch": 304.44444444444446, "eval_loss": 0.7427972555160522, "eval_runtime": 1.387, "eval_samples_per_second": 28.839, "eval_steps_per_second": 14.419, "num_input_tokens_seen": 7672000, "step": 27400 }, { "epoch": 304.5, "grad_norm": 0.00017624975589569658, "learning_rate": 1.1267865811818701e-05, "loss": 0.0, "num_input_tokens_seen": 7673360, "step": 27405 }, { "epoch": 304.55555555555554, "grad_norm": 0.0005901865079067647, "learning_rate": 1.1259663045010513e-05, "loss": 0.0, "num_input_tokens_seen": 7674720, "step": 27410 }, { "epoch": 304.6111111111111, "grad_norm": 9.849787602433935e-05, "learning_rate": 1.1251462397134957e-05, "loss": 0.0, "num_input_tokens_seen": 7676144, "step": 27415 }, { "epoch": 304.6666666666667, "grad_norm": 7.356438436545432e-05, "learning_rate": 1.1243263869456664e-05, "loss": 0.0, "num_input_tokens_seen": 7677568, "step": 27420 }, { "epoch": 304.72222222222223, "grad_norm": 0.00015137935406528413, "learning_rate": 1.1235067463239967e-05, "loss": 0.0, "num_input_tokens_seen": 7678992, "step": 27425 }, { "epoch": 304.77777777777777, "grad_norm": 0.0001851945708040148, "learning_rate": 1.122687317974884e-05, "loss": 0.0, "num_input_tokens_seen": 7680384, "step": 27430 }, { "epoch": 304.8333333333333, "grad_norm": 0.00012106833310099319, "learning_rate": 1.1218681020246963e-05, "loss": 0.0, "num_input_tokens_seen": 7681760, "step": 27435 }, { "epoch": 304.8888888888889, "grad_norm": 0.00029216791153885424, "learning_rate": 1.1210490985997652e-05, "loss": 0.0, "num_input_tokens_seen": 7683168, "step": 27440 }, { "epoch": 304.94444444444446, "grad_norm": 0.0005884245038032532, "learning_rate": 1.1202303078263917e-05, "loss": 0.0, "num_input_tokens_seen": 7684576, "step": 27445 }, { "epoch": 305.0, "grad_norm": 1.0057269719254691e-05, "learning_rate": 1.1194117298308451e-05, "loss": 0.0, "num_input_tokens_seen": 7685984, "step": 27450 }, { "epoch": 305.05555555555554, "grad_norm": 0.00024742356617935, "learning_rate": 1.1185933647393585e-05, "loss": 0.0, "num_input_tokens_seen": 7687376, "step": 27455 }, { "epoch": 305.1111111111111, "grad_norm": 0.00017203387687914073, "learning_rate": 1.1177752126781354e-05, "loss": 0.0, "num_input_tokens_seen": 7688816, "step": 27460 }, { "epoch": 305.1666666666667, "grad_norm": 0.0007502458756789565, "learning_rate": 1.1169572737733441e-05, "loss": 0.0, "num_input_tokens_seen": 7690224, "step": 27465 }, { "epoch": 305.22222222222223, "grad_norm": 6.523515185108408e-05, "learning_rate": 1.1161395481511216e-05, "loss": 0.0, "num_input_tokens_seen": 7691616, "step": 27470 }, { "epoch": 305.27777777777777, "grad_norm": 0.00011777404870372266, "learning_rate": 1.1153220359375722e-05, "loss": 0.0, "num_input_tokens_seen": 7692992, "step": 27475 }, { "epoch": 305.3333333333333, "grad_norm": 0.0002150691143469885, "learning_rate": 1.114504737258765e-05, "loss": 0.0, "num_input_tokens_seen": 7694384, "step": 27480 }, { "epoch": 305.3888888888889, "grad_norm": 0.00022101688955444843, "learning_rate": 1.1136876522407393e-05, "loss": 0.0, "num_input_tokens_seen": 7695824, "step": 27485 }, { "epoch": 305.44444444444446, "grad_norm": 0.00048737230827100575, "learning_rate": 1.1128707810094985e-05, "loss": 0.0, "num_input_tokens_seen": 7697216, "step": 27490 }, { "epoch": 305.5, "grad_norm": 6.376419332809746e-05, "learning_rate": 1.1120541236910157e-05, "loss": 0.0, "num_input_tokens_seen": 7698608, "step": 27495 }, { "epoch": 305.55555555555554, "grad_norm": 0.0004863613867200911, "learning_rate": 1.111237680411229e-05, "loss": 0.0, "num_input_tokens_seen": 7699968, "step": 27500 }, { "epoch": 305.6111111111111, "grad_norm": 0.00023328849056269974, "learning_rate": 1.1104214512960433e-05, "loss": 0.0, "num_input_tokens_seen": 7701360, "step": 27505 }, { "epoch": 305.6666666666667, "grad_norm": 2.7476877221488394e-05, "learning_rate": 1.1096054364713327e-05, "loss": 0.0, "num_input_tokens_seen": 7702752, "step": 27510 }, { "epoch": 305.72222222222223, "grad_norm": 0.00012223521480336785, "learning_rate": 1.1087896360629371e-05, "loss": 0.0, "num_input_tokens_seen": 7704128, "step": 27515 }, { "epoch": 305.77777777777777, "grad_norm": 1.9517867258400656e-05, "learning_rate": 1.107974050196662e-05, "loss": 0.0, "num_input_tokens_seen": 7705520, "step": 27520 }, { "epoch": 305.8333333333333, "grad_norm": 3.2995882065733895e-05, "learning_rate": 1.1071586789982816e-05, "loss": 0.0, "num_input_tokens_seen": 7706928, "step": 27525 }, { "epoch": 305.8888888888889, "grad_norm": 1.772410178091377e-05, "learning_rate": 1.1063435225935373e-05, "loss": 0.0, "num_input_tokens_seen": 7708336, "step": 27530 }, { "epoch": 305.94444444444446, "grad_norm": 0.00011574962991289794, "learning_rate": 1.1055285811081348e-05, "loss": 0.0, "num_input_tokens_seen": 7709728, "step": 27535 }, { "epoch": 306.0, "grad_norm": 0.00027490651700645685, "learning_rate": 1.1047138546677499e-05, "loss": 0.0, "num_input_tokens_seen": 7711104, "step": 27540 }, { "epoch": 306.05555555555554, "grad_norm": 2.554710408730898e-05, "learning_rate": 1.1038993433980219e-05, "loss": 0.0, "num_input_tokens_seen": 7712512, "step": 27545 }, { "epoch": 306.1111111111111, "grad_norm": 4.125846680835821e-05, "learning_rate": 1.1030850474245597e-05, "loss": 0.0, "num_input_tokens_seen": 7713936, "step": 27550 }, { "epoch": 306.1666666666667, "grad_norm": 5.5592692660866305e-05, "learning_rate": 1.102270966872939e-05, "loss": 0.0, "num_input_tokens_seen": 7715312, "step": 27555 }, { "epoch": 306.22222222222223, "grad_norm": 1.742724452924449e-05, "learning_rate": 1.1014571018687e-05, "loss": 0.0, "num_input_tokens_seen": 7716656, "step": 27560 }, { "epoch": 306.27777777777777, "grad_norm": 0.00020896503701806068, "learning_rate": 1.1006434525373502e-05, "loss": 0.0, "num_input_tokens_seen": 7718064, "step": 27565 }, { "epoch": 306.3333333333333, "grad_norm": 0.00013765714538749307, "learning_rate": 1.0998300190043664e-05, "loss": 0.0, "num_input_tokens_seen": 7719456, "step": 27570 }, { "epoch": 306.3888888888889, "grad_norm": 0.0002132035733666271, "learning_rate": 1.0990168013951882e-05, "loss": 0.0, "num_input_tokens_seen": 7720864, "step": 27575 }, { "epoch": 306.44444444444446, "grad_norm": 0.00022499088663607836, "learning_rate": 1.0982037998352263e-05, "loss": 0.0, "num_input_tokens_seen": 7722240, "step": 27580 }, { "epoch": 306.5, "grad_norm": 8.266284567071125e-05, "learning_rate": 1.0973910144498534e-05, "loss": 0.0, "num_input_tokens_seen": 7723648, "step": 27585 }, { "epoch": 306.55555555555554, "grad_norm": 0.0005498593091033399, "learning_rate": 1.0965784453644123e-05, "loss": 0.0, "num_input_tokens_seen": 7725040, "step": 27590 }, { "epoch": 306.6111111111111, "grad_norm": 0.0001309213403146714, "learning_rate": 1.0957660927042127e-05, "loss": 0.0, "num_input_tokens_seen": 7726400, "step": 27595 }, { "epoch": 306.6666666666667, "grad_norm": 0.00010076182661578059, "learning_rate": 1.094953956594527e-05, "loss": 0.0, "num_input_tokens_seen": 7727808, "step": 27600 }, { "epoch": 306.6666666666667, "eval_loss": 0.7945920825004578, "eval_runtime": 1.3892, "eval_samples_per_second": 28.794, "eval_steps_per_second": 14.397, "num_input_tokens_seen": 7727808, "step": 27600 }, { "epoch": 306.72222222222223, "grad_norm": 0.0001577354851178825, "learning_rate": 1.0941420371605981e-05, "loss": 0.0, "num_input_tokens_seen": 7729200, "step": 27605 }, { "epoch": 306.77777777777777, "grad_norm": 2.9180548153817654e-05, "learning_rate": 1.0933303345276354e-05, "loss": 0.0, "num_input_tokens_seen": 7730640, "step": 27610 }, { "epoch": 306.8333333333333, "grad_norm": 0.00012208797852508724, "learning_rate": 1.0925188488208112e-05, "loss": 0.0, "num_input_tokens_seen": 7732016, "step": 27615 }, { "epoch": 306.8888888888889, "grad_norm": 0.0001496079348726198, "learning_rate": 1.0917075801652694e-05, "loss": 0.0, "num_input_tokens_seen": 7733440, "step": 27620 }, { "epoch": 306.94444444444446, "grad_norm": 0.0005601139273494482, "learning_rate": 1.0908965286861151e-05, "loss": 0.0, "num_input_tokens_seen": 7734832, "step": 27625 }, { "epoch": 307.0, "grad_norm": 2.4007765659916913e-06, "learning_rate": 1.090085694508425e-05, "loss": 0.0, "num_input_tokens_seen": 7736224, "step": 27630 }, { "epoch": 307.05555555555554, "grad_norm": 0.0007302479352802038, "learning_rate": 1.089275077757238e-05, "loss": 0.0, "num_input_tokens_seen": 7737616, "step": 27635 }, { "epoch": 307.1111111111111, "grad_norm": 0.00017330834816675633, "learning_rate": 1.0884646785575633e-05, "loss": 0.0, "num_input_tokens_seen": 7739040, "step": 27640 }, { "epoch": 307.1666666666667, "grad_norm": 6.377084355335683e-05, "learning_rate": 1.0876544970343728e-05, "loss": 0.0, "num_input_tokens_seen": 7740464, "step": 27645 }, { "epoch": 307.22222222222223, "grad_norm": 0.00017129775369539857, "learning_rate": 1.0868445333126082e-05, "loss": 0.0, "num_input_tokens_seen": 7741872, "step": 27650 }, { "epoch": 307.27777777777777, "grad_norm": 1.4678266779810656e-05, "learning_rate": 1.0860347875171745e-05, "loss": 0.0, "num_input_tokens_seen": 7743264, "step": 27655 }, { "epoch": 307.3333333333333, "grad_norm": 0.00010987922723870724, "learning_rate": 1.0852252597729465e-05, "loss": 0.0, "num_input_tokens_seen": 7744672, "step": 27660 }, { "epoch": 307.3888888888889, "grad_norm": 3.420119537622668e-05, "learning_rate": 1.0844159502047615e-05, "loss": 0.0, "num_input_tokens_seen": 7746048, "step": 27665 }, { "epoch": 307.44444444444446, "grad_norm": 2.8458627639338374e-05, "learning_rate": 1.0836068589374265e-05, "loss": 0.0, "num_input_tokens_seen": 7747456, "step": 27670 }, { "epoch": 307.5, "grad_norm": 3.3308533602394164e-05, "learning_rate": 1.0827979860957144e-05, "loss": 0.0, "num_input_tokens_seen": 7748864, "step": 27675 }, { "epoch": 307.55555555555554, "grad_norm": 2.9269342121551745e-05, "learning_rate": 1.0819893318043615e-05, "loss": 0.0, "num_input_tokens_seen": 7750272, "step": 27680 }, { "epoch": 307.6111111111111, "grad_norm": 7.091283623594791e-05, "learning_rate": 1.0811808961880734e-05, "loss": 0.0, "num_input_tokens_seen": 7751616, "step": 27685 }, { "epoch": 307.6666666666667, "grad_norm": 7.160233508329839e-05, "learning_rate": 1.080372679371522e-05, "loss": 0.0, "num_input_tokens_seen": 7753024, "step": 27690 }, { "epoch": 307.72222222222223, "grad_norm": 0.0009695137850940228, "learning_rate": 1.0795646814793428e-05, "loss": 0.0, "num_input_tokens_seen": 7754400, "step": 27695 }, { "epoch": 307.77777777777777, "grad_norm": 0.0005725354421883821, "learning_rate": 1.078756902636141e-05, "loss": 0.0, "num_input_tokens_seen": 7755840, "step": 27700 }, { "epoch": 307.8333333333333, "grad_norm": 3.969411409343593e-05, "learning_rate": 1.077949342966485e-05, "loss": 0.0, "num_input_tokens_seen": 7757248, "step": 27705 }, { "epoch": 307.8888888888889, "grad_norm": 2.6996269298251718e-05, "learning_rate": 1.0771420025949103e-05, "loss": 0.0, "num_input_tokens_seen": 7758640, "step": 27710 }, { "epoch": 307.94444444444446, "grad_norm": 4.083068415638991e-05, "learning_rate": 1.0763348816459204e-05, "loss": 0.0, "num_input_tokens_seen": 7759968, "step": 27715 }, { "epoch": 308.0, "grad_norm": 4.2725539969978854e-05, "learning_rate": 1.0755279802439816e-05, "loss": 0.0, "num_input_tokens_seen": 7761408, "step": 27720 }, { "epoch": 308.05555555555554, "grad_norm": 8.460289245704189e-05, "learning_rate": 1.0747212985135293e-05, "loss": 0.0, "num_input_tokens_seen": 7762816, "step": 27725 }, { "epoch": 308.1111111111111, "grad_norm": 3.750878749997355e-05, "learning_rate": 1.073914836578965e-05, "loss": 0.0, "num_input_tokens_seen": 7764208, "step": 27730 }, { "epoch": 308.1666666666667, "grad_norm": 4.975012052454986e-05, "learning_rate": 1.0731085945646529e-05, "loss": 0.0, "num_input_tokens_seen": 7765584, "step": 27735 }, { "epoch": 308.22222222222223, "grad_norm": 0.00011312808783259243, "learning_rate": 1.0723025725949285e-05, "loss": 0.0, "num_input_tokens_seen": 7766976, "step": 27740 }, { "epoch": 308.27777777777777, "grad_norm": 0.0001433149736840278, "learning_rate": 1.0714967707940875e-05, "loss": 0.0, "num_input_tokens_seen": 7768336, "step": 27745 }, { "epoch": 308.3333333333333, "grad_norm": 8.024342969292775e-05, "learning_rate": 1.0706911892863963e-05, "loss": 0.0, "num_input_tokens_seen": 7769776, "step": 27750 }, { "epoch": 308.3888888888889, "grad_norm": 9.145666263066232e-05, "learning_rate": 1.0698858281960866e-05, "loss": 0.0, "num_input_tokens_seen": 7771168, "step": 27755 }, { "epoch": 308.44444444444446, "grad_norm": 6.996115553192794e-05, "learning_rate": 1.069080687647353e-05, "loss": 0.0, "num_input_tokens_seen": 7772544, "step": 27760 }, { "epoch": 308.5, "grad_norm": 0.0009049249347299337, "learning_rate": 1.0682757677643596e-05, "loss": 0.0, "num_input_tokens_seen": 7773984, "step": 27765 }, { "epoch": 308.55555555555554, "grad_norm": 0.0004423370701260865, "learning_rate": 1.0674710686712359e-05, "loss": 0.0, "num_input_tokens_seen": 7775408, "step": 27770 }, { "epoch": 308.6111111111111, "grad_norm": 5.1035975047852844e-05, "learning_rate": 1.0666665904920756e-05, "loss": 0.0, "num_input_tokens_seen": 7776784, "step": 27775 }, { "epoch": 308.6666666666667, "grad_norm": 1.0572928658802994e-05, "learning_rate": 1.0658623333509385e-05, "loss": 0.0, "num_input_tokens_seen": 7778192, "step": 27780 }, { "epoch": 308.72222222222223, "grad_norm": 2.4901772121666e-05, "learning_rate": 1.0650582973718532e-05, "loss": 0.0, "num_input_tokens_seen": 7779632, "step": 27785 }, { "epoch": 308.77777777777777, "grad_norm": 0.0007465848466381431, "learning_rate": 1.0642544826788098e-05, "loss": 0.0, "num_input_tokens_seen": 7781008, "step": 27790 }, { "epoch": 308.8333333333333, "grad_norm": 0.00013745094474870712, "learning_rate": 1.063450889395769e-05, "loss": 0.0, "num_input_tokens_seen": 7782384, "step": 27795 }, { "epoch": 308.8888888888889, "grad_norm": 0.00024019040574785322, "learning_rate": 1.062647517646653e-05, "loss": 0.0, "num_input_tokens_seen": 7783744, "step": 27800 }, { "epoch": 308.8888888888889, "eval_loss": 0.8179023861885071, "eval_runtime": 1.3828, "eval_samples_per_second": 28.926, "eval_steps_per_second": 14.463, "num_input_tokens_seen": 7783744, "step": 27800 }, { "epoch": 308.94444444444446, "grad_norm": 0.000951891066506505, "learning_rate": 1.0618443675553527e-05, "loss": 0.0, "num_input_tokens_seen": 7785152, "step": 27805 }, { "epoch": 309.0, "grad_norm": 0.00014115829253569245, "learning_rate": 1.0610414392457247e-05, "loss": 0.0, "num_input_tokens_seen": 7786560, "step": 27810 }, { "epoch": 309.05555555555554, "grad_norm": 0.00012326514115557075, "learning_rate": 1.0602387328415888e-05, "loss": 0.0, "num_input_tokens_seen": 7787952, "step": 27815 }, { "epoch": 309.1111111111111, "grad_norm": 1.9537950720405206e-05, "learning_rate": 1.0594362484667347e-05, "loss": 0.0, "num_input_tokens_seen": 7789376, "step": 27820 }, { "epoch": 309.1666666666667, "grad_norm": 1.6963482266874053e-05, "learning_rate": 1.0586339862449132e-05, "loss": 0.0, "num_input_tokens_seen": 7790800, "step": 27825 }, { "epoch": 309.22222222222223, "grad_norm": 5.865440107299946e-05, "learning_rate": 1.0578319462998445e-05, "loss": 0.0, "num_input_tokens_seen": 7792192, "step": 27830 }, { "epoch": 309.27777777777777, "grad_norm": 3.932563413400203e-05, "learning_rate": 1.057030128755214e-05, "loss": 0.0, "num_input_tokens_seen": 7793584, "step": 27835 }, { "epoch": 309.3333333333333, "grad_norm": 8.190950029529631e-05, "learning_rate": 1.0562285337346703e-05, "loss": 0.0, "num_input_tokens_seen": 7795024, "step": 27840 }, { "epoch": 309.3888888888889, "grad_norm": 9.441626025363803e-05, "learning_rate": 1.0554271613618308e-05, "loss": 0.0, "num_input_tokens_seen": 7796432, "step": 27845 }, { "epoch": 309.44444444444446, "grad_norm": 6.014222526573576e-05, "learning_rate": 1.054626011760276e-05, "loss": 0.0, "num_input_tokens_seen": 7797824, "step": 27850 }, { "epoch": 309.5, "grad_norm": 5.260671969153918e-05, "learning_rate": 1.0538250850535549e-05, "loss": 0.0, "num_input_tokens_seen": 7799200, "step": 27855 }, { "epoch": 309.55555555555554, "grad_norm": 4.0215356420958415e-05, "learning_rate": 1.0530243813651794e-05, "loss": 0.0, "num_input_tokens_seen": 7800576, "step": 27860 }, { "epoch": 309.6111111111111, "grad_norm": 0.0002975553215947002, "learning_rate": 1.0522239008186271e-05, "loss": 0.0, "num_input_tokens_seen": 7801936, "step": 27865 }, { "epoch": 309.6666666666667, "grad_norm": 9.562168997945264e-05, "learning_rate": 1.0514236435373434e-05, "loss": 0.0, "num_input_tokens_seen": 7803312, "step": 27870 }, { "epoch": 309.72222222222223, "grad_norm": 0.00018062406161334366, "learning_rate": 1.0506236096447386e-05, "loss": 0.0, "num_input_tokens_seen": 7804736, "step": 27875 }, { "epoch": 309.77777777777777, "grad_norm": 1.1450836609583348e-05, "learning_rate": 1.049823799264186e-05, "loss": 0.0, "num_input_tokens_seen": 7806144, "step": 27880 }, { "epoch": 309.8333333333333, "grad_norm": 6.402206781785935e-05, "learning_rate": 1.049024212519028e-05, "loss": 0.0, "num_input_tokens_seen": 7807536, "step": 27885 }, { "epoch": 309.8888888888889, "grad_norm": 3.94010785385035e-05, "learning_rate": 1.0482248495325713e-05, "loss": 0.0, "num_input_tokens_seen": 7808944, "step": 27890 }, { "epoch": 309.94444444444446, "grad_norm": 0.0007629558676853776, "learning_rate": 1.047425710428086e-05, "loss": 0.0, "num_input_tokens_seen": 7810336, "step": 27895 }, { "epoch": 310.0, "grad_norm": 4.4603253627428785e-05, "learning_rate": 1.0466267953288114e-05, "loss": 0.0, "num_input_tokens_seen": 7811744, "step": 27900 }, { "epoch": 310.05555555555554, "grad_norm": 1.1294704563624691e-05, "learning_rate": 1.0458281043579482e-05, "loss": 0.0, "num_input_tokens_seen": 7813104, "step": 27905 }, { "epoch": 310.1111111111111, "grad_norm": 0.0003707159776240587, "learning_rate": 1.0450296376386657e-05, "loss": 0.0, "num_input_tokens_seen": 7814496, "step": 27910 }, { "epoch": 310.1666666666667, "grad_norm": 1.4461575119639747e-05, "learning_rate": 1.044231395294098e-05, "loss": 0.0, "num_input_tokens_seen": 7815888, "step": 27915 }, { "epoch": 310.22222222222223, "grad_norm": 0.0004038590705022216, "learning_rate": 1.0434333774473435e-05, "loss": 0.0, "num_input_tokens_seen": 7817328, "step": 27920 }, { "epoch": 310.27777777777777, "grad_norm": 9.373046850669198e-06, "learning_rate": 1.0426355842214657e-05, "loss": 0.0, "num_input_tokens_seen": 7818752, "step": 27925 }, { "epoch": 310.3333333333333, "grad_norm": 0.00025492036365903914, "learning_rate": 1.0418380157394963e-05, "loss": 0.0, "num_input_tokens_seen": 7820176, "step": 27930 }, { "epoch": 310.3888888888889, "grad_norm": 1.4779683624510653e-05, "learning_rate": 1.0410406721244281e-05, "loss": 0.0, "num_input_tokens_seen": 7821568, "step": 27935 }, { "epoch": 310.44444444444446, "grad_norm": 0.00041671114740893245, "learning_rate": 1.0402435534992238e-05, "loss": 0.0, "num_input_tokens_seen": 7823024, "step": 27940 }, { "epoch": 310.5, "grad_norm": 0.00038930686423555017, "learning_rate": 1.0394466599868071e-05, "loss": 0.0, "num_input_tokens_seen": 7824400, "step": 27945 }, { "epoch": 310.55555555555554, "grad_norm": 3.659037247416563e-05, "learning_rate": 1.0386499917100697e-05, "loss": 0.0, "num_input_tokens_seen": 7825792, "step": 27950 }, { "epoch": 310.6111111111111, "grad_norm": 4.080358485225588e-05, "learning_rate": 1.0378535487918692e-05, "loss": 0.0, "num_input_tokens_seen": 7827168, "step": 27955 }, { "epoch": 310.6666666666667, "grad_norm": 0.0009728171862661839, "learning_rate": 1.037057331355025e-05, "loss": 0.0, "num_input_tokens_seen": 7828624, "step": 27960 }, { "epoch": 310.72222222222223, "grad_norm": 1.9029690520255826e-05, "learning_rate": 1.0362613395223247e-05, "loss": 0.0, "num_input_tokens_seen": 7830000, "step": 27965 }, { "epoch": 310.77777777777777, "grad_norm": 7.877929419919383e-06, "learning_rate": 1.0354655734165212e-05, "loss": 0.0, "num_input_tokens_seen": 7831424, "step": 27970 }, { "epoch": 310.8333333333333, "grad_norm": 2.7476948162075132e-05, "learning_rate": 1.03467003316033e-05, "loss": 0.0, "num_input_tokens_seen": 7832848, "step": 27975 }, { "epoch": 310.8888888888889, "grad_norm": 3.526772707118653e-05, "learning_rate": 1.033874718876435e-05, "loss": 0.0, "num_input_tokens_seen": 7834224, "step": 27980 }, { "epoch": 310.94444444444446, "grad_norm": 7.56995941628702e-05, "learning_rate": 1.0330796306874818e-05, "loss": 0.0, "num_input_tokens_seen": 7835600, "step": 27985 }, { "epoch": 311.0, "grad_norm": 3.188339542248286e-05, "learning_rate": 1.032284768716085e-05, "loss": 0.0, "num_input_tokens_seen": 7837008, "step": 27990 }, { "epoch": 311.05555555555554, "grad_norm": 1.0887617463595234e-05, "learning_rate": 1.0314901330848206e-05, "loss": 0.0, "num_input_tokens_seen": 7838416, "step": 27995 }, { "epoch": 311.1111111111111, "grad_norm": 2.4017586838454008e-05, "learning_rate": 1.030695723916233e-05, "loss": 0.0, "num_input_tokens_seen": 7839808, "step": 28000 }, { "epoch": 311.1111111111111, "eval_loss": 0.8505862951278687, "eval_runtime": 1.3935, "eval_samples_per_second": 28.705, "eval_steps_per_second": 14.352, "num_input_tokens_seen": 7839808, "step": 28000 }, { "epoch": 311.1666666666667, "grad_norm": 4.1298484575236216e-05, "learning_rate": 1.0299015413328289e-05, "loss": 0.0, "num_input_tokens_seen": 7841184, "step": 28005 }, { "epoch": 311.22222222222223, "grad_norm": 6.183829827932641e-05, "learning_rate": 1.0291075854570809e-05, "loss": 0.0, "num_input_tokens_seen": 7842576, "step": 28010 }, { "epoch": 311.27777777777777, "grad_norm": 2.7924917958443984e-05, "learning_rate": 1.0283138564114275e-05, "loss": 0.0, "num_input_tokens_seen": 7844016, "step": 28015 }, { "epoch": 311.3333333333333, "grad_norm": 2.286022026964929e-05, "learning_rate": 1.027520354318273e-05, "loss": 0.0, "num_input_tokens_seen": 7845376, "step": 28020 }, { "epoch": 311.3888888888889, "grad_norm": 0.0016199415549635887, "learning_rate": 1.0267270792999828e-05, "loss": 0.0, "num_input_tokens_seen": 7846768, "step": 28025 }, { "epoch": 311.44444444444446, "grad_norm": 5.280313780531287e-05, "learning_rate": 1.0259340314788919e-05, "loss": 0.0, "num_input_tokens_seen": 7848192, "step": 28030 }, { "epoch": 311.5, "grad_norm": 7.922801887616515e-05, "learning_rate": 1.0251412109772979e-05, "loss": 0.0, "num_input_tokens_seen": 7849600, "step": 28035 }, { "epoch": 311.55555555555554, "grad_norm": 6.271780875977129e-05, "learning_rate": 1.0243486179174627e-05, "loss": 0.0, "num_input_tokens_seen": 7851024, "step": 28040 }, { "epoch": 311.6111111111111, "grad_norm": 0.00011605830513872206, "learning_rate": 1.0235562524216158e-05, "loss": 0.0, "num_input_tokens_seen": 7852432, "step": 28045 }, { "epoch": 311.6666666666667, "grad_norm": 2.232694168924354e-05, "learning_rate": 1.022764114611948e-05, "loss": 0.0, "num_input_tokens_seen": 7853824, "step": 28050 }, { "epoch": 311.72222222222223, "grad_norm": 0.00016817651339806616, "learning_rate": 1.0219722046106178e-05, "loss": 0.0, "num_input_tokens_seen": 7855216, "step": 28055 }, { "epoch": 311.77777777777777, "grad_norm": 4.047450056532398e-05, "learning_rate": 1.0211805225397486e-05, "loss": 0.0, "num_input_tokens_seen": 7856624, "step": 28060 }, { "epoch": 311.8333333333333, "grad_norm": 2.8716534870909527e-05, "learning_rate": 1.020389068521426e-05, "loss": 0.0, "num_input_tokens_seen": 7858016, "step": 28065 }, { "epoch": 311.8888888888889, "grad_norm": 0.00010749873035820201, "learning_rate": 1.0195978426777039e-05, "loss": 0.0, "num_input_tokens_seen": 7859424, "step": 28070 }, { "epoch": 311.94444444444446, "grad_norm": 1.507125489297323e-05, "learning_rate": 1.0188068451305982e-05, "loss": 0.0, "num_input_tokens_seen": 7860800, "step": 28075 }, { "epoch": 312.0, "grad_norm": 0.00047319557052105665, "learning_rate": 1.0180160760020902e-05, "loss": 0.0, "num_input_tokens_seen": 7862176, "step": 28080 }, { "epoch": 312.05555555555554, "grad_norm": 5.175996193429455e-05, "learning_rate": 1.0172255354141278e-05, "loss": 0.0, "num_input_tokens_seen": 7863568, "step": 28085 }, { "epoch": 312.1111111111111, "grad_norm": 0.00014567548350896686, "learning_rate": 1.0164352234886205e-05, "loss": 0.0, "num_input_tokens_seen": 7864944, "step": 28090 }, { "epoch": 312.1666666666667, "grad_norm": 0.0001452109863748774, "learning_rate": 1.0156451403474454e-05, "loss": 0.0, "num_input_tokens_seen": 7866336, "step": 28095 }, { "epoch": 312.22222222222223, "grad_norm": 4.413554415805265e-05, "learning_rate": 1.0148552861124443e-05, "loss": 0.0, "num_input_tokens_seen": 7867744, "step": 28100 }, { "epoch": 312.27777777777777, "grad_norm": 0.0005743435467593372, "learning_rate": 1.0140656609054205e-05, "loss": 0.0, "num_input_tokens_seen": 7869152, "step": 28105 }, { "epoch": 312.3333333333333, "grad_norm": 6.32058217888698e-05, "learning_rate": 1.0132762648481455e-05, "loss": 0.0, "num_input_tokens_seen": 7870560, "step": 28110 }, { "epoch": 312.3888888888889, "grad_norm": 1.7528425360069377e-06, "learning_rate": 1.0124870980623543e-05, "loss": 0.0, "num_input_tokens_seen": 7871968, "step": 28115 }, { "epoch": 312.44444444444446, "grad_norm": 0.00041552860056981444, "learning_rate": 1.0116981606697453e-05, "loss": 0.0, "num_input_tokens_seen": 7873376, "step": 28120 }, { "epoch": 312.5, "grad_norm": 1.496157801739173e-05, "learning_rate": 1.0109094527919838e-05, "loss": 0.0, "num_input_tokens_seen": 7874736, "step": 28125 }, { "epoch": 312.55555555555554, "grad_norm": 0.00024174092686735094, "learning_rate": 1.010120974550697e-05, "loss": 0.0, "num_input_tokens_seen": 7876096, "step": 28130 }, { "epoch": 312.6111111111111, "grad_norm": 2.5678200472611934e-05, "learning_rate": 1.0093327260674795e-05, "loss": 0.0, "num_input_tokens_seen": 7877536, "step": 28135 }, { "epoch": 312.6666666666667, "grad_norm": 6.871141522424296e-05, "learning_rate": 1.0085447074638878e-05, "loss": 0.0, "num_input_tokens_seen": 7878928, "step": 28140 }, { "epoch": 312.72222222222223, "grad_norm": 0.0003795597585849464, "learning_rate": 1.0077569188614461e-05, "loss": 0.0, "num_input_tokens_seen": 7880320, "step": 28145 }, { "epoch": 312.77777777777777, "grad_norm": 8.648014954815153e-06, "learning_rate": 1.0069693603816393e-05, "loss": 0.0, "num_input_tokens_seen": 7881760, "step": 28150 }, { "epoch": 312.8333333333333, "grad_norm": 6.905372720211744e-05, "learning_rate": 1.0061820321459204e-05, "loss": 0.0, "num_input_tokens_seen": 7883168, "step": 28155 }, { "epoch": 312.8888888888889, "grad_norm": 2.026065885729622e-05, "learning_rate": 1.0053949342757038e-05, "loss": 0.0, "num_input_tokens_seen": 7884592, "step": 28160 }, { "epoch": 312.94444444444446, "grad_norm": 4.5028758904663846e-05, "learning_rate": 1.0046080668923717e-05, "loss": 0.0, "num_input_tokens_seen": 7886032, "step": 28165 }, { "epoch": 313.0, "grad_norm": 0.00019306769536342472, "learning_rate": 1.003821430117267e-05, "loss": 0.0, "num_input_tokens_seen": 7887424, "step": 28170 }, { "epoch": 313.05555555555554, "grad_norm": 3.957882654503919e-05, "learning_rate": 1.0030350240716999e-05, "loss": 0.0, "num_input_tokens_seen": 7888816, "step": 28175 }, { "epoch": 313.1111111111111, "grad_norm": 1.2871920262114145e-05, "learning_rate": 1.0022488488769449e-05, "loss": 0.0, "num_input_tokens_seen": 7890192, "step": 28180 }, { "epoch": 313.1666666666667, "grad_norm": 5.3552335884887725e-05, "learning_rate": 1.0014629046542387e-05, "loss": 0.0, "num_input_tokens_seen": 7891632, "step": 28185 }, { "epoch": 313.22222222222223, "grad_norm": 4.4890810386277735e-05, "learning_rate": 1.0006771915247842e-05, "loss": 0.0, "num_input_tokens_seen": 7893056, "step": 28190 }, { "epoch": 313.27777777777777, "grad_norm": 2.5515204470138997e-05, "learning_rate": 9.998917096097495e-06, "loss": 0.0, "num_input_tokens_seen": 7894448, "step": 28195 }, { "epoch": 313.3333333333333, "grad_norm": 0.00019477619207464159, "learning_rate": 9.991064590302638e-06, "loss": 0.0, "num_input_tokens_seen": 7895872, "step": 28200 }, { "epoch": 313.3333333333333, "eval_loss": 0.8725196719169617, "eval_runtime": 1.395, "eval_samples_per_second": 28.674, "eval_steps_per_second": 14.337, "num_input_tokens_seen": 7895872, "step": 28200 }, { "epoch": 313.3888888888889, "grad_norm": 2.5258836103603244e-05, "learning_rate": 9.983214399074241e-06, "loss": 0.0, "num_input_tokens_seen": 7897264, "step": 28205 }, { "epoch": 313.44444444444446, "grad_norm": 1.8183485735789873e-05, "learning_rate": 9.975366523622893e-06, "loss": 0.0, "num_input_tokens_seen": 7898656, "step": 28210 }, { "epoch": 313.5, "grad_norm": 1.8521861420595087e-05, "learning_rate": 9.967520965158841e-06, "loss": 0.0, "num_input_tokens_seen": 7900048, "step": 28215 }, { "epoch": 313.55555555555554, "grad_norm": 7.282390470209066e-06, "learning_rate": 9.95967772489197e-06, "loss": 0.0, "num_input_tokens_seen": 7901472, "step": 28220 }, { "epoch": 313.6111111111111, "grad_norm": 1.3222031157056335e-05, "learning_rate": 9.951836804031794e-06, "loss": 0.0, "num_input_tokens_seen": 7902880, "step": 28225 }, { "epoch": 313.6666666666667, "grad_norm": 0.0003636404580902308, "learning_rate": 9.943998203787489e-06, "loss": 0.0, "num_input_tokens_seen": 7904288, "step": 28230 }, { "epoch": 313.72222222222223, "grad_norm": 8.759262527746614e-06, "learning_rate": 9.936161925367874e-06, "loss": 0.0, "num_input_tokens_seen": 7905696, "step": 28235 }, { "epoch": 313.77777777777777, "grad_norm": 0.00035574284265749156, "learning_rate": 9.928327969981386e-06, "loss": 0.0, "num_input_tokens_seen": 7907072, "step": 28240 }, { "epoch": 313.8333333333333, "grad_norm": 2.8922750061610714e-05, "learning_rate": 9.920496338836135e-06, "loss": 0.0, "num_input_tokens_seen": 7908480, "step": 28245 }, { "epoch": 313.8888888888889, "grad_norm": 0.00016108179988805205, "learning_rate": 9.912667033139844e-06, "loss": 0.0, "num_input_tokens_seen": 7909824, "step": 28250 }, { "epoch": 313.94444444444446, "grad_norm": 5.088607213110663e-05, "learning_rate": 9.904840054099893e-06, "loss": 0.0, "num_input_tokens_seen": 7911184, "step": 28255 }, { "epoch": 314.0, "grad_norm": 4.160875505476724e-06, "learning_rate": 9.897015402923312e-06, "loss": 0.0, "num_input_tokens_seen": 7912528, "step": 28260 }, { "epoch": 314.05555555555554, "grad_norm": 2.240431604150217e-05, "learning_rate": 9.889193080816744e-06, "loss": 0.0, "num_input_tokens_seen": 7913936, "step": 28265 }, { "epoch": 314.1111111111111, "grad_norm": 7.988962352101225e-06, "learning_rate": 9.881373088986498e-06, "loss": 0.0, "num_input_tokens_seen": 7915328, "step": 28270 }, { "epoch": 314.1666666666667, "grad_norm": 2.3143233192968182e-05, "learning_rate": 9.873555428638523e-06, "loss": 0.0, "num_input_tokens_seen": 7916688, "step": 28275 }, { "epoch": 314.22222222222223, "grad_norm": 9.073469846043736e-05, "learning_rate": 9.865740100978383e-06, "loss": 0.0, "num_input_tokens_seen": 7918128, "step": 28280 }, { "epoch": 314.27777777777777, "grad_norm": 1.9936337594117504e-06, "learning_rate": 9.857927107211315e-06, "loss": 0.0, "num_input_tokens_seen": 7919536, "step": 28285 }, { "epoch": 314.3333333333333, "grad_norm": 2.3230354599945713e-06, "learning_rate": 9.850116448542177e-06, "loss": 0.0, "num_input_tokens_seen": 7920896, "step": 28290 }, { "epoch": 314.3888888888889, "grad_norm": 5.8691424783319235e-05, "learning_rate": 9.842308126175457e-06, "loss": 0.0, "num_input_tokens_seen": 7922304, "step": 28295 }, { "epoch": 314.44444444444446, "grad_norm": 1.2292918654566165e-05, "learning_rate": 9.834502141315315e-06, "loss": 0.0, "num_input_tokens_seen": 7923744, "step": 28300 }, { "epoch": 314.5, "grad_norm": 1.6689331459929235e-05, "learning_rate": 9.82669849516552e-06, "loss": 0.0, "num_input_tokens_seen": 7925152, "step": 28305 }, { "epoch": 314.55555555555554, "grad_norm": 0.00019510289712343365, "learning_rate": 9.818897188929493e-06, "loss": 0.0, "num_input_tokens_seen": 7926544, "step": 28310 }, { "epoch": 314.6111111111111, "grad_norm": 9.758935448189732e-06, "learning_rate": 9.811098223810309e-06, "loss": 0.0, "num_input_tokens_seen": 7927936, "step": 28315 }, { "epoch": 314.6666666666667, "grad_norm": 2.752724503807258e-05, "learning_rate": 9.803301601010641e-06, "loss": 0.0, "num_input_tokens_seen": 7929312, "step": 28320 }, { "epoch": 314.72222222222223, "grad_norm": 0.00017252775433007628, "learning_rate": 9.795507321732853e-06, "loss": 0.0, "num_input_tokens_seen": 7930688, "step": 28325 }, { "epoch": 314.77777777777777, "grad_norm": 9.422603761777282e-05, "learning_rate": 9.787715387178898e-06, "loss": 0.0, "num_input_tokens_seen": 7932144, "step": 28330 }, { "epoch": 314.8333333333333, "grad_norm": 2.089017016260186e-06, "learning_rate": 9.779925798550399e-06, "loss": 0.0, "num_input_tokens_seen": 7933536, "step": 28335 }, { "epoch": 314.8888888888889, "grad_norm": 2.7648841296468163e-06, "learning_rate": 9.772138557048619e-06, "loss": 0.0, "num_input_tokens_seen": 7934896, "step": 28340 }, { "epoch": 314.94444444444446, "grad_norm": 1.5489140423596837e-05, "learning_rate": 9.764353663874426e-06, "loss": 0.0, "num_input_tokens_seen": 7936336, "step": 28345 }, { "epoch": 315.0, "grad_norm": 2.8957017548236763e-06, "learning_rate": 9.756571120228375e-06, "loss": 0.0, "num_input_tokens_seen": 7937744, "step": 28350 }, { "epoch": 315.05555555555554, "grad_norm": 0.0004423517093528062, "learning_rate": 9.748790927310605e-06, "loss": 0.0, "num_input_tokens_seen": 7939104, "step": 28355 }, { "epoch": 315.1111111111111, "grad_norm": 1.3427674275590107e-05, "learning_rate": 9.741013086320946e-06, "loss": 0.0, "num_input_tokens_seen": 7940496, "step": 28360 }, { "epoch": 315.1666666666667, "grad_norm": 1.3721846698899753e-05, "learning_rate": 9.733237598458821e-06, "loss": 0.0, "num_input_tokens_seen": 7941888, "step": 28365 }, { "epoch": 315.22222222222223, "grad_norm": 1.797709273887449e-06, "learning_rate": 9.725464464923308e-06, "loss": 0.0, "num_input_tokens_seen": 7943280, "step": 28370 }, { "epoch": 315.27777777777777, "grad_norm": 7.283834565896541e-05, "learning_rate": 9.717693686913123e-06, "loss": 0.0, "num_input_tokens_seen": 7944688, "step": 28375 }, { "epoch": 315.3333333333333, "grad_norm": 5.6913733715191483e-05, "learning_rate": 9.709925265626632e-06, "loss": 0.0, "num_input_tokens_seen": 7946064, "step": 28380 }, { "epoch": 315.3888888888889, "grad_norm": 4.097349119547289e-06, "learning_rate": 9.702159202261801e-06, "loss": 0.0, "num_input_tokens_seen": 7947472, "step": 28385 }, { "epoch": 315.44444444444446, "grad_norm": 2.0539389879559167e-05, "learning_rate": 9.694395498016268e-06, "loss": 0.0, "num_input_tokens_seen": 7948880, "step": 28390 }, { "epoch": 315.5, "grad_norm": 1.399872144247638e-05, "learning_rate": 9.686634154087298e-06, "loss": 0.0, "num_input_tokens_seen": 7950288, "step": 28395 }, { "epoch": 315.55555555555554, "grad_norm": 0.0003507899818941951, "learning_rate": 9.678875171671776e-06, "loss": 0.0, "num_input_tokens_seen": 7951664, "step": 28400 }, { "epoch": 315.55555555555554, "eval_loss": 0.8748008608818054, "eval_runtime": 1.3927, "eval_samples_per_second": 28.721, "eval_steps_per_second": 14.36, "num_input_tokens_seen": 7951664, "step": 28400 }, { "epoch": 315.6111111111111, "grad_norm": 5.070224688097369e-06, "learning_rate": 9.671118551966246e-06, "loss": 0.0, "num_input_tokens_seen": 7953120, "step": 28405 }, { "epoch": 315.6666666666667, "grad_norm": 1.7310335351794492e-06, "learning_rate": 9.66336429616686e-06, "loss": 0.0, "num_input_tokens_seen": 7954544, "step": 28410 }, { "epoch": 315.72222222222223, "grad_norm": 9.602675709174946e-05, "learning_rate": 9.655612405469436e-06, "loss": 0.0, "num_input_tokens_seen": 7955936, "step": 28415 }, { "epoch": 315.77777777777777, "grad_norm": 4.8611942474963143e-05, "learning_rate": 9.647862881069413e-06, "loss": 0.0, "num_input_tokens_seen": 7957328, "step": 28420 }, { "epoch": 315.8333333333333, "grad_norm": 1.8484014390196535e-06, "learning_rate": 9.640115724161855e-06, "loss": 0.0, "num_input_tokens_seen": 7958736, "step": 28425 }, { "epoch": 315.8888888888889, "grad_norm": 1.3604171726910863e-05, "learning_rate": 9.632370935941483e-06, "loss": 0.0, "num_input_tokens_seen": 7960128, "step": 28430 }, { "epoch": 315.94444444444446, "grad_norm": 3.989548076788196e-06, "learning_rate": 9.624628517602634e-06, "loss": 0.0, "num_input_tokens_seen": 7961536, "step": 28435 }, { "epoch": 316.0, "grad_norm": 9.970891369448509e-06, "learning_rate": 9.61688847033928e-06, "loss": 0.0, "num_input_tokens_seen": 7962928, "step": 28440 }, { "epoch": 316.05555555555554, "grad_norm": 3.7587327824439853e-05, "learning_rate": 9.609150795345051e-06, "loss": 0.0, "num_input_tokens_seen": 7964336, "step": 28445 }, { "epoch": 316.1111111111111, "grad_norm": 0.0011120425770059228, "learning_rate": 9.601415493813171e-06, "loss": 0.0, "num_input_tokens_seen": 7965728, "step": 28450 }, { "epoch": 316.1666666666667, "grad_norm": 0.00014461118553299457, "learning_rate": 9.593682566936533e-06, "loss": 0.0, "num_input_tokens_seen": 7967152, "step": 28455 }, { "epoch": 316.22222222222223, "grad_norm": 9.008478627947625e-06, "learning_rate": 9.58595201590766e-06, "loss": 0.0, "num_input_tokens_seen": 7968512, "step": 28460 }, { "epoch": 316.27777777777777, "grad_norm": 1.8407441530143842e-05, "learning_rate": 9.578223841918681e-06, "loss": 0.0, "num_input_tokens_seen": 7969952, "step": 28465 }, { "epoch": 316.3333333333333, "grad_norm": 1.4959803593228571e-05, "learning_rate": 9.570498046161389e-06, "loss": 0.0, "num_input_tokens_seen": 7971392, "step": 28470 }, { "epoch": 316.3888888888889, "grad_norm": 1.5498342690989375e-05, "learning_rate": 9.562774629827206e-06, "loss": 0.0, "num_input_tokens_seen": 7972800, "step": 28475 }, { "epoch": 316.44444444444446, "grad_norm": 1.0418476449558511e-05, "learning_rate": 9.555053594107163e-06, "loss": 0.0, "num_input_tokens_seen": 7974208, "step": 28480 }, { "epoch": 316.5, "grad_norm": 2.8077934075554367e-06, "learning_rate": 9.547334940191957e-06, "loss": 0.0, "num_input_tokens_seen": 7975584, "step": 28485 }, { "epoch": 316.55555555555554, "grad_norm": 0.0004544483090285212, "learning_rate": 9.539618669271886e-06, "loss": 0.0, "num_input_tokens_seen": 7977008, "step": 28490 }, { "epoch": 316.6111111111111, "grad_norm": 2.284672336827498e-05, "learning_rate": 9.531904782536904e-06, "loss": 0.0, "num_input_tokens_seen": 7978352, "step": 28495 }, { "epoch": 316.6666666666667, "grad_norm": 5.175991645955946e-06, "learning_rate": 9.524193281176597e-06, "loss": 0.0, "num_input_tokens_seen": 7979744, "step": 28500 }, { "epoch": 316.72222222222223, "grad_norm": 4.3679538066498935e-05, "learning_rate": 9.516484166380165e-06, "loss": 0.0, "num_input_tokens_seen": 7981120, "step": 28505 }, { "epoch": 316.77777777777777, "grad_norm": 5.891370528843254e-05, "learning_rate": 9.508777439336447e-06, "loss": 0.0, "num_input_tokens_seen": 7982480, "step": 28510 }, { "epoch": 316.8333333333333, "grad_norm": 1.0360291526012588e-05, "learning_rate": 9.50107310123393e-06, "loss": 0.0, "num_input_tokens_seen": 7983872, "step": 28515 }, { "epoch": 316.8888888888889, "grad_norm": 6.0474598285509273e-05, "learning_rate": 9.493371153260702e-06, "loss": 0.0, "num_input_tokens_seen": 7985328, "step": 28520 }, { "epoch": 316.94444444444446, "grad_norm": 6.370928167598322e-05, "learning_rate": 9.485671596604523e-06, "loss": 0.0, "num_input_tokens_seen": 7986704, "step": 28525 }, { "epoch": 317.0, "grad_norm": 2.989140739373397e-05, "learning_rate": 9.477974432452738e-06, "loss": 0.0, "num_input_tokens_seen": 7988096, "step": 28530 }, { "epoch": 317.05555555555554, "grad_norm": 9.286787644668948e-06, "learning_rate": 9.470279661992356e-06, "loss": 0.0, "num_input_tokens_seen": 7989472, "step": 28535 }, { "epoch": 317.1111111111111, "grad_norm": 2.4649716579006054e-06, "learning_rate": 9.462587286410021e-06, "loss": 0.0, "num_input_tokens_seen": 7990864, "step": 28540 }, { "epoch": 317.1666666666667, "grad_norm": 0.0001672737707849592, "learning_rate": 9.454897306891972e-06, "loss": 0.0, "num_input_tokens_seen": 7992224, "step": 28545 }, { "epoch": 317.22222222222223, "grad_norm": 2.4835313524818048e-05, "learning_rate": 9.44720972462411e-06, "loss": 0.0, "num_input_tokens_seen": 7993632, "step": 28550 }, { "epoch": 317.27777777777777, "grad_norm": 2.776517067104578e-05, "learning_rate": 9.439524540791964e-06, "loss": 0.0, "num_input_tokens_seen": 7995008, "step": 28555 }, { "epoch": 317.3333333333333, "grad_norm": 1.3027691238676198e-05, "learning_rate": 9.431841756580673e-06, "loss": 0.0, "num_input_tokens_seen": 7996432, "step": 28560 }, { "epoch": 317.3888888888889, "grad_norm": 1.838142634369433e-05, "learning_rate": 9.42416137317503e-06, "loss": 0.0, "num_input_tokens_seen": 7997840, "step": 28565 }, { "epoch": 317.44444444444446, "grad_norm": 6.9533502937702e-06, "learning_rate": 9.416483391759437e-06, "loss": 0.0, "num_input_tokens_seen": 7999200, "step": 28570 }, { "epoch": 317.5, "grad_norm": 1.168565404441324e-06, "learning_rate": 9.408807813517945e-06, "loss": 0.0, "num_input_tokens_seen": 8000560, "step": 28575 }, { "epoch": 317.55555555555554, "grad_norm": 0.00016053002036642283, "learning_rate": 9.401134639634221e-06, "loss": 0.0, "num_input_tokens_seen": 8002032, "step": 28580 }, { "epoch": 317.6111111111111, "grad_norm": 8.2131264207419e-06, "learning_rate": 9.393463871291555e-06, "loss": 0.0, "num_input_tokens_seen": 8003424, "step": 28585 }, { "epoch": 317.6666666666667, "grad_norm": 1.650304147915449e-05, "learning_rate": 9.385795509672881e-06, "loss": 0.0, "num_input_tokens_seen": 8004848, "step": 28590 }, { "epoch": 317.72222222222223, "grad_norm": 5.54550024389755e-06, "learning_rate": 9.378129555960771e-06, "loss": 0.0, "num_input_tokens_seen": 8006320, "step": 28595 }, { "epoch": 317.77777777777777, "grad_norm": 1.969448203453794e-05, "learning_rate": 9.370466011337392e-06, "loss": 0.0, "num_input_tokens_seen": 8007744, "step": 28600 }, { "epoch": 317.77777777777777, "eval_loss": 0.9427477121353149, "eval_runtime": 1.3981, "eval_samples_per_second": 28.609, "eval_steps_per_second": 14.305, "num_input_tokens_seen": 8007744, "step": 28600 }, { "epoch": 317.8333333333333, "grad_norm": 9.494424375589006e-06, "learning_rate": 9.362804876984573e-06, "loss": 0.0, "num_input_tokens_seen": 8009168, "step": 28605 }, { "epoch": 317.8888888888889, "grad_norm": 0.00011580307909753174, "learning_rate": 9.355146154083747e-06, "loss": 0.0, "num_input_tokens_seen": 8010544, "step": 28610 }, { "epoch": 317.94444444444446, "grad_norm": 6.078303727008461e-07, "learning_rate": 9.347489843815987e-06, "loss": 0.0, "num_input_tokens_seen": 8011936, "step": 28615 }, { "epoch": 318.0, "grad_norm": 7.521819043176947e-06, "learning_rate": 9.339835947362002e-06, "loss": 0.0, "num_input_tokens_seen": 8013328, "step": 28620 }, { "epoch": 318.05555555555554, "grad_norm": 9.15581222216133e-06, "learning_rate": 9.332184465902105e-06, "loss": 0.0, "num_input_tokens_seen": 8014736, "step": 28625 }, { "epoch": 318.1111111111111, "grad_norm": 1.599268216523342e-05, "learning_rate": 9.324535400616266e-06, "loss": 0.0, "num_input_tokens_seen": 8016160, "step": 28630 }, { "epoch": 318.1666666666667, "grad_norm": 2.8378539354889654e-05, "learning_rate": 9.31688875268405e-06, "loss": 0.0, "num_input_tokens_seen": 8017536, "step": 28635 }, { "epoch": 318.22222222222223, "grad_norm": 3.995935185230337e-05, "learning_rate": 9.309244523284674e-06, "loss": 0.0, "num_input_tokens_seen": 8018928, "step": 28640 }, { "epoch": 318.27777777777777, "grad_norm": 0.00014460767852142453, "learning_rate": 9.301602713596982e-06, "loss": 0.0, "num_input_tokens_seen": 8020368, "step": 28645 }, { "epoch": 318.3333333333333, "grad_norm": 6.4580813159409445e-06, "learning_rate": 9.293963324799432e-06, "loss": 0.0, "num_input_tokens_seen": 8021712, "step": 28650 }, { "epoch": 318.3888888888889, "grad_norm": 7.028480013104854e-06, "learning_rate": 9.286326358070104e-06, "loss": 0.0, "num_input_tokens_seen": 8023104, "step": 28655 }, { "epoch": 318.44444444444446, "grad_norm": 2.3601334760314785e-05, "learning_rate": 9.278691814586729e-06, "loss": 0.0, "num_input_tokens_seen": 8024512, "step": 28660 }, { "epoch": 318.5, "grad_norm": 6.957317236810923e-05, "learning_rate": 9.271059695526635e-06, "loss": 0.0, "num_input_tokens_seen": 8025952, "step": 28665 }, { "epoch": 318.55555555555554, "grad_norm": 7.914419256849214e-06, "learning_rate": 9.263430002066805e-06, "loss": 0.0, "num_input_tokens_seen": 8027312, "step": 28670 }, { "epoch": 318.6111111111111, "grad_norm": 1.2801528100681026e-05, "learning_rate": 9.25580273538382e-06, "loss": 0.0, "num_input_tokens_seen": 8028720, "step": 28675 }, { "epoch": 318.6666666666667, "grad_norm": 0.00010466124513186514, "learning_rate": 9.248177896653907e-06, "loss": 0.0, "num_input_tokens_seen": 8030160, "step": 28680 }, { "epoch": 318.72222222222223, "grad_norm": 4.864879883825779e-05, "learning_rate": 9.240555487052918e-06, "loss": 0.0, "num_input_tokens_seen": 8031552, "step": 28685 }, { "epoch": 318.77777777777777, "grad_norm": 2.279569116581115e-06, "learning_rate": 9.232935507756313e-06, "loss": 0.0, "num_input_tokens_seen": 8032944, "step": 28690 }, { "epoch": 318.8333333333333, "grad_norm": 2.2691999674862018e-06, "learning_rate": 9.225317959939193e-06, "loss": 0.0, "num_input_tokens_seen": 8034352, "step": 28695 }, { "epoch": 318.8888888888889, "grad_norm": 2.5399782316526398e-05, "learning_rate": 9.217702844776287e-06, "loss": 0.0, "num_input_tokens_seen": 8035808, "step": 28700 }, { "epoch": 318.94444444444446, "grad_norm": 7.747824383841362e-06, "learning_rate": 9.210090163441929e-06, "loss": 0.0, "num_input_tokens_seen": 8037152, "step": 28705 }, { "epoch": 319.0, "grad_norm": 6.6403801611159e-05, "learning_rate": 9.202479917110105e-06, "loss": 0.0, "num_input_tokens_seen": 8038512, "step": 28710 }, { "epoch": 319.05555555555554, "grad_norm": 1.2354207683529239e-05, "learning_rate": 9.194872106954392e-06, "loss": 0.0, "num_input_tokens_seen": 8039888, "step": 28715 }, { "epoch": 319.1111111111111, "grad_norm": 2.2313190584100084e-06, "learning_rate": 9.187266734148029e-06, "loss": 0.0, "num_input_tokens_seen": 8041280, "step": 28720 }, { "epoch": 319.1666666666667, "grad_norm": 4.724858081317507e-05, "learning_rate": 9.179663799863849e-06, "loss": 0.0, "num_input_tokens_seen": 8042656, "step": 28725 }, { "epoch": 319.22222222222223, "grad_norm": 6.394936463038903e-06, "learning_rate": 9.172063305274317e-06, "loss": 0.0, "num_input_tokens_seen": 8044048, "step": 28730 }, { "epoch": 319.27777777777777, "grad_norm": 4.456142778508365e-05, "learning_rate": 9.164465251551527e-06, "loss": 0.0, "num_input_tokens_seen": 8045440, "step": 28735 }, { "epoch": 319.3333333333333, "grad_norm": 2.3889788280939683e-06, "learning_rate": 9.156869639867205e-06, "loss": 0.0, "num_input_tokens_seen": 8046848, "step": 28740 }, { "epoch": 319.3888888888889, "grad_norm": 1.824954233597964e-05, "learning_rate": 9.149276471392677e-06, "loss": 0.0, "num_input_tokens_seen": 8048256, "step": 28745 }, { "epoch": 319.44444444444446, "grad_norm": 6.210060746525414e-06, "learning_rate": 9.141685747298914e-06, "loss": 0.0, "num_input_tokens_seen": 8049664, "step": 28750 }, { "epoch": 319.5, "grad_norm": 1.5631196220056154e-05, "learning_rate": 9.13409746875649e-06, "loss": 0.0, "num_input_tokens_seen": 8051056, "step": 28755 }, { "epoch": 319.55555555555554, "grad_norm": 1.3842451153323054e-05, "learning_rate": 9.12651163693562e-06, "loss": 0.0, "num_input_tokens_seen": 8052480, "step": 28760 }, { "epoch": 319.6111111111111, "grad_norm": 5.7700584875419736e-05, "learning_rate": 9.11892825300614e-06, "loss": 0.0, "num_input_tokens_seen": 8053856, "step": 28765 }, { "epoch": 319.6666666666667, "grad_norm": 5.0341986934654415e-06, "learning_rate": 9.111347318137491e-06, "loss": 0.0, "num_input_tokens_seen": 8055232, "step": 28770 }, { "epoch": 319.72222222222223, "grad_norm": 7.339013791352045e-06, "learning_rate": 9.103768833498755e-06, "loss": 0.0, "num_input_tokens_seen": 8056624, "step": 28775 }, { "epoch": 319.77777777777777, "grad_norm": 0.00010867640230571851, "learning_rate": 9.096192800258639e-06, "loss": 0.0, "num_input_tokens_seen": 8057984, "step": 28780 }, { "epoch": 319.8333333333333, "grad_norm": 5.3418671086546965e-06, "learning_rate": 9.088619219585443e-06, "loss": 0.0, "num_input_tokens_seen": 8059392, "step": 28785 }, { "epoch": 319.8888888888889, "grad_norm": 1.1578789781196974e-05, "learning_rate": 9.081048092647127e-06, "loss": 0.0, "num_input_tokens_seen": 8060800, "step": 28790 }, { "epoch": 319.94444444444446, "grad_norm": 3.8262105590547435e-06, "learning_rate": 9.073479420611245e-06, "loss": 0.0, "num_input_tokens_seen": 8062224, "step": 28795 }, { "epoch": 320.0, "grad_norm": 4.673507282859646e-05, "learning_rate": 9.065913204644974e-06, "loss": 0.0, "num_input_tokens_seen": 8063616, "step": 28800 }, { "epoch": 320.0, "eval_loss": 0.9171139001846313, "eval_runtime": 1.3999, "eval_samples_per_second": 28.573, "eval_steps_per_second": 14.286, "num_input_tokens_seen": 8063616, "step": 28800 }, { "epoch": 320.05555555555554, "grad_norm": 2.926371234934777e-05, "learning_rate": 9.058349445915135e-06, "loss": 0.0, "num_input_tokens_seen": 8065024, "step": 28805 }, { "epoch": 320.1111111111111, "grad_norm": 5.900503765587928e-06, "learning_rate": 9.050788145588138e-06, "loss": 0.0, "num_input_tokens_seen": 8066416, "step": 28810 }, { "epoch": 320.1666666666667, "grad_norm": 1.8693498304855893e-06, "learning_rate": 9.043229304830039e-06, "loss": 0.0, "num_input_tokens_seen": 8067856, "step": 28815 }, { "epoch": 320.22222222222223, "grad_norm": 2.7449121262179688e-05, "learning_rate": 9.035672924806515e-06, "loss": 0.0, "num_input_tokens_seen": 8069232, "step": 28820 }, { "epoch": 320.27777777777777, "grad_norm": 4.749943400383927e-06, "learning_rate": 9.028119006682839e-06, "loss": 0.0, "num_input_tokens_seen": 8070624, "step": 28825 }, { "epoch": 320.3333333333333, "grad_norm": 4.7577060513503966e-07, "learning_rate": 9.020567551623935e-06, "loss": 0.0, "num_input_tokens_seen": 8071984, "step": 28830 }, { "epoch": 320.3888888888889, "grad_norm": 5.426067673397483e-06, "learning_rate": 9.013018560794318e-06, "loss": 0.0, "num_input_tokens_seen": 8073360, "step": 28835 }, { "epoch": 320.44444444444446, "grad_norm": 6.0091620980529115e-05, "learning_rate": 9.005472035358139e-06, "loss": 0.0, "num_input_tokens_seen": 8074784, "step": 28840 }, { "epoch": 320.5, "grad_norm": 2.905918881879188e-05, "learning_rate": 8.997927976479185e-06, "loss": 0.0, "num_input_tokens_seen": 8076160, "step": 28845 }, { "epoch": 320.55555555555554, "grad_norm": 0.0005948003963567317, "learning_rate": 8.99038638532082e-06, "loss": 0.0, "num_input_tokens_seen": 8077600, "step": 28850 }, { "epoch": 320.6111111111111, "grad_norm": 2.309925775989541e-06, "learning_rate": 8.982847263046065e-06, "loss": 0.0, "num_input_tokens_seen": 8079040, "step": 28855 }, { "epoch": 320.6666666666667, "grad_norm": 2.3402144506690092e-05, "learning_rate": 8.975310610817555e-06, "loss": 0.0, "num_input_tokens_seen": 8080400, "step": 28860 }, { "epoch": 320.72222222222223, "grad_norm": 9.061067430593539e-06, "learning_rate": 8.967776429797528e-06, "loss": 0.0, "num_input_tokens_seen": 8081824, "step": 28865 }, { "epoch": 320.77777777777777, "grad_norm": 7.233387441374362e-05, "learning_rate": 8.960244721147842e-06, "loss": 0.0, "num_input_tokens_seen": 8083216, "step": 28870 }, { "epoch": 320.8333333333333, "grad_norm": 1.7929820614881464e-06, "learning_rate": 8.952715486029995e-06, "loss": 0.0, "num_input_tokens_seen": 8084640, "step": 28875 }, { "epoch": 320.8888888888889, "grad_norm": 3.2113734050653875e-05, "learning_rate": 8.945188725605075e-06, "loss": 0.0, "num_input_tokens_seen": 8086016, "step": 28880 }, { "epoch": 320.94444444444446, "grad_norm": 3.856550756609067e-05, "learning_rate": 8.937664441033817e-06, "loss": 0.0, "num_input_tokens_seen": 8087392, "step": 28885 }, { "epoch": 321.0, "grad_norm": 4.375544904178241e-06, "learning_rate": 8.930142633476549e-06, "loss": 0.0, "num_input_tokens_seen": 8088784, "step": 28890 }, { "epoch": 321.05555555555554, "grad_norm": 0.00024607640807516873, "learning_rate": 8.92262330409323e-06, "loss": 0.0, "num_input_tokens_seen": 8090144, "step": 28895 }, { "epoch": 321.1111111111111, "grad_norm": 2.9169903427828103e-05, "learning_rate": 8.915106454043448e-06, "loss": 0.0, "num_input_tokens_seen": 8091536, "step": 28900 }, { "epoch": 321.1666666666667, "grad_norm": 4.771809472003952e-06, "learning_rate": 8.90759208448638e-06, "loss": 0.0, "num_input_tokens_seen": 8092928, "step": 28905 }, { "epoch": 321.22222222222223, "grad_norm": 4.436327344592428e-06, "learning_rate": 8.900080196580848e-06, "loss": 0.0, "num_input_tokens_seen": 8094336, "step": 28910 }, { "epoch": 321.27777777777777, "grad_norm": 2.021249929384794e-05, "learning_rate": 8.892570791485267e-06, "loss": 0.0, "num_input_tokens_seen": 8095792, "step": 28915 }, { "epoch": 321.3333333333333, "grad_norm": 0.000202822862775065, "learning_rate": 8.885063870357688e-06, "loss": 0.0, "num_input_tokens_seen": 8097216, "step": 28920 }, { "epoch": 321.3888888888889, "grad_norm": 9.622067409509327e-06, "learning_rate": 8.87755943435578e-06, "loss": 0.0, "num_input_tokens_seen": 8098608, "step": 28925 }, { "epoch": 321.44444444444446, "grad_norm": 2.0138922991463915e-06, "learning_rate": 8.87005748463681e-06, "loss": 0.0, "num_input_tokens_seen": 8100000, "step": 28930 }, { "epoch": 321.5, "grad_norm": 1.8527468910178868e-06, "learning_rate": 8.862558022357681e-06, "loss": 0.0, "num_input_tokens_seen": 8101360, "step": 28935 }, { "epoch": 321.55555555555554, "grad_norm": 0.00014125094457995147, "learning_rate": 8.855061048674903e-06, "loss": 0.0, "num_input_tokens_seen": 8102736, "step": 28940 }, { "epoch": 321.6111111111111, "grad_norm": 5.504243290488375e-06, "learning_rate": 8.847566564744595e-06, "loss": 0.0, "num_input_tokens_seen": 8104176, "step": 28945 }, { "epoch": 321.6666666666667, "grad_norm": 3.078047484450508e-06, "learning_rate": 8.840074571722512e-06, "loss": 0.0, "num_input_tokens_seen": 8105616, "step": 28950 }, { "epoch": 321.72222222222223, "grad_norm": 1.2839218470617197e-05, "learning_rate": 8.832585070764002e-06, "loss": 0.0, "num_input_tokens_seen": 8106992, "step": 28955 }, { "epoch": 321.77777777777777, "grad_norm": 1.5018249541753903e-05, "learning_rate": 8.825098063024045e-06, "loss": 0.0, "num_input_tokens_seen": 8108416, "step": 28960 }, { "epoch": 321.8333333333333, "grad_norm": 2.7060012143920176e-05, "learning_rate": 8.817613549657244e-06, "loss": 0.0, "num_input_tokens_seen": 8109856, "step": 28965 }, { "epoch": 321.8888888888889, "grad_norm": 1.4829885003564414e-05, "learning_rate": 8.810131531817783e-06, "loss": 0.0, "num_input_tokens_seen": 8111248, "step": 28970 }, { "epoch": 321.94444444444446, "grad_norm": 4.380061454867246e-06, "learning_rate": 8.802652010659496e-06, "loss": 0.0, "num_input_tokens_seen": 8112640, "step": 28975 }, { "epoch": 322.0, "grad_norm": 0.0002157487760996446, "learning_rate": 8.795174987335827e-06, "loss": 0.0, "num_input_tokens_seen": 8114000, "step": 28980 }, { "epoch": 322.05555555555554, "grad_norm": 1.397614960296778e-05, "learning_rate": 8.787700462999807e-06, "loss": 0.0, "num_input_tokens_seen": 8115424, "step": 28985 }, { "epoch": 322.1111111111111, "grad_norm": 3.3570199775567744e-06, "learning_rate": 8.780228438804122e-06, "loss": 0.0, "num_input_tokens_seen": 8116800, "step": 28990 }, { "epoch": 322.1666666666667, "grad_norm": 1.6871100569915143e-06, "learning_rate": 8.772758915901032e-06, "loss": 0.0, "num_input_tokens_seen": 8118192, "step": 28995 }, { "epoch": 322.22222222222223, "grad_norm": 5.54815778741613e-05, "learning_rate": 8.765291895442443e-06, "loss": 0.0, "num_input_tokens_seen": 8119520, "step": 29000 }, { "epoch": 322.22222222222223, "eval_loss": 0.9533809423446655, "eval_runtime": 1.3867, "eval_samples_per_second": 28.845, "eval_steps_per_second": 14.423, "num_input_tokens_seen": 8119520, "step": 29000 }, { "epoch": 322.27777777777777, "grad_norm": 3.519444362609647e-05, "learning_rate": 8.75782737857987e-06, "loss": 0.0, "num_input_tokens_seen": 8120944, "step": 29005 }, { "epoch": 322.3333333333333, "grad_norm": 8.386417903238907e-06, "learning_rate": 8.750365366464425e-06, "loss": 0.0, "num_input_tokens_seen": 8122384, "step": 29010 }, { "epoch": 322.3888888888889, "grad_norm": 3.692033715196885e-05, "learning_rate": 8.742905860246838e-06, "loss": 0.0, "num_input_tokens_seen": 8123808, "step": 29015 }, { "epoch": 322.44444444444446, "grad_norm": 0.00020582956494763494, "learning_rate": 8.735448861077478e-06, "loss": 0.0, "num_input_tokens_seen": 8125200, "step": 29020 }, { "epoch": 322.5, "grad_norm": 0.00015743070980533957, "learning_rate": 8.727994370106288e-06, "loss": 0.0, "num_input_tokens_seen": 8126560, "step": 29025 }, { "epoch": 322.55555555555554, "grad_norm": 1.5139371498662513e-05, "learning_rate": 8.720542388482861e-06, "loss": 0.0, "num_input_tokens_seen": 8127968, "step": 29030 }, { "epoch": 322.6111111111111, "grad_norm": 2.7493053494254127e-05, "learning_rate": 8.71309291735637e-06, "loss": 0.0, "num_input_tokens_seen": 8129328, "step": 29035 }, { "epoch": 322.6666666666667, "grad_norm": 4.360451384854969e-06, "learning_rate": 8.705645957875621e-06, "loss": 0.0, "num_input_tokens_seen": 8130688, "step": 29040 }, { "epoch": 322.72222222222223, "grad_norm": 2.701925768633373e-05, "learning_rate": 8.698201511189048e-06, "loss": 0.0, "num_input_tokens_seen": 8132096, "step": 29045 }, { "epoch": 322.77777777777777, "grad_norm": 1.6988482457236387e-05, "learning_rate": 8.690759578444649e-06, "loss": 0.0, "num_input_tokens_seen": 8133536, "step": 29050 }, { "epoch": 322.8333333333333, "grad_norm": 1.1332838766975328e-05, "learning_rate": 8.68332016079008e-06, "loss": 0.0, "num_input_tokens_seen": 8134960, "step": 29055 }, { "epoch": 322.8888888888889, "grad_norm": 6.831565769971348e-06, "learning_rate": 8.6758832593726e-06, "loss": 0.0, "num_input_tokens_seen": 8136384, "step": 29060 }, { "epoch": 322.94444444444446, "grad_norm": 3.089976189585286e-06, "learning_rate": 8.668448875339053e-06, "loss": 0.0, "num_input_tokens_seen": 8137776, "step": 29065 }, { "epoch": 323.0, "grad_norm": 7.435579027514905e-05, "learning_rate": 8.661017009835933e-06, "loss": 0.0, "num_input_tokens_seen": 8139152, "step": 29070 }, { "epoch": 323.05555555555554, "grad_norm": 4.578753305395367e-06, "learning_rate": 8.653587664009311e-06, "loss": 0.0, "num_input_tokens_seen": 8140544, "step": 29075 }, { "epoch": 323.1111111111111, "grad_norm": 5.951300045126118e-05, "learning_rate": 8.646160839004902e-06, "loss": 0.0, "num_input_tokens_seen": 8141952, "step": 29080 }, { "epoch": 323.1666666666667, "grad_norm": 1.110071661969414e-05, "learning_rate": 8.638736535967998e-06, "loss": 0.0, "num_input_tokens_seen": 8143376, "step": 29085 }, { "epoch": 323.22222222222223, "grad_norm": 7.764404290355742e-05, "learning_rate": 8.631314756043535e-06, "loss": 0.0, "num_input_tokens_seen": 8144736, "step": 29090 }, { "epoch": 323.27777777777777, "grad_norm": 1.9973458620370366e-05, "learning_rate": 8.62389550037603e-06, "loss": 0.0, "num_input_tokens_seen": 8146144, "step": 29095 }, { "epoch": 323.3333333333333, "grad_norm": 9.14341057978163e-07, "learning_rate": 8.616478770109646e-06, "loss": 0.0, "num_input_tokens_seen": 8147536, "step": 29100 }, { "epoch": 323.3888888888889, "grad_norm": 1.9509674530127086e-05, "learning_rate": 8.609064566388111e-06, "loss": 0.0, "num_input_tokens_seen": 8148928, "step": 29105 }, { "epoch": 323.44444444444446, "grad_norm": 9.442175041840528e-07, "learning_rate": 8.601652890354815e-06, "loss": 0.0, "num_input_tokens_seen": 8150304, "step": 29110 }, { "epoch": 323.5, "grad_norm": 8.550737220502924e-06, "learning_rate": 8.594243743152705e-06, "loss": 0.0, "num_input_tokens_seen": 8151712, "step": 29115 }, { "epoch": 323.55555555555554, "grad_norm": 1.8794820789480582e-05, "learning_rate": 8.58683712592438e-06, "loss": 0.0, "num_input_tokens_seen": 8153104, "step": 29120 }, { "epoch": 323.6111111111111, "grad_norm": 2.6276500193489483e-06, "learning_rate": 8.579433039812037e-06, "loss": 0.0, "num_input_tokens_seen": 8154496, "step": 29125 }, { "epoch": 323.6666666666667, "grad_norm": 4.54122982773697e-06, "learning_rate": 8.572031485957466e-06, "loss": 0.0, "num_input_tokens_seen": 8155936, "step": 29130 }, { "epoch": 323.72222222222223, "grad_norm": 7.94410370872356e-05, "learning_rate": 8.564632465502084e-06, "loss": 0.0, "num_input_tokens_seen": 8157312, "step": 29135 }, { "epoch": 323.77777777777777, "grad_norm": 2.5718525648699142e-05, "learning_rate": 8.557235979586928e-06, "loss": 0.0, "num_input_tokens_seen": 8158720, "step": 29140 }, { "epoch": 323.8333333333333, "grad_norm": 8.357099432032555e-06, "learning_rate": 8.549842029352606e-06, "loss": 0.0, "num_input_tokens_seen": 8160144, "step": 29145 }, { "epoch": 323.8888888888889, "grad_norm": 3.6293781704443973e-06, "learning_rate": 8.542450615939376e-06, "loss": 0.0, "num_input_tokens_seen": 8161536, "step": 29150 }, { "epoch": 323.94444444444446, "grad_norm": 5.157055966265034e-06, "learning_rate": 8.535061740487082e-06, "loss": 0.0, "num_input_tokens_seen": 8162976, "step": 29155 }, { "epoch": 324.0, "grad_norm": 2.2703945433022454e-05, "learning_rate": 8.527675404135168e-06, "loss": 0.0, "num_input_tokens_seen": 8164384, "step": 29160 }, { "epoch": 324.05555555555554, "grad_norm": 3.699959779623896e-05, "learning_rate": 8.520291608022724e-06, "loss": 0.0, "num_input_tokens_seen": 8165760, "step": 29165 }, { "epoch": 324.1111111111111, "grad_norm": 9.567056986270472e-05, "learning_rate": 8.512910353288398e-06, "loss": 0.0, "num_input_tokens_seen": 8167168, "step": 29170 }, { "epoch": 324.1666666666667, "grad_norm": 2.2092779545346275e-05, "learning_rate": 8.505531641070486e-06, "loss": 0.0, "num_input_tokens_seen": 8168544, "step": 29175 }, { "epoch": 324.22222222222223, "grad_norm": 4.366888970253058e-05, "learning_rate": 8.498155472506885e-06, "loss": 0.0, "num_input_tokens_seen": 8169952, "step": 29180 }, { "epoch": 324.27777777777777, "grad_norm": 8.52588618727168e-06, "learning_rate": 8.49078184873508e-06, "loss": 0.0, "num_input_tokens_seen": 8171344, "step": 29185 }, { "epoch": 324.3333333333333, "grad_norm": 1.7186635886901058e-05, "learning_rate": 8.483410770892188e-06, "loss": 0.0, "num_input_tokens_seen": 8172832, "step": 29190 }, { "epoch": 324.3888888888889, "grad_norm": 5.62646564503666e-05, "learning_rate": 8.476042240114909e-06, "loss": 0.0, "num_input_tokens_seen": 8174208, "step": 29195 }, { "epoch": 324.44444444444446, "grad_norm": 8.96942219696939e-06, "learning_rate": 8.468676257539568e-06, "loss": 0.0, "num_input_tokens_seen": 8175584, "step": 29200 }, { "epoch": 324.44444444444446, "eval_loss": 0.9510575532913208, "eval_runtime": 1.4129, "eval_samples_per_second": 28.311, "eval_steps_per_second": 14.156, "num_input_tokens_seen": 8175584, "step": 29200 }, { "epoch": 324.5, "grad_norm": 4.7553317017445806e-07, "learning_rate": 8.4613128243021e-06, "loss": 0.0, "num_input_tokens_seen": 8176976, "step": 29205 }, { "epoch": 324.55555555555554, "grad_norm": 1.3292959010868799e-05, "learning_rate": 8.453951941538028e-06, "loss": 0.0, "num_input_tokens_seen": 8178368, "step": 29210 }, { "epoch": 324.6111111111111, "grad_norm": 3.829925390164135e-06, "learning_rate": 8.446593610382495e-06, "loss": 0.0, "num_input_tokens_seen": 8179760, "step": 29215 }, { "epoch": 324.6666666666667, "grad_norm": 9.494689948041923e-06, "learning_rate": 8.439237831970259e-06, "loss": 0.0, "num_input_tokens_seen": 8181152, "step": 29220 }, { "epoch": 324.72222222222223, "grad_norm": 6.161329110909719e-06, "learning_rate": 8.431884607435667e-06, "loss": 0.0, "num_input_tokens_seen": 8182576, "step": 29225 }, { "epoch": 324.77777777777777, "grad_norm": 1.2797282806786825e-06, "learning_rate": 8.424533937912665e-06, "loss": 0.0, "num_input_tokens_seen": 8183968, "step": 29230 }, { "epoch": 324.8333333333333, "grad_norm": 1.0326099982194137e-05, "learning_rate": 8.41718582453484e-06, "loss": 0.0, "num_input_tokens_seen": 8185344, "step": 29235 }, { "epoch": 324.8888888888889, "grad_norm": 0.00018484819156583399, "learning_rate": 8.409840268435346e-06, "loss": 0.0, "num_input_tokens_seen": 8186752, "step": 29240 }, { "epoch": 324.94444444444446, "grad_norm": 5.077549576526508e-05, "learning_rate": 8.402497270746976e-06, "loss": 0.0, "num_input_tokens_seen": 8188208, "step": 29245 }, { "epoch": 325.0, "grad_norm": 1.429635267413687e-05, "learning_rate": 8.395156832602095e-06, "loss": 0.0, "num_input_tokens_seen": 8189616, "step": 29250 }, { "epoch": 325.05555555555554, "grad_norm": 2.38962093135342e-06, "learning_rate": 8.387818955132707e-06, "loss": 0.0, "num_input_tokens_seen": 8191056, "step": 29255 }, { "epoch": 325.1111111111111, "grad_norm": 7.030302640487207e-06, "learning_rate": 8.38048363947039e-06, "loss": 0.0, "num_input_tokens_seen": 8192416, "step": 29260 }, { "epoch": 325.1666666666667, "grad_norm": 1.5309009313568822e-06, "learning_rate": 8.373150886746351e-06, "loss": 0.0, "num_input_tokens_seen": 8193824, "step": 29265 }, { "epoch": 325.22222222222223, "grad_norm": 1.7812419173424132e-05, "learning_rate": 8.365820698091397e-06, "loss": 0.0, "num_input_tokens_seen": 8195216, "step": 29270 }, { "epoch": 325.27777777777777, "grad_norm": 2.8866650154668605e-06, "learning_rate": 8.358493074635922e-06, "loss": 0.0, "num_input_tokens_seen": 8196624, "step": 29275 }, { "epoch": 325.3333333333333, "grad_norm": 3.009830606970354e-06, "learning_rate": 8.351168017509948e-06, "loss": 0.0, "num_input_tokens_seen": 8198016, "step": 29280 }, { "epoch": 325.3888888888889, "grad_norm": 2.7563915864448063e-05, "learning_rate": 8.343845527843094e-06, "loss": 0.0, "num_input_tokens_seen": 8199456, "step": 29285 }, { "epoch": 325.44444444444446, "grad_norm": 1.5483279639738612e-05, "learning_rate": 8.336525606764566e-06, "loss": 0.0, "num_input_tokens_seen": 8200864, "step": 29290 }, { "epoch": 325.5, "grad_norm": 7.614240075781709e-06, "learning_rate": 8.329208255403204e-06, "loss": 0.0, "num_input_tokens_seen": 8202304, "step": 29295 }, { "epoch": 325.55555555555554, "grad_norm": 2.3694556148257107e-05, "learning_rate": 8.321893474887426e-06, "loss": 0.0, "num_input_tokens_seen": 8203712, "step": 29300 }, { "epoch": 325.6111111111111, "grad_norm": 1.5960340533638373e-05, "learning_rate": 8.31458126634526e-06, "loss": 0.0, "num_input_tokens_seen": 8205104, "step": 29305 }, { "epoch": 325.6666666666667, "grad_norm": 3.553406259015901e-06, "learning_rate": 8.30727163090435e-06, "loss": 0.0, "num_input_tokens_seen": 8206496, "step": 29310 }, { "epoch": 325.72222222222223, "grad_norm": 1.959491100933519e-06, "learning_rate": 8.29996456969192e-06, "loss": 0.0, "num_input_tokens_seen": 8207888, "step": 29315 }, { "epoch": 325.77777777777777, "grad_norm": 6.0552461036422756e-06, "learning_rate": 8.292660083834818e-06, "loss": 0.0, "num_input_tokens_seen": 8209296, "step": 29320 }, { "epoch": 325.8333333333333, "grad_norm": 3.0812010663794354e-05, "learning_rate": 8.2853581744595e-06, "loss": 0.0, "num_input_tokens_seen": 8210704, "step": 29325 }, { "epoch": 325.8888888888889, "grad_norm": 6.118369810792501e-07, "learning_rate": 8.278058842691991e-06, "loss": 0.0, "num_input_tokens_seen": 8212112, "step": 29330 }, { "epoch": 325.94444444444446, "grad_norm": 1.8386924693913897e-06, "learning_rate": 8.27076208965796e-06, "loss": 0.0, "num_input_tokens_seen": 8213504, "step": 29335 }, { "epoch": 326.0, "grad_norm": 1.1182479283888824e-05, "learning_rate": 8.263467916482637e-06, "loss": 0.0, "num_input_tokens_seen": 8214896, "step": 29340 }, { "epoch": 326.05555555555554, "grad_norm": 6.152303626549838e-07, "learning_rate": 8.256176324290885e-06, "loss": 0.0, "num_input_tokens_seen": 8216272, "step": 29345 }, { "epoch": 326.1111111111111, "grad_norm": 6.6452580540499184e-06, "learning_rate": 8.248887314207168e-06, "loss": 0.0, "num_input_tokens_seen": 8217648, "step": 29350 }, { "epoch": 326.1666666666667, "grad_norm": 6.191321153892204e-05, "learning_rate": 8.24160088735553e-06, "loss": 0.0, "num_input_tokens_seen": 8219088, "step": 29355 }, { "epoch": 326.22222222222223, "grad_norm": 7.888623258622829e-06, "learning_rate": 8.234317044859629e-06, "loss": 0.0, "num_input_tokens_seen": 8220480, "step": 29360 }, { "epoch": 326.27777777777777, "grad_norm": 6.820279668318108e-05, "learning_rate": 8.227035787842744e-06, "loss": 0.0, "num_input_tokens_seen": 8221888, "step": 29365 }, { "epoch": 326.3333333333333, "grad_norm": 5.5527121730847284e-05, "learning_rate": 8.219757117427721e-06, "loss": 0.0, "num_input_tokens_seen": 8223328, "step": 29370 }, { "epoch": 326.3888888888889, "grad_norm": 2.0289415260776877e-05, "learning_rate": 8.212481034737014e-06, "loss": 0.0, "num_input_tokens_seen": 8224688, "step": 29375 }, { "epoch": 326.44444444444446, "grad_norm": 5.133822469360894e-06, "learning_rate": 8.205207540892707e-06, "loss": 0.0, "num_input_tokens_seen": 8226128, "step": 29380 }, { "epoch": 326.5, "grad_norm": 7.659858169972722e-07, "learning_rate": 8.197936637016442e-06, "loss": 0.0, "num_input_tokens_seen": 8227568, "step": 29385 }, { "epoch": 326.55555555555554, "grad_norm": 1.3689625120605342e-05, "learning_rate": 8.190668324229508e-06, "loss": 0.0, "num_input_tokens_seen": 8228944, "step": 29390 }, { "epoch": 326.6111111111111, "grad_norm": 7.765844429741264e-07, "learning_rate": 8.183402603652749e-06, "loss": 0.0, "num_input_tokens_seen": 8230352, "step": 29395 }, { "epoch": 326.6666666666667, "grad_norm": 0.00012044896720908582, "learning_rate": 8.176139476406635e-06, "loss": 0.0, "num_input_tokens_seen": 8231760, "step": 29400 }, { "epoch": 326.6666666666667, "eval_loss": 0.9983776211738586, "eval_runtime": 1.3921, "eval_samples_per_second": 28.735, "eval_steps_per_second": 14.367, "num_input_tokens_seen": 8231760, "step": 29400 }, { "epoch": 326.72222222222223, "grad_norm": 1.941242953762412e-05, "learning_rate": 8.16887894361125e-06, "loss": 0.0, "num_input_tokens_seen": 8233120, "step": 29405 }, { "epoch": 326.77777777777777, "grad_norm": 1.7814514876590692e-06, "learning_rate": 8.161621006386233e-06, "loss": 0.0, "num_input_tokens_seen": 8234480, "step": 29410 }, { "epoch": 326.8333333333333, "grad_norm": 5.736956609325716e-06, "learning_rate": 8.154365665850869e-06, "loss": 0.0, "num_input_tokens_seen": 8235872, "step": 29415 }, { "epoch": 326.8888888888889, "grad_norm": 1.7383008525939658e-05, "learning_rate": 8.147112923124005e-06, "loss": 0.0, "num_input_tokens_seen": 8237280, "step": 29420 }, { "epoch": 326.94444444444446, "grad_norm": 7.127293429221027e-06, "learning_rate": 8.13986277932412e-06, "loss": 0.0, "num_input_tokens_seen": 8238672, "step": 29425 }, { "epoch": 327.0, "grad_norm": 8.318805157614406e-06, "learning_rate": 8.132615235569277e-06, "loss": 0.0, "num_input_tokens_seen": 8240048, "step": 29430 }, { "epoch": 327.05555555555554, "grad_norm": 3.922142241208348e-06, "learning_rate": 8.125370292977124e-06, "loss": 0.0, "num_input_tokens_seen": 8241424, "step": 29435 }, { "epoch": 327.1111111111111, "grad_norm": 1.106054469346418e-06, "learning_rate": 8.118127952664944e-06, "loss": 0.0, "num_input_tokens_seen": 8242864, "step": 29440 }, { "epoch": 327.1666666666667, "grad_norm": 3.132915935566416e-06, "learning_rate": 8.110888215749574e-06, "loss": 0.0, "num_input_tokens_seen": 8244256, "step": 29445 }, { "epoch": 327.22222222222223, "grad_norm": 1.3509126802091487e-05, "learning_rate": 8.10365108334749e-06, "loss": 0.0, "num_input_tokens_seen": 8245696, "step": 29450 }, { "epoch": 327.27777777777777, "grad_norm": 3.327946978970431e-05, "learning_rate": 8.096416556574743e-06, "loss": 0.0, "num_input_tokens_seen": 8247072, "step": 29455 }, { "epoch": 327.3333333333333, "grad_norm": 3.4024540127575165e-06, "learning_rate": 8.08918463654698e-06, "loss": 0.0, "num_input_tokens_seen": 8248480, "step": 29460 }, { "epoch": 327.3888888888889, "grad_norm": 8.208276085497346e-07, "learning_rate": 8.081955324379458e-06, "loss": 0.0, "num_input_tokens_seen": 8249872, "step": 29465 }, { "epoch": 327.44444444444446, "grad_norm": 2.2483822249341756e-05, "learning_rate": 8.074728621187039e-06, "loss": 0.0, "num_input_tokens_seen": 8251248, "step": 29470 }, { "epoch": 327.5, "grad_norm": 0.0001644549920456484, "learning_rate": 8.067504528084158e-06, "loss": 0.0, "num_input_tokens_seen": 8252672, "step": 29475 }, { "epoch": 327.55555555555554, "grad_norm": 7.2988618740055244e-06, "learning_rate": 8.060283046184861e-06, "loss": 0.0, "num_input_tokens_seen": 8254080, "step": 29480 }, { "epoch": 327.6111111111111, "grad_norm": 5.470415999297984e-06, "learning_rate": 8.053064176602806e-06, "loss": 0.0, "num_input_tokens_seen": 8255504, "step": 29485 }, { "epoch": 327.6666666666667, "grad_norm": 2.027623850153759e-06, "learning_rate": 8.045847920451216e-06, "loss": 0.0, "num_input_tokens_seen": 8256912, "step": 29490 }, { "epoch": 327.72222222222223, "grad_norm": 3.176217433065176e-06, "learning_rate": 8.038634278842944e-06, "loss": 0.0, "num_input_tokens_seen": 8258288, "step": 29495 }, { "epoch": 327.77777777777777, "grad_norm": 6.440162542276084e-05, "learning_rate": 8.031423252890408e-06, "loss": 0.0, "num_input_tokens_seen": 8259680, "step": 29500 }, { "epoch": 327.8333333333333, "grad_norm": 3.4333097573835403e-06, "learning_rate": 8.024214843705646e-06, "loss": 0.0, "num_input_tokens_seen": 8261072, "step": 29505 }, { "epoch": 327.8888888888889, "grad_norm": 3.1621113976143533e-06, "learning_rate": 8.017009052400295e-06, "loss": 0.0, "num_input_tokens_seen": 8262448, "step": 29510 }, { "epoch": 327.94444444444446, "grad_norm": 2.377326018176973e-05, "learning_rate": 8.00980588008557e-06, "loss": 0.0, "num_input_tokens_seen": 8263888, "step": 29515 }, { "epoch": 328.0, "grad_norm": 2.6609325232129777e-06, "learning_rate": 8.002605327872282e-06, "loss": 0.0, "num_input_tokens_seen": 8265328, "step": 29520 }, { "epoch": 328.05555555555554, "grad_norm": 0.00015557589358650148, "learning_rate": 7.995407396870862e-06, "loss": 0.0, "num_input_tokens_seen": 8266720, "step": 29525 }, { "epoch": 328.1111111111111, "grad_norm": 7.704320523771457e-06, "learning_rate": 7.988212088191307e-06, "loss": 0.0, "num_input_tokens_seen": 8268128, "step": 29530 }, { "epoch": 328.1666666666667, "grad_norm": 1.151459673565114e-05, "learning_rate": 7.98101940294324e-06, "loss": 0.0, "num_input_tokens_seen": 8269552, "step": 29535 }, { "epoch": 328.22222222222223, "grad_norm": 1.595564572198782e-05, "learning_rate": 7.973829342235847e-06, "loss": 0.0, "num_input_tokens_seen": 8270960, "step": 29540 }, { "epoch": 328.27777777777777, "grad_norm": 2.878593295463361e-05, "learning_rate": 7.966641907177936e-06, "loss": 0.0, "num_input_tokens_seen": 8272384, "step": 29545 }, { "epoch": 328.3333333333333, "grad_norm": 6.14161035628058e-05, "learning_rate": 7.959457098877901e-06, "loss": 0.0, "num_input_tokens_seen": 8273792, "step": 29550 }, { "epoch": 328.3888888888889, "grad_norm": 6.053210654499708e-06, "learning_rate": 7.952274918443719e-06, "loss": 0.0, "num_input_tokens_seen": 8275184, "step": 29555 }, { "epoch": 328.44444444444446, "grad_norm": 2.3202378542919178e-06, "learning_rate": 7.945095366982983e-06, "loss": 0.0, "num_input_tokens_seen": 8276576, "step": 29560 }, { "epoch": 328.5, "grad_norm": 7.020754310360644e-06, "learning_rate": 7.937918445602871e-06, "loss": 0.0, "num_input_tokens_seen": 8278000, "step": 29565 }, { "epoch": 328.55555555555554, "grad_norm": 8.218340212806652e-07, "learning_rate": 7.930744155410145e-06, "loss": 0.0, "num_input_tokens_seen": 8279392, "step": 29570 }, { "epoch": 328.6111111111111, "grad_norm": 6.654062872257782e-06, "learning_rate": 7.923572497511181e-06, "loss": 0.0, "num_input_tokens_seen": 8280768, "step": 29575 }, { "epoch": 328.6666666666667, "grad_norm": 1.8004637240665033e-05, "learning_rate": 7.916403473011927e-06, "loss": 0.0, "num_input_tokens_seen": 8282176, "step": 29580 }, { "epoch": 328.72222222222223, "grad_norm": 1.3108289067531587e-06, "learning_rate": 7.909237083017953e-06, "loss": 0.0, "num_input_tokens_seen": 8283584, "step": 29585 }, { "epoch": 328.77777777777777, "grad_norm": 7.963364623719826e-06, "learning_rate": 7.902073328634389e-06, "loss": 0.0, "num_input_tokens_seen": 8284944, "step": 29590 }, { "epoch": 328.8333333333333, "grad_norm": 7.234868007799378e-06, "learning_rate": 7.894912210965987e-06, "loss": 0.0, "num_input_tokens_seen": 8286352, "step": 29595 }, { "epoch": 328.8888888888889, "grad_norm": 3.150402699247934e-05, "learning_rate": 7.887753731117075e-06, "loss": 0.0, "num_input_tokens_seen": 8287696, "step": 29600 }, { "epoch": 328.8888888888889, "eval_loss": 0.9725330471992493, "eval_runtime": 1.3867, "eval_samples_per_second": 28.845, "eval_steps_per_second": 14.422, "num_input_tokens_seen": 8287696, "step": 29600 }, { "epoch": 328.94444444444446, "grad_norm": 6.672213203273714e-06, "learning_rate": 7.880597890191587e-06, "loss": 0.0, "num_input_tokens_seen": 8289120, "step": 29605 }, { "epoch": 329.0, "grad_norm": 1.6216101357713342e-05, "learning_rate": 7.873444689293036e-06, "loss": 0.0, "num_input_tokens_seen": 8290560, "step": 29610 }, { "epoch": 329.05555555555554, "grad_norm": 1.468366554036038e-05, "learning_rate": 7.866294129524548e-06, "loss": 0.0, "num_input_tokens_seen": 8291968, "step": 29615 }, { "epoch": 329.1111111111111, "grad_norm": 7.411095793941058e-06, "learning_rate": 7.859146211988811e-06, "loss": 0.0, "num_input_tokens_seen": 8293376, "step": 29620 }, { "epoch": 329.1666666666667, "grad_norm": 4.1868934204103425e-05, "learning_rate": 7.852000937788134e-06, "loss": 0.0, "num_input_tokens_seen": 8294720, "step": 29625 }, { "epoch": 329.22222222222223, "grad_norm": 1.918319867399987e-06, "learning_rate": 7.844858308024416e-06, "loss": 0.0, "num_input_tokens_seen": 8296160, "step": 29630 }, { "epoch": 329.27777777777777, "grad_norm": 7.068115792208118e-06, "learning_rate": 7.837718323799122e-06, "loss": 0.0, "num_input_tokens_seen": 8297520, "step": 29635 }, { "epoch": 329.3333333333333, "grad_norm": 2.865041778932209e-06, "learning_rate": 7.83058098621334e-06, "loss": 0.0, "num_input_tokens_seen": 8298896, "step": 29640 }, { "epoch": 329.3888888888889, "grad_norm": 9.51056517806137e-06, "learning_rate": 7.823446296367739e-06, "loss": 0.0, "num_input_tokens_seen": 8300320, "step": 29645 }, { "epoch": 329.44444444444446, "grad_norm": 1.2715839147858787e-05, "learning_rate": 7.81631425536257e-06, "loss": 0.0, "num_input_tokens_seen": 8301680, "step": 29650 }, { "epoch": 329.5, "grad_norm": 6.69248493068153e-06, "learning_rate": 7.809184864297689e-06, "loss": 0.0, "num_input_tokens_seen": 8303104, "step": 29655 }, { "epoch": 329.55555555555554, "grad_norm": 6.599655989703024e-06, "learning_rate": 7.802058124272532e-06, "loss": 0.0, "num_input_tokens_seen": 8304528, "step": 29660 }, { "epoch": 329.6111111111111, "grad_norm": 4.5027009036857635e-06, "learning_rate": 7.79493403638614e-06, "loss": 0.0, "num_input_tokens_seen": 8305888, "step": 29665 }, { "epoch": 329.6666666666667, "grad_norm": 3.749662937480025e-05, "learning_rate": 7.787812601737132e-06, "loss": 0.0, "num_input_tokens_seen": 8307328, "step": 29670 }, { "epoch": 329.72222222222223, "grad_norm": 9.015064279083163e-06, "learning_rate": 7.780693821423715e-06, "loss": 0.0, "num_input_tokens_seen": 8308752, "step": 29675 }, { "epoch": 329.77777777777777, "grad_norm": 3.299506943221786e-06, "learning_rate": 7.773577696543705e-06, "loss": 0.0, "num_input_tokens_seen": 8310160, "step": 29680 }, { "epoch": 329.8333333333333, "grad_norm": 9.376683465234237e-07, "learning_rate": 7.7664642281945e-06, "loss": 0.0, "num_input_tokens_seen": 8311520, "step": 29685 }, { "epoch": 329.8888888888889, "grad_norm": 5.1653564696607646e-06, "learning_rate": 7.759353417473072e-06, "loss": 0.0, "num_input_tokens_seen": 8312944, "step": 29690 }, { "epoch": 329.94444444444446, "grad_norm": 3.1345054594567046e-05, "learning_rate": 7.752245265476016e-06, "loss": 0.0, "num_input_tokens_seen": 8314400, "step": 29695 }, { "epoch": 330.0, "grad_norm": 2.1583107354672393e-06, "learning_rate": 7.745139773299481e-06, "loss": 0.0, "num_input_tokens_seen": 8315840, "step": 29700 }, { "epoch": 330.05555555555554, "grad_norm": 7.100072707544314e-06, "learning_rate": 7.738036942039232e-06, "loss": 0.0, "num_input_tokens_seen": 8317248, "step": 29705 }, { "epoch": 330.1111111111111, "grad_norm": 5.046301885158755e-05, "learning_rate": 7.73093677279062e-06, "loss": 0.0, "num_input_tokens_seen": 8318592, "step": 29710 }, { "epoch": 330.1666666666667, "grad_norm": 1.0749680768640246e-05, "learning_rate": 7.72383926664857e-06, "loss": 0.0, "num_input_tokens_seen": 8320000, "step": 29715 }, { "epoch": 330.22222222222223, "grad_norm": 1.5672081644879654e-05, "learning_rate": 7.716744424707606e-06, "loss": 0.0, "num_input_tokens_seen": 8321392, "step": 29720 }, { "epoch": 330.27777777777777, "grad_norm": 9.42391579883406e-06, "learning_rate": 7.709652248061858e-06, "loss": 0.0, "num_input_tokens_seen": 8322768, "step": 29725 }, { "epoch": 330.3333333333333, "grad_norm": 1.1240472304052673e-05, "learning_rate": 7.702562737805017e-06, "loss": 0.0, "num_input_tokens_seen": 8324144, "step": 29730 }, { "epoch": 330.3888888888889, "grad_norm": 1.790007809177041e-05, "learning_rate": 7.695475895030365e-06, "loss": 0.0, "num_input_tokens_seen": 8325584, "step": 29735 }, { "epoch": 330.44444444444446, "grad_norm": 8.104655717033893e-05, "learning_rate": 7.6883917208308e-06, "loss": 0.0, "num_input_tokens_seen": 8327072, "step": 29740 }, { "epoch": 330.5, "grad_norm": 6.177464001666522e-06, "learning_rate": 7.681310216298778e-06, "loss": 0.0, "num_input_tokens_seen": 8328448, "step": 29745 }, { "epoch": 330.55555555555554, "grad_norm": 1.2200489436509088e-06, "learning_rate": 7.674231382526367e-06, "loss": 0.0, "num_input_tokens_seen": 8329872, "step": 29750 }, { "epoch": 330.6111111111111, "grad_norm": 0.00010359846783103421, "learning_rate": 7.667155220605198e-06, "loss": 0.0, "num_input_tokens_seen": 8331280, "step": 29755 }, { "epoch": 330.6666666666667, "grad_norm": 5.78621074964758e-06, "learning_rate": 7.660081731626515e-06, "loss": 0.0, "num_input_tokens_seen": 8332720, "step": 29760 }, { "epoch": 330.72222222222223, "grad_norm": 4.299667489249259e-05, "learning_rate": 7.653010916681141e-06, "loss": 0.0, "num_input_tokens_seen": 8334112, "step": 29765 }, { "epoch": 330.77777777777777, "grad_norm": 2.9092920158291236e-05, "learning_rate": 7.645942776859472e-06, "loss": 0.0, "num_input_tokens_seen": 8335472, "step": 29770 }, { "epoch": 330.8333333333333, "grad_norm": 2.9727325454587117e-06, "learning_rate": 7.63887731325152e-06, "loss": 0.0, "num_input_tokens_seen": 8336832, "step": 29775 }, { "epoch": 330.8888888888889, "grad_norm": 1.0229356348645524e-06, "learning_rate": 7.63181452694685e-06, "loss": 0.0, "num_input_tokens_seen": 8338192, "step": 29780 }, { "epoch": 330.94444444444446, "grad_norm": 0.00011194471153430641, "learning_rate": 7.624754419034644e-06, "loss": 0.0, "num_input_tokens_seen": 8339584, "step": 29785 }, { "epoch": 331.0, "grad_norm": 1.7089576431317255e-05, "learning_rate": 7.6176969906036645e-06, "loss": 0.0, "num_input_tokens_seen": 8340976, "step": 29790 }, { "epoch": 331.05555555555554, "grad_norm": 2.441319566059974e-06, "learning_rate": 7.610642242742242e-06, "loss": 0.0, "num_input_tokens_seen": 8342352, "step": 29795 }, { "epoch": 331.1111111111111, "grad_norm": 1.887337134576228e-06, "learning_rate": 7.603590176538322e-06, "loss": 0.0, "num_input_tokens_seen": 8343760, "step": 29800 }, { "epoch": 331.1111111111111, "eval_loss": 1.0153379440307617, "eval_runtime": 1.384, "eval_samples_per_second": 28.903, "eval_steps_per_second": 14.451, "num_input_tokens_seen": 8343760, "step": 29800 }, { "epoch": 331.1666666666667, "grad_norm": 1.3001043043914251e-05, "learning_rate": 7.596540793079404e-06, "loss": 0.0, "num_input_tokens_seen": 8345168, "step": 29805 }, { "epoch": 331.22222222222223, "grad_norm": 3.796124019572744e-06, "learning_rate": 7.5894940934526125e-06, "loss": 0.0, "num_input_tokens_seen": 8346576, "step": 29810 }, { "epoch": 331.27777777777777, "grad_norm": 7.964503311086446e-06, "learning_rate": 7.582450078744621e-06, "loss": 0.0, "num_input_tokens_seen": 8348000, "step": 29815 }, { "epoch": 331.3333333333333, "grad_norm": 1.0164363629883155e-05, "learning_rate": 7.575408750041707e-06, "loss": 0.0, "num_input_tokens_seen": 8349392, "step": 29820 }, { "epoch": 331.3888888888889, "grad_norm": 8.803864147921558e-06, "learning_rate": 7.568370108429732e-06, "loss": 0.0, "num_input_tokens_seen": 8350816, "step": 29825 }, { "epoch": 331.44444444444446, "grad_norm": 6.87659194227308e-05, "learning_rate": 7.561334154994154e-06, "loss": 0.0, "num_input_tokens_seen": 8352208, "step": 29830 }, { "epoch": 331.5, "grad_norm": 5.537531251320615e-06, "learning_rate": 7.55430089081999e-06, "loss": 0.0, "num_input_tokens_seen": 8353584, "step": 29835 }, { "epoch": 331.55555555555554, "grad_norm": 0.00010267386824125424, "learning_rate": 7.547270316991864e-06, "loss": 0.0, "num_input_tokens_seen": 8354944, "step": 29840 }, { "epoch": 331.6111111111111, "grad_norm": 1.6151858517332585e-06, "learning_rate": 7.5402424345939884e-06, "loss": 0.0, "num_input_tokens_seen": 8356320, "step": 29845 }, { "epoch": 331.6666666666667, "grad_norm": 1.1502083907544147e-06, "learning_rate": 7.533217244710133e-06, "loss": 0.0, "num_input_tokens_seen": 8357728, "step": 29850 }, { "epoch": 331.72222222222223, "grad_norm": 3.3916385291377082e-06, "learning_rate": 7.52619474842369e-06, "loss": 0.0, "num_input_tokens_seen": 8359152, "step": 29855 }, { "epoch": 331.77777777777777, "grad_norm": 6.855053129584121e-07, "learning_rate": 7.519174946817597e-06, "loss": 0.0, "num_input_tokens_seen": 8360544, "step": 29860 }, { "epoch": 331.8333333333333, "grad_norm": 2.451370619382942e-06, "learning_rate": 7.512157840974407e-06, "loss": 0.0, "num_input_tokens_seen": 8361968, "step": 29865 }, { "epoch": 331.8888888888889, "grad_norm": 4.421069377258391e-07, "learning_rate": 7.5051434319762496e-06, "loss": 0.0, "num_input_tokens_seen": 8363376, "step": 29870 }, { "epoch": 331.94444444444446, "grad_norm": 8.055374200921506e-05, "learning_rate": 7.498131720904822e-06, "loss": 0.0, "num_input_tokens_seen": 8364768, "step": 29875 }, { "epoch": 332.0, "grad_norm": 1.5868081391090527e-05, "learning_rate": 7.491122708841433e-06, "loss": 0.0, "num_input_tokens_seen": 8366160, "step": 29880 }, { "epoch": 332.05555555555554, "grad_norm": 1.4075051240070025e-06, "learning_rate": 7.4841163968669524e-06, "loss": 0.0, "num_input_tokens_seen": 8367584, "step": 29885 }, { "epoch": 332.1111111111111, "grad_norm": 5.420164598035626e-06, "learning_rate": 7.4771127860618355e-06, "loss": 0.0, "num_input_tokens_seen": 8368992, "step": 29890 }, { "epoch": 332.1666666666667, "grad_norm": 3.2892776289372705e-06, "learning_rate": 7.470111877506139e-06, "loss": 0.0, "num_input_tokens_seen": 8370384, "step": 29895 }, { "epoch": 332.22222222222223, "grad_norm": 7.476450718968408e-06, "learning_rate": 7.463113672279479e-06, "loss": 0.0, "num_input_tokens_seen": 8371760, "step": 29900 }, { "epoch": 332.27777777777777, "grad_norm": 1.4070944416744169e-05, "learning_rate": 7.456118171461071e-06, "loss": 0.0, "num_input_tokens_seen": 8373152, "step": 29905 }, { "epoch": 332.3333333333333, "grad_norm": 1.6115237713165698e-06, "learning_rate": 7.449125376129721e-06, "loss": 0.0, "num_input_tokens_seen": 8374576, "step": 29910 }, { "epoch": 332.3888888888889, "grad_norm": 2.560004986662534e-06, "learning_rate": 7.442135287363788e-06, "loss": 0.0, "num_input_tokens_seen": 8375968, "step": 29915 }, { "epoch": 332.44444444444446, "grad_norm": 1.0597295840852894e-05, "learning_rate": 7.435147906241247e-06, "loss": 0.0, "num_input_tokens_seen": 8377344, "step": 29920 }, { "epoch": 332.5, "grad_norm": 4.247550350555684e-06, "learning_rate": 7.428163233839624e-06, "loss": 0.0, "num_input_tokens_seen": 8378736, "step": 29925 }, { "epoch": 332.55555555555554, "grad_norm": 1.0280597052769735e-05, "learning_rate": 7.4211812712360525e-06, "loss": 0.0, "num_input_tokens_seen": 8380144, "step": 29930 }, { "epoch": 332.6111111111111, "grad_norm": 5.861435056431219e-06, "learning_rate": 7.4142020195072464e-06, "loss": 0.0, "num_input_tokens_seen": 8381536, "step": 29935 }, { "epoch": 332.6666666666667, "grad_norm": 1.1432030078140087e-06, "learning_rate": 7.407225479729479e-06, "loss": 0.0, "num_input_tokens_seen": 8382960, "step": 29940 }, { "epoch": 332.72222222222223, "grad_norm": 1.1864550288009923e-05, "learning_rate": 7.400251652978632e-06, "loss": 0.0, "num_input_tokens_seen": 8384368, "step": 29945 }, { "epoch": 332.77777777777777, "grad_norm": 1.8140917745768093e-06, "learning_rate": 7.393280540330147e-06, "loss": 0.0, "num_input_tokens_seen": 8385760, "step": 29950 }, { "epoch": 332.8333333333333, "grad_norm": 2.09105546673527e-05, "learning_rate": 7.386312142859069e-06, "loss": 0.0, "num_input_tokens_seen": 8387168, "step": 29955 }, { "epoch": 332.8888888888889, "grad_norm": 6.961366580071626e-06, "learning_rate": 7.379346461640008e-06, "loss": 0.0, "num_input_tokens_seen": 8388560, "step": 29960 }, { "epoch": 332.94444444444446, "grad_norm": 3.7773436361021595e-06, "learning_rate": 7.372383497747149e-06, "loss": 0.0, "num_input_tokens_seen": 8389920, "step": 29965 }, { "epoch": 333.0, "grad_norm": 6.682242315037001e-07, "learning_rate": 7.3654232522542775e-06, "loss": 0.0, "num_input_tokens_seen": 8391328, "step": 29970 }, { "epoch": 333.05555555555554, "grad_norm": 0.00014227672363631427, "learning_rate": 7.358465726234756e-06, "loss": 0.0, "num_input_tokens_seen": 8392720, "step": 29975 }, { "epoch": 333.1111111111111, "grad_norm": 6.035035653439991e-07, "learning_rate": 7.351510920761512e-06, "loss": 0.0, "num_input_tokens_seen": 8394112, "step": 29980 }, { "epoch": 333.1666666666667, "grad_norm": 2.720766815400566e-06, "learning_rate": 7.344558836907067e-06, "loss": 0.0, "num_input_tokens_seen": 8395504, "step": 29985 }, { "epoch": 333.22222222222223, "grad_norm": 1.9650581180030713e-06, "learning_rate": 7.3376094757435285e-06, "loss": 0.0, "num_input_tokens_seen": 8396928, "step": 29990 }, { "epoch": 333.27777777777777, "grad_norm": 2.79994378615811e-06, "learning_rate": 7.330662838342561e-06, "loss": 0.0, "num_input_tokens_seen": 8398320, "step": 29995 }, { "epoch": 333.3333333333333, "grad_norm": 4.8140856279133e-07, "learning_rate": 7.323718925775438e-06, "loss": 0.0, "num_input_tokens_seen": 8399696, "step": 30000 }, { "epoch": 333.3333333333333, "eval_loss": 1.0551388263702393, "eval_runtime": 1.3897, "eval_samples_per_second": 28.784, "eval_steps_per_second": 14.392, "num_input_tokens_seen": 8399696, "step": 30000 }, { "epoch": 333.3888888888889, "grad_norm": 6.996205570430902e-07, "learning_rate": 7.316777739112985e-06, "loss": 0.0, "num_input_tokens_seen": 8401104, "step": 30005 }, { "epoch": 333.44444444444446, "grad_norm": 5.695495019608643e-06, "learning_rate": 7.309839279425626e-06, "loss": 0.0, "num_input_tokens_seen": 8402512, "step": 30010 }, { "epoch": 333.5, "grad_norm": 5.451153356261784e-06, "learning_rate": 7.302903547783366e-06, "loss": 0.0, "num_input_tokens_seen": 8403888, "step": 30015 }, { "epoch": 333.55555555555554, "grad_norm": 5.493766366271302e-05, "learning_rate": 7.2959705452557644e-06, "loss": 0.0, "num_input_tokens_seen": 8405312, "step": 30020 }, { "epoch": 333.6111111111111, "grad_norm": 6.4276086959580425e-06, "learning_rate": 7.289040272911996e-06, "loss": 0.0, "num_input_tokens_seen": 8406752, "step": 30025 }, { "epoch": 333.6666666666667, "grad_norm": 1.1562625331862364e-05, "learning_rate": 7.282112731820789e-06, "loss": 0.0, "num_input_tokens_seen": 8408192, "step": 30030 }, { "epoch": 333.72222222222223, "grad_norm": 2.0088220480829477e-05, "learning_rate": 7.275187923050447e-06, "loss": 0.0, "num_input_tokens_seen": 8409568, "step": 30035 }, { "epoch": 333.77777777777777, "grad_norm": 8.540489943698049e-06, "learning_rate": 7.268265847668879e-06, "loss": 0.0, "num_input_tokens_seen": 8411008, "step": 30040 }, { "epoch": 333.8333333333333, "grad_norm": 1.9215147403883748e-05, "learning_rate": 7.261346506743538e-06, "loss": 0.0, "num_input_tokens_seen": 8412368, "step": 30045 }, { "epoch": 333.8888888888889, "grad_norm": 2.1906964775553206e-06, "learning_rate": 7.254429901341486e-06, "loss": 0.0, "num_input_tokens_seen": 8413744, "step": 30050 }, { "epoch": 333.94444444444446, "grad_norm": 2.2400070520234294e-05, "learning_rate": 7.247516032529356e-06, "loss": 0.0, "num_input_tokens_seen": 8415152, "step": 30055 }, { "epoch": 334.0, "grad_norm": 8.445164894510526e-06, "learning_rate": 7.240604901373338e-06, "loss": 0.0, "num_input_tokens_seen": 8416592, "step": 30060 }, { "epoch": 334.05555555555554, "grad_norm": 9.425264579476789e-06, "learning_rate": 7.233696508939223e-06, "loss": 0.0, "num_input_tokens_seen": 8418016, "step": 30065 }, { "epoch": 334.1111111111111, "grad_norm": 5.59683539904654e-05, "learning_rate": 7.226790856292376e-06, "loss": 0.0, "num_input_tokens_seen": 8419408, "step": 30070 }, { "epoch": 334.1666666666667, "grad_norm": 1.0235429499516613e-06, "learning_rate": 7.219887944497727e-06, "loss": 0.0, "num_input_tokens_seen": 8420784, "step": 30075 }, { "epoch": 334.22222222222223, "grad_norm": 8.349238669325132e-06, "learning_rate": 7.2129877746198e-06, "loss": 0.0, "num_input_tokens_seen": 8422192, "step": 30080 }, { "epoch": 334.27777777777777, "grad_norm": 2.4189756004489027e-05, "learning_rate": 7.20609034772268e-06, "loss": 0.0, "num_input_tokens_seen": 8423648, "step": 30085 }, { "epoch": 334.3333333333333, "grad_norm": 5.308556865202263e-06, "learning_rate": 7.19919566487004e-06, "loss": 0.0, "num_input_tokens_seen": 8425056, "step": 30090 }, { "epoch": 334.3888888888889, "grad_norm": 3.4490237794670975e-06, "learning_rate": 7.192303727125132e-06, "loss": 0.0, "num_input_tokens_seen": 8426464, "step": 30095 }, { "epoch": 334.44444444444446, "grad_norm": 8.823849384498317e-06, "learning_rate": 7.185414535550777e-06, "loss": 0.0, "num_input_tokens_seen": 8427872, "step": 30100 }, { "epoch": 334.5, "grad_norm": 5.393268293119036e-05, "learning_rate": 7.178528091209363e-06, "loss": 0.0, "num_input_tokens_seen": 8429280, "step": 30105 }, { "epoch": 334.55555555555554, "grad_norm": 1.0773156873256085e-06, "learning_rate": 7.171644395162888e-06, "loss": 0.0, "num_input_tokens_seen": 8430656, "step": 30110 }, { "epoch": 334.6111111111111, "grad_norm": 7.080895557010081e-06, "learning_rate": 7.164763448472881e-06, "loss": 0.0, "num_input_tokens_seen": 8432048, "step": 30115 }, { "epoch": 334.6666666666667, "grad_norm": 5.711129347218957e-07, "learning_rate": 7.157885252200491e-06, "loss": 0.0, "num_input_tokens_seen": 8433488, "step": 30120 }, { "epoch": 334.72222222222223, "grad_norm": 3.553691567503847e-05, "learning_rate": 7.151009807406403e-06, "loss": 0.0, "num_input_tokens_seen": 8434800, "step": 30125 }, { "epoch": 334.77777777777777, "grad_norm": 2.87028501588793e-06, "learning_rate": 7.144137115150909e-06, "loss": 0.0, "num_input_tokens_seen": 8436208, "step": 30130 }, { "epoch": 334.8333333333333, "grad_norm": 1.2519950587375206e-06, "learning_rate": 7.1372671764938725e-06, "loss": 0.0, "num_input_tokens_seen": 8437632, "step": 30135 }, { "epoch": 334.8888888888889, "grad_norm": 1.4945337625249522e-06, "learning_rate": 7.130399992494705e-06, "loss": 0.0, "num_input_tokens_seen": 8439056, "step": 30140 }, { "epoch": 334.94444444444446, "grad_norm": 3.112174817943014e-05, "learning_rate": 7.123535564212419e-06, "loss": 0.0, "num_input_tokens_seen": 8440480, "step": 30145 }, { "epoch": 335.0, "grad_norm": 8.656851605337579e-06, "learning_rate": 7.116673892705611e-06, "loss": 0.0, "num_input_tokens_seen": 8441872, "step": 30150 }, { "epoch": 335.05555555555554, "grad_norm": 6.248992576729506e-07, "learning_rate": 7.109814979032415e-06, "loss": 0.0, "num_input_tokens_seen": 8443296, "step": 30155 }, { "epoch": 335.1111111111111, "grad_norm": 3.289071628387319e-06, "learning_rate": 7.102958824250577e-06, "loss": 0.0, "num_input_tokens_seen": 8444640, "step": 30160 }, { "epoch": 335.1666666666667, "grad_norm": 3.6052807672604104e-07, "learning_rate": 7.096105429417393e-06, "loss": 0.0, "num_input_tokens_seen": 8446048, "step": 30165 }, { "epoch": 335.22222222222223, "grad_norm": 4.08661799156107e-06, "learning_rate": 7.0892547955897506e-06, "loss": 0.0, "num_input_tokens_seen": 8447488, "step": 30170 }, { "epoch": 335.27777777777777, "grad_norm": 1.3206310995883541e-06, "learning_rate": 7.0824069238241e-06, "loss": 0.0, "num_input_tokens_seen": 8448912, "step": 30175 }, { "epoch": 335.3333333333333, "grad_norm": 2.9520326279453002e-05, "learning_rate": 7.075561815176462e-06, "loss": 0.0, "num_input_tokens_seen": 8450240, "step": 30180 }, { "epoch": 335.3888888888889, "grad_norm": 6.752764875272987e-06, "learning_rate": 7.068719470702445e-06, "loss": 0.0, "num_input_tokens_seen": 8451632, "step": 30185 }, { "epoch": 335.44444444444446, "grad_norm": 2.8920751447003568e-06, "learning_rate": 7.061879891457229e-06, "loss": 0.0, "num_input_tokens_seen": 8452992, "step": 30190 }, { "epoch": 335.5, "grad_norm": 7.528412879764801e-06, "learning_rate": 7.0550430784955515e-06, "loss": 0.0, "num_input_tokens_seen": 8454400, "step": 30195 }, { "epoch": 335.55555555555554, "grad_norm": 1.8376896377958474e-06, "learning_rate": 7.048209032871752e-06, "loss": 0.0, "num_input_tokens_seen": 8455776, "step": 30200 }, { "epoch": 335.55555555555554, "eval_loss": 1.0328112840652466, "eval_runtime": 1.3894, "eval_samples_per_second": 28.789, "eval_steps_per_second": 14.395, "num_input_tokens_seen": 8455776, "step": 30200 }, { "epoch": 335.6111111111111, "grad_norm": 5.463666184368776e-06, "learning_rate": 7.0413777556397055e-06, "loss": 0.0, "num_input_tokens_seen": 8457232, "step": 30205 }, { "epoch": 335.6666666666667, "grad_norm": 6.597117135243025e-06, "learning_rate": 7.0345492478528925e-06, "loss": 0.0, "num_input_tokens_seen": 8458608, "step": 30210 }, { "epoch": 335.72222222222223, "grad_norm": 3.7443787732627243e-06, "learning_rate": 7.02772351056436e-06, "loss": 0.0, "num_input_tokens_seen": 8460016, "step": 30215 }, { "epoch": 335.77777777777777, "grad_norm": 2.239460400232929e-06, "learning_rate": 7.020900544826709e-06, "loss": 0.0, "num_input_tokens_seen": 8461360, "step": 30220 }, { "epoch": 335.8333333333333, "grad_norm": 1.2863846677646507e-05, "learning_rate": 7.014080351692134e-06, "loss": 0.0, "num_input_tokens_seen": 8462800, "step": 30225 }, { "epoch": 335.8888888888889, "grad_norm": 9.348976163892075e-06, "learning_rate": 7.0072629322124024e-06, "loss": 0.0, "num_input_tokens_seen": 8464192, "step": 30230 }, { "epoch": 335.94444444444446, "grad_norm": 3.146695962641388e-05, "learning_rate": 7.000448287438827e-06, "loss": 0.0, "num_input_tokens_seen": 8465616, "step": 30235 }, { "epoch": 336.0, "grad_norm": 6.434280749090249e-06, "learning_rate": 6.993636418422331e-06, "loss": 0.0, "num_input_tokens_seen": 8466992, "step": 30240 }, { "epoch": 336.05555555555554, "grad_norm": 2.1905307221459225e-06, "learning_rate": 6.986827326213383e-06, "loss": 0.0, "num_input_tokens_seen": 8468416, "step": 30245 }, { "epoch": 336.1111111111111, "grad_norm": 3.2566001095801766e-07, "learning_rate": 6.9800210118620205e-06, "loss": 0.0, "num_input_tokens_seen": 8469808, "step": 30250 }, { "epoch": 336.1666666666667, "grad_norm": 3.195475073880516e-06, "learning_rate": 6.973217476417876e-06, "loss": 0.0, "num_input_tokens_seen": 8471216, "step": 30255 }, { "epoch": 336.22222222222223, "grad_norm": 4.7195597289828584e-05, "learning_rate": 6.96641672093013e-06, "loss": 0.0, "num_input_tokens_seen": 8472608, "step": 30260 }, { "epoch": 336.27777777777777, "grad_norm": 2.0185401808703318e-05, "learning_rate": 6.95961874644755e-06, "loss": 0.0, "num_input_tokens_seen": 8473984, "step": 30265 }, { "epoch": 336.3333333333333, "grad_norm": 3.095103465966531e-06, "learning_rate": 6.952823554018476e-06, "loss": 0.0, "num_input_tokens_seen": 8475424, "step": 30270 }, { "epoch": 336.3888888888889, "grad_norm": 1.6163894542842172e-06, "learning_rate": 6.946031144690798e-06, "loss": 0.0, "num_input_tokens_seen": 8476848, "step": 30275 }, { "epoch": 336.44444444444446, "grad_norm": 4.933794571115868e-06, "learning_rate": 6.939241519512005e-06, "loss": 0.0, "num_input_tokens_seen": 8478256, "step": 30280 }, { "epoch": 336.5, "grad_norm": 8.482157340949925e-07, "learning_rate": 6.932454679529129e-06, "loss": 0.0, "num_input_tokens_seen": 8479696, "step": 30285 }, { "epoch": 336.55555555555554, "grad_norm": 1.0174776434723753e-05, "learning_rate": 6.925670625788791e-06, "loss": 0.0, "num_input_tokens_seen": 8481104, "step": 30290 }, { "epoch": 336.6111111111111, "grad_norm": 5.2577070164261386e-05, "learning_rate": 6.918889359337186e-06, "loss": 0.0, "num_input_tokens_seen": 8482496, "step": 30295 }, { "epoch": 336.6666666666667, "grad_norm": 4.231357252137968e-06, "learning_rate": 6.912110881220058e-06, "loss": 0.0, "num_input_tokens_seen": 8483856, "step": 30300 }, { "epoch": 336.72222222222223, "grad_norm": 2.055237564491108e-05, "learning_rate": 6.905335192482735e-06, "loss": 0.0, "num_input_tokens_seen": 8485248, "step": 30305 }, { "epoch": 336.77777777777777, "grad_norm": 1.632961539144162e-05, "learning_rate": 6.8985622941701275e-06, "loss": 0.0, "num_input_tokens_seen": 8486624, "step": 30310 }, { "epoch": 336.8333333333333, "grad_norm": 3.7694092043238925e-06, "learning_rate": 6.89179218732669e-06, "loss": 0.0, "num_input_tokens_seen": 8487984, "step": 30315 }, { "epoch": 336.8888888888889, "grad_norm": 5.759808118455112e-05, "learning_rate": 6.8850248729964595e-06, "loss": 0.0, "num_input_tokens_seen": 8489392, "step": 30320 }, { "epoch": 336.94444444444446, "grad_norm": 1.218267425429076e-05, "learning_rate": 6.8782603522230314e-06, "loss": 0.0, "num_input_tokens_seen": 8490768, "step": 30325 }, { "epoch": 337.0, "grad_norm": 6.422506430681096e-06, "learning_rate": 6.871498626049591e-06, "loss": 0.0, "num_input_tokens_seen": 8492160, "step": 30330 }, { "epoch": 337.05555555555554, "grad_norm": 8.721890480956063e-06, "learning_rate": 6.8647396955188875e-06, "loss": 0.0, "num_input_tokens_seen": 8493584, "step": 30335 }, { "epoch": 337.1111111111111, "grad_norm": 9.619653837944497e-07, "learning_rate": 6.857983561673218e-06, "loss": 0.0, "num_input_tokens_seen": 8494992, "step": 30340 }, { "epoch": 337.1666666666667, "grad_norm": 2.9836935937055387e-05, "learning_rate": 6.851230225554467e-06, "loss": 0.0, "num_input_tokens_seen": 8496384, "step": 30345 }, { "epoch": 337.22222222222223, "grad_norm": 1.5378841453639325e-06, "learning_rate": 6.8444796882040946e-06, "loss": 0.0, "num_input_tokens_seen": 8497792, "step": 30350 }, { "epoch": 337.27777777777777, "grad_norm": 5.193581273488235e-06, "learning_rate": 6.837731950663106e-06, "loss": 0.0, "num_input_tokens_seen": 8499168, "step": 30355 }, { "epoch": 337.3333333333333, "grad_norm": 5.552633865590906e-06, "learning_rate": 6.830987013972098e-06, "loss": 0.0, "num_input_tokens_seen": 8500544, "step": 30360 }, { "epoch": 337.3888888888889, "grad_norm": 1.8884340136082756e-07, "learning_rate": 6.82424487917121e-06, "loss": 0.0, "num_input_tokens_seen": 8501968, "step": 30365 }, { "epoch": 337.44444444444446, "grad_norm": 7.655301033082651e-07, "learning_rate": 6.8175055473001735e-06, "loss": 0.0, "num_input_tokens_seen": 8503360, "step": 30370 }, { "epoch": 337.5, "grad_norm": 3.239878424210474e-05, "learning_rate": 6.8107690193982855e-06, "loss": 0.0, "num_input_tokens_seen": 8504752, "step": 30375 }, { "epoch": 337.55555555555554, "grad_norm": 9.799249482966843e-07, "learning_rate": 6.804035296504385e-06, "loss": 0.0, "num_input_tokens_seen": 8506160, "step": 30380 }, { "epoch": 337.6111111111111, "grad_norm": 4.877835817751475e-05, "learning_rate": 6.797304379656916e-06, "loss": 0.0, "num_input_tokens_seen": 8507552, "step": 30385 }, { "epoch": 337.6666666666667, "grad_norm": 1.4481337302640895e-06, "learning_rate": 6.790576269893861e-06, "loss": 0.0, "num_input_tokens_seen": 8508960, "step": 30390 }, { "epoch": 337.72222222222223, "grad_norm": 5.594207777903648e-06, "learning_rate": 6.783850968252772e-06, "loss": 0.0, "num_input_tokens_seen": 8510352, "step": 30395 }, { "epoch": 337.77777777777777, "grad_norm": 5.115382919029798e-06, "learning_rate": 6.777128475770789e-06, "loss": 0.0, "num_input_tokens_seen": 8511760, "step": 30400 }, { "epoch": 337.77777777777777, "eval_loss": 1.0401995182037354, "eval_runtime": 1.3981, "eval_samples_per_second": 28.611, "eval_steps_per_second": 14.305, "num_input_tokens_seen": 8511760, "step": 30400 }, { "epoch": 337.8333333333333, "grad_norm": 5.49027172382921e-05, "learning_rate": 6.77040879348459e-06, "loss": 0.0, "num_input_tokens_seen": 8513184, "step": 30405 }, { "epoch": 337.8888888888889, "grad_norm": 1.6267151295323856e-05, "learning_rate": 6.763691922430443e-06, "loss": 0.0, "num_input_tokens_seen": 8514560, "step": 30410 }, { "epoch": 337.94444444444446, "grad_norm": 2.7702485567715485e-06, "learning_rate": 6.756977863644178e-06, "loss": 0.0, "num_input_tokens_seen": 8515952, "step": 30415 }, { "epoch": 338.0, "grad_norm": 1.8286095837538596e-06, "learning_rate": 6.7502666181611804e-06, "loss": 0.0, "num_input_tokens_seen": 8517360, "step": 30420 }, { "epoch": 338.05555555555554, "grad_norm": 2.023780552917742e-06, "learning_rate": 6.743558187016405e-06, "loss": 0.0, "num_input_tokens_seen": 8518800, "step": 30425 }, { "epoch": 338.1111111111111, "grad_norm": 1.056866403814638e-05, "learning_rate": 6.7368525712443925e-06, "loss": 0.0, "num_input_tokens_seen": 8520192, "step": 30430 }, { "epoch": 338.1666666666667, "grad_norm": 9.929948646458797e-06, "learning_rate": 6.7301497718792155e-06, "loss": 0.0, "num_input_tokens_seen": 8521600, "step": 30435 }, { "epoch": 338.22222222222223, "grad_norm": 3.94276003135019e-06, "learning_rate": 6.723449789954544e-06, "loss": 0.0, "num_input_tokens_seen": 8522960, "step": 30440 }, { "epoch": 338.27777777777777, "grad_norm": 3.171066055074334e-06, "learning_rate": 6.716752626503586e-06, "loss": 0.0, "num_input_tokens_seen": 8524384, "step": 30445 }, { "epoch": 338.3333333333333, "grad_norm": 4.820666390514816e-07, "learning_rate": 6.710058282559131e-06, "loss": 0.0, "num_input_tokens_seen": 8525824, "step": 30450 }, { "epoch": 338.3888888888889, "grad_norm": 2.4623113858979195e-06, "learning_rate": 6.703366759153545e-06, "loss": 0.0, "num_input_tokens_seen": 8527216, "step": 30455 }, { "epoch": 338.44444444444446, "grad_norm": 2.7162221272192255e-07, "learning_rate": 6.6966780573187335e-06, "loss": 0.0, "num_input_tokens_seen": 8528624, "step": 30460 }, { "epoch": 338.5, "grad_norm": 7.181969976954861e-06, "learning_rate": 6.689992178086174e-06, "loss": 0.0, "num_input_tokens_seen": 8530032, "step": 30465 }, { "epoch": 338.55555555555554, "grad_norm": 5.827905624755658e-06, "learning_rate": 6.683309122486925e-06, "loss": 0.0, "num_input_tokens_seen": 8531376, "step": 30470 }, { "epoch": 338.6111111111111, "grad_norm": 5.055734618508723e-06, "learning_rate": 6.676628891551584e-06, "loss": 0.0, "num_input_tokens_seen": 8532736, "step": 30475 }, { "epoch": 338.6666666666667, "grad_norm": 9.8992222774541e-07, "learning_rate": 6.6699514863103385e-06, "loss": 0.0, "num_input_tokens_seen": 8534160, "step": 30480 }, { "epoch": 338.72222222222223, "grad_norm": 1.1671910442601074e-06, "learning_rate": 6.663276907792921e-06, "loss": 0.0, "num_input_tokens_seen": 8535584, "step": 30485 }, { "epoch": 338.77777777777777, "grad_norm": 0.00023699404846411198, "learning_rate": 6.656605157028634e-06, "loss": 0.0, "num_input_tokens_seen": 8537024, "step": 30490 }, { "epoch": 338.8333333333333, "grad_norm": 2.9145799089747015e-06, "learning_rate": 6.649936235046358e-06, "loss": 0.0, "num_input_tokens_seen": 8538352, "step": 30495 }, { "epoch": 338.8888888888889, "grad_norm": 8.052249177126214e-05, "learning_rate": 6.643270142874508e-06, "loss": 0.0, "num_input_tokens_seen": 8539792, "step": 30500 }, { "epoch": 338.94444444444446, "grad_norm": 4.892705760539684e-07, "learning_rate": 6.636606881541094e-06, "loss": 0.0, "num_input_tokens_seen": 8541184, "step": 30505 }, { "epoch": 339.0, "grad_norm": 2.8213935365783982e-05, "learning_rate": 6.629946452073662e-06, "loss": 0.0, "num_input_tokens_seen": 8542576, "step": 30510 }, { "epoch": 339.05555555555554, "grad_norm": 4.74579223919136e-07, "learning_rate": 6.6232888554993375e-06, "loss": 0.0, "num_input_tokens_seen": 8543968, "step": 30515 }, { "epoch": 339.1111111111111, "grad_norm": 4.038250153826084e-06, "learning_rate": 6.616634092844817e-06, "loss": 0.0, "num_input_tokens_seen": 8545392, "step": 30520 }, { "epoch": 339.1666666666667, "grad_norm": 1.3171361388231162e-05, "learning_rate": 6.609982165136331e-06, "loss": 0.0, "num_input_tokens_seen": 8546784, "step": 30525 }, { "epoch": 339.22222222222223, "grad_norm": 2.234809244328062e-06, "learning_rate": 6.603333073399706e-06, "loss": 0.0, "num_input_tokens_seen": 8548192, "step": 30530 }, { "epoch": 339.27777777777777, "grad_norm": 3.3738183446985204e-06, "learning_rate": 6.596686818660308e-06, "loss": 0.0, "num_input_tokens_seen": 8549568, "step": 30535 }, { "epoch": 339.3333333333333, "grad_norm": 2.400527989721013e-07, "learning_rate": 6.590043401943066e-06, "loss": 0.0, "num_input_tokens_seen": 8550976, "step": 30540 }, { "epoch": 339.3888888888889, "grad_norm": 6.326894208541489e-07, "learning_rate": 6.583402824272494e-06, "loss": 0.0, "num_input_tokens_seen": 8552368, "step": 30545 }, { "epoch": 339.44444444444446, "grad_norm": 3.0491419238387607e-05, "learning_rate": 6.576765086672634e-06, "loss": 0.0, "num_input_tokens_seen": 8553792, "step": 30550 }, { "epoch": 339.5, "grad_norm": 8.092361895251088e-06, "learning_rate": 6.57013019016712e-06, "loss": 0.0, "num_input_tokens_seen": 8555216, "step": 30555 }, { "epoch": 339.55555555555554, "grad_norm": 1.5294044715119526e-06, "learning_rate": 6.563498135779142e-06, "loss": 0.0, "num_input_tokens_seen": 8556656, "step": 30560 }, { "epoch": 339.6111111111111, "grad_norm": 3.108532382611884e-07, "learning_rate": 6.556868924531431e-06, "loss": 0.0, "num_input_tokens_seen": 8558048, "step": 30565 }, { "epoch": 339.6666666666667, "grad_norm": 3.6588762668543495e-06, "learning_rate": 6.550242557446304e-06, "loss": 0.0, "num_input_tokens_seen": 8559408, "step": 30570 }, { "epoch": 339.72222222222223, "grad_norm": 8.837493510327477e-07, "learning_rate": 6.543619035545634e-06, "loss": 0.0, "num_input_tokens_seen": 8560768, "step": 30575 }, { "epoch": 339.77777777777777, "grad_norm": 4.844795967073878e-06, "learning_rate": 6.53699835985084e-06, "loss": 0.0, "num_input_tokens_seen": 8562208, "step": 30580 }, { "epoch": 339.8333333333333, "grad_norm": 3.500809543766081e-05, "learning_rate": 6.530380531382927e-06, "loss": 0.0, "num_input_tokens_seen": 8563584, "step": 30585 }, { "epoch": 339.8888888888889, "grad_norm": 1.0743207212726702e-06, "learning_rate": 6.523765551162433e-06, "loss": 0.0, "num_input_tokens_seen": 8564976, "step": 30590 }, { "epoch": 339.94444444444446, "grad_norm": 1.7171981880892417e-06, "learning_rate": 6.517153420209476e-06, "loss": 0.0, "num_input_tokens_seen": 8566400, "step": 30595 }, { "epoch": 340.0, "grad_norm": 1.3946068975201342e-05, "learning_rate": 6.510544139543739e-06, "loss": 0.0, "num_input_tokens_seen": 8567792, "step": 30600 }, { "epoch": 340.0, "eval_loss": 1.0617506504058838, "eval_runtime": 1.3859, "eval_samples_per_second": 28.862, "eval_steps_per_second": 14.431, "num_input_tokens_seen": 8567792, "step": 30600 }, { "epoch": 340.05555555555554, "grad_norm": 1.4192162325343816e-06, "learning_rate": 6.503937710184452e-06, "loss": 0.0, "num_input_tokens_seen": 8569168, "step": 30605 }, { "epoch": 340.1111111111111, "grad_norm": 2.9389304927462945e-06, "learning_rate": 6.4973341331503954e-06, "loss": 0.0, "num_input_tokens_seen": 8570576, "step": 30610 }, { "epoch": 340.1666666666667, "grad_norm": 2.7979633614449995e-06, "learning_rate": 6.490733409459942e-06, "loss": 0.0, "num_input_tokens_seen": 8572000, "step": 30615 }, { "epoch": 340.22222222222223, "grad_norm": 6.149011369416257e-06, "learning_rate": 6.484135540130995e-06, "loss": 0.0, "num_input_tokens_seen": 8573408, "step": 30620 }, { "epoch": 340.27777777777777, "grad_norm": 8.897039833755116e-07, "learning_rate": 6.4775405261810364e-06, "loss": 0.0, "num_input_tokens_seen": 8574816, "step": 30625 }, { "epoch": 340.3333333333333, "grad_norm": 5.8315217756899074e-05, "learning_rate": 6.470948368627092e-06, "loss": 0.0, "num_input_tokens_seen": 8576208, "step": 30630 }, { "epoch": 340.3888888888889, "grad_norm": 8.778963547229068e-07, "learning_rate": 6.464359068485756e-06, "loss": 0.0, "num_input_tokens_seen": 8577632, "step": 30635 }, { "epoch": 340.44444444444446, "grad_norm": 6.606033366551856e-06, "learning_rate": 6.457772626773195e-06, "loss": 0.0, "num_input_tokens_seen": 8579056, "step": 30640 }, { "epoch": 340.5, "grad_norm": 1.7778518667910248e-05, "learning_rate": 6.451189044505104e-06, "loss": 0.0, "num_input_tokens_seen": 8580416, "step": 30645 }, { "epoch": 340.55555555555554, "grad_norm": 3.1860415674600517e-06, "learning_rate": 6.44460832269676e-06, "loss": 0.0, "num_input_tokens_seen": 8581824, "step": 30650 }, { "epoch": 340.6111111111111, "grad_norm": 1.4729484064446297e-05, "learning_rate": 6.438030462363001e-06, "loss": 0.0, "num_input_tokens_seen": 8583200, "step": 30655 }, { "epoch": 340.6666666666667, "grad_norm": 1.2934458482050104e-07, "learning_rate": 6.431455464518205e-06, "loss": 0.0, "num_input_tokens_seen": 8584560, "step": 30660 }, { "epoch": 340.72222222222223, "grad_norm": 3.3529786378494464e-06, "learning_rate": 6.424883330176326e-06, "loss": 0.0, "num_input_tokens_seen": 8585936, "step": 30665 }, { "epoch": 340.77777777777777, "grad_norm": 1.7776860659068916e-06, "learning_rate": 6.418314060350864e-06, "loss": 0.0, "num_input_tokens_seen": 8587328, "step": 30670 }, { "epoch": 340.8333333333333, "grad_norm": 7.686693606956396e-06, "learning_rate": 6.4117476560548895e-06, "loss": 0.0, "num_input_tokens_seen": 8588688, "step": 30675 }, { "epoch": 340.8888888888889, "grad_norm": 2.0647426026698668e-06, "learning_rate": 6.405184118301016e-06, "loss": 0.0, "num_input_tokens_seen": 8590128, "step": 30680 }, { "epoch": 340.94444444444446, "grad_norm": 2.074174972221954e-06, "learning_rate": 6.398623448101434e-06, "loss": 0.0, "num_input_tokens_seen": 8591504, "step": 30685 }, { "epoch": 341.0, "grad_norm": 4.2068302718689665e-05, "learning_rate": 6.392065646467871e-06, "loss": 0.0, "num_input_tokens_seen": 8592944, "step": 30690 }, { "epoch": 341.05555555555554, "grad_norm": 2.2139461179904174e-06, "learning_rate": 6.385510714411632e-06, "loss": 0.0, "num_input_tokens_seen": 8594352, "step": 30695 }, { "epoch": 341.1111111111111, "grad_norm": 1.6506103293068008e-06, "learning_rate": 6.378958652943559e-06, "loss": 0.0, "num_input_tokens_seen": 8595776, "step": 30700 }, { "epoch": 341.1666666666667, "grad_norm": 2.2157037165015936e-06, "learning_rate": 6.3724094630740776e-06, "loss": 0.0, "num_input_tokens_seen": 8597152, "step": 30705 }, { "epoch": 341.22222222222223, "grad_norm": 8.622708946859348e-07, "learning_rate": 6.365863145813136e-06, "loss": 0.0, "num_input_tokens_seen": 8598608, "step": 30710 }, { "epoch": 341.27777777777777, "grad_norm": 1.6434573808510322e-06, "learning_rate": 6.359319702170269e-06, "loss": 0.0, "num_input_tokens_seen": 8600016, "step": 30715 }, { "epoch": 341.3333333333333, "grad_norm": 4.171331511315657e-07, "learning_rate": 6.352779133154566e-06, "loss": 0.0, "num_input_tokens_seen": 8601424, "step": 30720 }, { "epoch": 341.3888888888889, "grad_norm": 2.6518382583162747e-05, "learning_rate": 6.346241439774648e-06, "loss": 0.0, "num_input_tokens_seen": 8602848, "step": 30725 }, { "epoch": 341.44444444444446, "grad_norm": 4.238593646732625e-06, "learning_rate": 6.339706623038716e-06, "loss": 0.0, "num_input_tokens_seen": 8604256, "step": 30730 }, { "epoch": 341.5, "grad_norm": 9.989616955863312e-06, "learning_rate": 6.333174683954532e-06, "loss": 0.0, "num_input_tokens_seen": 8605616, "step": 30735 }, { "epoch": 341.55555555555554, "grad_norm": 2.6519695893512107e-05, "learning_rate": 6.326645623529387e-06, "loss": 0.0, "num_input_tokens_seen": 8606976, "step": 30740 }, { "epoch": 341.6111111111111, "grad_norm": 8.715563126315828e-06, "learning_rate": 6.320119442770156e-06, "loss": 0.0, "num_input_tokens_seen": 8608400, "step": 30745 }, { "epoch": 341.6666666666667, "grad_norm": 1.6526238368896884e-06, "learning_rate": 6.313596142683254e-06, "loss": 0.0, "num_input_tokens_seen": 8609792, "step": 30750 }, { "epoch": 341.72222222222223, "grad_norm": 7.903501000328106e-07, "learning_rate": 6.307075724274647e-06, "loss": 0.0, "num_input_tokens_seen": 8611200, "step": 30755 }, { "epoch": 341.77777777777777, "grad_norm": 2.5594158614694607e-06, "learning_rate": 6.300558188549882e-06, "loss": 0.0, "num_input_tokens_seen": 8612592, "step": 30760 }, { "epoch": 341.8333333333333, "grad_norm": 1.0572863402558141e-06, "learning_rate": 6.29404353651403e-06, "loss": 0.0, "num_input_tokens_seen": 8613984, "step": 30765 }, { "epoch": 341.8888888888889, "grad_norm": 4.245590389473364e-06, "learning_rate": 6.287531769171737e-06, "loss": 0.0, "num_input_tokens_seen": 8615376, "step": 30770 }, { "epoch": 341.94444444444446, "grad_norm": 8.802857678347209e-07, "learning_rate": 6.2810228875272045e-06, "loss": 0.0, "num_input_tokens_seen": 8616752, "step": 30775 }, { "epoch": 342.0, "grad_norm": 2.3968009372765664e-06, "learning_rate": 6.274516892584179e-06, "loss": 0.0, "num_input_tokens_seen": 8618160, "step": 30780 }, { "epoch": 342.05555555555554, "grad_norm": 1.0007097444031388e-05, "learning_rate": 6.268013785345969e-06, "loss": 0.0, "num_input_tokens_seen": 8619568, "step": 30785 }, { "epoch": 342.1111111111111, "grad_norm": 3.837382337223971e-06, "learning_rate": 6.26151356681543e-06, "loss": 0.0, "num_input_tokens_seen": 8620928, "step": 30790 }, { "epoch": 342.1666666666667, "grad_norm": 6.738667934769182e-07, "learning_rate": 6.255016237994981e-06, "loss": 0.0, "num_input_tokens_seen": 8622320, "step": 30795 }, { "epoch": 342.22222222222223, "grad_norm": 5.487775069923373e-07, "learning_rate": 6.248521799886603e-06, "loss": 0.0, "num_input_tokens_seen": 8623728, "step": 30800 }, { "epoch": 342.22222222222223, "eval_loss": 1.069406270980835, "eval_runtime": 1.396, "eval_samples_per_second": 28.654, "eval_steps_per_second": 14.327, "num_input_tokens_seen": 8623728, "step": 30800 }, { "epoch": 342.27777777777777, "grad_norm": 2.536579813749995e-06, "learning_rate": 6.242030253491798e-06, "loss": 0.0, "num_input_tokens_seen": 8625088, "step": 30805 }, { "epoch": 342.3333333333333, "grad_norm": 2.1630548872053623e-06, "learning_rate": 6.235541599811656e-06, "loss": 0.0, "num_input_tokens_seen": 8626528, "step": 30810 }, { "epoch": 342.3888888888889, "grad_norm": 0.0003261941601522267, "learning_rate": 6.229055839846814e-06, "loss": 0.0, "num_input_tokens_seen": 8627920, "step": 30815 }, { "epoch": 342.44444444444446, "grad_norm": 1.1959660923821502e-06, "learning_rate": 6.222572974597455e-06, "loss": 0.0, "num_input_tokens_seen": 8629312, "step": 30820 }, { "epoch": 342.5, "grad_norm": 2.6882809834205545e-05, "learning_rate": 6.216093005063306e-06, "loss": 0.0, "num_input_tokens_seen": 8630736, "step": 30825 }, { "epoch": 342.55555555555554, "grad_norm": 1.2072055142198224e-06, "learning_rate": 6.209615932243678e-06, "loss": 0.0, "num_input_tokens_seen": 8632112, "step": 30830 }, { "epoch": 342.6111111111111, "grad_norm": 4.0023238057074195e-07, "learning_rate": 6.203141757137399e-06, "loss": 0.0, "num_input_tokens_seen": 8633536, "step": 30835 }, { "epoch": 342.6666666666667, "grad_norm": 2.4969025957943813e-07, "learning_rate": 6.196670480742886e-06, "loss": 0.0, "num_input_tokens_seen": 8634976, "step": 30840 }, { "epoch": 342.72222222222223, "grad_norm": 7.221737519103044e-07, "learning_rate": 6.190202104058074e-06, "loss": 0.0, "num_input_tokens_seen": 8636400, "step": 30845 }, { "epoch": 342.77777777777777, "grad_norm": 7.942957381601445e-06, "learning_rate": 6.183736628080475e-06, "loss": 0.0, "num_input_tokens_seen": 8637792, "step": 30850 }, { "epoch": 342.8333333333333, "grad_norm": 1.5723044270998798e-06, "learning_rate": 6.177274053807155e-06, "loss": 0.0, "num_input_tokens_seen": 8639168, "step": 30855 }, { "epoch": 342.8888888888889, "grad_norm": 8.237063411797862e-06, "learning_rate": 6.170814382234713e-06, "loss": 0.0, "num_input_tokens_seen": 8640608, "step": 30860 }, { "epoch": 342.94444444444446, "grad_norm": 4.192839242023183e-06, "learning_rate": 6.16435761435932e-06, "loss": 0.0, "num_input_tokens_seen": 8642016, "step": 30865 }, { "epoch": 343.0, "grad_norm": 2.0500413029367337e-06, "learning_rate": 6.157903751176681e-06, "loss": 0.0, "num_input_tokens_seen": 8643456, "step": 30870 }, { "epoch": 343.05555555555554, "grad_norm": 6.801083600294078e-06, "learning_rate": 6.151452793682066e-06, "loss": 0.0, "num_input_tokens_seen": 8644880, "step": 30875 }, { "epoch": 343.1111111111111, "grad_norm": 2.2290953438641736e-07, "learning_rate": 6.145004742870305e-06, "loss": 0.0, "num_input_tokens_seen": 8646304, "step": 30880 }, { "epoch": 343.1666666666667, "grad_norm": 7.586099286527315e-07, "learning_rate": 6.138559599735752e-06, "loss": 0.0, "num_input_tokens_seen": 8647712, "step": 30885 }, { "epoch": 343.22222222222223, "grad_norm": 1.1320227713440545e-05, "learning_rate": 6.132117365272344e-06, "loss": 0.0, "num_input_tokens_seen": 8649104, "step": 30890 }, { "epoch": 343.27777777777777, "grad_norm": 6.958808171475539e-06, "learning_rate": 6.125678040473545e-06, "loss": 0.0, "num_input_tokens_seen": 8650512, "step": 30895 }, { "epoch": 343.3333333333333, "grad_norm": 2.1198673493927345e-05, "learning_rate": 6.1192416263323755e-06, "loss": 0.0, "num_input_tokens_seen": 8651888, "step": 30900 }, { "epoch": 343.3888888888889, "grad_norm": 1.8933330636627943e-07, "learning_rate": 6.112808123841424e-06, "loss": 0.0, "num_input_tokens_seen": 8653296, "step": 30905 }, { "epoch": 343.44444444444446, "grad_norm": 6.916070560691878e-07, "learning_rate": 6.106377533992805e-06, "loss": 0.0, "num_input_tokens_seen": 8654704, "step": 30910 }, { "epoch": 343.5, "grad_norm": 4.860129934058932e-07, "learning_rate": 6.099949857778204e-06, "loss": 0.0, "num_input_tokens_seen": 8656064, "step": 30915 }, { "epoch": 343.55555555555554, "grad_norm": 2.3013828467810526e-05, "learning_rate": 6.093525096188852e-06, "loss": 0.0, "num_input_tokens_seen": 8657424, "step": 30920 }, { "epoch": 343.6111111111111, "grad_norm": 3.983378064731369e-06, "learning_rate": 6.087103250215518e-06, "loss": 0.0, "num_input_tokens_seen": 8658864, "step": 30925 }, { "epoch": 343.6666666666667, "grad_norm": 6.780261628591688e-06, "learning_rate": 6.080684320848537e-06, "loss": 0.0, "num_input_tokens_seen": 8660272, "step": 30930 }, { "epoch": 343.72222222222223, "grad_norm": 0.00023899588268250227, "learning_rate": 6.074268309077794e-06, "loss": 0.0, "num_input_tokens_seen": 8661664, "step": 30935 }, { "epoch": 343.77777777777777, "grad_norm": 2.240888716187328e-05, "learning_rate": 6.067855215892709e-06, "loss": 0.0, "num_input_tokens_seen": 8663120, "step": 30940 }, { "epoch": 343.8333333333333, "grad_norm": 3.929575086658588e-06, "learning_rate": 6.061445042282271e-06, "loss": 0.0, "num_input_tokens_seen": 8664528, "step": 30945 }, { "epoch": 343.8888888888889, "grad_norm": 3.104251618424314e-06, "learning_rate": 6.055037789234999e-06, "loss": 0.0, "num_input_tokens_seen": 8665968, "step": 30950 }, { "epoch": 343.94444444444446, "grad_norm": 6.791741270717466e-06, "learning_rate": 6.048633457738975e-06, "loss": 0.0, "num_input_tokens_seen": 8667360, "step": 30955 }, { "epoch": 344.0, "grad_norm": 5.404649527918082e-06, "learning_rate": 6.042232048781837e-06, "loss": 0.0, "num_input_tokens_seen": 8668736, "step": 30960 }, { "epoch": 344.05555555555554, "grad_norm": 1.0605427860355121e-06, "learning_rate": 6.035833563350757e-06, "loss": 0.0, "num_input_tokens_seen": 8670112, "step": 30965 }, { "epoch": 344.1111111111111, "grad_norm": 4.206930952932453e-06, "learning_rate": 6.0294380024324525e-06, "loss": 0.0, "num_input_tokens_seen": 8671520, "step": 30970 }, { "epoch": 344.1666666666667, "grad_norm": 6.960974587855162e-07, "learning_rate": 6.023045367013213e-06, "loss": 0.0, "num_input_tokens_seen": 8672896, "step": 30975 }, { "epoch": 344.22222222222223, "grad_norm": 6.937489160918631e-07, "learning_rate": 6.016655658078851e-06, "loss": 0.0, "num_input_tokens_seen": 8674288, "step": 30980 }, { "epoch": 344.27777777777777, "grad_norm": 4.155648639425635e-06, "learning_rate": 6.010268876614753e-06, "loss": 0.0, "num_input_tokens_seen": 8675696, "step": 30985 }, { "epoch": 344.3333333333333, "grad_norm": 1.0360801070419257e-06, "learning_rate": 6.0038850236058266e-06, "loss": 0.0, "num_input_tokens_seen": 8677120, "step": 30990 }, { "epoch": 344.3888888888889, "grad_norm": 4.889124738838291e-06, "learning_rate": 5.997504100036549e-06, "loss": 0.0, "num_input_tokens_seen": 8678512, "step": 30995 }, { "epoch": 344.44444444444446, "grad_norm": 6.221472176548559e-06, "learning_rate": 5.991126106890949e-06, "loss": 0.0, "num_input_tokens_seen": 8679920, "step": 31000 }, { "epoch": 344.44444444444446, "eval_loss": 1.058841347694397, "eval_runtime": 1.3875, "eval_samples_per_second": 28.828, "eval_steps_per_second": 14.414, "num_input_tokens_seen": 8679920, "step": 31000 }, { "epoch": 344.5, "grad_norm": 1.2952963288626052e-06, "learning_rate": 5.984751045152576e-06, "loss": 0.0, "num_input_tokens_seen": 8681360, "step": 31005 }, { "epoch": 344.55555555555554, "grad_norm": 1.5769372794238734e-06, "learning_rate": 5.978378915804553e-06, "loss": 0.0, "num_input_tokens_seen": 8682736, "step": 31010 }, { "epoch": 344.6111111111111, "grad_norm": 1.8427112991048489e-06, "learning_rate": 5.972009719829547e-06, "loss": 0.0, "num_input_tokens_seen": 8684128, "step": 31015 }, { "epoch": 344.6666666666667, "grad_norm": 3.898359864251688e-06, "learning_rate": 5.965643458209755e-06, "loss": 0.0, "num_input_tokens_seen": 8685536, "step": 31020 }, { "epoch": 344.72222222222223, "grad_norm": 2.3202321131066128e-07, "learning_rate": 5.95928013192695e-06, "loss": 0.0, "num_input_tokens_seen": 8686912, "step": 31025 }, { "epoch": 344.77777777777777, "grad_norm": 4.248748882673681e-05, "learning_rate": 5.952919741962423e-06, "loss": 0.0, "num_input_tokens_seen": 8688336, "step": 31030 }, { "epoch": 344.8333333333333, "grad_norm": 2.678144937817706e-06, "learning_rate": 5.946562289297042e-06, "loss": 0.0, "num_input_tokens_seen": 8689760, "step": 31035 }, { "epoch": 344.8888888888889, "grad_norm": 3.640920340330922e-06, "learning_rate": 5.9402077749111855e-06, "loss": 0.0, "num_input_tokens_seen": 8691152, "step": 31040 }, { "epoch": 344.94444444444446, "grad_norm": 3.640325303422287e-06, "learning_rate": 5.933856199784821e-06, "loss": 0.0, "num_input_tokens_seen": 8692560, "step": 31045 }, { "epoch": 345.0, "grad_norm": 3.1955671602190705e-06, "learning_rate": 5.927507564897419e-06, "loss": 0.0, "num_input_tokens_seen": 8693984, "step": 31050 }, { "epoch": 345.05555555555554, "grad_norm": 3.5308071346662473e-06, "learning_rate": 5.9211618712280395e-06, "loss": 0.0, "num_input_tokens_seen": 8695376, "step": 31055 }, { "epoch": 345.1111111111111, "grad_norm": 1.3792542631563265e-05, "learning_rate": 5.914819119755255e-06, "loss": 0.0, "num_input_tokens_seen": 8696784, "step": 31060 }, { "epoch": 345.1666666666667, "grad_norm": 7.228673837289534e-08, "learning_rate": 5.908479311457205e-06, "loss": 0.0, "num_input_tokens_seen": 8698176, "step": 31065 }, { "epoch": 345.22222222222223, "grad_norm": 1.1239308150834404e-06, "learning_rate": 5.902142447311559e-06, "loss": 0.0, "num_input_tokens_seen": 8699616, "step": 31070 }, { "epoch": 345.27777777777777, "grad_norm": 4.910645657218993e-05, "learning_rate": 5.895808528295546e-06, "loss": 0.0, "num_input_tokens_seen": 8701040, "step": 31075 }, { "epoch": 345.3333333333333, "grad_norm": 2.8615691007871646e-06, "learning_rate": 5.889477555385941e-06, "loss": 0.0, "num_input_tokens_seen": 8702448, "step": 31080 }, { "epoch": 345.3888888888889, "grad_norm": 3.622558779170504e-06, "learning_rate": 5.883149529559051e-06, "loss": 0.0, "num_input_tokens_seen": 8703840, "step": 31085 }, { "epoch": 345.44444444444446, "grad_norm": 7.1973149715631735e-06, "learning_rate": 5.876824451790738e-06, "loss": 0.0, "num_input_tokens_seen": 8705200, "step": 31090 }, { "epoch": 345.5, "grad_norm": 7.865612587920623e-07, "learning_rate": 5.87050232305642e-06, "loss": 0.0, "num_input_tokens_seen": 8706592, "step": 31095 }, { "epoch": 345.55555555555554, "grad_norm": 1.449767569283722e-05, "learning_rate": 5.864183144331034e-06, "loss": 0.0, "num_input_tokens_seen": 8707984, "step": 31100 }, { "epoch": 345.6111111111111, "grad_norm": 8.078905921138357e-06, "learning_rate": 5.857866916589089e-06, "loss": 0.0, "num_input_tokens_seen": 8709360, "step": 31105 }, { "epoch": 345.6666666666667, "grad_norm": 6.174064424158132e-07, "learning_rate": 5.8515536408046216e-06, "loss": 0.0, "num_input_tokens_seen": 8710816, "step": 31110 }, { "epoch": 345.72222222222223, "grad_norm": 2.4571208996349014e-06, "learning_rate": 5.845243317951208e-06, "loss": 0.0, "num_input_tokens_seen": 8712224, "step": 31115 }, { "epoch": 345.77777777777777, "grad_norm": 1.0623842172208242e-05, "learning_rate": 5.838935949001997e-06, "loss": 0.0, "num_input_tokens_seen": 8713632, "step": 31120 }, { "epoch": 345.8333333333333, "grad_norm": 6.802566190344805e-07, "learning_rate": 5.8326315349296476e-06, "loss": 0.0, "num_input_tokens_seen": 8714992, "step": 31125 }, { "epoch": 345.8888888888889, "grad_norm": 3.732203822437441e-06, "learning_rate": 5.826330076706396e-06, "loss": 0.0, "num_input_tokens_seen": 8716400, "step": 31130 }, { "epoch": 345.94444444444446, "grad_norm": 1.683300411059463e-06, "learning_rate": 5.820031575303988e-06, "loss": 0.0, "num_input_tokens_seen": 8717792, "step": 31135 }, { "epoch": 346.0, "grad_norm": 3.0723792860953836e-06, "learning_rate": 5.813736031693745e-06, "loss": 0.0, "num_input_tokens_seen": 8719184, "step": 31140 }, { "epoch": 346.05555555555554, "grad_norm": 4.014950434338971e-07, "learning_rate": 5.807443446846522e-06, "loss": 0.0, "num_input_tokens_seen": 8720608, "step": 31145 }, { "epoch": 346.1111111111111, "grad_norm": 3.1428831448465644e-07, "learning_rate": 5.801153821732699e-06, "loss": 0.0, "num_input_tokens_seen": 8722032, "step": 31150 }, { "epoch": 346.1666666666667, "grad_norm": 3.7805427837156458e-06, "learning_rate": 5.794867157322229e-06, "loss": 0.0, "num_input_tokens_seen": 8723472, "step": 31155 }, { "epoch": 346.22222222222223, "grad_norm": 1.2652940313273575e-06, "learning_rate": 5.788583454584593e-06, "loss": 0.0, "num_input_tokens_seen": 8724880, "step": 31160 }, { "epoch": 346.27777777777777, "grad_norm": 7.811138402757933e-07, "learning_rate": 5.7823027144888075e-06, "loss": 0.0, "num_input_tokens_seen": 8726272, "step": 31165 }, { "epoch": 346.3333333333333, "grad_norm": 2.3787545160303125e-06, "learning_rate": 5.776024938003455e-06, "loss": 0.0, "num_input_tokens_seen": 8727712, "step": 31170 }, { "epoch": 346.3888888888889, "grad_norm": 2.119109922205098e-06, "learning_rate": 5.7697501260966345e-06, "loss": 0.0, "num_input_tokens_seen": 8729088, "step": 31175 }, { "epoch": 346.44444444444446, "grad_norm": 1.8554933376435656e-06, "learning_rate": 5.7634782797360145e-06, "loss": 0.0, "num_input_tokens_seen": 8730480, "step": 31180 }, { "epoch": 346.5, "grad_norm": 4.81897814097465e-06, "learning_rate": 5.757209399888777e-06, "loss": 0.0, "num_input_tokens_seen": 8731888, "step": 31185 }, { "epoch": 346.55555555555554, "grad_norm": 8.386623449041508e-07, "learning_rate": 5.750943487521679e-06, "loss": 0.0, "num_input_tokens_seen": 8733264, "step": 31190 }, { "epoch": 346.6111111111111, "grad_norm": 4.806767719855998e-06, "learning_rate": 5.744680543600986e-06, "loss": 0.0, "num_input_tokens_seen": 8734624, "step": 31195 }, { "epoch": 346.6666666666667, "grad_norm": 1.647078988753492e-06, "learning_rate": 5.738420569092537e-06, "loss": 0.0, "num_input_tokens_seen": 8736032, "step": 31200 }, { "epoch": 346.6666666666667, "eval_loss": 1.1038787364959717, "eval_runtime": 1.3915, "eval_samples_per_second": 28.747, "eval_steps_per_second": 14.373, "num_input_tokens_seen": 8736032, "step": 31200 }, { "epoch": 346.72222222222223, "grad_norm": 6.903576377226273e-07, "learning_rate": 5.732163564961684e-06, "loss": 0.0, "num_input_tokens_seen": 8737424, "step": 31205 }, { "epoch": 346.77777777777777, "grad_norm": 3.6229080251359846e-06, "learning_rate": 5.725909532173354e-06, "loss": 0.0, "num_input_tokens_seen": 8738784, "step": 31210 }, { "epoch": 346.8333333333333, "grad_norm": 1.544635961181484e-05, "learning_rate": 5.719658471691977e-06, "loss": 0.0, "num_input_tokens_seen": 8740208, "step": 31215 }, { "epoch": 346.8888888888889, "grad_norm": 8.426261047134176e-06, "learning_rate": 5.71341038448156e-06, "loss": 0.0, "num_input_tokens_seen": 8741584, "step": 31220 }, { "epoch": 346.94444444444446, "grad_norm": 5.065328878117725e-06, "learning_rate": 5.707165271505635e-06, "loss": 0.0, "num_input_tokens_seen": 8742976, "step": 31225 }, { "epoch": 347.0, "grad_norm": 3.4333057556068525e-05, "learning_rate": 5.700923133727271e-06, "loss": 0.0, "num_input_tokens_seen": 8744368, "step": 31230 }, { "epoch": 347.05555555555554, "grad_norm": 3.3935015153474524e-07, "learning_rate": 5.694683972109083e-06, "loss": 0.0, "num_input_tokens_seen": 8745760, "step": 31235 }, { "epoch": 347.1111111111111, "grad_norm": 1.303760086557304e-06, "learning_rate": 5.688447787613241e-06, "loss": 0.0, "num_input_tokens_seen": 8747136, "step": 31240 }, { "epoch": 347.1666666666667, "grad_norm": 4.3354373246984323e-07, "learning_rate": 5.6822145812014285e-06, "loss": 0.0, "num_input_tokens_seen": 8748528, "step": 31245 }, { "epoch": 347.22222222222223, "grad_norm": 9.184647751681041e-06, "learning_rate": 5.675984353834896e-06, "loss": 0.0, "num_input_tokens_seen": 8749952, "step": 31250 }, { "epoch": 347.27777777777777, "grad_norm": 5.319377578416606e-06, "learning_rate": 5.66975710647441e-06, "loss": 0.0, "num_input_tokens_seen": 8751392, "step": 31255 }, { "epoch": 347.3333333333333, "grad_norm": 9.197920917358715e-06, "learning_rate": 5.663532840080304e-06, "loss": 0.0, "num_input_tokens_seen": 8752816, "step": 31260 }, { "epoch": 347.3888888888889, "grad_norm": 6.665622549917316e-06, "learning_rate": 5.6573115556124325e-06, "loss": 0.0, "num_input_tokens_seen": 8754240, "step": 31265 }, { "epoch": 347.44444444444446, "grad_norm": 2.5563187591615133e-05, "learning_rate": 5.651093254030185e-06, "loss": 0.0, "num_input_tokens_seen": 8755616, "step": 31270 }, { "epoch": 347.5, "grad_norm": 1.7604284948902205e-05, "learning_rate": 5.644877936292514e-06, "loss": 0.0, "num_input_tokens_seen": 8756944, "step": 31275 }, { "epoch": 347.55555555555554, "grad_norm": 1.749425064190291e-05, "learning_rate": 5.638665603357901e-06, "loss": 0.0, "num_input_tokens_seen": 8758352, "step": 31280 }, { "epoch": 347.6111111111111, "grad_norm": 4.991935929865576e-05, "learning_rate": 5.632456256184357e-06, "loss": 0.0, "num_input_tokens_seen": 8759808, "step": 31285 }, { "epoch": 347.6666666666667, "grad_norm": 2.686642801563721e-06, "learning_rate": 5.626249895729452e-06, "loss": 0.0, "num_input_tokens_seen": 8761216, "step": 31290 }, { "epoch": 347.72222222222223, "grad_norm": 6.978591500228504e-06, "learning_rate": 5.620046522950273e-06, "loss": 0.0, "num_input_tokens_seen": 8762624, "step": 31295 }, { "epoch": 347.77777777777777, "grad_norm": 2.4019991542445496e-05, "learning_rate": 5.613846138803464e-06, "loss": 0.0, "num_input_tokens_seen": 8764032, "step": 31300 }, { "epoch": 347.8333333333333, "grad_norm": 1.2024963780277176e-06, "learning_rate": 5.607648744245206e-06, "loss": 0.0, "num_input_tokens_seen": 8765440, "step": 31305 }, { "epoch": 347.8888888888889, "grad_norm": 5.270588303574186e-07, "learning_rate": 5.601454340231207e-06, "loss": 0.0, "num_input_tokens_seen": 8766816, "step": 31310 }, { "epoch": 347.94444444444446, "grad_norm": 6.80690709486953e-06, "learning_rate": 5.595262927716724e-06, "loss": 0.0, "num_input_tokens_seen": 8768176, "step": 31315 }, { "epoch": 348.0, "grad_norm": 2.491754571565252e-07, "learning_rate": 5.589074507656561e-06, "loss": 0.0, "num_input_tokens_seen": 8769584, "step": 31320 }, { "epoch": 348.05555555555554, "grad_norm": 1.5198072560451692e-06, "learning_rate": 5.582889081005044e-06, "loss": 0.0, "num_input_tokens_seen": 8770928, "step": 31325 }, { "epoch": 348.1111111111111, "grad_norm": 3.685765932459617e-06, "learning_rate": 5.5767066487160316e-06, "loss": 0.0, "num_input_tokens_seen": 8772336, "step": 31330 }, { "epoch": 348.1666666666667, "grad_norm": 2.414821267393563e-07, "learning_rate": 5.570527211742949e-06, "loss": 0.0, "num_input_tokens_seen": 8773728, "step": 31335 }, { "epoch": 348.22222222222223, "grad_norm": 1.68197277616855e-06, "learning_rate": 5.564350771038731e-06, "loss": 0.0, "num_input_tokens_seen": 8775168, "step": 31340 }, { "epoch": 348.27777777777777, "grad_norm": 5.140058192409924e-07, "learning_rate": 5.558177327555875e-06, "loss": 0.0, "num_input_tokens_seen": 8776592, "step": 31345 }, { "epoch": 348.3333333333333, "grad_norm": 1.1048962278437102e-06, "learning_rate": 5.552006882246388e-06, "loss": 0.0, "num_input_tokens_seen": 8777952, "step": 31350 }, { "epoch": 348.3888888888889, "grad_norm": 4.936164259561338e-07, "learning_rate": 5.545839436061839e-06, "loss": 0.0, "num_input_tokens_seen": 8779360, "step": 31355 }, { "epoch": 348.44444444444446, "grad_norm": 3.3804251870606095e-05, "learning_rate": 5.539674989953331e-06, "loss": 0.0, "num_input_tokens_seen": 8780768, "step": 31360 }, { "epoch": 348.5, "grad_norm": 6.182862125569955e-05, "learning_rate": 5.533513544871488e-06, "loss": 0.0, "num_input_tokens_seen": 8782160, "step": 31365 }, { "epoch": 348.55555555555554, "grad_norm": 5.829496672049572e-07, "learning_rate": 5.527355101766493e-06, "loss": 0.0, "num_input_tokens_seen": 8783584, "step": 31370 }, { "epoch": 348.6111111111111, "grad_norm": 2.3731588498776546e-06, "learning_rate": 5.521199661588044e-06, "loss": 0.0, "num_input_tokens_seen": 8784960, "step": 31375 }, { "epoch": 348.6666666666667, "grad_norm": 1.5610388572895317e-06, "learning_rate": 5.5150472252853944e-06, "loss": 0.0, "num_input_tokens_seen": 8786384, "step": 31380 }, { "epoch": 348.72222222222223, "grad_norm": 2.058249265246559e-06, "learning_rate": 5.50889779380733e-06, "loss": 0.0, "num_input_tokens_seen": 8787776, "step": 31385 }, { "epoch": 348.77777777777777, "grad_norm": 4.3001200538128614e-06, "learning_rate": 5.5027513681021605e-06, "loss": 0.0, "num_input_tokens_seen": 8789120, "step": 31390 }, { "epoch": 348.8333333333333, "grad_norm": 7.978057965374319e-07, "learning_rate": 5.4966079491177545e-06, "loss": 0.0, "num_input_tokens_seen": 8790512, "step": 31395 }, { "epoch": 348.8888888888889, "grad_norm": 3.035108193216729e-06, "learning_rate": 5.490467537801491e-06, "loss": 0.0, "num_input_tokens_seen": 8791888, "step": 31400 }, { "epoch": 348.8888888888889, "eval_loss": 1.0880801677703857, "eval_runtime": 1.3869, "eval_samples_per_second": 28.841, "eval_steps_per_second": 14.421, "num_input_tokens_seen": 8791888, "step": 31400 }, { "epoch": 348.94444444444446, "grad_norm": 1.2058899301337078e-06, "learning_rate": 5.484330135100313e-06, "loss": 0.0, "num_input_tokens_seen": 8793312, "step": 31405 }, { "epoch": 349.0, "grad_norm": 6.979232239245903e-06, "learning_rate": 5.4781957419606785e-06, "loss": 0.0, "num_input_tokens_seen": 8794656, "step": 31410 }, { "epoch": 349.05555555555554, "grad_norm": 5.739275366067886e-06, "learning_rate": 5.472064359328577e-06, "loss": 0.0, "num_input_tokens_seen": 8796016, "step": 31415 }, { "epoch": 349.1111111111111, "grad_norm": 5.556552196139819e-07, "learning_rate": 5.4659359881495565e-06, "loss": 0.0, "num_input_tokens_seen": 8797424, "step": 31420 }, { "epoch": 349.1666666666667, "grad_norm": 3.832469701592345e-06, "learning_rate": 5.4598106293686916e-06, "loss": 0.0, "num_input_tokens_seen": 8798848, "step": 31425 }, { "epoch": 349.22222222222223, "grad_norm": 1.7473470848017314e-07, "learning_rate": 5.45368828393058e-06, "loss": 0.0, "num_input_tokens_seen": 8800256, "step": 31430 }, { "epoch": 349.27777777777777, "grad_norm": 5.236829565546941e-07, "learning_rate": 5.44756895277937e-06, "loss": 0.0, "num_input_tokens_seen": 8801632, "step": 31435 }, { "epoch": 349.3333333333333, "grad_norm": 5.72273347643204e-06, "learning_rate": 5.441452636858746e-06, "loss": 0.0, "num_input_tokens_seen": 8803040, "step": 31440 }, { "epoch": 349.3888888888889, "grad_norm": 5.365614015317988e-06, "learning_rate": 5.435339337111905e-06, "loss": 0.0, "num_input_tokens_seen": 8804432, "step": 31445 }, { "epoch": 349.44444444444446, "grad_norm": 6.025222774042049e-06, "learning_rate": 5.42922905448161e-06, "loss": 0.0, "num_input_tokens_seen": 8805824, "step": 31450 }, { "epoch": 349.5, "grad_norm": 1.3701091120310593e-05, "learning_rate": 5.423121789910129e-06, "loss": 0.0, "num_input_tokens_seen": 8807216, "step": 31455 }, { "epoch": 349.55555555555554, "grad_norm": 1.9119211174256634e-06, "learning_rate": 5.417017544339287e-06, "loss": 0.0, "num_input_tokens_seen": 8808640, "step": 31460 }, { "epoch": 349.6111111111111, "grad_norm": 2.110026343871141e-06, "learning_rate": 5.410916318710443e-06, "loss": 0.0, "num_input_tokens_seen": 8810048, "step": 31465 }, { "epoch": 349.6666666666667, "grad_norm": 7.183805337263038e-07, "learning_rate": 5.404818113964466e-06, "loss": 0.0, "num_input_tokens_seen": 8811472, "step": 31470 }, { "epoch": 349.72222222222223, "grad_norm": 3.908486974069092e-07, "learning_rate": 5.398722931041792e-06, "loss": 0.0, "num_input_tokens_seen": 8812864, "step": 31475 }, { "epoch": 349.77777777777777, "grad_norm": 9.163908316622837e-07, "learning_rate": 5.392630770882367e-06, "loss": 0.0, "num_input_tokens_seen": 8814272, "step": 31480 }, { "epoch": 349.8333333333333, "grad_norm": 1.0203924603047199e-06, "learning_rate": 5.3865416344256705e-06, "loss": 0.0, "num_input_tokens_seen": 8815680, "step": 31485 }, { "epoch": 349.8888888888889, "grad_norm": 1.1879192243213765e-06, "learning_rate": 5.380455522610742e-06, "loss": 0.0, "num_input_tokens_seen": 8817104, "step": 31490 }, { "epoch": 349.94444444444446, "grad_norm": 9.417809451406356e-06, "learning_rate": 5.374372436376116e-06, "loss": 0.0, "num_input_tokens_seen": 8818496, "step": 31495 }, { "epoch": 350.0, "grad_norm": 4.957310011377558e-07, "learning_rate": 5.368292376659895e-06, "loss": 0.0, "num_input_tokens_seen": 8819888, "step": 31500 }, { "epoch": 350.05555555555554, "grad_norm": 7.410883426928194e-07, "learning_rate": 5.362215344399701e-06, "loss": 0.0, "num_input_tokens_seen": 8821264, "step": 31505 }, { "epoch": 350.1111111111111, "grad_norm": 1.2051116300426656e-06, "learning_rate": 5.356141340532678e-06, "loss": 0.0, "num_input_tokens_seen": 8822640, "step": 31510 }, { "epoch": 350.1666666666667, "grad_norm": 7.201267635537079e-07, "learning_rate": 5.350070365995522e-06, "loss": 0.0, "num_input_tokens_seen": 8824048, "step": 31515 }, { "epoch": 350.22222222222223, "grad_norm": 8.089207881312177e-07, "learning_rate": 5.344002421724459e-06, "loss": 0.0, "num_input_tokens_seen": 8825456, "step": 31520 }, { "epoch": 350.27777777777777, "grad_norm": 4.006304607173661e-06, "learning_rate": 5.337937508655228e-06, "loss": 0.0, "num_input_tokens_seen": 8826864, "step": 31525 }, { "epoch": 350.3333333333333, "grad_norm": 4.81905999549781e-06, "learning_rate": 5.331875627723126e-06, "loss": 0.0, "num_input_tokens_seen": 8828256, "step": 31530 }, { "epoch": 350.3888888888889, "grad_norm": 4.4426328713598195e-06, "learning_rate": 5.325816779862963e-06, "loss": 0.0, "num_input_tokens_seen": 8829648, "step": 31535 }, { "epoch": 350.44444444444446, "grad_norm": 1.0502428267500363e-05, "learning_rate": 5.319760966009102e-06, "loss": 0.0, "num_input_tokens_seen": 8831024, "step": 31540 }, { "epoch": 350.5, "grad_norm": 1.502713899981245e-07, "learning_rate": 5.3137081870954096e-06, "loss": 0.0, "num_input_tokens_seen": 8832448, "step": 31545 }, { "epoch": 350.55555555555554, "grad_norm": 4.964957042830065e-06, "learning_rate": 5.307658444055313e-06, "loss": 0.0, "num_input_tokens_seen": 8833856, "step": 31550 }, { "epoch": 350.6111111111111, "grad_norm": 7.603070116601884e-06, "learning_rate": 5.301611737821749e-06, "loss": 0.0, "num_input_tokens_seen": 8835216, "step": 31555 }, { "epoch": 350.6666666666667, "grad_norm": 2.2825293854111806e-05, "learning_rate": 5.295568069327206e-06, "loss": 0.0, "num_input_tokens_seen": 8836592, "step": 31560 }, { "epoch": 350.72222222222223, "grad_norm": 3.930108221084083e-07, "learning_rate": 5.289527439503683e-06, "loss": 0.0, "num_input_tokens_seen": 8837984, "step": 31565 }, { "epoch": 350.77777777777777, "grad_norm": 9.638378060117248e-07, "learning_rate": 5.28348984928273e-06, "loss": 0.0, "num_input_tokens_seen": 8839360, "step": 31570 }, { "epoch": 350.8333333333333, "grad_norm": 2.6321165933040902e-05, "learning_rate": 5.27745529959541e-06, "loss": 0.0, "num_input_tokens_seen": 8840752, "step": 31575 }, { "epoch": 350.8888888888889, "grad_norm": 5.994047569402028e-06, "learning_rate": 5.271423791372335e-06, "loss": 0.0, "num_input_tokens_seen": 8842160, "step": 31580 }, { "epoch": 350.94444444444446, "grad_norm": 4.025596808787668e-06, "learning_rate": 5.26539532554364e-06, "loss": 0.0, "num_input_tokens_seen": 8843568, "step": 31585 }, { "epoch": 351.0, "grad_norm": 3.1662302717450075e-06, "learning_rate": 5.25936990303898e-06, "loss": 0.0, "num_input_tokens_seen": 8844960, "step": 31590 }, { "epoch": 351.05555555555554, "grad_norm": 5.19160403200658e-06, "learning_rate": 5.253347524787555e-06, "loss": 0.0, "num_input_tokens_seen": 8846320, "step": 31595 }, { "epoch": 351.1111111111111, "grad_norm": 5.028741156820615e-07, "learning_rate": 5.2473281917181035e-06, "loss": 0.0, "num_input_tokens_seen": 8847728, "step": 31600 }, { "epoch": 351.1111111111111, "eval_loss": 1.1246286630630493, "eval_runtime": 1.3907, "eval_samples_per_second": 28.762, "eval_steps_per_second": 14.381, "num_input_tokens_seen": 8847728, "step": 31600 }, { "epoch": 351.1666666666667, "grad_norm": 0.00011287104280199856, "learning_rate": 5.241311904758864e-06, "loss": 0.0, "num_input_tokens_seen": 8849184, "step": 31605 }, { "epoch": 351.22222222222223, "grad_norm": 6.613685400225222e-05, "learning_rate": 5.23529866483764e-06, "loss": 0.0, "num_input_tokens_seen": 8850608, "step": 31610 }, { "epoch": 351.27777777777777, "grad_norm": 9.945505553332623e-06, "learning_rate": 5.229288472881732e-06, "loss": 0.0, "num_input_tokens_seen": 8852000, "step": 31615 }, { "epoch": 351.3333333333333, "grad_norm": 1.8588081729831174e-05, "learning_rate": 5.2232813298180025e-06, "loss": 0.0, "num_input_tokens_seen": 8853408, "step": 31620 }, { "epoch": 351.3888888888889, "grad_norm": 5.436573928818689e-07, "learning_rate": 5.217277236572824e-06, "loss": 0.0, "num_input_tokens_seen": 8854816, "step": 31625 }, { "epoch": 351.44444444444446, "grad_norm": 2.131638865421337e-07, "learning_rate": 5.211276194072093e-06, "loss": 0.0, "num_input_tokens_seen": 8856208, "step": 31630 }, { "epoch": 351.5, "grad_norm": 2.7581195354287047e-06, "learning_rate": 5.205278203241254e-06, "loss": 0.0, "num_input_tokens_seen": 8857632, "step": 31635 }, { "epoch": 351.55555555555554, "grad_norm": 1.9118087948299944e-06, "learning_rate": 5.199283265005278e-06, "loss": 0.0, "num_input_tokens_seen": 8859008, "step": 31640 }, { "epoch": 351.6111111111111, "grad_norm": 2.047318503173301e-06, "learning_rate": 5.193291380288648e-06, "loss": 0.0, "num_input_tokens_seen": 8860384, "step": 31645 }, { "epoch": 351.6666666666667, "grad_norm": 1.1186684787389822e-06, "learning_rate": 5.1873025500153995e-06, "loss": 0.0, "num_input_tokens_seen": 8861792, "step": 31650 }, { "epoch": 351.72222222222223, "grad_norm": 2.1874226149520837e-05, "learning_rate": 5.181316775109071e-06, "loss": 0.0, "num_input_tokens_seen": 8863184, "step": 31655 }, { "epoch": 351.77777777777777, "grad_norm": 1.7389556887792423e-05, "learning_rate": 5.1753340564927564e-06, "loss": 0.0, "num_input_tokens_seen": 8864608, "step": 31660 }, { "epoch": 351.8333333333333, "grad_norm": 1.0527311360419844e-06, "learning_rate": 5.169354395089068e-06, "loss": 0.0, "num_input_tokens_seen": 8865984, "step": 31665 }, { "epoch": 351.8888888888889, "grad_norm": 1.0708798754421878e-06, "learning_rate": 5.1633777918201346e-06, "loss": 0.0, "num_input_tokens_seen": 8867424, "step": 31670 }, { "epoch": 351.94444444444446, "grad_norm": 8.971081228992261e-07, "learning_rate": 5.157404247607625e-06, "loss": 0.0, "num_input_tokens_seen": 8868832, "step": 31675 }, { "epoch": 352.0, "grad_norm": 2.8554225082189078e-06, "learning_rate": 5.1514337633727454e-06, "loss": 0.0, "num_input_tokens_seen": 8870240, "step": 31680 }, { "epoch": 352.05555555555554, "grad_norm": 5.785792041024251e-07, "learning_rate": 5.145466340036206e-06, "loss": 0.0, "num_input_tokens_seen": 8871680, "step": 31685 }, { "epoch": 352.1111111111111, "grad_norm": 1.0189606655330863e-05, "learning_rate": 5.139501978518274e-06, "loss": 0.0, "num_input_tokens_seen": 8873056, "step": 31690 }, { "epoch": 352.1666666666667, "grad_norm": 3.216850927856285e-06, "learning_rate": 5.133540679738716e-06, "loss": 0.0, "num_input_tokens_seen": 8874496, "step": 31695 }, { "epoch": 352.22222222222223, "grad_norm": 2.5516044843243435e-05, "learning_rate": 5.127582444616838e-06, "loss": 0.0, "num_input_tokens_seen": 8875904, "step": 31700 }, { "epoch": 352.27777777777777, "grad_norm": 1.6748426787671633e-05, "learning_rate": 5.121627274071486e-06, "loss": 0.0, "num_input_tokens_seen": 8877296, "step": 31705 }, { "epoch": 352.3333333333333, "grad_norm": 2.7919322747038677e-05, "learning_rate": 5.115675169021009e-06, "loss": 0.0, "num_input_tokens_seen": 8878720, "step": 31710 }, { "epoch": 352.3888888888889, "grad_norm": 1.5322187891797512e-06, "learning_rate": 5.1097261303832994e-06, "loss": 0.0, "num_input_tokens_seen": 8880080, "step": 31715 }, { "epoch": 352.44444444444446, "grad_norm": 1.3716243074668455e-06, "learning_rate": 5.103780159075788e-06, "loss": 0.0, "num_input_tokens_seen": 8881488, "step": 31720 }, { "epoch": 352.5, "grad_norm": 2.8741155233547033e-07, "learning_rate": 5.0978372560154e-06, "loss": 0.0, "num_input_tokens_seen": 8882912, "step": 31725 }, { "epoch": 352.55555555555554, "grad_norm": 5.92043591041147e-07, "learning_rate": 5.091897422118619e-06, "loss": 0.0, "num_input_tokens_seen": 8884320, "step": 31730 }, { "epoch": 352.6111111111111, "grad_norm": 1.2664193491218612e-05, "learning_rate": 5.0859606583014305e-06, "loss": 0.0, "num_input_tokens_seen": 8885744, "step": 31735 }, { "epoch": 352.6666666666667, "grad_norm": 3.918119091395056e-06, "learning_rate": 5.080026965479365e-06, "loss": 0.0, "num_input_tokens_seen": 8887104, "step": 31740 }, { "epoch": 352.72222222222223, "grad_norm": 4.266197720426135e-06, "learning_rate": 5.074096344567475e-06, "loss": 0.0, "num_input_tokens_seen": 8888512, "step": 31745 }, { "epoch": 352.77777777777777, "grad_norm": 9.479022992309183e-06, "learning_rate": 5.0681687964803294e-06, "loss": 0.0, "num_input_tokens_seen": 8889904, "step": 31750 }, { "epoch": 352.8333333333333, "grad_norm": 2.034364570135949e-06, "learning_rate": 5.06224432213204e-06, "loss": 0.0, "num_input_tokens_seen": 8891328, "step": 31755 }, { "epoch": 352.8888888888889, "grad_norm": 3.04892091662623e-06, "learning_rate": 5.056322922436224e-06, "loss": 0.0, "num_input_tokens_seen": 8892720, "step": 31760 }, { "epoch": 352.94444444444446, "grad_norm": 5.191644163460296e-07, "learning_rate": 5.0504045983060465e-06, "loss": 0.0, "num_input_tokens_seen": 8894112, "step": 31765 }, { "epoch": 353.0, "grad_norm": 3.9303326957451645e-06, "learning_rate": 5.044489350654183e-06, "loss": 0.0, "num_input_tokens_seen": 8895488, "step": 31770 }, { "epoch": 353.05555555555554, "grad_norm": 5.541556333810149e-07, "learning_rate": 5.038577180392831e-06, "loss": 0.0, "num_input_tokens_seen": 8896928, "step": 31775 }, { "epoch": 353.1111111111111, "grad_norm": 1.715347207209561e-05, "learning_rate": 5.032668088433729e-06, "loss": 0.0, "num_input_tokens_seen": 8898320, "step": 31780 }, { "epoch": 353.1666666666667, "grad_norm": 1.7175502762256656e-06, "learning_rate": 5.02676207568814e-06, "loss": 0.0, "num_input_tokens_seen": 8899744, "step": 31785 }, { "epoch": 353.22222222222223, "grad_norm": 1.0079010053232196e-06, "learning_rate": 5.02085914306683e-06, "loss": 0.0, "num_input_tokens_seen": 8901168, "step": 31790 }, { "epoch": 353.27777777777777, "grad_norm": 1.7393338112015044e-06, "learning_rate": 5.014959291480123e-06, "loss": 0.0, "num_input_tokens_seen": 8902544, "step": 31795 }, { "epoch": 353.3333333333333, "grad_norm": 1.0633707461238373e-05, "learning_rate": 5.009062521837835e-06, "loss": 0.0, "num_input_tokens_seen": 8903952, "step": 31800 }, { "epoch": 353.3333333333333, "eval_loss": 1.0899066925048828, "eval_runtime": 1.3918, "eval_samples_per_second": 28.74, "eval_steps_per_second": 14.37, "num_input_tokens_seen": 8903952, "step": 31800 }, { "epoch": 353.3888888888889, "grad_norm": 8.25456424990989e-07, "learning_rate": 5.003168835049324e-06, "loss": 0.0, "num_input_tokens_seen": 8905344, "step": 31805 }, { "epoch": 353.44444444444446, "grad_norm": 2.1506271878024563e-05, "learning_rate": 4.997278232023483e-06, "loss": 0.0, "num_input_tokens_seen": 8906768, "step": 31810 }, { "epoch": 353.5, "grad_norm": 5.092731498734793e-07, "learning_rate": 4.9913907136687036e-06, "loss": 0.0, "num_input_tokens_seen": 8908192, "step": 31815 }, { "epoch": 353.55555555555554, "grad_norm": 6.869384833407821e-06, "learning_rate": 4.985506280892918e-06, "loss": 0.0, "num_input_tokens_seen": 8909568, "step": 31820 }, { "epoch": 353.6111111111111, "grad_norm": 6.638226750510512e-06, "learning_rate": 4.979624934603589e-06, "loss": 0.0, "num_input_tokens_seen": 8910960, "step": 31825 }, { "epoch": 353.6666666666667, "grad_norm": 3.7411173252621666e-06, "learning_rate": 4.97374667570768e-06, "loss": 0.0, "num_input_tokens_seen": 8912352, "step": 31830 }, { "epoch": 353.72222222222223, "grad_norm": 8.8798078650143e-06, "learning_rate": 4.967871505111704e-06, "loss": 0.0, "num_input_tokens_seen": 8913728, "step": 31835 }, { "epoch": 353.77777777777777, "grad_norm": 7.790078484504193e-07, "learning_rate": 4.961999423721686e-06, "loss": 0.0, "num_input_tokens_seen": 8915136, "step": 31840 }, { "epoch": 353.8333333333333, "grad_norm": 1.7815750652516726e-06, "learning_rate": 4.956130432443159e-06, "loss": 0.0, "num_input_tokens_seen": 8916544, "step": 31845 }, { "epoch": 353.8888888888889, "grad_norm": 3.981515874329489e-06, "learning_rate": 4.950264532181215e-06, "loss": 0.0, "num_input_tokens_seen": 8917872, "step": 31850 }, { "epoch": 353.94444444444446, "grad_norm": 1.38435007102089e-06, "learning_rate": 4.944401723840433e-06, "loss": 0.0, "num_input_tokens_seen": 8919296, "step": 31855 }, { "epoch": 354.0, "grad_norm": 2.7259648049948737e-07, "learning_rate": 4.938542008324942e-06, "loss": 0.0, "num_input_tokens_seen": 8920688, "step": 31860 }, { "epoch": 354.05555555555554, "grad_norm": 1.547666579426732e-05, "learning_rate": 4.9326853865383855e-06, "loss": 0.0, "num_input_tokens_seen": 8922080, "step": 31865 }, { "epoch": 354.1111111111111, "grad_norm": 1.479838260820543e-06, "learning_rate": 4.926831859383918e-06, "loss": 0.0, "num_input_tokens_seen": 8923472, "step": 31870 }, { "epoch": 354.1666666666667, "grad_norm": 5.066752919447026e-07, "learning_rate": 4.92098142776424e-06, "loss": 0.0, "num_input_tokens_seen": 8924880, "step": 31875 }, { "epoch": 354.22222222222223, "grad_norm": 8.806857749732444e-07, "learning_rate": 4.91513409258155e-06, "loss": 0.0, "num_input_tokens_seen": 8926272, "step": 31880 }, { "epoch": 354.27777777777777, "grad_norm": 1.0452474725752836e-06, "learning_rate": 4.909289854737581e-06, "loss": 0.0, "num_input_tokens_seen": 8927664, "step": 31885 }, { "epoch": 354.3333333333333, "grad_norm": 4.442298177309567e-06, "learning_rate": 4.903448715133602e-06, "loss": 0.0, "num_input_tokens_seen": 8929056, "step": 31890 }, { "epoch": 354.3888888888889, "grad_norm": 2.549425516917836e-05, "learning_rate": 4.897610674670372e-06, "loss": 0.0, "num_input_tokens_seen": 8930496, "step": 31895 }, { "epoch": 354.44444444444446, "grad_norm": 9.801963187783258e-07, "learning_rate": 4.8917757342482e-06, "loss": 0.0, "num_input_tokens_seen": 8931888, "step": 31900 }, { "epoch": 354.5, "grad_norm": 1.5461480415979167e-06, "learning_rate": 4.885943894766909e-06, "loss": 0.0, "num_input_tokens_seen": 8933264, "step": 31905 }, { "epoch": 354.55555555555554, "grad_norm": 2.7164098355569877e-05, "learning_rate": 4.880115157125842e-06, "loss": 0.0, "num_input_tokens_seen": 8934640, "step": 31910 }, { "epoch": 354.6111111111111, "grad_norm": 2.604187784527312e-06, "learning_rate": 4.874289522223857e-06, "loss": 0.0, "num_input_tokens_seen": 8936080, "step": 31915 }, { "epoch": 354.6666666666667, "grad_norm": 4.865922619501362e-06, "learning_rate": 4.868466990959339e-06, "loss": 0.0, "num_input_tokens_seen": 8937520, "step": 31920 }, { "epoch": 354.72222222222223, "grad_norm": 3.8017230963305337e-06, "learning_rate": 4.8626475642301964e-06, "loss": 0.0, "num_input_tokens_seen": 8938896, "step": 31925 }, { "epoch": 354.77777777777777, "grad_norm": 1.1046932257841036e-07, "learning_rate": 4.856831242933871e-06, "loss": 0.0, "num_input_tokens_seen": 8940272, "step": 31930 }, { "epoch": 354.8333333333333, "grad_norm": 1.1272071560597396e-06, "learning_rate": 4.851018027967294e-06, "loss": 0.0, "num_input_tokens_seen": 8941680, "step": 31935 }, { "epoch": 354.8888888888889, "grad_norm": 9.843560292210896e-06, "learning_rate": 4.845207920226946e-06, "loss": 0.0, "num_input_tokens_seen": 8943056, "step": 31940 }, { "epoch": 354.94444444444446, "grad_norm": 2.540314380894415e-05, "learning_rate": 4.839400920608825e-06, "loss": 0.0, "num_input_tokens_seen": 8944496, "step": 31945 }, { "epoch": 355.0, "grad_norm": 3.1041563488543034e-06, "learning_rate": 4.83359703000843e-06, "loss": 0.0, "num_input_tokens_seen": 8945936, "step": 31950 }, { "epoch": 355.05555555555554, "grad_norm": 3.1101961894819397e-07, "learning_rate": 4.827796249320804e-06, "loss": 0.0, "num_input_tokens_seen": 8947344, "step": 31955 }, { "epoch": 355.1111111111111, "grad_norm": 1.0905891940637957e-06, "learning_rate": 4.82199857944049e-06, "loss": 0.0, "num_input_tokens_seen": 8948752, "step": 31960 }, { "epoch": 355.1666666666667, "grad_norm": 6.180840500746854e-06, "learning_rate": 4.8162040212615695e-06, "loss": 0.0, "num_input_tokens_seen": 8950160, "step": 31965 }, { "epoch": 355.22222222222223, "grad_norm": 2.7891696845472325e-06, "learning_rate": 4.810412575677639e-06, "loss": 0.0, "num_input_tokens_seen": 8951568, "step": 31970 }, { "epoch": 355.27777777777777, "grad_norm": 1.031834813147725e-06, "learning_rate": 4.804624243581801e-06, "loss": 0.0, "num_input_tokens_seen": 8952944, "step": 31975 }, { "epoch": 355.3333333333333, "grad_norm": 2.8150116122560576e-05, "learning_rate": 4.798839025866703e-06, "loss": 0.0, "num_input_tokens_seen": 8954368, "step": 31980 }, { "epoch": 355.3888888888889, "grad_norm": 9.639147720008623e-07, "learning_rate": 4.793056923424491e-06, "loss": 0.0, "num_input_tokens_seen": 8955776, "step": 31985 }, { "epoch": 355.44444444444446, "grad_norm": 1.0524355502639082e-06, "learning_rate": 4.78727793714683e-06, "loss": 0.0, "num_input_tokens_seen": 8957120, "step": 31990 }, { "epoch": 355.5, "grad_norm": 6.928211746526358e-07, "learning_rate": 4.7815020679249285e-06, "loss": 0.0, "num_input_tokens_seen": 8958512, "step": 31995 }, { "epoch": 355.55555555555554, "grad_norm": 3.3459450605732854e-06, "learning_rate": 4.775729316649483e-06, "loss": 0.0, "num_input_tokens_seen": 8959920, "step": 32000 }, { "epoch": 355.55555555555554, "eval_loss": 1.1147997379302979, "eval_runtime": 1.3927, "eval_samples_per_second": 28.721, "eval_steps_per_second": 14.361, "num_input_tokens_seen": 8959920, "step": 32000 }, { "epoch": 355.6111111111111, "grad_norm": 1.993343175854534e-06, "learning_rate": 4.769959684210728e-06, "loss": 0.0, "num_input_tokens_seen": 8961296, "step": 32005 }, { "epoch": 355.6666666666667, "grad_norm": 3.1000756735011237e-06, "learning_rate": 4.764193171498426e-06, "loss": 0.0, "num_input_tokens_seen": 8962688, "step": 32010 }, { "epoch": 355.72222222222223, "grad_norm": 1.0849474847418605e-06, "learning_rate": 4.75842977940183e-06, "loss": 0.0, "num_input_tokens_seen": 8964112, "step": 32015 }, { "epoch": 355.77777777777777, "grad_norm": 3.6251594792702235e-06, "learning_rate": 4.752669508809729e-06, "loss": 0.0, "num_input_tokens_seen": 8965552, "step": 32020 }, { "epoch": 355.8333333333333, "grad_norm": 1.0212064580628066e-06, "learning_rate": 4.746912360610445e-06, "loss": 0.0, "num_input_tokens_seen": 8966928, "step": 32025 }, { "epoch": 355.8888888888889, "grad_norm": 3.156687853334006e-06, "learning_rate": 4.741158335691781e-06, "loss": 0.0, "num_input_tokens_seen": 8968304, "step": 32030 }, { "epoch": 355.94444444444446, "grad_norm": 1.1256489784727819e-07, "learning_rate": 4.7354074349410994e-06, "loss": 0.0, "num_input_tokens_seen": 8969728, "step": 32035 }, { "epoch": 356.0, "grad_norm": 3.1908462005958427e-06, "learning_rate": 4.729659659245245e-06, "loss": 0.0, "num_input_tokens_seen": 8971152, "step": 32040 }, { "epoch": 356.05555555555554, "grad_norm": 6.524013542730245e-07, "learning_rate": 4.723915009490601e-06, "loss": 0.0, "num_input_tokens_seen": 8972576, "step": 32045 }, { "epoch": 356.1111111111111, "grad_norm": 3.7975739815010456e-06, "learning_rate": 4.718173486563077e-06, "loss": 0.0, "num_input_tokens_seen": 8974016, "step": 32050 }, { "epoch": 356.1666666666667, "grad_norm": 7.390190717160294e-07, "learning_rate": 4.71243509134808e-06, "loss": 0.0, "num_input_tokens_seen": 8975408, "step": 32055 }, { "epoch": 356.22222222222223, "grad_norm": 9.833763215283398e-06, "learning_rate": 4.706699824730532e-06, "loss": 0.0, "num_input_tokens_seen": 8976800, "step": 32060 }, { "epoch": 356.27777777777777, "grad_norm": 4.457081104192184e-06, "learning_rate": 4.700967687594901e-06, "loss": 0.0, "num_input_tokens_seen": 8978192, "step": 32065 }, { "epoch": 356.3333333333333, "grad_norm": 7.288505912583787e-06, "learning_rate": 4.69523868082514e-06, "loss": 0.0, "num_input_tokens_seen": 8979568, "step": 32070 }, { "epoch": 356.3888888888889, "grad_norm": 1.7181110933961463e-07, "learning_rate": 4.689512805304747e-06, "loss": 0.0, "num_input_tokens_seen": 8980976, "step": 32075 }, { "epoch": 356.44444444444446, "grad_norm": 4.260203354533587e-07, "learning_rate": 4.683790061916707e-06, "loss": 0.0, "num_input_tokens_seen": 8982400, "step": 32080 }, { "epoch": 356.5, "grad_norm": 9.453524398850277e-07, "learning_rate": 4.678070451543551e-06, "loss": 0.0, "num_input_tokens_seen": 8983792, "step": 32085 }, { "epoch": 356.55555555555554, "grad_norm": 4.5986513441675925e-07, "learning_rate": 4.6723539750673204e-06, "loss": 0.0, "num_input_tokens_seen": 8985216, "step": 32090 }, { "epoch": 356.6111111111111, "grad_norm": 2.124518914570217e-06, "learning_rate": 4.666640633369551e-06, "loss": 0.0, "num_input_tokens_seen": 8986624, "step": 32095 }, { "epoch": 356.6666666666667, "grad_norm": 8.347182301804423e-06, "learning_rate": 4.660930427331323e-06, "loss": 0.0, "num_input_tokens_seen": 8988000, "step": 32100 }, { "epoch": 356.72222222222223, "grad_norm": 9.840940720096114e-07, "learning_rate": 4.6552233578332244e-06, "loss": 0.0, "num_input_tokens_seen": 8989344, "step": 32105 }, { "epoch": 356.77777777777777, "grad_norm": 1.977640522454749e-06, "learning_rate": 4.649519425755347e-06, "loss": 0.0, "num_input_tokens_seen": 8990800, "step": 32110 }, { "epoch": 356.8333333333333, "grad_norm": 5.939009497524239e-07, "learning_rate": 4.64381863197732e-06, "loss": 0.0, "num_input_tokens_seen": 8992176, "step": 32115 }, { "epoch": 356.8888888888889, "grad_norm": 1.571095140207035e-06, "learning_rate": 4.638120977378269e-06, "loss": 0.0, "num_input_tokens_seen": 8993552, "step": 32120 }, { "epoch": 356.94444444444446, "grad_norm": 1.5756906577735208e-05, "learning_rate": 4.632426462836848e-06, "loss": 0.0, "num_input_tokens_seen": 8994976, "step": 32125 }, { "epoch": 357.0, "grad_norm": 4.771362000610679e-06, "learning_rate": 4.626735089231224e-06, "loss": 0.0, "num_input_tokens_seen": 8996384, "step": 32130 }, { "epoch": 357.05555555555554, "grad_norm": 6.343151994769869e-07, "learning_rate": 4.621046857439068e-06, "loss": 0.0, "num_input_tokens_seen": 8997824, "step": 32135 }, { "epoch": 357.1111111111111, "grad_norm": 1.0076452781504486e-05, "learning_rate": 4.615361768337587e-06, "loss": 0.0, "num_input_tokens_seen": 8999216, "step": 32140 }, { "epoch": 357.1666666666667, "grad_norm": 1.7096275541916839e-06, "learning_rate": 4.6096798228034946e-06, "loss": 0.0, "num_input_tokens_seen": 9000624, "step": 32145 }, { "epoch": 357.22222222222223, "grad_norm": 9.907245157592115e-07, "learning_rate": 4.604001021713008e-06, "loss": 0.0, "num_input_tokens_seen": 9002080, "step": 32150 }, { "epoch": 357.27777777777777, "grad_norm": 8.763299774727784e-06, "learning_rate": 4.598325365941883e-06, "loss": 0.0, "num_input_tokens_seen": 9003488, "step": 32155 }, { "epoch": 357.3333333333333, "grad_norm": 7.255846412590472e-07, "learning_rate": 4.5926528563653645e-06, "loss": 0.0, "num_input_tokens_seen": 9004896, "step": 32160 }, { "epoch": 357.3888888888889, "grad_norm": 2.0936007786076516e-05, "learning_rate": 4.5869834938582295e-06, "loss": 0.0, "num_input_tokens_seen": 9006256, "step": 32165 }, { "epoch": 357.44444444444446, "grad_norm": 4.883329438598594e-07, "learning_rate": 4.581317279294772e-06, "loss": 0.0, "num_input_tokens_seen": 9007680, "step": 32170 }, { "epoch": 357.5, "grad_norm": 4.669519512390252e-06, "learning_rate": 4.57565421354878e-06, "loss": 0.0, "num_input_tokens_seen": 9009120, "step": 32175 }, { "epoch": 357.55555555555554, "grad_norm": 1.097212589229457e-06, "learning_rate": 4.569994297493579e-06, "loss": 0.0, "num_input_tokens_seen": 9010528, "step": 32180 }, { "epoch": 357.6111111111111, "grad_norm": 3.006988663400989e-06, "learning_rate": 4.564337532002002e-06, "loss": 0.0, "num_input_tokens_seen": 9011920, "step": 32185 }, { "epoch": 357.6666666666667, "grad_norm": 5.35650451638503e-06, "learning_rate": 4.55868391794638e-06, "loss": 0.0, "num_input_tokens_seen": 9013328, "step": 32190 }, { "epoch": 357.72222222222223, "grad_norm": 4.124149654671783e-06, "learning_rate": 4.553033456198588e-06, "loss": 0.0, "num_input_tokens_seen": 9014704, "step": 32195 }, { "epoch": 357.77777777777777, "grad_norm": 1.4895260846969904e-06, "learning_rate": 4.54738614762999e-06, "loss": 0.0, "num_input_tokens_seen": 9016096, "step": 32200 }, { "epoch": 357.77777777777777, "eval_loss": 1.1080487966537476, "eval_runtime": 1.3979, "eval_samples_per_second": 28.615, "eval_steps_per_second": 14.308, "num_input_tokens_seen": 9016096, "step": 32200 }, { "epoch": 357.8333333333333, "grad_norm": 6.67490837713558e-07, "learning_rate": 4.541741993111465e-06, "loss": 0.0, "num_input_tokens_seen": 9017456, "step": 32205 }, { "epoch": 357.8888888888889, "grad_norm": 5.990189606563945e-07, "learning_rate": 4.536100993513423e-06, "loss": 0.0, "num_input_tokens_seen": 9018848, "step": 32210 }, { "epoch": 357.94444444444446, "grad_norm": 1.406745155918543e-07, "learning_rate": 4.530463149705768e-06, "loss": 0.0, "num_input_tokens_seen": 9020288, "step": 32215 }, { "epoch": 358.0, "grad_norm": 1.8826047210040997e-07, "learning_rate": 4.524828462557934e-06, "loss": 0.0, "num_input_tokens_seen": 9021664, "step": 32220 }, { "epoch": 358.05555555555554, "grad_norm": 4.1179864638252184e-05, "learning_rate": 4.5191969329388625e-06, "loss": 0.0, "num_input_tokens_seen": 9023088, "step": 32225 }, { "epoch": 358.1111111111111, "grad_norm": 1.1741559546862845e-06, "learning_rate": 4.5135685617169965e-06, "loss": 0.0, "num_input_tokens_seen": 9024528, "step": 32230 }, { "epoch": 358.1666666666667, "grad_norm": 1.1920807310161763e-06, "learning_rate": 4.507943349760313e-06, "loss": 0.0, "num_input_tokens_seen": 9025952, "step": 32235 }, { "epoch": 358.22222222222223, "grad_norm": 0.00011592888768063858, "learning_rate": 4.502321297936277e-06, "loss": 0.0, "num_input_tokens_seen": 9027360, "step": 32240 }, { "epoch": 358.27777777777777, "grad_norm": 7.064282726787496e-06, "learning_rate": 4.496702407111888e-06, "loss": 0.0, "num_input_tokens_seen": 9028768, "step": 32245 }, { "epoch": 358.3333333333333, "grad_norm": 2.9284537959028967e-06, "learning_rate": 4.491086678153653e-06, "loss": 0.0, "num_input_tokens_seen": 9030192, "step": 32250 }, { "epoch": 358.3888888888889, "grad_norm": 1.3860615126759512e-06, "learning_rate": 4.485474111927579e-06, "loss": 0.0, "num_input_tokens_seen": 9031584, "step": 32255 }, { "epoch": 358.44444444444446, "grad_norm": 7.57464590606105e-07, "learning_rate": 4.479864709299197e-06, "loss": 0.0, "num_input_tokens_seen": 9032976, "step": 32260 }, { "epoch": 358.5, "grad_norm": 7.180142347351648e-07, "learning_rate": 4.474258471133555e-06, "loss": 0.0, "num_input_tokens_seen": 9034400, "step": 32265 }, { "epoch": 358.55555555555554, "grad_norm": 2.0415916424099123e-06, "learning_rate": 4.4686553982952014e-06, "loss": 0.0, "num_input_tokens_seen": 9035776, "step": 32270 }, { "epoch": 358.6111111111111, "grad_norm": 5.937810669820465e-07, "learning_rate": 4.463055491648191e-06, "loss": 0.0, "num_input_tokens_seen": 9037168, "step": 32275 }, { "epoch": 358.6666666666667, "grad_norm": 2.6948225695377914e-06, "learning_rate": 4.457458752056112e-06, "loss": 0.0, "num_input_tokens_seen": 9038576, "step": 32280 }, { "epoch": 358.72222222222223, "grad_norm": 3.984436034443206e-07, "learning_rate": 4.451865180382042e-06, "loss": 0.0, "num_input_tokens_seen": 9039968, "step": 32285 }, { "epoch": 358.77777777777777, "grad_norm": 1.2832255436023843e-07, "learning_rate": 4.4462747774885936e-06, "loss": 0.0, "num_input_tokens_seen": 9041328, "step": 32290 }, { "epoch": 358.8333333333333, "grad_norm": 1.7909016492012597e-07, "learning_rate": 4.440687544237859e-06, "loss": 0.0, "num_input_tokens_seen": 9042736, "step": 32295 }, { "epoch": 358.8888888888889, "grad_norm": 3.451423708611401e-06, "learning_rate": 4.435103481491471e-06, "loss": 0.0, "num_input_tokens_seen": 9044176, "step": 32300 }, { "epoch": 358.94444444444446, "grad_norm": 1.1141624781885184e-05, "learning_rate": 4.429522590110569e-06, "loss": 0.0, "num_input_tokens_seen": 9045552, "step": 32305 }, { "epoch": 359.0, "grad_norm": 1.4702531814236863e-07, "learning_rate": 4.423944870955779e-06, "loss": 0.0, "num_input_tokens_seen": 9046960, "step": 32310 }, { "epoch": 359.05555555555554, "grad_norm": 5.172299552214099e-06, "learning_rate": 4.418370324887272e-06, "loss": 0.0, "num_input_tokens_seen": 9048384, "step": 32315 }, { "epoch": 359.1111111111111, "grad_norm": 1.4568065580533585e-06, "learning_rate": 4.412798952764699e-06, "loss": 0.0, "num_input_tokens_seen": 9049824, "step": 32320 }, { "epoch": 359.1666666666667, "grad_norm": 3.74906539946096e-06, "learning_rate": 4.407230755447245e-06, "loss": 0.0, "num_input_tokens_seen": 9051232, "step": 32325 }, { "epoch": 359.22222222222223, "grad_norm": 3.090417294515646e-06, "learning_rate": 4.401665733793598e-06, "loss": 0.0, "num_input_tokens_seen": 9052608, "step": 32330 }, { "epoch": 359.27777777777777, "grad_norm": 2.691751319616742e-07, "learning_rate": 4.3961038886619425e-06, "loss": 0.0, "num_input_tokens_seen": 9054064, "step": 32335 }, { "epoch": 359.3333333333333, "grad_norm": 6.242757990548853e-07, "learning_rate": 4.39054522091e-06, "loss": 0.0, "num_input_tokens_seen": 9055456, "step": 32340 }, { "epoch": 359.3888888888889, "grad_norm": 2.741915352544311e-07, "learning_rate": 4.384989731394979e-06, "loss": 0.0, "num_input_tokens_seen": 9056864, "step": 32345 }, { "epoch": 359.44444444444446, "grad_norm": 5.181768301554257e-06, "learning_rate": 4.379437420973598e-06, "loss": 0.0, "num_input_tokens_seen": 9058272, "step": 32350 }, { "epoch": 359.5, "grad_norm": 1.1384349818399642e-07, "learning_rate": 4.373888290502107e-06, "loss": 0.0, "num_input_tokens_seen": 9059680, "step": 32355 }, { "epoch": 359.55555555555554, "grad_norm": 4.5792168634761765e-07, "learning_rate": 4.36834234083624e-06, "loss": 0.0, "num_input_tokens_seen": 9061040, "step": 32360 }, { "epoch": 359.6111111111111, "grad_norm": 4.250954077633651e-07, "learning_rate": 4.362799572831258e-06, "loss": 0.0, "num_input_tokens_seen": 9062432, "step": 32365 }, { "epoch": 359.6666666666667, "grad_norm": 6.493174851129879e-07, "learning_rate": 4.35725998734193e-06, "loss": 0.0, "num_input_tokens_seen": 9063808, "step": 32370 }, { "epoch": 359.72222222222223, "grad_norm": 1.5695891306677368e-06, "learning_rate": 4.3517235852225195e-06, "loss": 0.0, "num_input_tokens_seen": 9065200, "step": 32375 }, { "epoch": 359.77777777777777, "grad_norm": 5.027800398238469e-07, "learning_rate": 4.346190367326822e-06, "loss": 0.0, "num_input_tokens_seen": 9066576, "step": 32380 }, { "epoch": 359.8333333333333, "grad_norm": 1.322228285971505e-06, "learning_rate": 4.340660334508115e-06, "loss": 0.0, "num_input_tokens_seen": 9067968, "step": 32385 }, { "epoch": 359.8888888888889, "grad_norm": 2.5025300942616013e-07, "learning_rate": 4.335133487619206e-06, "loss": 0.0, "num_input_tokens_seen": 9069344, "step": 32390 }, { "epoch": 359.94444444444446, "grad_norm": 1.8860362160921795e-06, "learning_rate": 4.329609827512409e-06, "loss": 0.0, "num_input_tokens_seen": 9070784, "step": 32395 }, { "epoch": 360.0, "grad_norm": 2.605003828648478e-06, "learning_rate": 4.324089355039531e-06, "loss": 0.0, "num_input_tokens_seen": 9072192, "step": 32400 }, { "epoch": 360.0, "eval_loss": 1.1077423095703125, "eval_runtime": 1.3856, "eval_samples_per_second": 28.868, "eval_steps_per_second": 14.434, "num_input_tokens_seen": 9072192, "step": 32400 }, { "epoch": 360.05555555555554, "grad_norm": 3.289449750809581e-06, "learning_rate": 4.3185720710519075e-06, "loss": 0.0, "num_input_tokens_seen": 9073584, "step": 32405 }, { "epoch": 360.1111111111111, "grad_norm": 3.699739681906067e-05, "learning_rate": 4.3130579764003724e-06, "loss": 0.0, "num_input_tokens_seen": 9075008, "step": 32410 }, { "epoch": 360.1666666666667, "grad_norm": 7.18867977411719e-06, "learning_rate": 4.307547071935267e-06, "loss": 0.0, "num_input_tokens_seen": 9076400, "step": 32415 }, { "epoch": 360.22222222222223, "grad_norm": 2.8542306154122343e-06, "learning_rate": 4.302039358506435e-06, "loss": 0.0, "num_input_tokens_seen": 9077792, "step": 32420 }, { "epoch": 360.27777777777777, "grad_norm": 6.785699042666238e-06, "learning_rate": 4.296534836963245e-06, "loss": 0.0, "num_input_tokens_seen": 9079200, "step": 32425 }, { "epoch": 360.3333333333333, "grad_norm": 2.100577148667071e-06, "learning_rate": 4.291033508154555e-06, "loss": 0.0, "num_input_tokens_seen": 9080592, "step": 32430 }, { "epoch": 360.3888888888889, "grad_norm": 3.6415069644135656e-06, "learning_rate": 4.285535372928748e-06, "loss": 0.0, "num_input_tokens_seen": 9082000, "step": 32435 }, { "epoch": 360.44444444444446, "grad_norm": 2.858934067262453e-07, "learning_rate": 4.280040432133695e-06, "loss": 0.0, "num_input_tokens_seen": 9083360, "step": 32440 }, { "epoch": 360.5, "grad_norm": 4.4155078171570494e-07, "learning_rate": 4.274548686616789e-06, "loss": 0.0, "num_input_tokens_seen": 9084736, "step": 32445 }, { "epoch": 360.55555555555554, "grad_norm": 7.752961295182104e-08, "learning_rate": 4.2690601372249364e-06, "loss": 0.0, "num_input_tokens_seen": 9086176, "step": 32450 }, { "epoch": 360.6111111111111, "grad_norm": 8.941073019741452e-07, "learning_rate": 4.263574784804525e-06, "loss": 0.0, "num_input_tokens_seen": 9087584, "step": 32455 }, { "epoch": 360.6666666666667, "grad_norm": 5.768147843809857e-07, "learning_rate": 4.258092630201479e-06, "loss": 0.0, "num_input_tokens_seen": 9088960, "step": 32460 }, { "epoch": 360.72222222222223, "grad_norm": 1.2282976058486383e-06, "learning_rate": 4.252613674261202e-06, "loss": 0.0, "num_input_tokens_seen": 9090352, "step": 32465 }, { "epoch": 360.77777777777777, "grad_norm": 4.529615011961141e-07, "learning_rate": 4.2471379178286224e-06, "loss": 0.0, "num_input_tokens_seen": 9091712, "step": 32470 }, { "epoch": 360.8333333333333, "grad_norm": 8.599033094469632e-07, "learning_rate": 4.241665361748181e-06, "loss": 0.0, "num_input_tokens_seen": 9093120, "step": 32475 }, { "epoch": 360.8888888888889, "grad_norm": 6.495625939351157e-07, "learning_rate": 4.2361960068637994e-06, "loss": 0.0, "num_input_tokens_seen": 9094544, "step": 32480 }, { "epoch": 360.94444444444446, "grad_norm": 4.2255896914866753e-07, "learning_rate": 4.230729854018933e-06, "loss": 0.0, "num_input_tokens_seen": 9095968, "step": 32485 }, { "epoch": 361.0, "grad_norm": 5.906478008910199e-07, "learning_rate": 4.225266904056521e-06, "loss": 0.0, "num_input_tokens_seen": 9097424, "step": 32490 }, { "epoch": 361.05555555555554, "grad_norm": 1.0204037153016543e-06, "learning_rate": 4.21980715781903e-06, "loss": 0.0, "num_input_tokens_seen": 9098800, "step": 32495 }, { "epoch": 361.1111111111111, "grad_norm": 7.894963118815213e-07, "learning_rate": 4.214350616148416e-06, "loss": 0.0, "num_input_tokens_seen": 9100192, "step": 32500 }, { "epoch": 361.1666666666667, "grad_norm": 1.3008543930936867e-07, "learning_rate": 4.20889727988614e-06, "loss": 0.0, "num_input_tokens_seen": 9101568, "step": 32505 }, { "epoch": 361.22222222222223, "grad_norm": 4.349122718849685e-06, "learning_rate": 4.20344714987318e-06, "loss": 0.0, "num_input_tokens_seen": 9103024, "step": 32510 }, { "epoch": 361.27777777777777, "grad_norm": 3.252964916100609e-06, "learning_rate": 4.198000226950022e-06, "loss": 0.0, "num_input_tokens_seen": 9104432, "step": 32515 }, { "epoch": 361.3333333333333, "grad_norm": 5.569094696511456e-07, "learning_rate": 4.192556511956635e-06, "loss": 0.0, "num_input_tokens_seen": 9105824, "step": 32520 }, { "epoch": 361.3888888888889, "grad_norm": 7.620055839652196e-07, "learning_rate": 4.18711600573252e-06, "loss": 0.0, "num_input_tokens_seen": 9107232, "step": 32525 }, { "epoch": 361.44444444444446, "grad_norm": 9.465801440455834e-07, "learning_rate": 4.181678709116671e-06, "loss": 0.0, "num_input_tokens_seen": 9108576, "step": 32530 }, { "epoch": 361.5, "grad_norm": 1.656560755236569e-07, "learning_rate": 4.1762446229475785e-06, "loss": 0.0, "num_input_tokens_seen": 9109968, "step": 32535 }, { "epoch": 361.55555555555554, "grad_norm": 1.9982697097020719e-07, "learning_rate": 4.17081374806326e-06, "loss": 0.0, "num_input_tokens_seen": 9111360, "step": 32540 }, { "epoch": 361.6111111111111, "grad_norm": 2.969977401789947e-07, "learning_rate": 4.165386085301212e-06, "loss": 0.0, "num_input_tokens_seen": 9112784, "step": 32545 }, { "epoch": 361.6666666666667, "grad_norm": 2.466695605107816e-06, "learning_rate": 4.1599616354984525e-06, "loss": 0.0, "num_input_tokens_seen": 9114208, "step": 32550 }, { "epoch": 361.72222222222223, "grad_norm": 2.4669657250342425e-06, "learning_rate": 4.154540399491508e-06, "loss": 0.0, "num_input_tokens_seen": 9115648, "step": 32555 }, { "epoch": 361.77777777777777, "grad_norm": 7.3956462074420415e-06, "learning_rate": 4.149122378116394e-06, "loss": 0.0, "num_input_tokens_seen": 9117008, "step": 32560 }, { "epoch": 361.8333333333333, "grad_norm": 3.4417050756019307e-06, "learning_rate": 4.14370757220863e-06, "loss": 0.0, "num_input_tokens_seen": 9118400, "step": 32565 }, { "epoch": 361.8888888888889, "grad_norm": 1.0133934665645938e-05, "learning_rate": 4.138295982603263e-06, "loss": 0.0, "num_input_tokens_seen": 9119808, "step": 32570 }, { "epoch": 361.94444444444446, "grad_norm": 1.7105558924868092e-07, "learning_rate": 4.132887610134814e-06, "loss": 0.0, "num_input_tokens_seen": 9121200, "step": 32575 }, { "epoch": 362.0, "grad_norm": 1.144402745012485e-06, "learning_rate": 4.127482455637335e-06, "loss": 0.0, "num_input_tokens_seen": 9122624, "step": 32580 }, { "epoch": 362.05555555555554, "grad_norm": 5.540399001802143e-07, "learning_rate": 4.1220805199443545e-06, "loss": 0.0, "num_input_tokens_seen": 9124032, "step": 32585 }, { "epoch": 362.1111111111111, "grad_norm": 2.648734334798064e-05, "learning_rate": 4.116681803888925e-06, "loss": 0.0, "num_input_tokens_seen": 9125424, "step": 32590 }, { "epoch": 362.1666666666667, "grad_norm": 7.690008715144359e-06, "learning_rate": 4.111286308303605e-06, "loss": 0.0, "num_input_tokens_seen": 9126832, "step": 32595 }, { "epoch": 362.22222222222223, "grad_norm": 4.406037533044582e-06, "learning_rate": 4.105894034020433e-06, "loss": 0.0, "num_input_tokens_seen": 9128272, "step": 32600 }, { "epoch": 362.22222222222223, "eval_loss": 1.1189839839935303, "eval_runtime": 1.3984, "eval_samples_per_second": 28.604, "eval_steps_per_second": 14.302, "num_input_tokens_seen": 9128272, "step": 32600 }, { "epoch": 362.27777777777777, "grad_norm": 7.846456355764531e-07, "learning_rate": 4.100504981870975e-06, "loss": 0.0, "num_input_tokens_seen": 9129664, "step": 32605 }, { "epoch": 362.3333333333333, "grad_norm": 9.075806701730471e-07, "learning_rate": 4.0951191526862915e-06, "loss": 0.0, "num_input_tokens_seen": 9131008, "step": 32610 }, { "epoch": 362.3888888888889, "grad_norm": 3.7762478655167797e-07, "learning_rate": 4.089736547296938e-06, "loss": 0.0, "num_input_tokens_seen": 9132368, "step": 32615 }, { "epoch": 362.44444444444446, "grad_norm": 1.7279722669627517e-05, "learning_rate": 4.08435716653299e-06, "loss": 0.0, "num_input_tokens_seen": 9133728, "step": 32620 }, { "epoch": 362.5, "grad_norm": 1.9025112578674452e-06, "learning_rate": 4.0789810112240005e-06, "loss": 0.0, "num_input_tokens_seen": 9135152, "step": 32625 }, { "epoch": 362.55555555555554, "grad_norm": 1.446636161972492e-07, "learning_rate": 4.073608082199057e-06, "loss": 0.0, "num_input_tokens_seen": 9136544, "step": 32630 }, { "epoch": 362.6111111111111, "grad_norm": 3.1685939916314965e-07, "learning_rate": 4.068238380286718e-06, "loss": 0.0, "num_input_tokens_seen": 9137952, "step": 32635 }, { "epoch": 362.6666666666667, "grad_norm": 1.407639018680129e-07, "learning_rate": 4.062871906315072e-06, "loss": 0.0, "num_input_tokens_seen": 9139344, "step": 32640 }, { "epoch": 362.72222222222223, "grad_norm": 1.3146881201464566e-06, "learning_rate": 4.057508661111686e-06, "loss": 0.0, "num_input_tokens_seen": 9140704, "step": 32645 }, { "epoch": 362.77777777777777, "grad_norm": 5.990375029796269e-06, "learning_rate": 4.052148645503648e-06, "loss": 0.0, "num_input_tokens_seen": 9142128, "step": 32650 }, { "epoch": 362.8333333333333, "grad_norm": 1.274115675187204e-06, "learning_rate": 4.046791860317531e-06, "loss": 0.0, "num_input_tokens_seen": 9143520, "step": 32655 }, { "epoch": 362.8888888888889, "grad_norm": 4.7515334244963014e-07, "learning_rate": 4.041438306379431e-06, "loss": 0.0, "num_input_tokens_seen": 9144928, "step": 32660 }, { "epoch": 362.94444444444446, "grad_norm": 4.1718567445059307e-07, "learning_rate": 4.036087984514916e-06, "loss": 0.0, "num_input_tokens_seen": 9146336, "step": 32665 }, { "epoch": 363.0, "grad_norm": 2.1072328593163547e-07, "learning_rate": 4.030740895549084e-06, "loss": 0.0, "num_input_tokens_seen": 9147776, "step": 32670 }, { "epoch": 363.05555555555554, "grad_norm": 1.7879106906093511e-07, "learning_rate": 4.025397040306531e-06, "loss": 0.0, "num_input_tokens_seen": 9149200, "step": 32675 }, { "epoch": 363.1111111111111, "grad_norm": 2.3190148112917086e-06, "learning_rate": 4.0200564196113285e-06, "loss": 0.0, "num_input_tokens_seen": 9150576, "step": 32680 }, { "epoch": 363.1666666666667, "grad_norm": 4.932914634991903e-06, "learning_rate": 4.014719034287079e-06, "loss": 0.0, "num_input_tokens_seen": 9152000, "step": 32685 }, { "epoch": 363.22222222222223, "grad_norm": 2.8896474759676494e-06, "learning_rate": 4.0093848851568775e-06, "loss": 0.0, "num_input_tokens_seen": 9153424, "step": 32690 }, { "epoch": 363.27777777777777, "grad_norm": 9.353608220408205e-06, "learning_rate": 4.004053973043304e-06, "loss": 0.0, "num_input_tokens_seen": 9154800, "step": 32695 }, { "epoch": 363.3333333333333, "grad_norm": 1.3955325357528636e-06, "learning_rate": 3.998726298768465e-06, "loss": 0.0, "num_input_tokens_seen": 9156192, "step": 32700 }, { "epoch": 363.3888888888889, "grad_norm": 3.623847248945822e-07, "learning_rate": 3.99340186315395e-06, "loss": 0.0, "num_input_tokens_seen": 9157600, "step": 32705 }, { "epoch": 363.44444444444446, "grad_norm": 6.600716915272642e-07, "learning_rate": 3.988080667020849e-06, "loss": 0.0, "num_input_tokens_seen": 9159024, "step": 32710 }, { "epoch": 363.5, "grad_norm": 3.4971853324350377e-07, "learning_rate": 3.982762711189766e-06, "loss": 0.0, "num_input_tokens_seen": 9160416, "step": 32715 }, { "epoch": 363.55555555555554, "grad_norm": 2.0495052410751669e-07, "learning_rate": 3.977447996480785e-06, "loss": 0.0, "num_input_tokens_seen": 9161776, "step": 32720 }, { "epoch": 363.6111111111111, "grad_norm": 7.88104716775706e-06, "learning_rate": 3.97213652371351e-06, "loss": 0.0, "num_input_tokens_seen": 9163168, "step": 32725 }, { "epoch": 363.6666666666667, "grad_norm": 5.15000010636868e-07, "learning_rate": 3.966828293707042e-06, "loss": 0.0, "num_input_tokens_seen": 9164592, "step": 32730 }, { "epoch": 363.72222222222223, "grad_norm": 1.9732497094082646e-06, "learning_rate": 3.961523307279963e-06, "loss": 0.0, "num_input_tokens_seen": 9165952, "step": 32735 }, { "epoch": 363.77777777777777, "grad_norm": 1.5960487189659034e-06, "learning_rate": 3.956221565250382e-06, "loss": 0.0, "num_input_tokens_seen": 9167360, "step": 32740 }, { "epoch": 363.8333333333333, "grad_norm": 3.7086596194058075e-07, "learning_rate": 3.950923068435883e-06, "loss": 0.0, "num_input_tokens_seen": 9168768, "step": 32745 }, { "epoch": 363.8888888888889, "grad_norm": 6.215372650331119e-06, "learning_rate": 3.945627817653566e-06, "loss": 0.0, "num_input_tokens_seen": 9170160, "step": 32750 }, { "epoch": 363.94444444444446, "grad_norm": 3.6158621696813498e-06, "learning_rate": 3.9403358137200335e-06, "loss": 0.0, "num_input_tokens_seen": 9171600, "step": 32755 }, { "epoch": 364.0, "grad_norm": 1.8200405520474305e-06, "learning_rate": 3.9350470574513605e-06, "loss": 0.0, "num_input_tokens_seen": 9172992, "step": 32760 }, { "epoch": 364.05555555555554, "grad_norm": 3.0479588986054296e-07, "learning_rate": 3.9297615496631525e-06, "loss": 0.0, "num_input_tokens_seen": 9174400, "step": 32765 }, { "epoch": 364.1111111111111, "grad_norm": 5.794276489723416e-07, "learning_rate": 3.924479291170505e-06, "loss": 0.0, "num_input_tokens_seen": 9175872, "step": 32770 }, { "epoch": 364.1666666666667, "grad_norm": 1.2513190768004279e-06, "learning_rate": 3.919200282788002e-06, "loss": 0.0, "num_input_tokens_seen": 9177232, "step": 32775 }, { "epoch": 364.22222222222223, "grad_norm": 3.4923513680951146e-07, "learning_rate": 3.913924525329726e-06, "loss": 0.0, "num_input_tokens_seen": 9178608, "step": 32780 }, { "epoch": 364.27777777777777, "grad_norm": 2.273817926834454e-06, "learning_rate": 3.908652019609279e-06, "loss": 0.0, "num_input_tokens_seen": 9180032, "step": 32785 }, { "epoch": 364.3333333333333, "grad_norm": 4.686396337660881e-08, "learning_rate": 3.9033827664397364e-06, "loss": 0.0, "num_input_tokens_seen": 9181408, "step": 32790 }, { "epoch": 364.3888888888889, "grad_norm": 2.940388981187425e-07, "learning_rate": 3.898116766633694e-06, "loss": 0.0, "num_input_tokens_seen": 9182816, "step": 32795 }, { "epoch": 364.44444444444446, "grad_norm": 4.202944978715095e-07, "learning_rate": 3.8928540210032225e-06, "loss": 0.0, "num_input_tokens_seen": 9184240, "step": 32800 }, { "epoch": 364.44444444444446, "eval_loss": 1.1359033584594727, "eval_runtime": 1.3935, "eval_samples_per_second": 28.706, "eval_steps_per_second": 14.353, "num_input_tokens_seen": 9184240, "step": 32800 }, { "epoch": 364.5, "grad_norm": 5.724438096876838e-07, "learning_rate": 3.887594530359909e-06, "loss": 0.0, "num_input_tokens_seen": 9185664, "step": 32805 }, { "epoch": 364.55555555555554, "grad_norm": 5.623086849482206e-07, "learning_rate": 3.88233829551484e-06, "loss": 0.0, "num_input_tokens_seen": 9187056, "step": 32810 }, { "epoch": 364.6111111111111, "grad_norm": 1.9813769540633075e-06, "learning_rate": 3.877085317278581e-06, "loss": 0.0, "num_input_tokens_seen": 9188448, "step": 32815 }, { "epoch": 364.6666666666667, "grad_norm": 3.61417960448307e-07, "learning_rate": 3.87183559646122e-06, "loss": 0.0, "num_input_tokens_seen": 9189808, "step": 32820 }, { "epoch": 364.72222222222223, "grad_norm": 1.3562800631916616e-05, "learning_rate": 3.866589133872317e-06, "loss": 0.0, "num_input_tokens_seen": 9191216, "step": 32825 }, { "epoch": 364.77777777777777, "grad_norm": 2.2391021047951654e-05, "learning_rate": 3.861345930320948e-06, "loss": 0.0, "num_input_tokens_seen": 9192608, "step": 32830 }, { "epoch": 364.8333333333333, "grad_norm": 4.6605382522102445e-06, "learning_rate": 3.856105986615688e-06, "loss": 0.0, "num_input_tokens_seen": 9194016, "step": 32835 }, { "epoch": 364.8888888888889, "grad_norm": 1.36792863258961e-07, "learning_rate": 3.850869303564589e-06, "loss": 0.0, "num_input_tokens_seen": 9195456, "step": 32840 }, { "epoch": 364.94444444444446, "grad_norm": 2.488807467670995e-06, "learning_rate": 3.845635881975226e-06, "loss": 0.0, "num_input_tokens_seen": 9196800, "step": 32845 }, { "epoch": 365.0, "grad_norm": 1.5293265676064038e-07, "learning_rate": 3.840405722654647e-06, "loss": 0.0, "num_input_tokens_seen": 9198192, "step": 32850 }, { "epoch": 365.05555555555554, "grad_norm": 3.9599220258423884e-07, "learning_rate": 3.835178826409419e-06, "loss": 0.0, "num_input_tokens_seen": 9199584, "step": 32855 }, { "epoch": 365.1111111111111, "grad_norm": 8.108320344035747e-07, "learning_rate": 3.8299551940455895e-06, "loss": 0.0, "num_input_tokens_seen": 9200960, "step": 32860 }, { "epoch": 365.1666666666667, "grad_norm": 1.6406179383920971e-06, "learning_rate": 3.824734826368703e-06, "loss": 0.0, "num_input_tokens_seen": 9202336, "step": 32865 }, { "epoch": 365.22222222222223, "grad_norm": 2.1632665720972e-06, "learning_rate": 3.819517724183813e-06, "loss": 0.0, "num_input_tokens_seen": 9203760, "step": 32870 }, { "epoch": 365.27777777777777, "grad_norm": 5.572853297053371e-06, "learning_rate": 3.8143038882954648e-06, "loss": 0.0, "num_input_tokens_seen": 9205168, "step": 32875 }, { "epoch": 365.3333333333333, "grad_norm": 7.41690485028812e-07, "learning_rate": 3.8090933195076867e-06, "loss": 0.0, "num_input_tokens_seen": 9206640, "step": 32880 }, { "epoch": 365.3888888888889, "grad_norm": 6.345447332023468e-07, "learning_rate": 3.8038860186240198e-06, "loss": 0.0, "num_input_tokens_seen": 9208048, "step": 32885 }, { "epoch": 365.44444444444446, "grad_norm": 4.312689270591363e-06, "learning_rate": 3.7986819864475026e-06, "loss": 0.0, "num_input_tokens_seen": 9209392, "step": 32890 }, { "epoch": 365.5, "grad_norm": 4.2148582224399433e-07, "learning_rate": 3.793481223780651e-06, "loss": 0.0, "num_input_tokens_seen": 9210752, "step": 32895 }, { "epoch": 365.55555555555554, "grad_norm": 3.1825577480049105e-06, "learning_rate": 3.788283731425496e-06, "loss": 0.0, "num_input_tokens_seen": 9212144, "step": 32900 }, { "epoch": 365.6111111111111, "grad_norm": 4.803410433851241e-07, "learning_rate": 3.7830895101835488e-06, "loss": 0.0, "num_input_tokens_seen": 9213568, "step": 32905 }, { "epoch": 365.6666666666667, "grad_norm": 5.988457587591256e-07, "learning_rate": 3.7778985608558274e-06, "loss": 0.0, "num_input_tokens_seen": 9214992, "step": 32910 }, { "epoch": 365.72222222222223, "grad_norm": 2.613716958421719e-07, "learning_rate": 3.7727108842428443e-06, "loss": 0.0, "num_input_tokens_seen": 9216384, "step": 32915 }, { "epoch": 365.77777777777777, "grad_norm": 8.302885703415086e-07, "learning_rate": 3.7675264811446065e-06, "loss": 0.0, "num_input_tokens_seen": 9217792, "step": 32920 }, { "epoch": 365.8333333333333, "grad_norm": 5.556336191148148e-07, "learning_rate": 3.7623453523605994e-06, "loss": 0.0, "num_input_tokens_seen": 9219216, "step": 32925 }, { "epoch": 365.8888888888889, "grad_norm": 1.9096876258117845e-06, "learning_rate": 3.757167498689834e-06, "loss": 0.0, "num_input_tokens_seen": 9220592, "step": 32930 }, { "epoch": 365.94444444444446, "grad_norm": 1.1239426385145634e-06, "learning_rate": 3.7519929209307914e-06, "loss": 0.0, "num_input_tokens_seen": 9221936, "step": 32935 }, { "epoch": 366.0, "grad_norm": 2.001765551540302e-06, "learning_rate": 3.746821619881463e-06, "loss": 0.0, "num_input_tokens_seen": 9223312, "step": 32940 }, { "epoch": 366.05555555555554, "grad_norm": 6.903001121827401e-06, "learning_rate": 3.74165359633932e-06, "loss": 0.0, "num_input_tokens_seen": 9224720, "step": 32945 }, { "epoch": 366.1111111111111, "grad_norm": 2.6844345484278165e-05, "learning_rate": 3.736488851101341e-06, "loss": 0.0, "num_input_tokens_seen": 9226128, "step": 32950 }, { "epoch": 366.1666666666667, "grad_norm": 2.2973104307766334e-07, "learning_rate": 3.7313273849640035e-06, "loss": 0.0, "num_input_tokens_seen": 9227520, "step": 32955 }, { "epoch": 366.22222222222223, "grad_norm": 1.9720134787348798e-06, "learning_rate": 3.7261691987232533e-06, "loss": 0.0, "num_input_tokens_seen": 9228880, "step": 32960 }, { "epoch": 366.27777777777777, "grad_norm": 6.769271294615464e-07, "learning_rate": 3.7210142931745575e-06, "loss": 0.0, "num_input_tokens_seen": 9230240, "step": 32965 }, { "epoch": 366.3333333333333, "grad_norm": 2.5510632895020535e-07, "learning_rate": 3.7158626691128712e-06, "loss": 0.0, "num_input_tokens_seen": 9231648, "step": 32970 }, { "epoch": 366.3888888888889, "grad_norm": 2.1582438591849495e-07, "learning_rate": 3.710714327332629e-06, "loss": 0.0, "num_input_tokens_seen": 9233008, "step": 32975 }, { "epoch": 366.44444444444446, "grad_norm": 3.142668447253527e-06, "learning_rate": 3.7055692686277815e-06, "loss": 0.0, "num_input_tokens_seen": 9234400, "step": 32980 }, { "epoch": 366.5, "grad_norm": 3.5289195920995553e-07, "learning_rate": 3.70042749379175e-06, "loss": 0.0, "num_input_tokens_seen": 9235776, "step": 32985 }, { "epoch": 366.55555555555554, "grad_norm": 3.0937601991354313e-07, "learning_rate": 3.6952890036174693e-06, "loss": 0.0, "num_input_tokens_seen": 9237216, "step": 32990 }, { "epoch": 366.6111111111111, "grad_norm": 4.3413882622189703e-07, "learning_rate": 3.690153798897353e-06, "loss": 0.0, "num_input_tokens_seen": 9238656, "step": 32995 }, { "epoch": 366.6666666666667, "grad_norm": 2.799777973905293e-07, "learning_rate": 3.6850218804233225e-06, "loss": 0.0, "num_input_tokens_seen": 9240064, "step": 33000 }, { "epoch": 366.6666666666667, "eval_loss": 1.1823246479034424, "eval_runtime": 1.3949, "eval_samples_per_second": 28.676, "eval_steps_per_second": 14.338, "num_input_tokens_seen": 9240064, "step": 33000 }, { "epoch": 366.72222222222223, "grad_norm": 4.751226470034453e-07, "learning_rate": 3.679893248986779e-06, "loss": 0.0, "num_input_tokens_seen": 9241456, "step": 33005 }, { "epoch": 366.77777777777777, "grad_norm": 1.5106920159269066e-07, "learning_rate": 3.6747679053786147e-06, "loss": 0.0, "num_input_tokens_seen": 9242864, "step": 33010 }, { "epoch": 366.8333333333333, "grad_norm": 2.262559519294882e-06, "learning_rate": 3.669645850389228e-06, "loss": 0.0, "num_input_tokens_seen": 9244272, "step": 33015 }, { "epoch": 366.8888888888889, "grad_norm": 2.463526982410258e-07, "learning_rate": 3.664527084808514e-06, "loss": 0.0, "num_input_tokens_seen": 9245664, "step": 33020 }, { "epoch": 366.94444444444446, "grad_norm": 1.2258419701538514e-06, "learning_rate": 3.6594116094258337e-06, "loss": 0.0, "num_input_tokens_seen": 9247056, "step": 33025 }, { "epoch": 367.0, "grad_norm": 5.642690936724648e-08, "learning_rate": 3.6542994250300665e-06, "loss": 0.0, "num_input_tokens_seen": 9248496, "step": 33030 }, { "epoch": 367.05555555555554, "grad_norm": 6.939854415577429e-07, "learning_rate": 3.6491905324095825e-06, "loss": 0.0, "num_input_tokens_seen": 9249888, "step": 33035 }, { "epoch": 367.1111111111111, "grad_norm": 5.927689699092298e-07, "learning_rate": 3.644084932352221e-06, "loss": 0.0, "num_input_tokens_seen": 9251248, "step": 33040 }, { "epoch": 367.1666666666667, "grad_norm": 5.575874411078985e-07, "learning_rate": 3.6389826256453457e-06, "loss": 0.0, "num_input_tokens_seen": 9252672, "step": 33045 }, { "epoch": 367.22222222222223, "grad_norm": 3.3470234939159127e-07, "learning_rate": 3.633883613075781e-06, "loss": 0.0, "num_input_tokens_seen": 9254080, "step": 33050 }, { "epoch": 367.27777777777777, "grad_norm": 6.958464382478269e-07, "learning_rate": 3.6287878954298693e-06, "loss": 0.0, "num_input_tokens_seen": 9255488, "step": 33055 }, { "epoch": 367.3333333333333, "grad_norm": 3.1040087833389407e-07, "learning_rate": 3.6236954734934354e-06, "loss": 0.0, "num_input_tokens_seen": 9256912, "step": 33060 }, { "epoch": 367.3888888888889, "grad_norm": 1.3961046079202788e-06, "learning_rate": 3.618606348051784e-06, "loss": 0.0, "num_input_tokens_seen": 9258288, "step": 33065 }, { "epoch": 367.44444444444446, "grad_norm": 2.447249300985277e-07, "learning_rate": 3.6135205198897376e-06, "loss": 0.0, "num_input_tokens_seen": 9259664, "step": 33070 }, { "epoch": 367.5, "grad_norm": 7.634952794433048e-07, "learning_rate": 3.6084379897915854e-06, "loss": 0.0, "num_input_tokens_seen": 9261056, "step": 33075 }, { "epoch": 367.55555555555554, "grad_norm": 2.6888540105574066e-06, "learning_rate": 3.6033587585411115e-06, "loss": 0.0, "num_input_tokens_seen": 9262448, "step": 33080 }, { "epoch": 367.6111111111111, "grad_norm": 2.1521809685509652e-05, "learning_rate": 3.5982828269216117e-06, "loss": 0.0, "num_input_tokens_seen": 9263840, "step": 33085 }, { "epoch": 367.6666666666667, "grad_norm": 2.3727920961391646e-06, "learning_rate": 3.593210195715843e-06, "loss": 0.0, "num_input_tokens_seen": 9265232, "step": 33090 }, { "epoch": 367.72222222222223, "grad_norm": 3.3186839232257626e-07, "learning_rate": 3.5881408657060773e-06, "loss": 0.0, "num_input_tokens_seen": 9266640, "step": 33095 }, { "epoch": 367.77777777777777, "grad_norm": 2.2724880182067864e-05, "learning_rate": 3.583074837674075e-06, "loss": 0.0, "num_input_tokens_seen": 9268000, "step": 33100 }, { "epoch": 367.8333333333333, "grad_norm": 9.094007396015513e-07, "learning_rate": 3.578012112401069e-06, "loss": 0.0, "num_input_tokens_seen": 9269424, "step": 33105 }, { "epoch": 367.8888888888889, "grad_norm": 5.000514420316904e-07, "learning_rate": 3.5729526906677996e-06, "loss": 0.0, "num_input_tokens_seen": 9270880, "step": 33110 }, { "epoch": 367.94444444444446, "grad_norm": 4.619899300450925e-06, "learning_rate": 3.5678965732545007e-06, "loss": 0.0, "num_input_tokens_seen": 9272272, "step": 33115 }, { "epoch": 368.0, "grad_norm": 2.565536476595298e-07, "learning_rate": 3.562843760940876e-06, "loss": 0.0, "num_input_tokens_seen": 9273664, "step": 33120 }, { "epoch": 368.05555555555554, "grad_norm": 2.5357269350934075e-07, "learning_rate": 3.5577942545061473e-06, "loss": 0.0, "num_input_tokens_seen": 9275088, "step": 33125 }, { "epoch": 368.1111111111111, "grad_norm": 2.7514809630702075e-07, "learning_rate": 3.5527480547289967e-06, "loss": 0.0, "num_input_tokens_seen": 9276464, "step": 33130 }, { "epoch": 368.1666666666667, "grad_norm": 5.895046228943102e-07, "learning_rate": 3.547705162387624e-06, "loss": 0.0, "num_input_tokens_seen": 9277872, "step": 33135 }, { "epoch": 368.22222222222223, "grad_norm": 5.316008355293889e-06, "learning_rate": 3.542665578259699e-06, "loss": 0.0, "num_input_tokens_seen": 9279248, "step": 33140 }, { "epoch": 368.27777777777777, "grad_norm": 1.0449317642269307e-06, "learning_rate": 3.5376293031223945e-06, "loss": 0.0, "num_input_tokens_seen": 9280640, "step": 33145 }, { "epoch": 368.3333333333333, "grad_norm": 1.352687633016103e-07, "learning_rate": 3.5325963377523614e-06, "loss": 0.0, "num_input_tokens_seen": 9282000, "step": 33150 }, { "epoch": 368.3888888888889, "grad_norm": 2.1529534421915741e-07, "learning_rate": 3.5275666829257536e-06, "loss": 0.0, "num_input_tokens_seen": 9283424, "step": 33155 }, { "epoch": 368.44444444444446, "grad_norm": 1.733317276375601e-07, "learning_rate": 3.5225403394181955e-06, "loss": 0.0, "num_input_tokens_seen": 9284816, "step": 33160 }, { "epoch": 368.5, "grad_norm": 1.009102675197937e-06, "learning_rate": 3.517517308004828e-06, "loss": 0.0, "num_input_tokens_seen": 9286240, "step": 33165 }, { "epoch": 368.55555555555554, "grad_norm": 2.7700664304575184e-06, "learning_rate": 3.512497589460251e-06, "loss": 0.0, "num_input_tokens_seen": 9287664, "step": 33170 }, { "epoch": 368.6111111111111, "grad_norm": 8.851762913764105e-07, "learning_rate": 3.5074811845585727e-06, "loss": 0.0, "num_input_tokens_seen": 9289040, "step": 33175 }, { "epoch": 368.6666666666667, "grad_norm": 1.506291141595284e-06, "learning_rate": 3.5024680940733937e-06, "loss": 0.0, "num_input_tokens_seen": 9290416, "step": 33180 }, { "epoch": 368.72222222222223, "grad_norm": 2.953038574560196e-06, "learning_rate": 3.4974583187777852e-06, "loss": 0.0, "num_input_tokens_seen": 9291824, "step": 33185 }, { "epoch": 368.77777777777777, "grad_norm": 5.879847435608099e-07, "learning_rate": 3.4924518594443204e-06, "loss": 0.0, "num_input_tokens_seen": 9293200, "step": 33190 }, { "epoch": 368.8333333333333, "grad_norm": 4.155188548793376e-07, "learning_rate": 3.4874487168450682e-06, "loss": 0.0, "num_input_tokens_seen": 9294560, "step": 33195 }, { "epoch": 368.8888888888889, "grad_norm": 2.9339330467337277e-06, "learning_rate": 3.482448891751558e-06, "loss": 0.0, "num_input_tokens_seen": 9295952, "step": 33200 }, { "epoch": 368.8888888888889, "eval_loss": 1.1525328159332275, "eval_runtime": 1.4045, "eval_samples_per_second": 28.479, "eval_steps_per_second": 14.239, "num_input_tokens_seen": 9295952, "step": 33200 }, { "epoch": 368.94444444444446, "grad_norm": 5.34130458618165e-06, "learning_rate": 3.477452384934843e-06, "loss": 0.0, "num_input_tokens_seen": 9297408, "step": 33205 }, { "epoch": 369.0, "grad_norm": 1.5615303254890023e-06, "learning_rate": 3.472459197165434e-06, "loss": 0.0, "num_input_tokens_seen": 9298832, "step": 33210 }, { "epoch": 369.05555555555554, "grad_norm": 3.363690211699577e-06, "learning_rate": 3.4674693292133518e-06, "loss": 0.0, "num_input_tokens_seen": 9300224, "step": 33215 }, { "epoch": 369.1111111111111, "grad_norm": 6.679047714897024e-07, "learning_rate": 3.4624827818480977e-06, "loss": 0.0, "num_input_tokens_seen": 9301584, "step": 33220 }, { "epoch": 369.1666666666667, "grad_norm": 3.164137467592809e-07, "learning_rate": 3.4574995558386474e-06, "loss": 0.0, "num_input_tokens_seen": 9302976, "step": 33225 }, { "epoch": 369.22222222222223, "grad_norm": 3.505513404888916e-06, "learning_rate": 3.452519651953487e-06, "loss": 0.0, "num_input_tokens_seen": 9304384, "step": 33230 }, { "epoch": 369.27777777777777, "grad_norm": 2.5175916107400553e-06, "learning_rate": 3.447543070960585e-06, "loss": 0.0, "num_input_tokens_seen": 9305792, "step": 33235 }, { "epoch": 369.3333333333333, "grad_norm": 7.5365533120930195e-06, "learning_rate": 3.4425698136273778e-06, "loss": 0.0, "num_input_tokens_seen": 9307184, "step": 33240 }, { "epoch": 369.3888888888889, "grad_norm": 8.140922318489174e-07, "learning_rate": 3.437599880720821e-06, "loss": 0.0, "num_input_tokens_seen": 9308592, "step": 33245 }, { "epoch": 369.44444444444446, "grad_norm": 1.4840472431387752e-05, "learning_rate": 3.4326332730073267e-06, "loss": 0.0, "num_input_tokens_seen": 9309984, "step": 33250 }, { "epoch": 369.5, "grad_norm": 4.0795356426315266e-07, "learning_rate": 3.427669991252813e-06, "loss": 0.0, "num_input_tokens_seen": 9311360, "step": 33255 }, { "epoch": 369.55555555555554, "grad_norm": 1.0712489029174321e-06, "learning_rate": 3.42271003622269e-06, "loss": 0.0, "num_input_tokens_seen": 9312736, "step": 33260 }, { "epoch": 369.6111111111111, "grad_norm": 2.867123384930892e-06, "learning_rate": 3.4177534086818286e-06, "loss": 0.0, "num_input_tokens_seen": 9314112, "step": 33265 }, { "epoch": 369.6666666666667, "grad_norm": 6.406299917216529e-07, "learning_rate": 3.412800109394612e-06, "loss": 0.0, "num_input_tokens_seen": 9315584, "step": 33270 }, { "epoch": 369.72222222222223, "grad_norm": 6.1197961258585565e-06, "learning_rate": 3.4078501391249044e-06, "loss": 0.0, "num_input_tokens_seen": 9317008, "step": 33275 }, { "epoch": 369.77777777777777, "grad_norm": 6.311071842901583e-07, "learning_rate": 3.4029034986360453e-06, "loss": 0.0, "num_input_tokens_seen": 9318416, "step": 33280 }, { "epoch": 369.8333333333333, "grad_norm": 2.8833599685640365e-07, "learning_rate": 3.397960188690877e-06, "loss": 0.0, "num_input_tokens_seen": 9319776, "step": 33285 }, { "epoch": 369.8888888888889, "grad_norm": 2.0814450181205757e-05, "learning_rate": 3.393020210051717e-06, "loss": 0.0, "num_input_tokens_seen": 9321200, "step": 33290 }, { "epoch": 369.94444444444446, "grad_norm": 0.00013838245649822056, "learning_rate": 3.3880835634803655e-06, "loss": 0.0, "num_input_tokens_seen": 9322624, "step": 33295 }, { "epoch": 370.0, "grad_norm": 3.8258059475992923e-07, "learning_rate": 3.383150249738126e-06, "loss": 0.0, "num_input_tokens_seen": 9324064, "step": 33300 }, { "epoch": 370.05555555555554, "grad_norm": 2.92226627607306e-06, "learning_rate": 3.3782202695857663e-06, "loss": 0.0, "num_input_tokens_seen": 9325472, "step": 33305 }, { "epoch": 370.1111111111111, "grad_norm": 1.1617959216891904e-06, "learning_rate": 3.373293623783558e-06, "loss": 0.0, "num_input_tokens_seen": 9326880, "step": 33310 }, { "epoch": 370.1666666666667, "grad_norm": 4.166833491581201e-07, "learning_rate": 3.368370313091257e-06, "loss": 0.0, "num_input_tokens_seen": 9328304, "step": 33315 }, { "epoch": 370.22222222222223, "grad_norm": 3.670315322779061e-07, "learning_rate": 3.363450338268087e-06, "loss": 0.0, "num_input_tokens_seen": 9329648, "step": 33320 }, { "epoch": 370.27777777777777, "grad_norm": 2.7564159609028138e-05, "learning_rate": 3.358533700072783e-06, "loss": 0.0, "num_input_tokens_seen": 9331040, "step": 33325 }, { "epoch": 370.3333333333333, "grad_norm": 2.1977616881940776e-07, "learning_rate": 3.3536203992635377e-06, "loss": 0.0, "num_input_tokens_seen": 9332448, "step": 33330 }, { "epoch": 370.3888888888889, "grad_norm": 4.291349853247084e-07, "learning_rate": 3.348710436598057e-06, "loss": 0.0, "num_input_tokens_seen": 9333872, "step": 33335 }, { "epoch": 370.44444444444446, "grad_norm": 7.021808414719999e-07, "learning_rate": 3.3438038128335155e-06, "loss": 0.0, "num_input_tokens_seen": 9335264, "step": 33340 }, { "epoch": 370.5, "grad_norm": 5.996453182888217e-07, "learning_rate": 3.338900528726571e-06, "loss": 0.0, "num_input_tokens_seen": 9336672, "step": 33345 }, { "epoch": 370.55555555555554, "grad_norm": 2.6846068976738025e-06, "learning_rate": 3.3340005850333812e-06, "loss": 0.0, "num_input_tokens_seen": 9338000, "step": 33350 }, { "epoch": 370.6111111111111, "grad_norm": 3.7379675177362515e-07, "learning_rate": 3.329103982509568e-06, "loss": 0.0, "num_input_tokens_seen": 9339376, "step": 33355 }, { "epoch": 370.6666666666667, "grad_norm": 4.455741304809635e-07, "learning_rate": 3.324210721910259e-06, "loss": 0.0, "num_input_tokens_seen": 9340784, "step": 33360 }, { "epoch": 370.72222222222223, "grad_norm": 6.053107313164219e-07, "learning_rate": 3.319320803990053e-06, "loss": 0.0, "num_input_tokens_seen": 9342192, "step": 33365 }, { "epoch": 370.77777777777777, "grad_norm": 2.5244438006666314e-07, "learning_rate": 3.3144342295030274e-06, "loss": 0.0, "num_input_tokens_seen": 9343568, "step": 33370 }, { "epoch": 370.8333333333333, "grad_norm": 1.7855500118457712e-05, "learning_rate": 3.309550999202765e-06, "loss": 0.0, "num_input_tokens_seen": 9344928, "step": 33375 }, { "epoch": 370.8888888888889, "grad_norm": 1.788000372471288e-05, "learning_rate": 3.3046711138423197e-06, "loss": 0.0, "num_input_tokens_seen": 9346336, "step": 33380 }, { "epoch": 370.94444444444446, "grad_norm": 2.1807441896726232e-07, "learning_rate": 3.2997945741742255e-06, "loss": 0.0, "num_input_tokens_seen": 9347776, "step": 33385 }, { "epoch": 371.0, "grad_norm": 2.708101192183676e-07, "learning_rate": 3.2949213809505082e-06, "loss": 0.0, "num_input_tokens_seen": 9349184, "step": 33390 }, { "epoch": 371.05555555555554, "grad_norm": 3.6979673723180895e-07, "learning_rate": 3.2900515349226834e-06, "loss": 0.0, "num_input_tokens_seen": 9350592, "step": 33395 }, { "epoch": 371.1111111111111, "grad_norm": 5.280006689645234e-07, "learning_rate": 3.285185036841731e-06, "loss": 0.0, "num_input_tokens_seen": 9352016, "step": 33400 }, { "epoch": 371.1111111111111, "eval_loss": 1.1642444133758545, "eval_runtime": 1.402, "eval_samples_per_second": 28.531, "eval_steps_per_second": 14.265, "num_input_tokens_seen": 9352016, "step": 33400 }, { "epoch": 371.1666666666667, "grad_norm": 7.902108336566016e-07, "learning_rate": 3.2803218874581377e-06, "loss": 0.0, "num_input_tokens_seen": 9353392, "step": 33405 }, { "epoch": 371.22222222222223, "grad_norm": 1.6076020870059438e-07, "learning_rate": 3.2754620875218494e-06, "loss": 0.0, "num_input_tokens_seen": 9354816, "step": 33410 }, { "epoch": 371.27777777777777, "grad_norm": 8.893385370356555e-07, "learning_rate": 3.2706056377823146e-06, "loss": 0.0, "num_input_tokens_seen": 9356160, "step": 33415 }, { "epoch": 371.3333333333333, "grad_norm": 4.627332145901164e-06, "learning_rate": 3.2657525389884647e-06, "loss": 0.0, "num_input_tokens_seen": 9357568, "step": 33420 }, { "epoch": 371.3888888888889, "grad_norm": 3.357247635449312e-07, "learning_rate": 3.260902791888698e-06, "loss": 0.0, "num_input_tokens_seen": 9358992, "step": 33425 }, { "epoch": 371.44444444444446, "grad_norm": 9.70055589277763e-06, "learning_rate": 3.2560563972309166e-06, "loss": 0.0, "num_input_tokens_seen": 9360432, "step": 33430 }, { "epoch": 371.5, "grad_norm": 6.485822723334422e-08, "learning_rate": 3.251213355762489e-06, "loss": 0.0, "num_input_tokens_seen": 9361840, "step": 33435 }, { "epoch": 371.55555555555554, "grad_norm": 1.8106031518527743e-07, "learning_rate": 3.2463736682302707e-06, "loss": 0.0, "num_input_tokens_seen": 9363232, "step": 33440 }, { "epoch": 371.6111111111111, "grad_norm": 7.06427556451672e-07, "learning_rate": 3.2415373353806124e-06, "loss": 0.0, "num_input_tokens_seen": 9364640, "step": 33445 }, { "epoch": 371.6666666666667, "grad_norm": 2.5006425858009607e-05, "learning_rate": 3.236704357959322e-06, "loss": 0.0, "num_input_tokens_seen": 9365984, "step": 33450 }, { "epoch": 371.72222222222223, "grad_norm": 4.3724725173888146e-07, "learning_rate": 3.2318747367117154e-06, "loss": 0.0, "num_input_tokens_seen": 9367376, "step": 33455 }, { "epoch": 371.77777777777777, "grad_norm": 4.3213171352363133e-07, "learning_rate": 3.227048472382585e-06, "loss": 0.0, "num_input_tokens_seen": 9368768, "step": 33460 }, { "epoch": 371.8333333333333, "grad_norm": 5.926270318923343e-07, "learning_rate": 3.2222255657161915e-06, "loss": 0.0, "num_input_tokens_seen": 9370192, "step": 33465 }, { "epoch": 371.8888888888889, "grad_norm": 5.543113843486935e-07, "learning_rate": 3.2174060174562924e-06, "loss": 0.0, "num_input_tokens_seen": 9371568, "step": 33470 }, { "epoch": 371.94444444444446, "grad_norm": 2.8494193884398555e-06, "learning_rate": 3.2125898283461298e-06, "loss": 0.0, "num_input_tokens_seen": 9372992, "step": 33475 }, { "epoch": 372.0, "grad_norm": 2.070963773803669e-07, "learning_rate": 3.207776999128406e-06, "loss": 0.0, "num_input_tokens_seen": 9374416, "step": 33480 }, { "epoch": 372.05555555555554, "grad_norm": 6.676345520872928e-08, "learning_rate": 3.202967530545331e-06, "loss": 0.0, "num_input_tokens_seen": 9375824, "step": 33485 }, { "epoch": 372.1111111111111, "grad_norm": 1.5868283753661672e-06, "learning_rate": 3.1981614233385778e-06, "loss": 0.0, "num_input_tokens_seen": 9377168, "step": 33490 }, { "epoch": 372.1666666666667, "grad_norm": 1.615778160157788e-07, "learning_rate": 3.1933586782493115e-06, "loss": 0.0, "num_input_tokens_seen": 9378560, "step": 33495 }, { "epoch": 372.22222222222223, "grad_norm": 3.982935368185281e-07, "learning_rate": 3.188559296018184e-06, "loss": 0.0, "num_input_tokens_seen": 9379936, "step": 33500 }, { "epoch": 372.27777777777777, "grad_norm": 6.266717718972359e-06, "learning_rate": 3.1837632773853098e-06, "loss": 0.0, "num_input_tokens_seen": 9381312, "step": 33505 }, { "epoch": 372.3333333333333, "grad_norm": 1.0417700195830548e-06, "learning_rate": 3.178970623090294e-06, "loss": 0.0, "num_input_tokens_seen": 9382688, "step": 33510 }, { "epoch": 372.3888888888889, "grad_norm": 6.279179160628701e-06, "learning_rate": 3.174181333872234e-06, "loss": 0.0, "num_input_tokens_seen": 9384048, "step": 33515 }, { "epoch": 372.44444444444446, "grad_norm": 7.82179176894715e-06, "learning_rate": 3.169395410469686e-06, "loss": 0.0, "num_input_tokens_seen": 9385472, "step": 33520 }, { "epoch": 372.5, "grad_norm": 3.295297688055143e-07, "learning_rate": 3.164612853620713e-06, "loss": 0.0, "num_input_tokens_seen": 9386896, "step": 33525 }, { "epoch": 372.55555555555554, "grad_norm": 7.883780881456914e-07, "learning_rate": 3.1598336640628333e-06, "loss": 0.0, "num_input_tokens_seen": 9388272, "step": 33530 }, { "epoch": 372.6111111111111, "grad_norm": 5.280998607304355e-07, "learning_rate": 3.155057842533063e-06, "loss": 0.0, "num_input_tokens_seen": 9389696, "step": 33535 }, { "epoch": 372.6666666666667, "grad_norm": 6.407294677046593e-07, "learning_rate": 3.1502853897678984e-06, "loss": 0.0, "num_input_tokens_seen": 9391104, "step": 33540 }, { "epoch": 372.72222222222223, "grad_norm": 5.054193366049731e-07, "learning_rate": 3.1455163065033017e-06, "loss": 0.0, "num_input_tokens_seen": 9392512, "step": 33545 }, { "epoch": 372.77777777777777, "grad_norm": 2.6298306465832866e-07, "learning_rate": 3.140750593474734e-06, "loss": 0.0, "num_input_tokens_seen": 9393872, "step": 33550 }, { "epoch": 372.8333333333333, "grad_norm": 3.1510029430137365e-07, "learning_rate": 3.1359882514171294e-06, "loss": 0.0, "num_input_tokens_seen": 9395280, "step": 33555 }, { "epoch": 372.8888888888889, "grad_norm": 2.4516180019418243e-07, "learning_rate": 3.1312292810648903e-06, "loss": 0.0, "num_input_tokens_seen": 9396704, "step": 33560 }, { "epoch": 372.94444444444446, "grad_norm": 1.3657514728038223e-06, "learning_rate": 3.1264736831519204e-06, "loss": 0.0, "num_input_tokens_seen": 9398144, "step": 33565 }, { "epoch": 373.0, "grad_norm": 8.675906428834423e-06, "learning_rate": 3.1217214584115863e-06, "loss": 0.0, "num_input_tokens_seen": 9399584, "step": 33570 }, { "epoch": 373.05555555555554, "grad_norm": 6.4560945247649215e-06, "learning_rate": 3.116972607576746e-06, "loss": 0.0, "num_input_tokens_seen": 9400992, "step": 33575 }, { "epoch": 373.1111111111111, "grad_norm": 2.418180201857467e-06, "learning_rate": 3.1122271313797303e-06, "loss": 0.0, "num_input_tokens_seen": 9402368, "step": 33580 }, { "epoch": 373.1666666666667, "grad_norm": 1.9496164327392762e-07, "learning_rate": 3.107485030552343e-06, "loss": 0.0, "num_input_tokens_seen": 9403744, "step": 33585 }, { "epoch": 373.22222222222223, "grad_norm": 7.636157533852383e-06, "learning_rate": 3.1027463058258848e-06, "loss": 0.0, "num_input_tokens_seen": 9405136, "step": 33590 }, { "epoch": 373.27777777777777, "grad_norm": 3.0021312795724953e-06, "learning_rate": 3.0980109579311273e-06, "loss": 0.0, "num_input_tokens_seen": 9406512, "step": 33595 }, { "epoch": 373.3333333333333, "grad_norm": 4.280369694242836e-07, "learning_rate": 3.093278987598314e-06, "loss": 0.0, "num_input_tokens_seen": 9407968, "step": 33600 }, { "epoch": 373.3333333333333, "eval_loss": 1.1804169416427612, "eval_runtime": 1.3913, "eval_samples_per_second": 28.749, "eval_steps_per_second": 14.375, "num_input_tokens_seen": 9407968, "step": 33600 }, { "epoch": 373.3888888888889, "grad_norm": 1.375892253463462e-07, "learning_rate": 3.0885503955571826e-06, "loss": 0.0, "num_input_tokens_seen": 9409376, "step": 33605 }, { "epoch": 373.44444444444446, "grad_norm": 7.761021265650925e-07, "learning_rate": 3.0838251825369313e-06, "loss": 0.0, "num_input_tokens_seen": 9410752, "step": 33610 }, { "epoch": 373.5, "grad_norm": 1.600578980287537e-05, "learning_rate": 3.0791033492662517e-06, "loss": 0.0, "num_input_tokens_seen": 9412144, "step": 33615 }, { "epoch": 373.55555555555554, "grad_norm": 6.1249306781974155e-06, "learning_rate": 3.0743848964733203e-06, "loss": 0.0, "num_input_tokens_seen": 9413552, "step": 33620 }, { "epoch": 373.6111111111111, "grad_norm": 1.903950277437616e-07, "learning_rate": 3.0696698248857625e-06, "loss": 0.0, "num_input_tokens_seen": 9414912, "step": 33625 }, { "epoch": 373.6666666666667, "grad_norm": 1.8709688447415829e-06, "learning_rate": 3.0649581352307192e-06, "loss": 0.0, "num_input_tokens_seen": 9416336, "step": 33630 }, { "epoch": 373.72222222222223, "grad_norm": 7.4515205596981104e-06, "learning_rate": 3.060249828234776e-06, "loss": 0.0, "num_input_tokens_seen": 9417728, "step": 33635 }, { "epoch": 373.77777777777777, "grad_norm": 1.0659356348696747e-06, "learning_rate": 3.055544904624025e-06, "loss": 0.0, "num_input_tokens_seen": 9419120, "step": 33640 }, { "epoch": 373.8333333333333, "grad_norm": 4.720234755950514e-06, "learning_rate": 3.050843365124026e-06, "loss": 0.0, "num_input_tokens_seen": 9420560, "step": 33645 }, { "epoch": 373.8888888888889, "grad_norm": 5.13189161210903e-06, "learning_rate": 3.0461452104598083e-06, "loss": 0.0, "num_input_tokens_seen": 9421904, "step": 33650 }, { "epoch": 373.94444444444446, "grad_norm": 9.571738701197319e-06, "learning_rate": 3.0414504413558836e-06, "loss": 0.0, "num_input_tokens_seen": 9423312, "step": 33655 }, { "epoch": 374.0, "grad_norm": 6.414405362420439e-08, "learning_rate": 3.0367590585362564e-06, "loss": 0.0, "num_input_tokens_seen": 9424752, "step": 33660 }, { "epoch": 374.05555555555554, "grad_norm": 9.174385695587262e-07, "learning_rate": 3.0320710627243813e-06, "loss": 0.0, "num_input_tokens_seen": 9426208, "step": 33665 }, { "epoch": 374.1111111111111, "grad_norm": 2.1036394173279405e-05, "learning_rate": 3.027386454643222e-06, "loss": 0.0, "num_input_tokens_seen": 9427552, "step": 33670 }, { "epoch": 374.1666666666667, "grad_norm": 2.325360219401773e-05, "learning_rate": 3.0227052350151914e-06, "loss": 0.0, "num_input_tokens_seen": 9428928, "step": 33675 }, { "epoch": 374.22222222222223, "grad_norm": 7.535119266321999e-07, "learning_rate": 3.0180274045621957e-06, "loss": 0.0, "num_input_tokens_seen": 9430288, "step": 33680 }, { "epoch": 374.27777777777777, "grad_norm": 3.736509768259566e-08, "learning_rate": 3.013352964005625e-06, "loss": 0.0, "num_input_tokens_seen": 9431712, "step": 33685 }, { "epoch": 374.3333333333333, "grad_norm": 5.075658577879949e-07, "learning_rate": 3.0086819140663218e-06, "loss": 0.0, "num_input_tokens_seen": 9433104, "step": 33690 }, { "epoch": 374.3888888888889, "grad_norm": 9.660752766649239e-06, "learning_rate": 3.0040142554646265e-06, "loss": 0.0, "num_input_tokens_seen": 9434496, "step": 33695 }, { "epoch": 374.44444444444446, "grad_norm": 3.8660999734929646e-07, "learning_rate": 2.999349988920361e-06, "loss": 0.0, "num_input_tokens_seen": 9435888, "step": 33700 }, { "epoch": 374.5, "grad_norm": 2.8695701530523365e-06, "learning_rate": 2.994689115152796e-06, "loss": 0.0, "num_input_tokens_seen": 9437296, "step": 33705 }, { "epoch": 374.55555555555554, "grad_norm": 5.331339707481675e-06, "learning_rate": 2.9900316348807105e-06, "loss": 0.0, "num_input_tokens_seen": 9438688, "step": 33710 }, { "epoch": 374.6111111111111, "grad_norm": 3.5330850778336753e-07, "learning_rate": 2.985377548822338e-06, "loss": 0.0, "num_input_tokens_seen": 9440096, "step": 33715 }, { "epoch": 374.6666666666667, "grad_norm": 1.6215867617574986e-06, "learning_rate": 2.980726857695404e-06, "loss": 0.0, "num_input_tokens_seen": 9441488, "step": 33720 }, { "epoch": 374.72222222222223, "grad_norm": 5.324054654920474e-07, "learning_rate": 2.9760795622171017e-06, "loss": 0.0, "num_input_tokens_seen": 9442880, "step": 33725 }, { "epoch": 374.77777777777777, "grad_norm": 6.424437515306636e-07, "learning_rate": 2.971435663104094e-06, "loss": 0.0, "num_input_tokens_seen": 9444288, "step": 33730 }, { "epoch": 374.8333333333333, "grad_norm": 1.0187203258738009e-07, "learning_rate": 2.9667951610725385e-06, "loss": 0.0, "num_input_tokens_seen": 9445664, "step": 33735 }, { "epoch": 374.8888888888889, "grad_norm": 6.716740585943626e-07, "learning_rate": 2.9621580568380575e-06, "loss": 0.0, "num_input_tokens_seen": 9447008, "step": 33740 }, { "epoch": 374.94444444444446, "grad_norm": 2.4370563096454134e-06, "learning_rate": 2.9575243511157453e-06, "loss": 0.0, "num_input_tokens_seen": 9448464, "step": 33745 }, { "epoch": 375.0, "grad_norm": 1.2526568582416076e-07, "learning_rate": 2.952894044620186e-06, "loss": 0.0, "num_input_tokens_seen": 9449888, "step": 33750 }, { "epoch": 375.05555555555554, "grad_norm": 8.747442734602373e-07, "learning_rate": 2.948267138065419e-06, "loss": 0.0, "num_input_tokens_seen": 9451264, "step": 33755 }, { "epoch": 375.1111111111111, "grad_norm": 5.868013772669656e-07, "learning_rate": 2.943643632164983e-06, "loss": 0.0, "num_input_tokens_seen": 9452624, "step": 33760 }, { "epoch": 375.1666666666667, "grad_norm": 1.979730086532072e-06, "learning_rate": 2.939023527631879e-06, "loss": 0.0, "num_input_tokens_seen": 9454096, "step": 33765 }, { "epoch": 375.22222222222223, "grad_norm": 1.835582565945515e-07, "learning_rate": 2.934406825178576e-06, "loss": 0.0, "num_input_tokens_seen": 9455504, "step": 33770 }, { "epoch": 375.27777777777777, "grad_norm": 4.38995101603723e-07, "learning_rate": 2.9297935255170357e-06, "loss": 0.0, "num_input_tokens_seen": 9456912, "step": 33775 }, { "epoch": 375.3333333333333, "grad_norm": 6.489974111900665e-06, "learning_rate": 2.925183629358691e-06, "loss": 0.0, "num_input_tokens_seen": 9458304, "step": 33780 }, { "epoch": 375.3888888888889, "grad_norm": 6.065746447347919e-07, "learning_rate": 2.9205771374144346e-06, "loss": 0.0, "num_input_tokens_seen": 9459744, "step": 33785 }, { "epoch": 375.44444444444446, "grad_norm": 1.3073567117771745e-07, "learning_rate": 2.915974050394657e-06, "loss": 0.0, "num_input_tokens_seen": 9461120, "step": 33790 }, { "epoch": 375.5, "grad_norm": 9.665169500294724e-07, "learning_rate": 2.9113743690092067e-06, "loss": 0.0, "num_input_tokens_seen": 9462512, "step": 33795 }, { "epoch": 375.55555555555554, "grad_norm": 2.8045346880389843e-06, "learning_rate": 2.906778093967402e-06, "loss": 0.0, "num_input_tokens_seen": 9463920, "step": 33800 }, { "epoch": 375.55555555555554, "eval_loss": 1.1850050687789917, "eval_runtime": 1.3897, "eval_samples_per_second": 28.783, "eval_steps_per_second": 14.391, "num_input_tokens_seen": 9463920, "step": 33800 }, { "epoch": 375.6111111111111, "grad_norm": 2.1817530068801716e-05, "learning_rate": 2.9021852259780656e-06, "loss": 0.0, "num_input_tokens_seen": 9465312, "step": 33805 }, { "epoch": 375.6666666666667, "grad_norm": 2.218919917140738e-06, "learning_rate": 2.8975957657494583e-06, "loss": 0.0, "num_input_tokens_seen": 9466736, "step": 33810 }, { "epoch": 375.72222222222223, "grad_norm": 5.183291705179727e-06, "learning_rate": 2.8930097139893417e-06, "loss": 0.0, "num_input_tokens_seen": 9468112, "step": 33815 }, { "epoch": 375.77777777777777, "grad_norm": 2.0096159403237834e-07, "learning_rate": 2.888427071404945e-06, "loss": 0.0, "num_input_tokens_seen": 9469536, "step": 33820 }, { "epoch": 375.8333333333333, "grad_norm": 5.205446882428078e-07, "learning_rate": 2.8838478387029606e-06, "loss": 0.0, "num_input_tokens_seen": 9470928, "step": 33825 }, { "epoch": 375.8888888888889, "grad_norm": 6.419453058015279e-08, "learning_rate": 2.8792720165895737e-06, "loss": 0.0, "num_input_tokens_seen": 9472304, "step": 33830 }, { "epoch": 375.94444444444446, "grad_norm": 1.9136479750159197e-05, "learning_rate": 2.874699605770423e-06, "loss": 0.0, "num_input_tokens_seen": 9473712, "step": 33835 }, { "epoch": 376.0, "grad_norm": 2.3006312233064818e-07, "learning_rate": 2.8701306069506383e-06, "loss": 0.0, "num_input_tokens_seen": 9475136, "step": 33840 }, { "epoch": 376.05555555555554, "grad_norm": 1.5791390239883185e-07, "learning_rate": 2.8655650208348178e-06, "loss": 0.0, "num_input_tokens_seen": 9476544, "step": 33845 }, { "epoch": 376.1111111111111, "grad_norm": 7.4752078944584355e-06, "learning_rate": 2.8610028481270257e-06, "loss": 0.0, "num_input_tokens_seen": 9477936, "step": 33850 }, { "epoch": 376.1666666666667, "grad_norm": 8.623397889095941e-07, "learning_rate": 2.856444089530813e-06, "loss": 0.0, "num_input_tokens_seen": 9479312, "step": 33855 }, { "epoch": 376.22222222222223, "grad_norm": 1.1035315594654094e-07, "learning_rate": 2.8518887457491955e-06, "loss": 0.0, "num_input_tokens_seen": 9480688, "step": 33860 }, { "epoch": 376.27777777777777, "grad_norm": 7.784497574903071e-05, "learning_rate": 2.8473368174846666e-06, "loss": 0.0, "num_input_tokens_seen": 9482128, "step": 33865 }, { "epoch": 376.3333333333333, "grad_norm": 9.896122719510458e-06, "learning_rate": 2.842788305439184e-06, "loss": 0.0, "num_input_tokens_seen": 9483568, "step": 33870 }, { "epoch": 376.3888888888889, "grad_norm": 8.407133833543412e-08, "learning_rate": 2.8382432103141925e-06, "loss": 0.0, "num_input_tokens_seen": 9484928, "step": 33875 }, { "epoch": 376.44444444444446, "grad_norm": 1.526279334029823e-06, "learning_rate": 2.833701532810598e-06, "loss": 0.0, "num_input_tokens_seen": 9486336, "step": 33880 }, { "epoch": 376.5, "grad_norm": 1.12362286017742e-05, "learning_rate": 2.8291632736287877e-06, "loss": 0.0, "num_input_tokens_seen": 9487744, "step": 33885 }, { "epoch": 376.55555555555554, "grad_norm": 2.250139914394822e-06, "learning_rate": 2.824628433468615e-06, "loss": 0.0, "num_input_tokens_seen": 9489088, "step": 33890 }, { "epoch": 376.6111111111111, "grad_norm": 4.447120716122299e-07, "learning_rate": 2.8200970130294073e-06, "loss": 0.0, "num_input_tokens_seen": 9490480, "step": 33895 }, { "epoch": 376.6666666666667, "grad_norm": 1.3143697287887335e-05, "learning_rate": 2.8155690130099775e-06, "loss": 0.0, "num_input_tokens_seen": 9491872, "step": 33900 }, { "epoch": 376.72222222222223, "grad_norm": 3.683609293148038e-06, "learning_rate": 2.8110444341085895e-06, "loss": 0.0, "num_input_tokens_seen": 9493280, "step": 33905 }, { "epoch": 376.77777777777777, "grad_norm": 1.2040048602557363e-07, "learning_rate": 2.806523277022996e-06, "loss": 0.0, "num_input_tokens_seen": 9494672, "step": 33910 }, { "epoch": 376.8333333333333, "grad_norm": 6.38300946320669e-07, "learning_rate": 2.802005542450409e-06, "loss": 0.0, "num_input_tokens_seen": 9496080, "step": 33915 }, { "epoch": 376.8888888888889, "grad_norm": 3.6796288895857288e-06, "learning_rate": 2.797491231087526e-06, "loss": 0.0, "num_input_tokens_seen": 9497504, "step": 33920 }, { "epoch": 376.94444444444446, "grad_norm": 3.170614490954904e-07, "learning_rate": 2.7929803436305137e-06, "loss": 0.0, "num_input_tokens_seen": 9498928, "step": 33925 }, { "epoch": 377.0, "grad_norm": 5.389813964029599e-07, "learning_rate": 2.788472880774998e-06, "loss": 0.0, "num_input_tokens_seen": 9500336, "step": 33930 }, { "epoch": 377.05555555555554, "grad_norm": 8.755567932894337e-07, "learning_rate": 2.7839688432160977e-06, "loss": 0.0, "num_input_tokens_seen": 9501760, "step": 33935 }, { "epoch": 377.1111111111111, "grad_norm": 1.897070234235798e-07, "learning_rate": 2.779468231648383e-06, "loss": 0.0, "num_input_tokens_seen": 9503136, "step": 33940 }, { "epoch": 377.1666666666667, "grad_norm": 2.1499836577731912e-07, "learning_rate": 2.774971046765906e-06, "loss": 0.0, "num_input_tokens_seen": 9504560, "step": 33945 }, { "epoch": 377.22222222222223, "grad_norm": 6.779467867090716e-07, "learning_rate": 2.770477289262194e-06, "loss": 0.0, "num_input_tokens_seen": 9505936, "step": 33950 }, { "epoch": 377.27777777777777, "grad_norm": 7.590439849991526e-07, "learning_rate": 2.765986959830233e-06, "loss": 0.0, "num_input_tokens_seen": 9507328, "step": 33955 }, { "epoch": 377.3333333333333, "grad_norm": 1.2630188166440348e-06, "learning_rate": 2.761500059162492e-06, "loss": 0.0, "num_input_tokens_seen": 9508768, "step": 33960 }, { "epoch": 377.3888888888889, "grad_norm": 6.485694825641986e-07, "learning_rate": 2.757016587950914e-06, "loss": 0.0, "num_input_tokens_seen": 9510112, "step": 33965 }, { "epoch": 377.44444444444446, "grad_norm": 8.996881661005318e-05, "learning_rate": 2.752536546886897e-06, "loss": 0.0, "num_input_tokens_seen": 9511536, "step": 33970 }, { "epoch": 377.5, "grad_norm": 1.4804453485339764e-07, "learning_rate": 2.7480599366613234e-06, "loss": 0.0, "num_input_tokens_seen": 9512976, "step": 33975 }, { "epoch": 377.55555555555554, "grad_norm": 9.9463909464248e-07, "learning_rate": 2.7435867579645473e-06, "loss": 0.0, "num_input_tokens_seen": 9514368, "step": 33980 }, { "epoch": 377.6111111111111, "grad_norm": 2.1169582851143787e-06, "learning_rate": 2.739117011486378e-06, "loss": 0.0, "num_input_tokens_seen": 9515776, "step": 33985 }, { "epoch": 377.6666666666667, "grad_norm": 2.4866832859515853e-07, "learning_rate": 2.7346506979161216e-06, "loss": 0.0, "num_input_tokens_seen": 9517152, "step": 33990 }, { "epoch": 377.72222222222223, "grad_norm": 6.149407028033238e-08, "learning_rate": 2.7301878179425227e-06, "loss": 0.0, "num_input_tokens_seen": 9518544, "step": 33995 }, { "epoch": 377.77777777777777, "grad_norm": 5.384300038713263e-06, "learning_rate": 2.7257283722538244e-06, "loss": 0.0, "num_input_tokens_seen": 9519984, "step": 34000 }, { "epoch": 377.77777777777777, "eval_loss": 1.1751899719238281, "eval_runtime": 1.3915, "eval_samples_per_second": 28.747, "eval_steps_per_second": 14.373, "num_input_tokens_seen": 9519984, "step": 34000 }, { "epoch": 377.8333333333333, "grad_norm": 5.660392048412177e-07, "learning_rate": 2.7212723615377326e-06, "loss": 0.0, "num_input_tokens_seen": 9521376, "step": 34005 }, { "epoch": 377.8888888888889, "grad_norm": 3.3549110867170384e-07, "learning_rate": 2.7168197864814145e-06, "loss": 0.0, "num_input_tokens_seen": 9522784, "step": 34010 }, { "epoch": 377.94444444444446, "grad_norm": 2.0912226261771139e-07, "learning_rate": 2.712370647771509e-06, "loss": 0.0, "num_input_tokens_seen": 9524192, "step": 34015 }, { "epoch": 378.0, "grad_norm": 3.682735183474506e-08, "learning_rate": 2.707924946094137e-06, "loss": 0.0, "num_input_tokens_seen": 9525568, "step": 34020 }, { "epoch": 378.05555555555554, "grad_norm": 1.113640166749974e-07, "learning_rate": 2.7034826821348723e-06, "loss": 0.0, "num_input_tokens_seen": 9526944, "step": 34025 }, { "epoch": 378.1111111111111, "grad_norm": 1.8222964399683406e-06, "learning_rate": 2.6990438565787786e-06, "loss": 0.0, "num_input_tokens_seen": 9528272, "step": 34030 }, { "epoch": 378.1666666666667, "grad_norm": 1.209547093594665e-07, "learning_rate": 2.6946084701103714e-06, "loss": 0.0, "num_input_tokens_seen": 9529680, "step": 34035 }, { "epoch": 378.22222222222223, "grad_norm": 7.477317467419198e-06, "learning_rate": 2.6901765234136428e-06, "loss": 0.0, "num_input_tokens_seen": 9531104, "step": 34040 }, { "epoch": 378.27777777777777, "grad_norm": 3.37553137796931e-05, "learning_rate": 2.685748017172063e-06, "loss": 0.0, "num_input_tokens_seen": 9532512, "step": 34045 }, { "epoch": 378.3333333333333, "grad_norm": 9.715139412946883e-07, "learning_rate": 2.681322952068549e-06, "loss": 0.0, "num_input_tokens_seen": 9533920, "step": 34050 }, { "epoch": 378.3888888888889, "grad_norm": 8.399027819905314e-07, "learning_rate": 2.6769013287855137e-06, "loss": 0.0, "num_input_tokens_seen": 9535344, "step": 34055 }, { "epoch": 378.44444444444446, "grad_norm": 8.100184345494199e-08, "learning_rate": 2.6724831480048286e-06, "loss": 0.0, "num_input_tokens_seen": 9536704, "step": 34060 }, { "epoch": 378.5, "grad_norm": 1.830557630455587e-05, "learning_rate": 2.66806841040782e-06, "loss": 0.0, "num_input_tokens_seen": 9538128, "step": 34065 }, { "epoch": 378.55555555555554, "grad_norm": 3.65466735274822e-06, "learning_rate": 2.6636571166753083e-06, "loss": 0.0, "num_input_tokens_seen": 9539552, "step": 34070 }, { "epoch": 378.6111111111111, "grad_norm": 1.2368282114039175e-06, "learning_rate": 2.6592492674875598e-06, "loss": 0.0, "num_input_tokens_seen": 9540960, "step": 34075 }, { "epoch": 378.6666666666667, "grad_norm": 1.1652831233277539e-07, "learning_rate": 2.6548448635243305e-06, "loss": 0.0, "num_input_tokens_seen": 9542336, "step": 34080 }, { "epoch": 378.72222222222223, "grad_norm": 2.800896652388474e-07, "learning_rate": 2.650443905464828e-06, "loss": 0.0, "num_input_tokens_seen": 9543712, "step": 34085 }, { "epoch": 378.77777777777777, "grad_norm": 3.7843449263164075e-07, "learning_rate": 2.646046393987739e-06, "loss": 0.0, "num_input_tokens_seen": 9545104, "step": 34090 }, { "epoch": 378.8333333333333, "grad_norm": 1.1340945320625906e-06, "learning_rate": 2.64165232977121e-06, "loss": 0.0, "num_input_tokens_seen": 9546512, "step": 34095 }, { "epoch": 378.8888888888889, "grad_norm": 6.599146900043706e-07, "learning_rate": 2.6372617134928695e-06, "loss": 0.0, "num_input_tokens_seen": 9547920, "step": 34100 }, { "epoch": 378.94444444444446, "grad_norm": 2.700506627206778e-07, "learning_rate": 2.6328745458297943e-06, "loss": 0.0, "num_input_tokens_seen": 9549328, "step": 34105 }, { "epoch": 379.0, "grad_norm": 3.498524279166304e-07, "learning_rate": 2.6284908274585546e-06, "loss": 0.0, "num_input_tokens_seen": 9550704, "step": 34110 }, { "epoch": 379.05555555555554, "grad_norm": 2.2551143047166988e-05, "learning_rate": 2.6241105590551595e-06, "loss": 0.0, "num_input_tokens_seen": 9552128, "step": 34115 }, { "epoch": 379.1111111111111, "grad_norm": 1.88044361948414e-06, "learning_rate": 2.6197337412951105e-06, "loss": 0.0, "num_input_tokens_seen": 9553568, "step": 34120 }, { "epoch": 379.1666666666667, "grad_norm": 8.911873692341032e-07, "learning_rate": 2.6153603748533705e-06, "loss": 0.0, "num_input_tokens_seen": 9554976, "step": 34125 }, { "epoch": 379.22222222222223, "grad_norm": 3.014434696524404e-06, "learning_rate": 2.6109904604043585e-06, "loss": 0.0, "num_input_tokens_seen": 9556416, "step": 34130 }, { "epoch": 379.27777777777777, "grad_norm": 2.3594718356889643e-07, "learning_rate": 2.6066239986219765e-06, "loss": 0.0, "num_input_tokens_seen": 9557824, "step": 34135 }, { "epoch": 379.3333333333333, "grad_norm": 7.830060440028319e-07, "learning_rate": 2.602260990179592e-06, "loss": 0.0, "num_input_tokens_seen": 9559216, "step": 34140 }, { "epoch": 379.3888888888889, "grad_norm": 1.0819358976732474e-05, "learning_rate": 2.5979014357500248e-06, "loss": 0.0, "num_input_tokens_seen": 9560624, "step": 34145 }, { "epoch": 379.44444444444446, "grad_norm": 3.491375991870882e-06, "learning_rate": 2.5935453360055844e-06, "loss": 0.0, "num_input_tokens_seen": 9562032, "step": 34150 }, { "epoch": 379.5, "grad_norm": 8.210989221879572e-07, "learning_rate": 2.5891926916180283e-06, "loss": 0.0, "num_input_tokens_seen": 9563376, "step": 34155 }, { "epoch": 379.55555555555554, "grad_norm": 1.6209283160151244e-07, "learning_rate": 2.5848435032585883e-06, "loss": 0.0, "num_input_tokens_seen": 9564736, "step": 34160 }, { "epoch": 379.6111111111111, "grad_norm": 7.872520200180588e-07, "learning_rate": 2.58049777159797e-06, "loss": 0.0, "num_input_tokens_seen": 9566096, "step": 34165 }, { "epoch": 379.6666666666667, "grad_norm": 1.5781699858052889e-06, "learning_rate": 2.576155497306332e-06, "loss": 0.0, "num_input_tokens_seen": 9567504, "step": 34170 }, { "epoch": 379.72222222222223, "grad_norm": 2.1309354281129345e-07, "learning_rate": 2.57181668105331e-06, "loss": 0.0, "num_input_tokens_seen": 9568896, "step": 34175 }, { "epoch": 379.77777777777777, "grad_norm": 5.5056966630218085e-06, "learning_rate": 2.567481323508014e-06, "loss": 0.0, "num_input_tokens_seen": 9570304, "step": 34180 }, { "epoch": 379.8333333333333, "grad_norm": 9.429831493434904e-07, "learning_rate": 2.5631494253389954e-06, "loss": 0.0, "num_input_tokens_seen": 9571680, "step": 34185 }, { "epoch": 379.8888888888889, "grad_norm": 6.211868708305701e-07, "learning_rate": 2.5588209872142997e-06, "loss": 0.0, "num_input_tokens_seen": 9573104, "step": 34190 }, { "epoch": 379.94444444444446, "grad_norm": 4.925852863379987e-07, "learning_rate": 2.5544960098014186e-06, "loss": 0.0, "num_input_tokens_seen": 9574544, "step": 34195 }, { "epoch": 380.0, "grad_norm": 1.8005003994403523e-06, "learning_rate": 2.550174493767318e-06, "loss": 0.0, "num_input_tokens_seen": 9575936, "step": 34200 }, { "epoch": 380.0, "eval_loss": 1.2010749578475952, "eval_runtime": 1.3848, "eval_samples_per_second": 28.885, "eval_steps_per_second": 14.443, "num_input_tokens_seen": 9575936, "step": 34200 }, { "epoch": 380.05555555555554, "grad_norm": 1.3963750689072185e-06, "learning_rate": 2.545856439778438e-06, "loss": 0.0, "num_input_tokens_seen": 9577312, "step": 34205 }, { "epoch": 380.1111111111111, "grad_norm": 1.1704297548931208e-06, "learning_rate": 2.541541848500667e-06, "loss": 0.0, "num_input_tokens_seen": 9578704, "step": 34210 }, { "epoch": 380.1666666666667, "grad_norm": 2.6329649926992715e-07, "learning_rate": 2.5372307205993733e-06, "loss": 0.0, "num_input_tokens_seen": 9580112, "step": 34215 }, { "epoch": 380.22222222222223, "grad_norm": 2.798526793412748e-06, "learning_rate": 2.5329230567393917e-06, "loss": 0.0, "num_input_tokens_seen": 9581536, "step": 34220 }, { "epoch": 380.27777777777777, "grad_norm": 1.9952958041358215e-07, "learning_rate": 2.5286188575850164e-06, "loss": 0.0, "num_input_tokens_seen": 9582944, "step": 34225 }, { "epoch": 380.3333333333333, "grad_norm": 4.424491066856717e-07, "learning_rate": 2.5243181237999984e-06, "loss": 0.0, "num_input_tokens_seen": 9584352, "step": 34230 }, { "epoch": 380.3888888888889, "grad_norm": 3.8055480899856775e-07, "learning_rate": 2.520020856047578e-06, "loss": 0.0, "num_input_tokens_seen": 9585776, "step": 34235 }, { "epoch": 380.44444444444446, "grad_norm": 3.9014437902551435e-07, "learning_rate": 2.515727054990438e-06, "loss": 0.0, "num_input_tokens_seen": 9587136, "step": 34240 }, { "epoch": 380.5, "grad_norm": 7.74365762481466e-05, "learning_rate": 2.511436721290747e-06, "loss": 0.0, "num_input_tokens_seen": 9588576, "step": 34245 }, { "epoch": 380.55555555555554, "grad_norm": 4.3504249447323673e-07, "learning_rate": 2.5071498556101164e-06, "loss": 0.0, "num_input_tokens_seen": 9589968, "step": 34250 }, { "epoch": 380.6111111111111, "grad_norm": 2.069217543976265e-06, "learning_rate": 2.5028664586096485e-06, "loss": 0.0, "num_input_tokens_seen": 9591344, "step": 34255 }, { "epoch": 380.6666666666667, "grad_norm": 1.7548820324009284e-05, "learning_rate": 2.498586530949881e-06, "loss": 0.0, "num_input_tokens_seen": 9592752, "step": 34260 }, { "epoch": 380.72222222222223, "grad_norm": 3.805535797596349e-08, "learning_rate": 2.4943100732908427e-06, "loss": 0.0, "num_input_tokens_seen": 9594208, "step": 34265 }, { "epoch": 380.77777777777777, "grad_norm": 2.712709488150722e-07, "learning_rate": 2.4900370862920188e-06, "loss": 0.0, "num_input_tokens_seen": 9595568, "step": 34270 }, { "epoch": 380.8333333333333, "grad_norm": 3.716231162798067e-07, "learning_rate": 2.4857675706123518e-06, "loss": 0.0, "num_input_tokens_seen": 9596944, "step": 34275 }, { "epoch": 380.8888888888889, "grad_norm": 8.104310609269305e-07, "learning_rate": 2.4815015269102543e-06, "loss": 0.0, "num_input_tokens_seen": 9598336, "step": 34280 }, { "epoch": 380.94444444444446, "grad_norm": 2.5640456442488357e-05, "learning_rate": 2.477238955843611e-06, "loss": 0.0, "num_input_tokens_seen": 9599776, "step": 34285 }, { "epoch": 381.0, "grad_norm": 5.850548063790484e-07, "learning_rate": 2.4729798580697573e-06, "loss": 0.0, "num_input_tokens_seen": 9601152, "step": 34290 }, { "epoch": 381.05555555555554, "grad_norm": 7.996771955731674e-07, "learning_rate": 2.4687242342455034e-06, "loss": 0.0, "num_input_tokens_seen": 9602512, "step": 34295 }, { "epoch": 381.1111111111111, "grad_norm": 2.1038999875599984e-06, "learning_rate": 2.4644720850271196e-06, "loss": 0.0, "num_input_tokens_seen": 9603920, "step": 34300 }, { "epoch": 381.1666666666667, "grad_norm": 8.838916301101563e-07, "learning_rate": 2.4602234110703364e-06, "loss": 0.0, "num_input_tokens_seen": 9605344, "step": 34305 }, { "epoch": 381.22222222222223, "grad_norm": 5.09140932081209e-07, "learning_rate": 2.4559782130303576e-06, "loss": 0.0, "num_input_tokens_seen": 9606752, "step": 34310 }, { "epoch": 381.27777777777777, "grad_norm": 5.788188559563423e-07, "learning_rate": 2.451736491561843e-06, "loss": 0.0, "num_input_tokens_seen": 9608176, "step": 34315 }, { "epoch": 381.3333333333333, "grad_norm": 5.865298362550675e-07, "learning_rate": 2.4474982473189163e-06, "loss": 0.0, "num_input_tokens_seen": 9609552, "step": 34320 }, { "epoch": 381.3888888888889, "grad_norm": 1.8666153209778713e-06, "learning_rate": 2.4432634809551796e-06, "loss": 0.0, "num_input_tokens_seen": 9610944, "step": 34325 }, { "epoch": 381.44444444444446, "grad_norm": 8.782122904449352e-07, "learning_rate": 2.439032193123675e-06, "loss": 0.0, "num_input_tokens_seen": 9612352, "step": 34330 }, { "epoch": 381.5, "grad_norm": 2.932123379650875e-06, "learning_rate": 2.4348043844769297e-06, "loss": 0.0, "num_input_tokens_seen": 9613760, "step": 34335 }, { "epoch": 381.55555555555554, "grad_norm": 1.5295279354177183e-07, "learning_rate": 2.4305800556669146e-06, "loss": 0.0, "num_input_tokens_seen": 9615136, "step": 34340 }, { "epoch": 381.6111111111111, "grad_norm": 1.4561899206455564e-06, "learning_rate": 2.426359207345083e-06, "loss": 0.0, "num_input_tokens_seen": 9616544, "step": 34345 }, { "epoch": 381.6666666666667, "grad_norm": 7.861480071369442e-07, "learning_rate": 2.4221418401623396e-06, "loss": 0.0, "num_input_tokens_seen": 9617936, "step": 34350 }, { "epoch": 381.72222222222223, "grad_norm": 6.833988095422683e-07, "learning_rate": 2.4179279547690557e-06, "loss": 0.0, "num_input_tokens_seen": 9619344, "step": 34355 }, { "epoch": 381.77777777777777, "grad_norm": 2.4531667804694735e-05, "learning_rate": 2.413717551815062e-06, "loss": 0.0, "num_input_tokens_seen": 9620720, "step": 34360 }, { "epoch": 381.8333333333333, "grad_norm": 8.752541361900512e-06, "learning_rate": 2.409510631949666e-06, "loss": 0.0, "num_input_tokens_seen": 9622144, "step": 34365 }, { "epoch": 381.8888888888889, "grad_norm": 5.126322548676399e-07, "learning_rate": 2.405307195821618e-06, "loss": 0.0, "num_input_tokens_seen": 9623520, "step": 34370 }, { "epoch": 381.94444444444446, "grad_norm": 2.3431842066656827e-07, "learning_rate": 2.4011072440791372e-06, "loss": 0.0, "num_input_tokens_seen": 9624960, "step": 34375 }, { "epoch": 382.0, "grad_norm": 6.123048592598934e-07, "learning_rate": 2.3969107773699233e-06, "loss": 0.0, "num_input_tokens_seen": 9626368, "step": 34380 }, { "epoch": 382.05555555555554, "grad_norm": 1.920824388434994e-06, "learning_rate": 2.3927177963411096e-06, "loss": 0.0, "num_input_tokens_seen": 9627792, "step": 34385 }, { "epoch": 382.1111111111111, "grad_norm": 6.683039100607857e-05, "learning_rate": 2.3885283016393144e-06, "loss": 0.0, "num_input_tokens_seen": 9629168, "step": 34390 }, { "epoch": 382.1666666666667, "grad_norm": 2.1937295002771862e-07, "learning_rate": 2.3843422939106076e-06, "loss": 0.0, "num_input_tokens_seen": 9630544, "step": 34395 }, { "epoch": 382.22222222222223, "grad_norm": 1.4674725434815628e-07, "learning_rate": 2.380159773800525e-06, "loss": 0.0, "num_input_tokens_seen": 9631952, "step": 34400 }, { "epoch": 382.22222222222223, "eval_loss": 1.1527990102767944, "eval_runtime": 1.3878, "eval_samples_per_second": 28.822, "eval_steps_per_second": 14.411, "num_input_tokens_seen": 9631952, "step": 34400 }, { "epoch": 382.27777777777777, "grad_norm": 1.5461164366570301e-06, "learning_rate": 2.3759807419540675e-06, "loss": 0.0, "num_input_tokens_seen": 9633392, "step": 34405 }, { "epoch": 382.3333333333333, "grad_norm": 5.638175366584619e-07, "learning_rate": 2.3718051990156835e-06, "loss": 0.0, "num_input_tokens_seen": 9634816, "step": 34410 }, { "epoch": 382.3888888888889, "grad_norm": 5.963228886685101e-06, "learning_rate": 2.367633145629311e-06, "loss": 0.0, "num_input_tokens_seen": 9636176, "step": 34415 }, { "epoch": 382.44444444444446, "grad_norm": 1.3859498437795992e-07, "learning_rate": 2.363464582438316e-06, "loss": 0.0, "num_input_tokens_seen": 9637568, "step": 34420 }, { "epoch": 382.5, "grad_norm": 1.2752067846122372e-07, "learning_rate": 2.3592995100855526e-06, "loss": 0.0, "num_input_tokens_seen": 9638992, "step": 34425 }, { "epoch": 382.55555555555554, "grad_norm": 9.284558899480544e-08, "learning_rate": 2.3551379292133273e-06, "loss": 0.0, "num_input_tokens_seen": 9640400, "step": 34430 }, { "epoch": 382.6111111111111, "grad_norm": 6.392016530298861e-06, "learning_rate": 2.3509798404634047e-06, "loss": 0.0, "num_input_tokens_seen": 9641760, "step": 34435 }, { "epoch": 382.6666666666667, "grad_norm": 1.351034910612725e-07, "learning_rate": 2.346825244477019e-06, "loss": 0.0, "num_input_tokens_seen": 9643184, "step": 34440 }, { "epoch": 382.72222222222223, "grad_norm": 1.7212173020197952e-07, "learning_rate": 2.3426741418948545e-06, "loss": 0.0, "num_input_tokens_seen": 9644592, "step": 34445 }, { "epoch": 382.77777777777777, "grad_norm": 3.4999652598344255e-07, "learning_rate": 2.3385265333570715e-06, "loss": 0.0, "num_input_tokens_seen": 9645984, "step": 34450 }, { "epoch": 382.8333333333333, "grad_norm": 3.5985596014143084e-07, "learning_rate": 2.334382419503278e-06, "loss": 0.0, "num_input_tokens_seen": 9647376, "step": 34455 }, { "epoch": 382.8888888888889, "grad_norm": 9.273997420677915e-06, "learning_rate": 2.3302418009725465e-06, "loss": 0.0, "num_input_tokens_seen": 9648752, "step": 34460 }, { "epoch": 382.94444444444446, "grad_norm": 1.0236667549179401e-06, "learning_rate": 2.326104678403415e-06, "loss": 0.0, "num_input_tokens_seen": 9650160, "step": 34465 }, { "epoch": 383.0, "grad_norm": 1.6358264076643536e-07, "learning_rate": 2.321971052433883e-06, "loss": 0.0, "num_input_tokens_seen": 9651584, "step": 34470 }, { "epoch": 383.05555555555554, "grad_norm": 1.1233272516619763e-06, "learning_rate": 2.3178409237014004e-06, "loss": 0.0, "num_input_tokens_seen": 9653008, "step": 34475 }, { "epoch": 383.1111111111111, "grad_norm": 5.65210484637646e-06, "learning_rate": 2.313714292842889e-06, "loss": 0.0, "num_input_tokens_seen": 9654400, "step": 34480 }, { "epoch": 383.1666666666667, "grad_norm": 4.496688461586018e-07, "learning_rate": 2.309591160494734e-06, "loss": 0.0, "num_input_tokens_seen": 9655840, "step": 34485 }, { "epoch": 383.22222222222223, "grad_norm": 3.34870719598257e-07, "learning_rate": 2.305471527292763e-06, "loss": 0.0, "num_input_tokens_seen": 9657216, "step": 34490 }, { "epoch": 383.27777777777777, "grad_norm": 2.816394271576428e-06, "learning_rate": 2.3013553938722817e-06, "loss": 0.0, "num_input_tokens_seen": 9658624, "step": 34495 }, { "epoch": 383.3333333333333, "grad_norm": 2.1842558339812967e-07, "learning_rate": 2.297242760868043e-06, "loss": 0.0, "num_input_tokens_seen": 9660032, "step": 34500 }, { "epoch": 383.3888888888889, "grad_norm": 8.266468398687721e-07, "learning_rate": 2.2931336289142735e-06, "loss": 0.0, "num_input_tokens_seen": 9661424, "step": 34505 }, { "epoch": 383.44444444444446, "grad_norm": 9.613867405278143e-06, "learning_rate": 2.289027998644655e-06, "loss": 0.0, "num_input_tokens_seen": 9662816, "step": 34510 }, { "epoch": 383.5, "grad_norm": 5.573438102146611e-07, "learning_rate": 2.2849258706923228e-06, "loss": 0.0, "num_input_tokens_seen": 9664192, "step": 34515 }, { "epoch": 383.55555555555554, "grad_norm": 7.607966949763068e-07, "learning_rate": 2.2808272456898705e-06, "loss": 0.0, "num_input_tokens_seen": 9665600, "step": 34520 }, { "epoch": 383.6111111111111, "grad_norm": 7.392094403257943e-07, "learning_rate": 2.2767321242693707e-06, "loss": 0.0, "num_input_tokens_seen": 9667008, "step": 34525 }, { "epoch": 383.6666666666667, "grad_norm": 5.52911274098733e-07, "learning_rate": 2.272640507062329e-06, "loss": 0.0, "num_input_tokens_seen": 9668400, "step": 34530 }, { "epoch": 383.72222222222223, "grad_norm": 1.2548966878966894e-06, "learning_rate": 2.2685523946997382e-06, "loss": 0.0, "num_input_tokens_seen": 9669840, "step": 34535 }, { "epoch": 383.77777777777777, "grad_norm": 3.2233583624474704e-07, "learning_rate": 2.2644677878120245e-06, "loss": 0.0, "num_input_tokens_seen": 9671216, "step": 34540 }, { "epoch": 383.8333333333333, "grad_norm": 2.125417779552663e-07, "learning_rate": 2.2603866870290897e-06, "loss": 0.0, "num_input_tokens_seen": 9672624, "step": 34545 }, { "epoch": 383.8888888888889, "grad_norm": 3.93825274613846e-07, "learning_rate": 2.256309092980294e-06, "loss": 0.0, "num_input_tokens_seen": 9674032, "step": 34550 }, { "epoch": 383.94444444444446, "grad_norm": 1.0097654978835635e-07, "learning_rate": 2.252235006294448e-06, "loss": 0.0, "num_input_tokens_seen": 9675456, "step": 34555 }, { "epoch": 384.0, "grad_norm": 4.1169604969581997e-07, "learning_rate": 2.2481644275998333e-06, "loss": 0.0, "num_input_tokens_seen": 9676864, "step": 34560 }, { "epoch": 384.05555555555554, "grad_norm": 4.668742121793912e-07, "learning_rate": 2.2440973575241832e-06, "loss": 0.0, "num_input_tokens_seen": 9678256, "step": 34565 }, { "epoch": 384.1111111111111, "grad_norm": 1.4020527316915832e-07, "learning_rate": 2.240033796694685e-06, "loss": 0.0, "num_input_tokens_seen": 9679632, "step": 34570 }, { "epoch": 384.1666666666667, "grad_norm": 7.032159032860363e-07, "learning_rate": 2.235973745737999e-06, "loss": 0.0, "num_input_tokens_seen": 9680992, "step": 34575 }, { "epoch": 384.22222222222223, "grad_norm": 4.2390820453874767e-07, "learning_rate": 2.2319172052802263e-06, "loss": 0.0, "num_input_tokens_seen": 9682416, "step": 34580 }, { "epoch": 384.27777777777777, "grad_norm": 1.0190331067860825e-06, "learning_rate": 2.2278641759469477e-06, "loss": 0.0, "num_input_tokens_seen": 9683808, "step": 34585 }, { "epoch": 384.3333333333333, "grad_norm": 1.0071881661133375e-05, "learning_rate": 2.2238146583631825e-06, "loss": 0.0, "num_input_tokens_seen": 9685232, "step": 34590 }, { "epoch": 384.3888888888889, "grad_norm": 2.0055629192938795e-06, "learning_rate": 2.2197686531534256e-06, "loss": 0.0, "num_input_tokens_seen": 9686592, "step": 34595 }, { "epoch": 384.44444444444446, "grad_norm": 4.818020329366846e-07, "learning_rate": 2.2157261609416087e-06, "loss": 0.0, "num_input_tokens_seen": 9687936, "step": 34600 }, { "epoch": 384.44444444444446, "eval_loss": 1.1655941009521484, "eval_runtime": 1.3912, "eval_samples_per_second": 28.751, "eval_steps_per_second": 14.376, "num_input_tokens_seen": 9687936, "step": 34600 }, { "epoch": 384.5, "grad_norm": 1.5693105126501905e-07, "learning_rate": 2.211687182351149e-06, "loss": 0.0, "num_input_tokens_seen": 9689328, "step": 34605 }, { "epoch": 384.55555555555554, "grad_norm": 2.7233525656811253e-07, "learning_rate": 2.2076517180048993e-06, "loss": 0.0, "num_input_tokens_seen": 9690768, "step": 34610 }, { "epoch": 384.6111111111111, "grad_norm": 3.59641290970103e-07, "learning_rate": 2.2036197685251834e-06, "loss": 0.0, "num_input_tokens_seen": 9692144, "step": 34615 }, { "epoch": 384.6666666666667, "grad_norm": 8.143864960175051e-08, "learning_rate": 2.199591334533771e-06, "loss": 0.0, "num_input_tokens_seen": 9693552, "step": 34620 }, { "epoch": 384.72222222222223, "grad_norm": 8.909287316782866e-07, "learning_rate": 2.1955664166519036e-06, "loss": 0.0, "num_input_tokens_seen": 9694992, "step": 34625 }, { "epoch": 384.77777777777777, "grad_norm": 2.0587769711255532e-07, "learning_rate": 2.1915450155002793e-06, "loss": 0.0, "num_input_tokens_seen": 9696384, "step": 34630 }, { "epoch": 384.8333333333333, "grad_norm": 8.161362643477332e-07, "learning_rate": 2.187527131699038e-06, "loss": 0.0, "num_input_tokens_seen": 9697808, "step": 34635 }, { "epoch": 384.8888888888889, "grad_norm": 8.98718042208202e-08, "learning_rate": 2.18351276586779e-06, "loss": 0.0, "num_input_tokens_seen": 9699264, "step": 34640 }, { "epoch": 384.94444444444446, "grad_norm": 1.638864972619558e-07, "learning_rate": 2.1795019186256092e-06, "loss": 0.0, "num_input_tokens_seen": 9700672, "step": 34645 }, { "epoch": 385.0, "grad_norm": 1.3067809732092428e-06, "learning_rate": 2.1754945905910094e-06, "loss": 0.0, "num_input_tokens_seen": 9702048, "step": 34650 }, { "epoch": 385.05555555555554, "grad_norm": 5.258174837763363e-07, "learning_rate": 2.171490782381977e-06, "loss": 0.0, "num_input_tokens_seen": 9703440, "step": 34655 }, { "epoch": 385.1111111111111, "grad_norm": 5.137443395142327e-07, "learning_rate": 2.1674904946159425e-06, "loss": 0.0, "num_input_tokens_seen": 9704784, "step": 34660 }, { "epoch": 385.1666666666667, "grad_norm": 6.438852437895548e-07, "learning_rate": 2.16349372790981e-06, "loss": 0.0, "num_input_tokens_seen": 9706192, "step": 34665 }, { "epoch": 385.22222222222223, "grad_norm": 8.152758255164372e-07, "learning_rate": 2.159500482879928e-06, "loss": 0.0, "num_input_tokens_seen": 9707616, "step": 34670 }, { "epoch": 385.27777777777777, "grad_norm": 3.428733180044219e-06, "learning_rate": 2.155510760142096e-06, "loss": 0.0, "num_input_tokens_seen": 9708960, "step": 34675 }, { "epoch": 385.3333333333333, "grad_norm": 3.404276753826707e-07, "learning_rate": 2.151524560311588e-06, "loss": 0.0, "num_input_tokens_seen": 9710352, "step": 34680 }, { "epoch": 385.3888888888889, "grad_norm": 1.7061935864148836e-07, "learning_rate": 2.147541884003129e-06, "loss": 0.0, "num_input_tokens_seen": 9711744, "step": 34685 }, { "epoch": 385.44444444444446, "grad_norm": 2.411930495327397e-07, "learning_rate": 2.1435627318308895e-06, "loss": 0.0, "num_input_tokens_seen": 9713120, "step": 34690 }, { "epoch": 385.5, "grad_norm": 3.949924121116055e-06, "learning_rate": 2.139587104408511e-06, "loss": 0.0, "num_input_tokens_seen": 9714512, "step": 34695 }, { "epoch": 385.55555555555554, "grad_norm": 1.0919645774265518e-06, "learning_rate": 2.1356150023490783e-06, "loss": 0.0, "num_input_tokens_seen": 9715968, "step": 34700 }, { "epoch": 385.6111111111111, "grad_norm": 4.7046523832250386e-07, "learning_rate": 2.1316464262651464e-06, "loss": 0.0, "num_input_tokens_seen": 9717344, "step": 34705 }, { "epoch": 385.6666666666667, "grad_norm": 2.5831629955064273e-06, "learning_rate": 2.1276813767687224e-06, "loss": 0.0, "num_input_tokens_seen": 9718752, "step": 34710 }, { "epoch": 385.72222222222223, "grad_norm": 1.0282151379215065e-06, "learning_rate": 2.123719854471254e-06, "loss": 0.0, "num_input_tokens_seen": 9720160, "step": 34715 }, { "epoch": 385.77777777777777, "grad_norm": 4.3121545445501397e-07, "learning_rate": 2.119761859983668e-06, "loss": 0.0, "num_input_tokens_seen": 9721552, "step": 34720 }, { "epoch": 385.8333333333333, "grad_norm": 4.337829864198284e-07, "learning_rate": 2.1158073939163386e-06, "loss": 0.0, "num_input_tokens_seen": 9722976, "step": 34725 }, { "epoch": 385.8888888888889, "grad_norm": 1.060236769490075e-07, "learning_rate": 2.111856456879088e-06, "loss": 0.0, "num_input_tokens_seen": 9724352, "step": 34730 }, { "epoch": 385.94444444444446, "grad_norm": 6.609742285945686e-06, "learning_rate": 2.1079090494811993e-06, "loss": 0.0, "num_input_tokens_seen": 9725776, "step": 34735 }, { "epoch": 386.0, "grad_norm": 2.1017862650296593e-07, "learning_rate": 2.103965172331418e-06, "loss": 0.0, "num_input_tokens_seen": 9727216, "step": 34740 }, { "epoch": 386.05555555555554, "grad_norm": 3.3814677635746193e-07, "learning_rate": 2.100024826037933e-06, "loss": 0.0, "num_input_tokens_seen": 9728592, "step": 34745 }, { "epoch": 386.1111111111111, "grad_norm": 9.575850867804547e-08, "learning_rate": 2.0960880112084027e-06, "loss": 0.0, "num_input_tokens_seen": 9730016, "step": 34750 }, { "epoch": 386.1666666666667, "grad_norm": 6.20176990651089e-07, "learning_rate": 2.092154728449927e-06, "loss": 0.0, "num_input_tokens_seen": 9731424, "step": 34755 }, { "epoch": 386.22222222222223, "grad_norm": 3.120339044926368e-07, "learning_rate": 2.0882249783690687e-06, "loss": 0.0, "num_input_tokens_seen": 9732800, "step": 34760 }, { "epoch": 386.27777777777777, "grad_norm": 7.446565177815501e-07, "learning_rate": 2.084298761571851e-06, "loss": 0.0, "num_input_tokens_seen": 9734208, "step": 34765 }, { "epoch": 386.3333333333333, "grad_norm": 3.1218436902236135e-07, "learning_rate": 2.080376078663737e-06, "loss": 0.0, "num_input_tokens_seen": 9735616, "step": 34770 }, { "epoch": 386.3888888888889, "grad_norm": 2.673961496668653e-08, "learning_rate": 2.0764569302496593e-06, "loss": 0.0, "num_input_tokens_seen": 9736992, "step": 34775 }, { "epoch": 386.44444444444446, "grad_norm": 6.155367486826435e-07, "learning_rate": 2.0725413169339957e-06, "loss": 0.0, "num_input_tokens_seen": 9738400, "step": 34780 }, { "epoch": 386.5, "grad_norm": 4.2701415736701165e-08, "learning_rate": 2.068629239320588e-06, "loss": 0.0, "num_input_tokens_seen": 9739744, "step": 34785 }, { "epoch": 386.55555555555554, "grad_norm": 2.279939508298412e-07, "learning_rate": 2.064720698012726e-06, "loss": 0.0, "num_input_tokens_seen": 9741168, "step": 34790 }, { "epoch": 386.6111111111111, "grad_norm": 3.982025376103593e-08, "learning_rate": 2.0608156936131522e-06, "loss": 0.0, "num_input_tokens_seen": 9742544, "step": 34795 }, { "epoch": 386.6666666666667, "grad_norm": 6.642870289397251e-07, "learning_rate": 2.056914226724074e-06, "loss": 0.0, "num_input_tokens_seen": 9743968, "step": 34800 }, { "epoch": 386.6666666666667, "eval_loss": 1.16143000125885, "eval_runtime": 1.3956, "eval_samples_per_second": 28.662, "eval_steps_per_second": 14.331, "num_input_tokens_seen": 9743968, "step": 34800 }, { "epoch": 386.72222222222223, "grad_norm": 6.980952093726955e-06, "learning_rate": 2.0530162979471385e-06, "loss": 0.0, "num_input_tokens_seen": 9745376, "step": 34805 }, { "epoch": 386.77777777777777, "grad_norm": 2.9924897262390004e-06, "learning_rate": 2.0491219078834667e-06, "loss": 0.0, "num_input_tokens_seen": 9746704, "step": 34810 }, { "epoch": 386.8333333333333, "grad_norm": 8.373478340217844e-05, "learning_rate": 2.045231057133612e-06, "loss": 0.0, "num_input_tokens_seen": 9748096, "step": 34815 }, { "epoch": 386.8888888888889, "grad_norm": 3.4182860417786287e-06, "learning_rate": 2.0413437462975944e-06, "loss": 0.0, "num_input_tokens_seen": 9749488, "step": 34820 }, { "epoch": 386.94444444444446, "grad_norm": 2.1723499799009005e-07, "learning_rate": 2.0374599759748843e-06, "loss": 0.0, "num_input_tokens_seen": 9750912, "step": 34825 }, { "epoch": 387.0, "grad_norm": 3.485368779365672e-06, "learning_rate": 2.033579746764419e-06, "loss": 0.0, "num_input_tokens_seen": 9752320, "step": 34830 }, { "epoch": 387.05555555555554, "grad_norm": 1.6010949366318528e-07, "learning_rate": 2.029703059264565e-06, "loss": 0.0, "num_input_tokens_seen": 9753712, "step": 34835 }, { "epoch": 387.1111111111111, "grad_norm": 5.062297077529365e-06, "learning_rate": 2.02582991407316e-06, "loss": 0.0, "num_input_tokens_seen": 9755088, "step": 34840 }, { "epoch": 387.1666666666667, "grad_norm": 3.909738552465569e-06, "learning_rate": 2.0219603117874992e-06, "loss": 0.0, "num_input_tokens_seen": 9756464, "step": 34845 }, { "epoch": 387.22222222222223, "grad_norm": 4.2835617364289647e-07, "learning_rate": 2.0180942530043156e-06, "loss": 0.0, "num_input_tokens_seen": 9757824, "step": 34850 }, { "epoch": 387.27777777777777, "grad_norm": 2.526345724618295e-06, "learning_rate": 2.0142317383198107e-06, "loss": 0.0, "num_input_tokens_seen": 9759232, "step": 34855 }, { "epoch": 387.3333333333333, "grad_norm": 1.1136835098568554e-07, "learning_rate": 2.0103727683296243e-06, "loss": 0.0, "num_input_tokens_seen": 9760656, "step": 34860 }, { "epoch": 387.3888888888889, "grad_norm": 3.5453479085845174e-07, "learning_rate": 2.0065173436288636e-06, "loss": 0.0, "num_input_tokens_seen": 9762032, "step": 34865 }, { "epoch": 387.44444444444446, "grad_norm": 3.3159699341922533e-07, "learning_rate": 2.002665464812087e-06, "loss": 0.0, "num_input_tokens_seen": 9763424, "step": 34870 }, { "epoch": 387.5, "grad_norm": 8.094299914773728e-07, "learning_rate": 1.998817132473291e-06, "loss": 0.0, "num_input_tokens_seen": 9764880, "step": 34875 }, { "epoch": 387.55555555555554, "grad_norm": 6.810615360564043e-08, "learning_rate": 1.9949723472059507e-06, "loss": 0.0, "num_input_tokens_seen": 9766304, "step": 34880 }, { "epoch": 387.6111111111111, "grad_norm": 8.666125381751044e-07, "learning_rate": 1.9911311096029726e-06, "loss": 0.0, "num_input_tokens_seen": 9767728, "step": 34885 }, { "epoch": 387.6666666666667, "grad_norm": 7.203331620075915e-07, "learning_rate": 1.9872934202567224e-06, "loss": 0.0, "num_input_tokens_seen": 9769120, "step": 34890 }, { "epoch": 387.72222222222223, "grad_norm": 1.841471828356589e-07, "learning_rate": 1.9834592797590257e-06, "loss": 0.0, "num_input_tokens_seen": 9770544, "step": 34895 }, { "epoch": 387.77777777777777, "grad_norm": 4.9668369683786295e-06, "learning_rate": 1.979628688701149e-06, "loss": 0.0, "num_input_tokens_seen": 9771984, "step": 34900 }, { "epoch": 387.8333333333333, "grad_norm": 2.0739294370741845e-07, "learning_rate": 1.9758016476738193e-06, "loss": 0.0, "num_input_tokens_seen": 9773424, "step": 34905 }, { "epoch": 387.8888888888889, "grad_norm": 9.456835670107466e-08, "learning_rate": 1.971978157267221e-06, "loss": 0.0, "num_input_tokens_seen": 9774816, "step": 34910 }, { "epoch": 387.94444444444446, "grad_norm": 1.9567792719499266e-07, "learning_rate": 1.968158218070973e-06, "loss": 0.0, "num_input_tokens_seen": 9776192, "step": 34915 }, { "epoch": 388.0, "grad_norm": 9.884044516184076e-08, "learning_rate": 1.9643418306741682e-06, "loss": 0.0, "num_input_tokens_seen": 9777568, "step": 34920 }, { "epoch": 388.05555555555554, "grad_norm": 1.3913846714785905e-06, "learning_rate": 1.9605289956653337e-06, "loss": 0.0, "num_input_tokens_seen": 9778976, "step": 34925 }, { "epoch": 388.1111111111111, "grad_norm": 5.94081029703375e-07, "learning_rate": 1.9567197136324626e-06, "loss": 0.0, "num_input_tokens_seen": 9780416, "step": 34930 }, { "epoch": 388.1666666666667, "grad_norm": 9.50415028455609e-07, "learning_rate": 1.9529139851629935e-06, "loss": 0.0, "num_input_tokens_seen": 9781840, "step": 34935 }, { "epoch": 388.22222222222223, "grad_norm": 1.1688881329519063e-07, "learning_rate": 1.949111810843812e-06, "loss": 0.0, "num_input_tokens_seen": 9783232, "step": 34940 }, { "epoch": 388.27777777777777, "grad_norm": 6.373002747750434e-07, "learning_rate": 1.9453131912612694e-06, "loss": 0.0, "num_input_tokens_seen": 9784656, "step": 34945 }, { "epoch": 388.3333333333333, "grad_norm": 1.2225833643242368e-06, "learning_rate": 1.941518127001149e-06, "loss": 0.0, "num_input_tokens_seen": 9786080, "step": 34950 }, { "epoch": 388.3888888888889, "grad_norm": 8.431549076703959e-07, "learning_rate": 1.9377266186487107e-06, "loss": 0.0, "num_input_tokens_seen": 9787440, "step": 34955 }, { "epoch": 388.44444444444446, "grad_norm": 8.587087904743385e-06, "learning_rate": 1.9339386667886483e-06, "loss": 0.0, "num_input_tokens_seen": 9788832, "step": 34960 }, { "epoch": 388.5, "grad_norm": 1.1014923302354873e-06, "learning_rate": 1.9301542720051024e-06, "loss": 0.0, "num_input_tokens_seen": 9790240, "step": 34965 }, { "epoch": 388.55555555555554, "grad_norm": 1.840997538238298e-05, "learning_rate": 1.926373434881684e-06, "loss": 0.0, "num_input_tokens_seen": 9791600, "step": 34970 }, { "epoch": 388.6111111111111, "grad_norm": 4.696897519806953e-07, "learning_rate": 1.9225961560014468e-06, "loss": 0.0, "num_input_tokens_seen": 9793008, "step": 34975 }, { "epoch": 388.6666666666667, "grad_norm": 2.866525846911827e-06, "learning_rate": 1.918822435946885e-06, "loss": 0.0, "num_input_tokens_seen": 9794416, "step": 34980 }, { "epoch": 388.72222222222223, "grad_norm": 1.677522760701322e-07, "learning_rate": 1.915052275299961e-06, "loss": 0.0, "num_input_tokens_seen": 9795792, "step": 34985 }, { "epoch": 388.77777777777777, "grad_norm": 1.2290237805245852e-07, "learning_rate": 1.9112856746420854e-06, "loss": 0.0, "num_input_tokens_seen": 9797232, "step": 34990 }, { "epoch": 388.8333333333333, "grad_norm": 1.5964423027980956e-06, "learning_rate": 1.907522634554104e-06, "loss": 0.0, "num_input_tokens_seen": 9798624, "step": 34995 }, { "epoch": 388.8888888888889, "grad_norm": 3.618124992499361e-06, "learning_rate": 1.9037631556163337e-06, "loss": 0.0, "num_input_tokens_seen": 9800016, "step": 35000 }, { "epoch": 388.8888888888889, "eval_loss": 1.194258689880371, "eval_runtime": 1.3944, "eval_samples_per_second": 28.686, "eval_steps_per_second": 14.343, "num_input_tokens_seen": 9800016, "step": 35000 }, { "epoch": 388.94444444444446, "grad_norm": 6.553070306836162e-06, "learning_rate": 1.9000072384085272e-06, "loss": 0.0, "num_input_tokens_seen": 9801376, "step": 35005 }, { "epoch": 389.0, "grad_norm": 4.28953825348799e-07, "learning_rate": 1.8962548835098987e-06, "loss": 0.0, "num_input_tokens_seen": 9802816, "step": 35010 }, { "epoch": 389.05555555555554, "grad_norm": 1.0841246194104315e-06, "learning_rate": 1.8925060914991077e-06, "loss": 0.0, "num_input_tokens_seen": 9804256, "step": 35015 }, { "epoch": 389.1111111111111, "grad_norm": 1.0421008482808247e-06, "learning_rate": 1.888760862954264e-06, "loss": 0.0, "num_input_tokens_seen": 9805648, "step": 35020 }, { "epoch": 389.1666666666667, "grad_norm": 1.0800822991541281e-07, "learning_rate": 1.8850191984529309e-06, "loss": 0.0, "num_input_tokens_seen": 9807040, "step": 35025 }, { "epoch": 389.22222222222223, "grad_norm": 6.62975367049512e-07, "learning_rate": 1.8812810985721186e-06, "loss": 0.0, "num_input_tokens_seen": 9808416, "step": 35030 }, { "epoch": 389.27777777777777, "grad_norm": 6.100075893300527e-07, "learning_rate": 1.8775465638882856e-06, "loss": 0.0, "num_input_tokens_seen": 9809808, "step": 35035 }, { "epoch": 389.3333333333333, "grad_norm": 1.213438594049876e-07, "learning_rate": 1.8738155949773517e-06, "loss": 0.0, "num_input_tokens_seen": 9811232, "step": 35040 }, { "epoch": 389.3888888888889, "grad_norm": 7.091388454227854e-08, "learning_rate": 1.8700881924146707e-06, "loss": 0.0, "num_input_tokens_seen": 9812640, "step": 35045 }, { "epoch": 389.44444444444446, "grad_norm": 2.2001245270075742e-06, "learning_rate": 1.8663643567750577e-06, "loss": 0.0, "num_input_tokens_seen": 9814048, "step": 35050 }, { "epoch": 389.5, "grad_norm": 5.705212515749736e-06, "learning_rate": 1.8626440886327813e-06, "loss": 0.0, "num_input_tokens_seen": 9815424, "step": 35055 }, { "epoch": 389.55555555555554, "grad_norm": 5.476467208609392e-07, "learning_rate": 1.8589273885615432e-06, "loss": 0.0, "num_input_tokens_seen": 9816816, "step": 35060 }, { "epoch": 389.6111111111111, "grad_norm": 2.271761019301266e-07, "learning_rate": 1.8552142571345133e-06, "loss": 0.0, "num_input_tokens_seen": 9818272, "step": 35065 }, { "epoch": 389.6666666666667, "grad_norm": 2.8878212106064893e-06, "learning_rate": 1.8515046949243025e-06, "loss": 0.0, "num_input_tokens_seen": 9819712, "step": 35070 }, { "epoch": 389.72222222222223, "grad_norm": 9.067983341992658e-07, "learning_rate": 1.8477987025029674e-06, "loss": 0.0, "num_input_tokens_seen": 9821056, "step": 35075 }, { "epoch": 389.77777777777777, "grad_norm": 2.860966219486727e-07, "learning_rate": 1.8440962804420232e-06, "loss": 0.0, "num_input_tokens_seen": 9822448, "step": 35080 }, { "epoch": 389.8333333333333, "grad_norm": 3.7770627159261494e-07, "learning_rate": 1.8403974293124265e-06, "loss": 0.0, "num_input_tokens_seen": 9823824, "step": 35085 }, { "epoch": 389.8888888888889, "grad_norm": 4.1028195596481964e-07, "learning_rate": 1.8367021496845854e-06, "loss": 0.0, "num_input_tokens_seen": 9825248, "step": 35090 }, { "epoch": 389.94444444444446, "grad_norm": 5.557829396707348e-08, "learning_rate": 1.8330104421283662e-06, "loss": 0.0, "num_input_tokens_seen": 9826624, "step": 35095 }, { "epoch": 390.0, "grad_norm": 2.6102472361344553e-07, "learning_rate": 1.8293223072130717e-06, "loss": 0.0, "num_input_tokens_seen": 9828016, "step": 35100 }, { "epoch": 390.05555555555554, "grad_norm": 1.3754701285506599e-05, "learning_rate": 1.8256377455074525e-06, "loss": 0.0, "num_input_tokens_seen": 9829408, "step": 35105 }, { "epoch": 390.1111111111111, "grad_norm": 2.272809979331214e-06, "learning_rate": 1.8219567575797263e-06, "loss": 0.0, "num_input_tokens_seen": 9830832, "step": 35110 }, { "epoch": 390.1666666666667, "grad_norm": 5.410215067058743e-07, "learning_rate": 1.8182793439975365e-06, "loss": 0.0, "num_input_tokens_seen": 9832288, "step": 35115 }, { "epoch": 390.22222222222223, "grad_norm": 1.7542864952702075e-05, "learning_rate": 1.8146055053279958e-06, "loss": 0.0, "num_input_tokens_seen": 9833664, "step": 35120 }, { "epoch": 390.27777777777777, "grad_norm": 5.092878723189642e-07, "learning_rate": 1.8109352421376486e-06, "loss": 0.0, "num_input_tokens_seen": 9835072, "step": 35125 }, { "epoch": 390.3333333333333, "grad_norm": 3.872399076954025e-07, "learning_rate": 1.8072685549924972e-06, "loss": 0.0, "num_input_tokens_seen": 9836464, "step": 35130 }, { "epoch": 390.3888888888889, "grad_norm": 1.1069960237364285e-07, "learning_rate": 1.8036054444579982e-06, "loss": 0.0, "num_input_tokens_seen": 9837840, "step": 35135 }, { "epoch": 390.44444444444446, "grad_norm": 6.614861831621965e-07, "learning_rate": 1.7999459110990407e-06, "loss": 0.0, "num_input_tokens_seen": 9839264, "step": 35140 }, { "epoch": 390.5, "grad_norm": 4.985637360732653e-07, "learning_rate": 1.7962899554799712e-06, "loss": 0.0, "num_input_tokens_seen": 9840640, "step": 35145 }, { "epoch": 390.55555555555554, "grad_norm": 2.551613533796626e-07, "learning_rate": 1.7926375781645937e-06, "loss": 0.0, "num_input_tokens_seen": 9842016, "step": 35150 }, { "epoch": 390.6111111111111, "grad_norm": 2.6282489784534846e-07, "learning_rate": 1.7889887797161359e-06, "loss": 0.0, "num_input_tokens_seen": 9843408, "step": 35155 }, { "epoch": 390.6666666666667, "grad_norm": 5.14744407098533e-08, "learning_rate": 1.7853435606973028e-06, "loss": 0.0, "num_input_tokens_seen": 9844800, "step": 35160 }, { "epoch": 390.72222222222223, "grad_norm": 3.413109652683488e-06, "learning_rate": 1.781701921670223e-06, "loss": 0.0, "num_input_tokens_seen": 9846160, "step": 35165 }, { "epoch": 390.77777777777777, "grad_norm": 3.030694699646119e-07, "learning_rate": 1.7780638631964886e-06, "loss": 0.0, "num_input_tokens_seen": 9847664, "step": 35170 }, { "epoch": 390.8333333333333, "grad_norm": 1.1228642051719362e-06, "learning_rate": 1.7744293858371314e-06, "loss": 0.0, "num_input_tokens_seen": 9849072, "step": 35175 }, { "epoch": 390.8888888888889, "grad_norm": 2.2062813798129355e-07, "learning_rate": 1.770798490152631e-06, "loss": 0.0, "num_input_tokens_seen": 9850480, "step": 35180 }, { "epoch": 390.94444444444446, "grad_norm": 9.878026503429282e-06, "learning_rate": 1.767171176702917e-06, "loss": 0.0, "num_input_tokens_seen": 9851840, "step": 35185 }, { "epoch": 391.0, "grad_norm": 1.0065763262900873e-06, "learning_rate": 1.7635474460473755e-06, "loss": 0.0, "num_input_tokens_seen": 9853232, "step": 35190 }, { "epoch": 391.05555555555554, "grad_norm": 3.4388227732051746e-07, "learning_rate": 1.7599272987448206e-06, "loss": 0.0, "num_input_tokens_seen": 9854640, "step": 35195 }, { "epoch": 391.1111111111111, "grad_norm": 6.679583748336881e-05, "learning_rate": 1.7563107353535362e-06, "loss": 0.0, "num_input_tokens_seen": 9856016, "step": 35200 }, { "epoch": 391.1111111111111, "eval_loss": 1.2018158435821533, "eval_runtime": 1.3928, "eval_samples_per_second": 28.72, "eval_steps_per_second": 14.36, "num_input_tokens_seen": 9856016, "step": 35200 }, { "epoch": 391.1666666666667, "grad_norm": 1.0285019698130782e-06, "learning_rate": 1.7526977564312263e-06, "loss": 0.0, "num_input_tokens_seen": 9857408, "step": 35205 }, { "epoch": 391.22222222222223, "grad_norm": 3.55083307113091e-06, "learning_rate": 1.7490883625350701e-06, "loss": 0.0, "num_input_tokens_seen": 9858784, "step": 35210 }, { "epoch": 391.27777777777777, "grad_norm": 1.9738146761483222e-07, "learning_rate": 1.7454825542216807e-06, "loss": 0.0, "num_input_tokens_seen": 9860192, "step": 35215 }, { "epoch": 391.3333333333333, "grad_norm": 3.706845291162608e-06, "learning_rate": 1.7418803320471105e-06, "loss": 0.0, "num_input_tokens_seen": 9861552, "step": 35220 }, { "epoch": 391.3888888888889, "grad_norm": 6.986641665207571e-08, "learning_rate": 1.7382816965668737e-06, "loss": 0.0, "num_input_tokens_seen": 9862960, "step": 35225 }, { "epoch": 391.44444444444446, "grad_norm": 6.38184337731218e-06, "learning_rate": 1.7346866483359285e-06, "loss": 0.0, "num_input_tokens_seen": 9864384, "step": 35230 }, { "epoch": 391.5, "grad_norm": 6.297689196799183e-07, "learning_rate": 1.7310951879086657e-06, "loss": 0.0, "num_input_tokens_seen": 9865824, "step": 35235 }, { "epoch": 391.55555555555554, "grad_norm": 2.9195845741014637e-07, "learning_rate": 1.7275073158389471e-06, "loss": 0.0, "num_input_tokens_seen": 9867232, "step": 35240 }, { "epoch": 391.6111111111111, "grad_norm": 4.791286869476608e-07, "learning_rate": 1.723923032680061e-06, "loss": 0.0, "num_input_tokens_seen": 9868624, "step": 35245 }, { "epoch": 391.6666666666667, "grad_norm": 4.972716283191403e-07, "learning_rate": 1.7203423389847428e-06, "loss": 0.0, "num_input_tokens_seen": 9870016, "step": 35250 }, { "epoch": 391.72222222222223, "grad_norm": 4.789463332599553e-07, "learning_rate": 1.7167652353051928e-06, "loss": 0.0, "num_input_tokens_seen": 9871424, "step": 35255 }, { "epoch": 391.77777777777777, "grad_norm": 5.509748461918207e-06, "learning_rate": 1.7131917221930333e-06, "loss": 0.0, "num_input_tokens_seen": 9872864, "step": 35260 }, { "epoch": 391.8333333333333, "grad_norm": 1.60382260219194e-05, "learning_rate": 1.7096218001993513e-06, "loss": 0.0, "num_input_tokens_seen": 9874240, "step": 35265 }, { "epoch": 391.8888888888889, "grad_norm": 1.0512979997656657e-07, "learning_rate": 1.706055469874676e-06, "loss": 0.0, "num_input_tokens_seen": 9875680, "step": 35270 }, { "epoch": 391.94444444444446, "grad_norm": 2.1216726509010186e-07, "learning_rate": 1.702492731768976e-06, "loss": 0.0, "num_input_tokens_seen": 9877056, "step": 35275 }, { "epoch": 392.0, "grad_norm": 3.18883394356817e-06, "learning_rate": 1.6989335864316724e-06, "loss": 0.0, "num_input_tokens_seen": 9878496, "step": 35280 }, { "epoch": 392.05555555555554, "grad_norm": 3.8506832424900495e-06, "learning_rate": 1.6953780344116265e-06, "loss": 0.0, "num_input_tokens_seen": 9879904, "step": 35285 }, { "epoch": 392.1111111111111, "grad_norm": 4.6366670858333237e-07, "learning_rate": 1.6918260762571497e-06, "loss": 0.0, "num_input_tokens_seen": 9881296, "step": 35290 }, { "epoch": 392.1666666666667, "grad_norm": 5.433703904600407e-07, "learning_rate": 1.6882777125160093e-06, "loss": 0.0, "num_input_tokens_seen": 9882624, "step": 35295 }, { "epoch": 392.22222222222223, "grad_norm": 2.0240501896751084e-07, "learning_rate": 1.6847329437353899e-06, "loss": 0.0, "num_input_tokens_seen": 9884032, "step": 35300 }, { "epoch": 392.27777777777777, "grad_norm": 1.594487730471883e-05, "learning_rate": 1.6811917704619511e-06, "loss": 0.0, "num_input_tokens_seen": 9885424, "step": 35305 }, { "epoch": 392.3333333333333, "grad_norm": 3.4677134408411803e-06, "learning_rate": 1.67765419324179e-06, "loss": 0.0, "num_input_tokens_seen": 9886832, "step": 35310 }, { "epoch": 392.3888888888889, "grad_norm": 7.621329132234678e-07, "learning_rate": 1.6741202126204364e-06, "loss": 0.0, "num_input_tokens_seen": 9888208, "step": 35315 }, { "epoch": 392.44444444444446, "grad_norm": 5.346664693206549e-05, "learning_rate": 1.6705898291428767e-06, "loss": 0.0, "num_input_tokens_seen": 9889616, "step": 35320 }, { "epoch": 392.5, "grad_norm": 2.4771131279521796e-07, "learning_rate": 1.6670630433535395e-06, "loss": 0.0, "num_input_tokens_seen": 9891072, "step": 35325 }, { "epoch": 392.55555555555554, "grad_norm": 1.1082522632932523e-06, "learning_rate": 1.6635398557962979e-06, "loss": 0.0, "num_input_tokens_seen": 9892416, "step": 35330 }, { "epoch": 392.6111111111111, "grad_norm": 4.40727859540857e-07, "learning_rate": 1.660020267014481e-06, "loss": 0.0, "num_input_tokens_seen": 9893840, "step": 35335 }, { "epoch": 392.6666666666667, "grad_norm": 1.8833016213193332e-07, "learning_rate": 1.6565042775508438e-06, "loss": 0.0, "num_input_tokens_seen": 9895200, "step": 35340 }, { "epoch": 392.72222222222223, "grad_norm": 2.6767230565383215e-07, "learning_rate": 1.6529918879475997e-06, "loss": 0.0, "num_input_tokens_seen": 9896640, "step": 35345 }, { "epoch": 392.77777777777777, "grad_norm": 1.1142149105580756e-06, "learning_rate": 1.6494830987464043e-06, "loss": 0.0, "num_input_tokens_seen": 9898000, "step": 35350 }, { "epoch": 392.8333333333333, "grad_norm": 6.464314878940058e-07, "learning_rate": 1.6459779104883555e-06, "loss": 0.0, "num_input_tokens_seen": 9899440, "step": 35355 }, { "epoch": 392.8888888888889, "grad_norm": 9.813046375484191e-08, "learning_rate": 1.6424763237140013e-06, "loss": 0.0, "num_input_tokens_seen": 9900864, "step": 35360 }, { "epoch": 392.94444444444446, "grad_norm": 1.9709145249180438e-07, "learning_rate": 1.6389783389633207e-06, "loss": 0.0, "num_input_tokens_seen": 9902272, "step": 35365 }, { "epoch": 393.0, "grad_norm": 5.749613137595588e-06, "learning_rate": 1.6354839567757546e-06, "loss": 0.0, "num_input_tokens_seen": 9903712, "step": 35370 }, { "epoch": 393.05555555555554, "grad_norm": 9.136869266512804e-06, "learning_rate": 1.6319931776901831e-06, "loss": 0.0, "num_input_tokens_seen": 9905072, "step": 35375 }, { "epoch": 393.1111111111111, "grad_norm": 5.070288580100168e-07, "learning_rate": 1.6285060022449229e-06, "loss": 0.0, "num_input_tokens_seen": 9906480, "step": 35380 }, { "epoch": 393.1666666666667, "grad_norm": 8.972610032742523e-08, "learning_rate": 1.6250224309777434e-06, "loss": 0.0, "num_input_tokens_seen": 9907856, "step": 35385 }, { "epoch": 393.22222222222223, "grad_norm": 7.141087507989141e-07, "learning_rate": 1.6215424644258515e-06, "loss": 0.0, "num_input_tokens_seen": 9909280, "step": 35390 }, { "epoch": 393.27777777777777, "grad_norm": 1.3409999155555852e-06, "learning_rate": 1.6180661031259036e-06, "loss": 0.0, "num_input_tokens_seen": 9910672, "step": 35395 }, { "epoch": 393.3333333333333, "grad_norm": 2.9950396651656774e-07, "learning_rate": 1.614593347613999e-06, "loss": 0.0, "num_input_tokens_seen": 9912112, "step": 35400 }, { "epoch": 393.3333333333333, "eval_loss": 1.2035990953445435, "eval_runtime": 1.3901, "eval_samples_per_second": 28.775, "eval_steps_per_second": 14.387, "num_input_tokens_seen": 9912112, "step": 35400 }, { "epoch": 393.3888888888889, "grad_norm": 1.0523825011432564e-07, "learning_rate": 1.6111241984256758e-06, "loss": 0.0, "num_input_tokens_seen": 9913472, "step": 35405 }, { "epoch": 393.44444444444446, "grad_norm": 3.0189764288479637e-07, "learning_rate": 1.6076586560959257e-06, "loss": 0.0, "num_input_tokens_seen": 9914864, "step": 35410 }, { "epoch": 393.5, "grad_norm": 3.7166083188822086e-07, "learning_rate": 1.604196721159182e-06, "loss": 0.0, "num_input_tokens_seen": 9916240, "step": 35415 }, { "epoch": 393.55555555555554, "grad_norm": 2.0184050697480416e-07, "learning_rate": 1.6007383941493092e-06, "loss": 0.0, "num_input_tokens_seen": 9917664, "step": 35420 }, { "epoch": 393.6111111111111, "grad_norm": 6.168529012029467e-07, "learning_rate": 1.5972836755996285e-06, "loss": 0.0, "num_input_tokens_seen": 9919056, "step": 35425 }, { "epoch": 393.6666666666667, "grad_norm": 6.52636344966595e-08, "learning_rate": 1.5938325660429076e-06, "loss": 0.0, "num_input_tokens_seen": 9920448, "step": 35430 }, { "epoch": 393.72222222222223, "grad_norm": 2.390828512943699e-07, "learning_rate": 1.5903850660113378e-06, "loss": 0.0, "num_input_tokens_seen": 9921872, "step": 35435 }, { "epoch": 393.77777777777777, "grad_norm": 3.5317819424562913e-07, "learning_rate": 1.5869411760365826e-06, "loss": 0.0, "num_input_tokens_seen": 9923280, "step": 35440 }, { "epoch": 393.8333333333333, "grad_norm": 1.1102255115247317e-07, "learning_rate": 1.58350089664972e-06, "loss": 0.0, "num_input_tokens_seen": 9924688, "step": 35445 }, { "epoch": 393.8888888888889, "grad_norm": 2.5251274564652704e-05, "learning_rate": 1.5800642283812865e-06, "loss": 0.0, "num_input_tokens_seen": 9926064, "step": 35450 }, { "epoch": 393.94444444444446, "grad_norm": 1.914325622465185e-07, "learning_rate": 1.5766311717612698e-06, "loss": 0.0, "num_input_tokens_seen": 9927456, "step": 35455 }, { "epoch": 394.0, "grad_norm": 1.422170470277706e-07, "learning_rate": 1.5732017273190818e-06, "loss": 0.0, "num_input_tokens_seen": 9928832, "step": 35460 }, { "epoch": 394.05555555555554, "grad_norm": 2.9123330023139715e-06, "learning_rate": 1.5697758955835806e-06, "loss": 0.0, "num_input_tokens_seen": 9930256, "step": 35465 }, { "epoch": 394.1111111111111, "grad_norm": 7.930469081429692e-08, "learning_rate": 1.566353677083085e-06, "loss": 0.0, "num_input_tokens_seen": 9931664, "step": 35470 }, { "epoch": 394.1666666666667, "grad_norm": 8.706697940397135e-07, "learning_rate": 1.562935072345334e-06, "loss": 0.0, "num_input_tokens_seen": 9933056, "step": 35475 }, { "epoch": 394.22222222222223, "grad_norm": 1.37909191835206e-05, "learning_rate": 1.5595200818975281e-06, "loss": 0.0, "num_input_tokens_seen": 9934432, "step": 35480 }, { "epoch": 394.27777777777777, "grad_norm": 1.1800969446085219e-07, "learning_rate": 1.5561087062662905e-06, "loss": 0.0, "num_input_tokens_seen": 9935808, "step": 35485 }, { "epoch": 394.3333333333333, "grad_norm": 8.849466439642129e-07, "learning_rate": 1.5527009459777087e-06, "loss": 0.0, "num_input_tokens_seen": 9937216, "step": 35490 }, { "epoch": 394.3888888888889, "grad_norm": 2.960046003863681e-07, "learning_rate": 1.5492968015572984e-06, "loss": 0.0, "num_input_tokens_seen": 9938672, "step": 35495 }, { "epoch": 394.44444444444446, "grad_norm": 8.11403197076288e-07, "learning_rate": 1.5458962735300203e-06, "loss": 0.0, "num_input_tokens_seen": 9940080, "step": 35500 }, { "epoch": 394.5, "grad_norm": 7.724547401721793e-08, "learning_rate": 1.54249936242028e-06, "loss": 0.0, "num_input_tokens_seen": 9941488, "step": 35505 }, { "epoch": 394.55555555555554, "grad_norm": 4.843741976401361e-07, "learning_rate": 1.5391060687519222e-06, "loss": 0.0, "num_input_tokens_seen": 9942832, "step": 35510 }, { "epoch": 394.6111111111111, "grad_norm": 3.4139659419452073e-06, "learning_rate": 1.5357163930482367e-06, "loss": 0.0, "num_input_tokens_seen": 9944224, "step": 35515 }, { "epoch": 394.6666666666667, "grad_norm": 4.024661279800057e-07, "learning_rate": 1.532330335831955e-06, "loss": 0.0, "num_input_tokens_seen": 9945648, "step": 35520 }, { "epoch": 394.72222222222223, "grad_norm": 4.115375304536428e-06, "learning_rate": 1.5289478976252491e-06, "loss": 0.0, "num_input_tokens_seen": 9947040, "step": 35525 }, { "epoch": 394.77777777777777, "grad_norm": 3.519198514823074e-07, "learning_rate": 1.5255690789497345e-06, "loss": 0.0, "num_input_tokens_seen": 9948480, "step": 35530 }, { "epoch": 394.8333333333333, "grad_norm": 3.833642495010281e-06, "learning_rate": 1.5221938803264641e-06, "loss": 0.0, "num_input_tokens_seen": 9949888, "step": 35535 }, { "epoch": 394.8888888888889, "grad_norm": 2.734426743700169e-06, "learning_rate": 1.518822302275938e-06, "loss": 0.0, "num_input_tokens_seen": 9951312, "step": 35540 }, { "epoch": 394.94444444444446, "grad_norm": 1.682978563621873e-06, "learning_rate": 1.5154543453180958e-06, "loss": 0.0, "num_input_tokens_seen": 9952720, "step": 35545 }, { "epoch": 395.0, "grad_norm": 4.790911134477938e-07, "learning_rate": 1.5120900099723167e-06, "loss": 0.0, "num_input_tokens_seen": 9954096, "step": 35550 }, { "epoch": 395.05555555555554, "grad_norm": 3.0990129573638114e-08, "learning_rate": 1.5087292967574273e-06, "loss": 0.0, "num_input_tokens_seen": 9955504, "step": 35555 }, { "epoch": 395.1111111111111, "grad_norm": 4.161766810284462e-06, "learning_rate": 1.5053722061916908e-06, "loss": 0.0, "num_input_tokens_seen": 9956864, "step": 35560 }, { "epoch": 395.1666666666667, "grad_norm": 2.96635516860988e-06, "learning_rate": 1.5020187387928124e-06, "loss": 0.0, "num_input_tokens_seen": 9958240, "step": 35565 }, { "epoch": 395.22222222222223, "grad_norm": 3.401313506401493e-07, "learning_rate": 1.4986688950779343e-06, "loss": 0.0, "num_input_tokens_seen": 9959696, "step": 35570 }, { "epoch": 395.27777777777777, "grad_norm": 8.78115599789453e-07, "learning_rate": 1.495322675563654e-06, "loss": 0.0, "num_input_tokens_seen": 9961104, "step": 35575 }, { "epoch": 395.3333333333333, "grad_norm": 3.629760954027006e-07, "learning_rate": 1.4919800807659922e-06, "loss": 0.0, "num_input_tokens_seen": 9962496, "step": 35580 }, { "epoch": 395.3888888888889, "grad_norm": 1.0879156064902418e-07, "learning_rate": 1.4886411112004255e-06, "loss": 0.0, "num_input_tokens_seen": 9963888, "step": 35585 }, { "epoch": 395.44444444444446, "grad_norm": 5.735376475968224e-07, "learning_rate": 1.4853057673818588e-06, "loss": 0.0, "num_input_tokens_seen": 9965312, "step": 35590 }, { "epoch": 395.5, "grad_norm": 1.4246782598092977e-07, "learning_rate": 1.481974049824647e-06, "loss": 0.0, "num_input_tokens_seen": 9966704, "step": 35595 }, { "epoch": 395.55555555555554, "grad_norm": 4.221152778427495e-07, "learning_rate": 1.4786459590425849e-06, "loss": 0.0, "num_input_tokens_seen": 9968112, "step": 35600 }, { "epoch": 395.55555555555554, "eval_loss": 1.1697461605072021, "eval_runtime": 1.3996, "eval_samples_per_second": 28.579, "eval_steps_per_second": 14.29, "num_input_tokens_seen": 9968112, "step": 35600 }, { "epoch": 395.6111111111111, "grad_norm": 4.203937464808405e-07, "learning_rate": 1.4753214955489036e-06, "loss": 0.0, "num_input_tokens_seen": 9969520, "step": 35605 }, { "epoch": 395.6666666666667, "grad_norm": 9.239329301635735e-08, "learning_rate": 1.4720006598562737e-06, "loss": 0.0, "num_input_tokens_seen": 9970912, "step": 35610 }, { "epoch": 395.72222222222223, "grad_norm": 3.4028109894279623e-06, "learning_rate": 1.4686834524768185e-06, "loss": 0.0, "num_input_tokens_seen": 9972304, "step": 35615 }, { "epoch": 395.77777777777777, "grad_norm": 1.5579528280795785e-06, "learning_rate": 1.4653698739220844e-06, "loss": 0.0, "num_input_tokens_seen": 9973712, "step": 35620 }, { "epoch": 395.8333333333333, "grad_norm": 3.9529783180114464e-07, "learning_rate": 1.4620599247030715e-06, "loss": 0.0, "num_input_tokens_seen": 9975072, "step": 35625 }, { "epoch": 395.8888888888889, "grad_norm": 9.544935863914361e-08, "learning_rate": 1.4587536053302125e-06, "loss": 0.0, "num_input_tokens_seen": 9976480, "step": 35630 }, { "epoch": 395.94444444444446, "grad_norm": 3.875943832554185e-07, "learning_rate": 1.4554509163133862e-06, "loss": 0.0, "num_input_tokens_seen": 9977888, "step": 35635 }, { "epoch": 396.0, "grad_norm": 9.929098609973153e-08, "learning_rate": 1.4521518581619098e-06, "loss": 0.0, "num_input_tokens_seen": 9979264, "step": 35640 }, { "epoch": 396.05555555555554, "grad_norm": 3.985140665463405e-06, "learning_rate": 1.4488564313845348e-06, "loss": 0.0, "num_input_tokens_seen": 9980640, "step": 35645 }, { "epoch": 396.1111111111111, "grad_norm": 1.0539340564719168e-06, "learning_rate": 1.4455646364894603e-06, "loss": 0.0, "num_input_tokens_seen": 9982048, "step": 35650 }, { "epoch": 396.1666666666667, "grad_norm": 1.55033731630283e-07, "learning_rate": 1.4422764739843247e-06, "loss": 0.0, "num_input_tokens_seen": 9983424, "step": 35655 }, { "epoch": 396.22222222222223, "grad_norm": 3.6441284123611695e-07, "learning_rate": 1.4389919443762e-06, "loss": 0.0, "num_input_tokens_seen": 9984816, "step": 35660 }, { "epoch": 396.27777777777777, "grad_norm": 1.2783164038410177e-07, "learning_rate": 1.4357110481716063e-06, "loss": 0.0, "num_input_tokens_seen": 9986256, "step": 35665 }, { "epoch": 396.3333333333333, "grad_norm": 5.1829561016347725e-06, "learning_rate": 1.4324337858764941e-06, "loss": 0.0, "num_input_tokens_seen": 9987664, "step": 35670 }, { "epoch": 396.3888888888889, "grad_norm": 1.5010584775154712e-07, "learning_rate": 1.4291601579962622e-06, "loss": 0.0, "num_input_tokens_seen": 9989120, "step": 35675 }, { "epoch": 396.44444444444446, "grad_norm": 1.4252373148337938e-05, "learning_rate": 1.42589016503574e-06, "loss": 0.0, "num_input_tokens_seen": 9990576, "step": 35680 }, { "epoch": 396.5, "grad_norm": 2.9710652142966865e-06, "learning_rate": 1.4226238074992099e-06, "loss": 0.0, "num_input_tokens_seen": 9991984, "step": 35685 }, { "epoch": 396.55555555555554, "grad_norm": 7.490648954444623e-07, "learning_rate": 1.4193610858903778e-06, "loss": 0.0, "num_input_tokens_seen": 9993392, "step": 35690 }, { "epoch": 396.6111111111111, "grad_norm": 2.7936928290728247e-06, "learning_rate": 1.416102000712402e-06, "loss": 0.0, "num_input_tokens_seen": 9994784, "step": 35695 }, { "epoch": 396.6666666666667, "grad_norm": 4.112283136237238e-07, "learning_rate": 1.4128465524678668e-06, "loss": 0.0, "num_input_tokens_seen": 9996160, "step": 35700 }, { "epoch": 396.72222222222223, "grad_norm": 5.637968456539966e-07, "learning_rate": 1.4095947416588124e-06, "loss": 0.0, "num_input_tokens_seen": 9997584, "step": 35705 }, { "epoch": 396.77777777777777, "grad_norm": 3.7385891005214944e-07, "learning_rate": 1.4063465687866983e-06, "loss": 0.0, "num_input_tokens_seen": 9998944, "step": 35710 }, { "epoch": 396.8333333333333, "grad_norm": 3.048039616260212e-07, "learning_rate": 1.4031020343524438e-06, "loss": 0.0, "num_input_tokens_seen": 10000368, "step": 35715 }, { "epoch": 396.8888888888889, "grad_norm": 4.344245780885103e-07, "learning_rate": 1.3998611388563926e-06, "loss": 0.0, "num_input_tokens_seen": 10001792, "step": 35720 }, { "epoch": 396.94444444444446, "grad_norm": 4.275215133020538e-07, "learning_rate": 1.3966238827983314e-06, "loss": 0.0, "num_input_tokens_seen": 10003184, "step": 35725 }, { "epoch": 397.0, "grad_norm": 7.203031060498688e-08, "learning_rate": 1.393390266677483e-06, "loss": 0.0, "num_input_tokens_seen": 10004528, "step": 35730 }, { "epoch": 397.05555555555554, "grad_norm": 1.3543934755944065e-06, "learning_rate": 1.3901602909925204e-06, "loss": 0.0, "num_input_tokens_seen": 10005936, "step": 35735 }, { "epoch": 397.1111111111111, "grad_norm": 6.167665560496971e-07, "learning_rate": 1.3869339562415373e-06, "loss": 0.0, "num_input_tokens_seen": 10007344, "step": 35740 }, { "epoch": 397.1666666666667, "grad_norm": 3.72606990595159e-07, "learning_rate": 1.38371126292208e-06, "loss": 0.0, "num_input_tokens_seen": 10008752, "step": 35745 }, { "epoch": 397.22222222222223, "grad_norm": 2.8969222398700367e-07, "learning_rate": 1.3804922115311286e-06, "loss": 0.0, "num_input_tokens_seen": 10010176, "step": 35750 }, { "epoch": 397.27777777777777, "grad_norm": 2.541165144975821e-07, "learning_rate": 1.3772768025650945e-06, "loss": 0.0, "num_input_tokens_seen": 10011536, "step": 35755 }, { "epoch": 397.3333333333333, "grad_norm": 2.3297921813991707e-07, "learning_rate": 1.3740650365198448e-06, "loss": 0.0, "num_input_tokens_seen": 10012928, "step": 35760 }, { "epoch": 397.3888888888889, "grad_norm": 5.138149958838767e-07, "learning_rate": 1.3708569138906612e-06, "loss": 0.0, "num_input_tokens_seen": 10014304, "step": 35765 }, { "epoch": 397.44444444444446, "grad_norm": 2.2717574665875873e-07, "learning_rate": 1.367652435172287e-06, "loss": 0.0, "num_input_tokens_seen": 10015728, "step": 35770 }, { "epoch": 397.5, "grad_norm": 3.560578306860407e-06, "learning_rate": 1.364451600858893e-06, "loss": 0.0, "num_input_tokens_seen": 10017152, "step": 35775 }, { "epoch": 397.55555555555554, "grad_norm": 2.40338579260424e-07, "learning_rate": 1.3612544114440823e-06, "loss": 0.0, "num_input_tokens_seen": 10018560, "step": 35780 }, { "epoch": 397.6111111111111, "grad_norm": 9.98705331767269e-07, "learning_rate": 1.3580608674209072e-06, "loss": 0.0, "num_input_tokens_seen": 10019920, "step": 35785 }, { "epoch": 397.6666666666667, "grad_norm": 4.689224795129121e-07, "learning_rate": 1.3548709692818434e-06, "loss": 0.0, "num_input_tokens_seen": 10021296, "step": 35790 }, { "epoch": 397.72222222222223, "grad_norm": 3.945012565509387e-07, "learning_rate": 1.3516847175188223e-06, "loss": 0.0, "num_input_tokens_seen": 10022704, "step": 35795 }, { "epoch": 397.77777777777777, "grad_norm": 2.4664413444952515e-07, "learning_rate": 1.348502112623204e-06, "loss": 0.0, "num_input_tokens_seen": 10024160, "step": 35800 }, { "epoch": 397.77777777777777, "eval_loss": 1.2243367433547974, "eval_runtime": 1.3936, "eval_samples_per_second": 28.703, "eval_steps_per_second": 14.352, "num_input_tokens_seen": 10024160, "step": 35800 }, { "epoch": 397.8333333333333, "grad_norm": 7.088190301374198e-08, "learning_rate": 1.3453231550857787e-06, "loss": 0.0, "num_input_tokens_seen": 10025536, "step": 35805 }, { "epoch": 397.8888888888889, "grad_norm": 2.296086904607364e-06, "learning_rate": 1.3421478453967878e-06, "loss": 0.0, "num_input_tokens_seen": 10026912, "step": 35810 }, { "epoch": 397.94444444444446, "grad_norm": 4.6374316298170015e-07, "learning_rate": 1.3389761840459065e-06, "loss": 0.0, "num_input_tokens_seen": 10028336, "step": 35815 }, { "epoch": 398.0, "grad_norm": 8.406124152315897e-07, "learning_rate": 1.3358081715222376e-06, "loss": 0.0, "num_input_tokens_seen": 10029728, "step": 35820 }, { "epoch": 398.05555555555554, "grad_norm": 6.34041555258591e-07, "learning_rate": 1.3326438083143295e-06, "loss": 0.0, "num_input_tokens_seen": 10031120, "step": 35825 }, { "epoch": 398.1111111111111, "grad_norm": 9.690824072094983e-07, "learning_rate": 1.3294830949101723e-06, "loss": 0.0, "num_input_tokens_seen": 10032544, "step": 35830 }, { "epoch": 398.1666666666667, "grad_norm": 6.299562187450647e-07, "learning_rate": 1.3263260317971815e-06, "loss": 0.0, "num_input_tokens_seen": 10033984, "step": 35835 }, { "epoch": 398.22222222222223, "grad_norm": 1.3772908857845323e-07, "learning_rate": 1.3231726194622208e-06, "loss": 0.0, "num_input_tokens_seen": 10035376, "step": 35840 }, { "epoch": 398.27777777777777, "grad_norm": 1.874420740932692e-05, "learning_rate": 1.3200228583915814e-06, "loss": 0.0, "num_input_tokens_seen": 10036800, "step": 35845 }, { "epoch": 398.3333333333333, "grad_norm": 1.9814140728158236e-07, "learning_rate": 1.3168767490709971e-06, "loss": 0.0, "num_input_tokens_seen": 10038176, "step": 35850 }, { "epoch": 398.3888888888889, "grad_norm": 4.6509751427947776e-07, "learning_rate": 1.3137342919856437e-06, "loss": 0.0, "num_input_tokens_seen": 10039600, "step": 35855 }, { "epoch": 398.44444444444446, "grad_norm": 3.5182840747438604e-06, "learning_rate": 1.310595487620117e-06, "loss": 0.0, "num_input_tokens_seen": 10040992, "step": 35860 }, { "epoch": 398.5, "grad_norm": 6.655864126514643e-08, "learning_rate": 1.3074603364584715e-06, "loss": 0.0, "num_input_tokens_seen": 10042416, "step": 35865 }, { "epoch": 398.55555555555554, "grad_norm": 2.8617799330277194e-07, "learning_rate": 1.3043288389841758e-06, "loss": 0.0, "num_input_tokens_seen": 10043760, "step": 35870 }, { "epoch": 398.6111111111111, "grad_norm": 1.2461149481168832e-06, "learning_rate": 1.3012009956801546e-06, "loss": 0.0, "num_input_tokens_seen": 10045216, "step": 35875 }, { "epoch": 398.6666666666667, "grad_norm": 7.814151103957556e-08, "learning_rate": 1.2980768070287586e-06, "loss": 0.0, "num_input_tokens_seen": 10046592, "step": 35880 }, { "epoch": 398.72222222222223, "grad_norm": 1.8118174693881883e-06, "learning_rate": 1.2949562735117716e-06, "loss": 0.0, "num_input_tokens_seen": 10047968, "step": 35885 }, { "epoch": 398.77777777777777, "grad_norm": 6.479240823864529e-07, "learning_rate": 1.291839395610428e-06, "loss": 0.0, "num_input_tokens_seen": 10049392, "step": 35890 }, { "epoch": 398.8333333333333, "grad_norm": 5.143274393049069e-07, "learning_rate": 1.2887261738053852e-06, "loss": 0.0, "num_input_tokens_seen": 10050832, "step": 35895 }, { "epoch": 398.8888888888889, "grad_norm": 6.815816959715448e-07, "learning_rate": 1.2856166085767396e-06, "loss": 0.0, "num_input_tokens_seen": 10052176, "step": 35900 }, { "epoch": 398.94444444444446, "grad_norm": 1.8396809764453792e-07, "learning_rate": 1.2825107004040272e-06, "loss": 0.0, "num_input_tokens_seen": 10053600, "step": 35905 }, { "epoch": 399.0, "grad_norm": 1.2510086889960803e-05, "learning_rate": 1.2794084497662146e-06, "loss": 0.0, "num_input_tokens_seen": 10055008, "step": 35910 }, { "epoch": 399.05555555555554, "grad_norm": 2.2959000034461496e-06, "learning_rate": 1.276309857141711e-06, "loss": 0.0, "num_input_tokens_seen": 10056432, "step": 35915 }, { "epoch": 399.1111111111111, "grad_norm": 1.7386583976986003e-06, "learning_rate": 1.273214923008359e-06, "loss": 0.0, "num_input_tokens_seen": 10057840, "step": 35920 }, { "epoch": 399.1666666666667, "grad_norm": 3.37498647695611e-07, "learning_rate": 1.2701236478434352e-06, "loss": 0.0, "num_input_tokens_seen": 10059248, "step": 35925 }, { "epoch": 399.22222222222223, "grad_norm": 8.476075663566007e-07, "learning_rate": 1.2670360321236502e-06, "loss": 0.0, "num_input_tokens_seen": 10060656, "step": 35930 }, { "epoch": 399.27777777777777, "grad_norm": 5.741948427839816e-08, "learning_rate": 1.2639520763251617e-06, "loss": 0.0, "num_input_tokens_seen": 10062080, "step": 35935 }, { "epoch": 399.3333333333333, "grad_norm": 1.3111795169606921e-06, "learning_rate": 1.2608717809235448e-06, "loss": 0.0, "num_input_tokens_seen": 10063456, "step": 35940 }, { "epoch": 399.3888888888889, "grad_norm": 8.256092769443057e-07, "learning_rate": 1.2577951463938282e-06, "loss": 0.0, "num_input_tokens_seen": 10064896, "step": 35945 }, { "epoch": 399.44444444444446, "grad_norm": 9.487014551723405e-08, "learning_rate": 1.2547221732104569e-06, "loss": 0.0, "num_input_tokens_seen": 10066320, "step": 35950 }, { "epoch": 399.5, "grad_norm": 4.91821083414834e-07, "learning_rate": 1.25165286184733e-06, "loss": 0.0, "num_input_tokens_seen": 10067696, "step": 35955 }, { "epoch": 399.55555555555554, "grad_norm": 4.675204650084197e-07, "learning_rate": 1.248587212777777e-06, "loss": 0.0, "num_input_tokens_seen": 10069120, "step": 35960 }, { "epoch": 399.6111111111111, "grad_norm": 6.388498832166079e-07, "learning_rate": 1.2455252264745532e-06, "loss": 0.0, "num_input_tokens_seen": 10070464, "step": 35965 }, { "epoch": 399.6666666666667, "grad_norm": 3.7599539837174234e-07, "learning_rate": 1.2424669034098528e-06, "loss": 0.0, "num_input_tokens_seen": 10071840, "step": 35970 }, { "epoch": 399.72222222222223, "grad_norm": 7.658212552996702e-08, "learning_rate": 1.2394122440553185e-06, "loss": 0.0, "num_input_tokens_seen": 10073248, "step": 35975 }, { "epoch": 399.77777777777777, "grad_norm": 1.274461374123348e-05, "learning_rate": 1.2363612488820037e-06, "loss": 0.0, "num_input_tokens_seen": 10074656, "step": 35980 }, { "epoch": 399.8333333333333, "grad_norm": 2.44798314952277e-07, "learning_rate": 1.2333139183604208e-06, "loss": 0.0, "num_input_tokens_seen": 10076064, "step": 35985 }, { "epoch": 399.8888888888889, "grad_norm": 7.047838721518929e-07, "learning_rate": 1.2302702529604998e-06, "loss": 0.0, "num_input_tokens_seen": 10077456, "step": 35990 }, { "epoch": 399.94444444444446, "grad_norm": 8.41961991682183e-07, "learning_rate": 1.227230253151615e-06, "loss": 0.0, "num_input_tokens_seen": 10078848, "step": 35995 }, { "epoch": 400.0, "grad_norm": 5.491542651725467e-07, "learning_rate": 1.2241939194025748e-06, "loss": 0.0, "num_input_tokens_seen": 10080240, "step": 36000 }, { "epoch": 400.0, "eval_loss": 1.1904724836349487, "eval_runtime": 1.3863, "eval_samples_per_second": 28.853, "eval_steps_per_second": 14.427, "num_input_tokens_seen": 10080240, "step": 36000 }, { "epoch": 400.05555555555554, "grad_norm": 4.3351587919460144e-07, "learning_rate": 1.2211612521816156e-06, "loss": 0.0, "num_input_tokens_seen": 10081712, "step": 36005 }, { "epoch": 400.1111111111111, "grad_norm": 8.043402743851402e-08, "learning_rate": 1.2181322519564137e-06, "loss": 0.0, "num_input_tokens_seen": 10083104, "step": 36010 }, { "epoch": 400.1666666666667, "grad_norm": 4.0090395714287297e-07, "learning_rate": 1.2151069191940839e-06, "loss": 0.0, "num_input_tokens_seen": 10084528, "step": 36015 }, { "epoch": 400.22222222222223, "grad_norm": 6.975001724640606e-06, "learning_rate": 1.2120852543611644e-06, "loss": 0.0, "num_input_tokens_seen": 10085984, "step": 36020 }, { "epoch": 400.27777777777777, "grad_norm": 1.0585796417217352e-07, "learning_rate": 1.2090672579236379e-06, "loss": 0.0, "num_input_tokens_seen": 10087392, "step": 36025 }, { "epoch": 400.3333333333333, "grad_norm": 2.8558929443534e-07, "learning_rate": 1.2060529303469126e-06, "loss": 0.0, "num_input_tokens_seen": 10088784, "step": 36030 }, { "epoch": 400.3888888888889, "grad_norm": 5.092432431297311e-08, "learning_rate": 1.2030422720958445e-06, "loss": 0.0, "num_input_tokens_seen": 10090160, "step": 36035 }, { "epoch": 400.44444444444446, "grad_norm": 8.181968951248564e-07, "learning_rate": 1.200035283634704e-06, "loss": 0.0, "num_input_tokens_seen": 10091536, "step": 36040 }, { "epoch": 400.5, "grad_norm": 1.2197075420772308e-06, "learning_rate": 1.1970319654272144e-06, "loss": 0.0, "num_input_tokens_seen": 10092928, "step": 36045 }, { "epoch": 400.55555555555554, "grad_norm": 7.035628641460789e-06, "learning_rate": 1.1940323179365192e-06, "loss": 0.0, "num_input_tokens_seen": 10094304, "step": 36050 }, { "epoch": 400.6111111111111, "grad_norm": 8.070583703556622e-07, "learning_rate": 1.1910363416252095e-06, "loss": 0.0, "num_input_tokens_seen": 10095680, "step": 36055 }, { "epoch": 400.6666666666667, "grad_norm": 1.3743510862695985e-07, "learning_rate": 1.1880440369552964e-06, "loss": 0.0, "num_input_tokens_seen": 10097072, "step": 36060 }, { "epoch": 400.72222222222223, "grad_norm": 6.885445458237882e-08, "learning_rate": 1.1850554043882328e-06, "loss": 0.0, "num_input_tokens_seen": 10098432, "step": 36065 }, { "epoch": 400.77777777777777, "grad_norm": 4.978642920150378e-08, "learning_rate": 1.1820704443849028e-06, "loss": 0.0, "num_input_tokens_seen": 10099856, "step": 36070 }, { "epoch": 400.8333333333333, "grad_norm": 5.335329689160062e-08, "learning_rate": 1.1790891574056219e-06, "loss": 0.0, "num_input_tokens_seen": 10101264, "step": 36075 }, { "epoch": 400.8888888888889, "grad_norm": 1.0073065226379185e-07, "learning_rate": 1.1761115439101523e-06, "loss": 0.0, "num_input_tokens_seen": 10102608, "step": 36080 }, { "epoch": 400.94444444444446, "grad_norm": 2.007888895150245e-07, "learning_rate": 1.1731376043576659e-06, "loss": 0.0, "num_input_tokens_seen": 10104032, "step": 36085 }, { "epoch": 401.0, "grad_norm": 2.9613599963340675e-06, "learning_rate": 1.1701673392067875e-06, "loss": 0.0, "num_input_tokens_seen": 10105376, "step": 36090 }, { "epoch": 401.05555555555554, "grad_norm": 3.547500227796263e-06, "learning_rate": 1.1672007489155757e-06, "loss": 0.0, "num_input_tokens_seen": 10106800, "step": 36095 }, { "epoch": 401.1111111111111, "grad_norm": 2.1611212730476836e-07, "learning_rate": 1.164237833941506e-06, "loss": 0.0, "num_input_tokens_seen": 10108240, "step": 36100 }, { "epoch": 401.1666666666667, "grad_norm": 2.1483000978150812e-07, "learning_rate": 1.1612785947415022e-06, "loss": 0.0, "num_input_tokens_seen": 10109648, "step": 36105 }, { "epoch": 401.22222222222223, "grad_norm": 4.6803558007013635e-07, "learning_rate": 1.1583230317719185e-06, "loss": 0.0, "num_input_tokens_seen": 10111072, "step": 36110 }, { "epoch": 401.27777777777777, "grad_norm": 2.164773036383849e-07, "learning_rate": 1.1553711454885318e-06, "loss": 0.0, "num_input_tokens_seen": 10112448, "step": 36115 }, { "epoch": 401.3333333333333, "grad_norm": 2.504574240447255e-06, "learning_rate": 1.152422936346567e-06, "loss": 0.0, "num_input_tokens_seen": 10113824, "step": 36120 }, { "epoch": 401.3888888888889, "grad_norm": 1.9640754089778056e-07, "learning_rate": 1.1494784048006718e-06, "loss": 0.0, "num_input_tokens_seen": 10115232, "step": 36125 }, { "epoch": 401.44444444444446, "grad_norm": 7.620642463734839e-07, "learning_rate": 1.1465375513049326e-06, "loss": 0.0, "num_input_tokens_seen": 10116656, "step": 36130 }, { "epoch": 401.5, "grad_norm": 5.934008413532865e-07, "learning_rate": 1.1436003763128616e-06, "loss": 0.0, "num_input_tokens_seen": 10118096, "step": 36135 }, { "epoch": 401.55555555555554, "grad_norm": 3.10776215428632e-07, "learning_rate": 1.1406668802774106e-06, "loss": 0.0, "num_input_tokens_seen": 10119488, "step": 36140 }, { "epoch": 401.6111111111111, "grad_norm": 8.446773449577449e-07, "learning_rate": 1.137737063650965e-06, "loss": 0.0, "num_input_tokens_seen": 10120896, "step": 36145 }, { "epoch": 401.6666666666667, "grad_norm": 8.256843386789114e-08, "learning_rate": 1.1348109268853323e-06, "loss": 0.0, "num_input_tokens_seen": 10122304, "step": 36150 }, { "epoch": 401.72222222222223, "grad_norm": 6.436230393092046e-08, "learning_rate": 1.1318884704317634e-06, "loss": 0.0, "num_input_tokens_seen": 10123712, "step": 36155 }, { "epoch": 401.77777777777777, "grad_norm": 2.22706717067922e-06, "learning_rate": 1.1289696947409417e-06, "loss": 0.0, "num_input_tokens_seen": 10125104, "step": 36160 }, { "epoch": 401.8333333333333, "grad_norm": 1.1193940707698857e-07, "learning_rate": 1.126054600262974e-06, "loss": 0.0, "num_input_tokens_seen": 10126496, "step": 36165 }, { "epoch": 401.8888888888889, "grad_norm": 3.5947884953202447e-06, "learning_rate": 1.1231431874474064e-06, "loss": 0.0, "num_input_tokens_seen": 10127840, "step": 36170 }, { "epoch": 401.94444444444446, "grad_norm": 5.0260243966704365e-08, "learning_rate": 1.12023545674321e-06, "loss": 0.0, "num_input_tokens_seen": 10129216, "step": 36175 }, { "epoch": 402.0, "grad_norm": 6.755269055247481e-07, "learning_rate": 1.117331408598804e-06, "loss": 0.0, "num_input_tokens_seen": 10130624, "step": 36180 }, { "epoch": 402.05555555555554, "grad_norm": 1.2164417739768396e-07, "learning_rate": 1.1144310434620191e-06, "loss": 0.0, "num_input_tokens_seen": 10132032, "step": 36185 }, { "epoch": 402.1111111111111, "grad_norm": 5.854059850207705e-07, "learning_rate": 1.1115343617801365e-06, "loss": 0.0, "num_input_tokens_seen": 10133392, "step": 36190 }, { "epoch": 402.1666666666667, "grad_norm": 2.2063056803744985e-06, "learning_rate": 1.1086413639998515e-06, "loss": 0.0, "num_input_tokens_seen": 10134800, "step": 36195 }, { "epoch": 402.22222222222223, "grad_norm": 2.9796817102578643e-07, "learning_rate": 1.1057520505673103e-06, "loss": 0.0, "num_input_tokens_seen": 10136208, "step": 36200 }, { "epoch": 402.22222222222223, "eval_loss": 1.2014354467391968, "eval_runtime": 1.3922, "eval_samples_per_second": 28.732, "eval_steps_per_second": 14.366, "num_input_tokens_seen": 10136208, "step": 36200 }, { "epoch": 402.27777777777777, "grad_norm": 3.898488216691476e-07, "learning_rate": 1.1028664219280727e-06, "loss": 0.0, "num_input_tokens_seen": 10137648, "step": 36205 }, { "epoch": 402.3333333333333, "grad_norm": 6.338123966997955e-06, "learning_rate": 1.0999844785271468e-06, "loss": 0.0, "num_input_tokens_seen": 10139056, "step": 36210 }, { "epoch": 402.3888888888889, "grad_norm": 1.3043540093349293e-05, "learning_rate": 1.097106220808955e-06, "loss": 0.0, "num_input_tokens_seen": 10140432, "step": 36215 }, { "epoch": 402.44444444444446, "grad_norm": 1.376296836497204e-07, "learning_rate": 1.0942316492173698e-06, "loss": 0.0, "num_input_tokens_seen": 10141824, "step": 36220 }, { "epoch": 402.5, "grad_norm": 1.4986005680839298e-07, "learning_rate": 1.0913607641956841e-06, "loss": 0.0, "num_input_tokens_seen": 10143216, "step": 36225 }, { "epoch": 402.55555555555554, "grad_norm": 2.850507144103176e-06, "learning_rate": 1.0884935661866213e-06, "loss": 0.0, "num_input_tokens_seen": 10144624, "step": 36230 }, { "epoch": 402.6111111111111, "grad_norm": 8.35962453038519e-07, "learning_rate": 1.0856300556323418e-06, "loss": 0.0, "num_input_tokens_seen": 10146032, "step": 36235 }, { "epoch": 402.6666666666667, "grad_norm": 1.6322644569299882e-07, "learning_rate": 1.0827702329744365e-06, "loss": 0.0, "num_input_tokens_seen": 10147408, "step": 36240 }, { "epoch": 402.72222222222223, "grad_norm": 1.3633795958867267e-07, "learning_rate": 1.0799140986539197e-06, "loss": 0.0, "num_input_tokens_seen": 10148816, "step": 36245 }, { "epoch": 402.77777777777777, "grad_norm": 2.2802431942636758e-07, "learning_rate": 1.0770616531112526e-06, "loss": 0.0, "num_input_tokens_seen": 10150224, "step": 36250 }, { "epoch": 402.8333333333333, "grad_norm": 1.0596238098514732e-05, "learning_rate": 1.0742128967863085e-06, "loss": 0.0, "num_input_tokens_seen": 10151648, "step": 36255 }, { "epoch": 402.8888888888889, "grad_norm": 1.1331574967243796e-07, "learning_rate": 1.071367830118411e-06, "loss": 0.0, "num_input_tokens_seen": 10152992, "step": 36260 }, { "epoch": 402.94444444444446, "grad_norm": 2.785224069157266e-07, "learning_rate": 1.068526453546298e-06, "loss": 0.0, "num_input_tokens_seen": 10154384, "step": 36265 }, { "epoch": 403.0, "grad_norm": 1.13848159344343e-06, "learning_rate": 1.0656887675081467e-06, "loss": 0.0, "num_input_tokens_seen": 10155744, "step": 36270 }, { "epoch": 403.05555555555554, "grad_norm": 3.804958481623544e-08, "learning_rate": 1.0628547724415628e-06, "loss": 0.0, "num_input_tokens_seen": 10157088, "step": 36275 }, { "epoch": 403.1111111111111, "grad_norm": 8.243335400948126e-07, "learning_rate": 1.0600244687835881e-06, "loss": 0.0, "num_input_tokens_seen": 10158512, "step": 36280 }, { "epoch": 403.1666666666667, "grad_norm": 3.499741296764114e-06, "learning_rate": 1.0571978569706876e-06, "loss": 0.0, "num_input_tokens_seen": 10159920, "step": 36285 }, { "epoch": 403.22222222222223, "grad_norm": 1.1081697266490664e-06, "learning_rate": 1.0543749374387652e-06, "loss": 0.0, "num_input_tokens_seen": 10161296, "step": 36290 }, { "epoch": 403.27777777777777, "grad_norm": 2.5133607550742454e-07, "learning_rate": 1.051555710623142e-06, "loss": 0.0, "num_input_tokens_seen": 10162704, "step": 36295 }, { "epoch": 403.3333333333333, "grad_norm": 1.7096407134431502e-07, "learning_rate": 1.0487401769585847e-06, "loss": 0.0, "num_input_tokens_seen": 10164128, "step": 36300 }, { "epoch": 403.3888888888889, "grad_norm": 5.793913828711084e-07, "learning_rate": 1.0459283368792845e-06, "loss": 0.0, "num_input_tokens_seen": 10165552, "step": 36305 }, { "epoch": 403.44444444444446, "grad_norm": 4.1700889141793596e-07, "learning_rate": 1.043120190818858e-06, "loss": 0.0, "num_input_tokens_seen": 10166976, "step": 36310 }, { "epoch": 403.5, "grad_norm": 1.271510257083719e-07, "learning_rate": 1.0403157392103596e-06, "loss": 0.0, "num_input_tokens_seen": 10168384, "step": 36315 }, { "epoch": 403.55555555555554, "grad_norm": 6.793267175453366e-07, "learning_rate": 1.0375149824862735e-06, "loss": 0.0, "num_input_tokens_seen": 10169760, "step": 36320 }, { "epoch": 403.6111111111111, "grad_norm": 2.795440650515957e-06, "learning_rate": 1.034717921078507e-06, "loss": 0.0, "num_input_tokens_seen": 10171200, "step": 36325 }, { "epoch": 403.6666666666667, "grad_norm": 1.3116044783600955e-07, "learning_rate": 1.0319245554184009e-06, "loss": 0.0, "num_input_tokens_seen": 10172624, "step": 36330 }, { "epoch": 403.72222222222223, "grad_norm": 9.590112313162535e-07, "learning_rate": 1.0291348859367361e-06, "loss": 0.0, "num_input_tokens_seen": 10174000, "step": 36335 }, { "epoch": 403.77777777777777, "grad_norm": 1.2451678230718244e-06, "learning_rate": 1.0263489130637016e-06, "loss": 0.0, "num_input_tokens_seen": 10175408, "step": 36340 }, { "epoch": 403.8333333333333, "grad_norm": 1.7703827381865267e-07, "learning_rate": 1.0235666372289427e-06, "loss": 0.0, "num_input_tokens_seen": 10176816, "step": 36345 }, { "epoch": 403.8888888888889, "grad_norm": 5.737029482588696e-07, "learning_rate": 1.0207880588615076e-06, "loss": 0.0, "num_input_tokens_seen": 10178208, "step": 36350 }, { "epoch": 403.94444444444446, "grad_norm": 6.878045866187676e-08, "learning_rate": 1.0180131783898984e-06, "loss": 0.0, "num_input_tokens_seen": 10179600, "step": 36355 }, { "epoch": 404.0, "grad_norm": 1.0034904107669718e-06, "learning_rate": 1.0152419962420362e-06, "loss": 0.0, "num_input_tokens_seen": 10180976, "step": 36360 }, { "epoch": 404.05555555555554, "grad_norm": 3.6905960314470576e-06, "learning_rate": 1.0124745128452685e-06, "loss": 0.0, "num_input_tokens_seen": 10182384, "step": 36365 }, { "epoch": 404.1111111111111, "grad_norm": 3.984415798186092e-06, "learning_rate": 1.0097107286263758e-06, "loss": 0.0, "num_input_tokens_seen": 10183728, "step": 36370 }, { "epoch": 404.1666666666667, "grad_norm": 6.328058930193947e-07, "learning_rate": 1.00695064401157e-06, "loss": 0.0, "num_input_tokens_seen": 10185152, "step": 36375 }, { "epoch": 404.22222222222223, "grad_norm": 6.99607198839658e-08, "learning_rate": 1.0041942594264886e-06, "loss": 0.0, "num_input_tokens_seen": 10186560, "step": 36380 }, { "epoch": 404.27777777777777, "grad_norm": 1.5905509371805238e-06, "learning_rate": 1.001441575296208e-06, "loss": 0.0, "num_input_tokens_seen": 10187984, "step": 36385 }, { "epoch": 404.3333333333333, "grad_norm": 1.0099170566491011e-07, "learning_rate": 9.986925920452139e-07, "loss": 0.0, "num_input_tokens_seen": 10189392, "step": 36390 }, { "epoch": 404.3888888888889, "grad_norm": 4.635878667613724e-07, "learning_rate": 9.959473100974475e-07, "loss": 0.0, "num_input_tokens_seen": 10190800, "step": 36395 }, { "epoch": 404.44444444444446, "grad_norm": 2.2291372658855835e-07, "learning_rate": 9.932057298762564e-07, "loss": 0.0, "num_input_tokens_seen": 10192208, "step": 36400 }, { "epoch": 404.44444444444446, "eval_loss": 1.2126741409301758, "eval_runtime": 1.4021, "eval_samples_per_second": 28.528, "eval_steps_per_second": 14.264, "num_input_tokens_seen": 10192208, "step": 36400 }, { "epoch": 404.5, "grad_norm": 3.2188938803301426e-07, "learning_rate": 9.90467851804433e-07, "loss": 0.0, "num_input_tokens_seen": 10193616, "step": 36405 }, { "epoch": 404.55555555555554, "grad_norm": 6.150340681188027e-08, "learning_rate": 9.877336763041895e-07, "loss": 0.0, "num_input_tokens_seen": 10195008, "step": 36410 }, { "epoch": 404.6111111111111, "grad_norm": 1.0073772500618361e-05, "learning_rate": 9.850032037971662e-07, "loss": 0.0, "num_input_tokens_seen": 10196432, "step": 36415 }, { "epoch": 404.6666666666667, "grad_norm": 3.920395101886243e-06, "learning_rate": 9.822764347044406e-07, "loss": 0.0, "num_input_tokens_seen": 10197856, "step": 36420 }, { "epoch": 404.72222222222223, "grad_norm": 4.8143014907964243e-08, "learning_rate": 9.795533694465175e-07, "loss": 0.0, "num_input_tokens_seen": 10199248, "step": 36425 }, { "epoch": 404.77777777777777, "grad_norm": 6.869855496915989e-06, "learning_rate": 9.768340084433197e-07, "loss": 0.0, "num_input_tokens_seen": 10200704, "step": 36430 }, { "epoch": 404.8333333333333, "grad_norm": 1.9531423731677933e-06, "learning_rate": 9.741183521142143e-07, "loss": 0.0, "num_input_tokens_seen": 10202128, "step": 36435 }, { "epoch": 404.8888888888889, "grad_norm": 9.604243444982785e-08, "learning_rate": 9.714064008779889e-07, "loss": 0.0, "num_input_tokens_seen": 10203488, "step": 36440 }, { "epoch": 404.94444444444446, "grad_norm": 2.116566975018941e-06, "learning_rate": 9.686981551528584e-07, "loss": 0.0, "num_input_tokens_seen": 10204880, "step": 36445 }, { "epoch": 405.0, "grad_norm": 6.484055603550587e-08, "learning_rate": 9.65993615356467e-07, "loss": 0.0, "num_input_tokens_seen": 10206272, "step": 36450 }, { "epoch": 405.05555555555554, "grad_norm": 1.198649073330671e-07, "learning_rate": 9.632927819058917e-07, "loss": 0.0, "num_input_tokens_seen": 10207648, "step": 36455 }, { "epoch": 405.1111111111111, "grad_norm": 5.945296166487424e-08, "learning_rate": 9.605956552176305e-07, "loss": 0.0, "num_input_tokens_seen": 10209040, "step": 36460 }, { "epoch": 405.1666666666667, "grad_norm": 3.757903641599114e-06, "learning_rate": 9.579022357076223e-07, "loss": 0.0, "num_input_tokens_seen": 10210416, "step": 36465 }, { "epoch": 405.22222222222223, "grad_norm": 8.22866468297434e-08, "learning_rate": 9.552125237912158e-07, "loss": 0.0, "num_input_tokens_seen": 10211792, "step": 36470 }, { "epoch": 405.27777777777777, "grad_norm": 4.0389622313341533e-07, "learning_rate": 9.525265198832096e-07, "loss": 0.0, "num_input_tokens_seen": 10213200, "step": 36475 }, { "epoch": 405.3333333333333, "grad_norm": 5.801951374451164e-06, "learning_rate": 9.498442243978112e-07, "loss": 0.0, "num_input_tokens_seen": 10214608, "step": 36480 }, { "epoch": 405.3888888888889, "grad_norm": 1.6266960756183835e-07, "learning_rate": 9.471656377486649e-07, "loss": 0.0, "num_input_tokens_seen": 10215968, "step": 36485 }, { "epoch": 405.44444444444446, "grad_norm": 5.405820857617982e-08, "learning_rate": 9.444907603488456e-07, "loss": 0.0, "num_input_tokens_seen": 10217376, "step": 36490 }, { "epoch": 405.5, "grad_norm": 5.291691650199937e-06, "learning_rate": 9.418195926108514e-07, "loss": 0.0, "num_input_tokens_seen": 10218784, "step": 36495 }, { "epoch": 405.55555555555554, "grad_norm": 3.1043188641888264e-07, "learning_rate": 9.391521349466053e-07, "loss": 0.0, "num_input_tokens_seen": 10220224, "step": 36500 }, { "epoch": 405.6111111111111, "grad_norm": 3.547836797679338e-07, "learning_rate": 9.364883877674758e-07, "loss": 0.0, "num_input_tokens_seen": 10221648, "step": 36505 }, { "epoch": 405.6666666666667, "grad_norm": 4.720850910189256e-08, "learning_rate": 9.33828351484231e-07, "loss": 0.0, "num_input_tokens_seen": 10223056, "step": 36510 }, { "epoch": 405.72222222222223, "grad_norm": 4.49164218707665e-07, "learning_rate": 9.311720265070906e-07, "loss": 0.0, "num_input_tokens_seen": 10224464, "step": 36515 }, { "epoch": 405.77777777777777, "grad_norm": 6.873488587189058e-07, "learning_rate": 9.285194132456931e-07, "loss": 0.0, "num_input_tokens_seen": 10225840, "step": 36520 }, { "epoch": 405.8333333333333, "grad_norm": 6.337955937851802e-07, "learning_rate": 9.258705121091032e-07, "loss": 0.0, "num_input_tokens_seen": 10227200, "step": 36525 }, { "epoch": 405.8888888888889, "grad_norm": 6.112141477387922e-07, "learning_rate": 9.232253235058136e-07, "loss": 0.0, "num_input_tokens_seen": 10228624, "step": 36530 }, { "epoch": 405.94444444444446, "grad_norm": 6.431147880903154e-07, "learning_rate": 9.205838478437478e-07, "loss": 0.0, "num_input_tokens_seen": 10230000, "step": 36535 }, { "epoch": 406.0, "grad_norm": 1.8555138581177744e-07, "learning_rate": 9.179460855302524e-07, "loss": 0.0, "num_input_tokens_seen": 10231360, "step": 36540 }, { "epoch": 406.05555555555554, "grad_norm": 1.2521662029030267e-05, "learning_rate": 9.153120369721046e-07, "loss": 0.0, "num_input_tokens_seen": 10232736, "step": 36545 }, { "epoch": 406.1111111111111, "grad_norm": 1.9376004445348372e-07, "learning_rate": 9.126817025755103e-07, "loss": 0.0, "num_input_tokens_seen": 10234144, "step": 36550 }, { "epoch": 406.1666666666667, "grad_norm": 2.448768441354332e-07, "learning_rate": 9.100550827460947e-07, "loss": 0.0, "num_input_tokens_seen": 10235520, "step": 36555 }, { "epoch": 406.22222222222223, "grad_norm": 5.348189233700396e-07, "learning_rate": 9.0743217788892e-07, "loss": 0.0, "num_input_tokens_seen": 10236960, "step": 36560 }, { "epoch": 406.27777777777777, "grad_norm": 1.0407516981558729e-07, "learning_rate": 9.048129884084683e-07, "loss": 0.0, "num_input_tokens_seen": 10238384, "step": 36565 }, { "epoch": 406.3333333333333, "grad_norm": 4.044961698923544e-08, "learning_rate": 9.021975147086553e-07, "loss": 0.0, "num_input_tokens_seen": 10239776, "step": 36570 }, { "epoch": 406.3888888888889, "grad_norm": 2.5212486320924654e-07, "learning_rate": 8.995857571928141e-07, "loss": 0.0, "num_input_tokens_seen": 10241184, "step": 36575 }, { "epoch": 406.44444444444446, "grad_norm": 8.613626789610862e-08, "learning_rate": 8.969777162637139e-07, "loss": 0.0, "num_input_tokens_seen": 10242576, "step": 36580 }, { "epoch": 406.5, "grad_norm": 1.9188419173588045e-05, "learning_rate": 8.943733923235525e-07, "loss": 0.0, "num_input_tokens_seen": 10244000, "step": 36585 }, { "epoch": 406.55555555555554, "grad_norm": 1.3864546417607926e-06, "learning_rate": 8.917727857739394e-07, "loss": 0.0, "num_input_tokens_seen": 10245392, "step": 36590 }, { "epoch": 406.6111111111111, "grad_norm": 4.852159690926783e-06, "learning_rate": 8.891758970159258e-07, "loss": 0.0, "num_input_tokens_seen": 10246816, "step": 36595 }, { "epoch": 406.6666666666667, "grad_norm": 8.145890717514703e-08, "learning_rate": 8.86582726449986e-07, "loss": 0.0, "num_input_tokens_seen": 10248192, "step": 36600 }, { "epoch": 406.6666666666667, "eval_loss": 1.2262276411056519, "eval_runtime": 1.3957, "eval_samples_per_second": 28.659, "eval_steps_per_second": 14.33, "num_input_tokens_seen": 10248192, "step": 36600 }, { "epoch": 406.72222222222223, "grad_norm": 2.3260929538082564e-07, "learning_rate": 8.839932744760165e-07, "loss": 0.0, "num_input_tokens_seen": 10249600, "step": 36605 }, { "epoch": 406.77777777777777, "grad_norm": 6.655975539615611e-07, "learning_rate": 8.814075414933482e-07, "loss": 0.0, "num_input_tokens_seen": 10250992, "step": 36610 }, { "epoch": 406.8333333333333, "grad_norm": 6.456133178289747e-06, "learning_rate": 8.788255279007257e-07, "loss": 0.0, "num_input_tokens_seen": 10252384, "step": 36615 }, { "epoch": 406.8888888888889, "grad_norm": 7.561284860457818e-07, "learning_rate": 8.762472340963362e-07, "loss": 0.0, "num_input_tokens_seen": 10253776, "step": 36620 }, { "epoch": 406.94444444444446, "grad_norm": 4.97836936119711e-08, "learning_rate": 8.736726604777811e-07, "loss": 0.0, "num_input_tokens_seen": 10255168, "step": 36625 }, { "epoch": 407.0, "grad_norm": 4.977372327630292e-07, "learning_rate": 8.711018074420901e-07, "loss": 0.0, "num_input_tokens_seen": 10256544, "step": 36630 }, { "epoch": 407.05555555555554, "grad_norm": 3.3304448265880637e-07, "learning_rate": 8.685346753857209e-07, "loss": 0.0, "num_input_tokens_seen": 10257952, "step": 36635 }, { "epoch": 407.1111111111111, "grad_norm": 1.3977277376397979e-05, "learning_rate": 8.659712647045654e-07, "loss": 0.0, "num_input_tokens_seen": 10259328, "step": 36640 }, { "epoch": 407.1666666666667, "grad_norm": 1.1130530452874154e-07, "learning_rate": 8.634115757939209e-07, "loss": 0.0, "num_input_tokens_seen": 10260704, "step": 36645 }, { "epoch": 407.22222222222223, "grad_norm": 9.120284971686488e-07, "learning_rate": 8.608556090485387e-07, "loss": 0.0, "num_input_tokens_seen": 10262096, "step": 36650 }, { "epoch": 407.27777777777777, "grad_norm": 8.241162845479266e-07, "learning_rate": 8.583033648625671e-07, "loss": 0.0, "num_input_tokens_seen": 10263488, "step": 36655 }, { "epoch": 407.3333333333333, "grad_norm": 1.9274290252724313e-07, "learning_rate": 8.557548436295998e-07, "loss": 0.0, "num_input_tokens_seen": 10264864, "step": 36660 }, { "epoch": 407.3888888888889, "grad_norm": 1.0398564853630887e-07, "learning_rate": 8.532100457426556e-07, "loss": 0.0, "num_input_tokens_seen": 10266320, "step": 36665 }, { "epoch": 407.44444444444446, "grad_norm": 4.663148956751684e-06, "learning_rate": 8.506689715941679e-07, "loss": 0.0, "num_input_tokens_seen": 10267728, "step": 36670 }, { "epoch": 407.5, "grad_norm": 6.384738071574247e-08, "learning_rate": 8.481316215760011e-07, "loss": 0.0, "num_input_tokens_seen": 10269104, "step": 36675 }, { "epoch": 407.55555555555554, "grad_norm": 1.0399198799859732e-06, "learning_rate": 8.455979960794558e-07, "loss": 0.0, "num_input_tokens_seen": 10270528, "step": 36680 }, { "epoch": 407.6111111111111, "grad_norm": 6.862606483082345e-07, "learning_rate": 8.430680954952364e-07, "loss": 0.0, "num_input_tokens_seen": 10271920, "step": 36685 }, { "epoch": 407.6666666666667, "grad_norm": 3.035802365047857e-07, "learning_rate": 8.405419202134974e-07, "loss": 0.0, "num_input_tokens_seen": 10273296, "step": 36690 }, { "epoch": 407.72222222222223, "grad_norm": 5.313935162121197e-06, "learning_rate": 8.380194706237993e-07, "loss": 0.0, "num_input_tokens_seen": 10274688, "step": 36695 }, { "epoch": 407.77777777777777, "grad_norm": 1.7427081502319197e-06, "learning_rate": 8.355007471151366e-07, "loss": 0.0, "num_input_tokens_seen": 10276128, "step": 36700 }, { "epoch": 407.8333333333333, "grad_norm": 3.644458956841845e-07, "learning_rate": 8.329857500759292e-07, "loss": 0.0, "num_input_tokens_seen": 10277520, "step": 36705 }, { "epoch": 407.8888888888889, "grad_norm": 3.0487694857583847e-07, "learning_rate": 8.304744798940194e-07, "loss": 0.0, "num_input_tokens_seen": 10278912, "step": 36710 }, { "epoch": 407.94444444444446, "grad_norm": 1.1175555414411065e-07, "learning_rate": 8.279669369566756e-07, "loss": 0.0, "num_input_tokens_seen": 10280288, "step": 36715 }, { "epoch": 408.0, "grad_norm": 2.4237974116658734e-07, "learning_rate": 8.254631216505993e-07, "loss": 0.0, "num_input_tokens_seen": 10281680, "step": 36720 }, { "epoch": 408.05555555555554, "grad_norm": 6.210158858266368e-07, "learning_rate": 8.229630343619038e-07, "loss": 0.0, "num_input_tokens_seen": 10283104, "step": 36725 }, { "epoch": 408.1111111111111, "grad_norm": 1.761676116984745e-06, "learning_rate": 8.204666754761392e-07, "loss": 0.0, "num_input_tokens_seen": 10284544, "step": 36730 }, { "epoch": 408.1666666666667, "grad_norm": 6.494096396636451e-06, "learning_rate": 8.179740453782669e-07, "loss": 0.0, "num_input_tokens_seen": 10285952, "step": 36735 }, { "epoch": 408.22222222222223, "grad_norm": 5.4889600420438e-08, "learning_rate": 8.154851444526907e-07, "loss": 0.0, "num_input_tokens_seen": 10287328, "step": 36740 }, { "epoch": 408.27777777777777, "grad_norm": 2.5283802074227424e-07, "learning_rate": 8.129999730832283e-07, "loss": 0.0, "num_input_tokens_seen": 10288688, "step": 36745 }, { "epoch": 408.3333333333333, "grad_norm": 2.887972812004591e-07, "learning_rate": 8.105185316531178e-07, "loss": 0.0, "num_input_tokens_seen": 10290064, "step": 36750 }, { "epoch": 408.3888888888889, "grad_norm": 3.7875593079661485e-06, "learning_rate": 8.08040820545039e-07, "loss": 0.0, "num_input_tokens_seen": 10291472, "step": 36755 }, { "epoch": 408.44444444444446, "grad_norm": 1.68123477806148e-07, "learning_rate": 8.055668401410782e-07, "loss": 0.0, "num_input_tokens_seen": 10292896, "step": 36760 }, { "epoch": 408.5, "grad_norm": 1.353281822957797e-05, "learning_rate": 8.030965908227578e-07, "loss": 0.0, "num_input_tokens_seen": 10294352, "step": 36765 }, { "epoch": 408.55555555555554, "grad_norm": 4.245974594141444e-07, "learning_rate": 8.006300729710203e-07, "loss": 0.0, "num_input_tokens_seen": 10295760, "step": 36770 }, { "epoch": 408.6111111111111, "grad_norm": 2.2352870132635871e-07, "learning_rate": 7.981672869662337e-07, "loss": 0.0, "num_input_tokens_seen": 10297168, "step": 36775 }, { "epoch": 408.6666666666667, "grad_norm": 1.3852674385361752e-07, "learning_rate": 7.957082331881888e-07, "loss": 0.0, "num_input_tokens_seen": 10298592, "step": 36780 }, { "epoch": 408.72222222222223, "grad_norm": 2.4548278361180564e-06, "learning_rate": 7.932529120161069e-07, "loss": 0.0, "num_input_tokens_seen": 10299952, "step": 36785 }, { "epoch": 408.77777777777777, "grad_norm": 1.4322927199827973e-07, "learning_rate": 7.908013238286243e-07, "loss": 0.0, "num_input_tokens_seen": 10301392, "step": 36790 }, { "epoch": 408.8333333333333, "grad_norm": 4.181148938187107e-07, "learning_rate": 7.883534690038136e-07, "loss": 0.0, "num_input_tokens_seen": 10302768, "step": 36795 }, { "epoch": 408.8888888888889, "grad_norm": 1.2687149819612387e-06, "learning_rate": 7.859093479191559e-07, "loss": 0.0, "num_input_tokens_seen": 10304144, "step": 36800 }, { "epoch": 408.8888888888889, "eval_loss": 1.2095892429351807, "eval_runtime": 1.3902, "eval_samples_per_second": 28.773, "eval_steps_per_second": 14.387, "num_input_tokens_seen": 10304144, "step": 36800 }, { "epoch": 408.94444444444446, "grad_norm": 1.2190012910195946e-07, "learning_rate": 7.834689609515722e-07, "loss": 0.0, "num_input_tokens_seen": 10305552, "step": 36805 }, { "epoch": 409.0, "grad_norm": 2.855068714779918e-07, "learning_rate": 7.810323084774002e-07, "loss": 0.0, "num_input_tokens_seen": 10306960, "step": 36810 }, { "epoch": 409.05555555555554, "grad_norm": 1.0419663567518e-06, "learning_rate": 7.785993908723976e-07, "loss": 0.0, "num_input_tokens_seen": 10308320, "step": 36815 }, { "epoch": 409.1111111111111, "grad_norm": 7.446784877629398e-08, "learning_rate": 7.761702085117534e-07, "loss": 0.0, "num_input_tokens_seen": 10309728, "step": 36820 }, { "epoch": 409.1666666666667, "grad_norm": 2.621673047542572e-07, "learning_rate": 7.737447617700844e-07, "loss": 0.0, "num_input_tokens_seen": 10311136, "step": 36825 }, { "epoch": 409.22222222222223, "grad_norm": 2.9605023428302957e-06, "learning_rate": 7.713230510214136e-07, "loss": 0.0, "num_input_tokens_seen": 10312544, "step": 36830 }, { "epoch": 409.27777777777777, "grad_norm": 3.4494412375352113e-06, "learning_rate": 7.689050766392092e-07, "loss": 0.0, "num_input_tokens_seen": 10313936, "step": 36835 }, { "epoch": 409.3333333333333, "grad_norm": 4.876926027463924e-07, "learning_rate": 7.664908389963477e-07, "loss": 0.0, "num_input_tokens_seen": 10315376, "step": 36840 }, { "epoch": 409.3888888888889, "grad_norm": 7.060921234369744e-07, "learning_rate": 7.64080338465134e-07, "loss": 0.0, "num_input_tokens_seen": 10316752, "step": 36845 }, { "epoch": 409.44444444444446, "grad_norm": 1.11078620079752e-07, "learning_rate": 7.616735754173043e-07, "loss": 0.0, "num_input_tokens_seen": 10318112, "step": 36850 }, { "epoch": 409.5, "grad_norm": 6.439677235903218e-08, "learning_rate": 7.592705502240005e-07, "loss": 0.0, "num_input_tokens_seen": 10319520, "step": 36855 }, { "epoch": 409.55555555555554, "grad_norm": 2.4485163407916843e-07, "learning_rate": 7.568712632558095e-07, "loss": 0.0, "num_input_tokens_seen": 10320880, "step": 36860 }, { "epoch": 409.6111111111111, "grad_norm": 3.62506534656859e-06, "learning_rate": 7.544757148827297e-07, "loss": 0.0, "num_input_tokens_seen": 10322304, "step": 36865 }, { "epoch": 409.6666666666667, "grad_norm": 1.9545343548088567e-06, "learning_rate": 7.520839054741797e-07, "loss": 0.0, "num_input_tokens_seen": 10323760, "step": 36870 }, { "epoch": 409.72222222222223, "grad_norm": 2.9836311910003133e-07, "learning_rate": 7.496958353990113e-07, "loss": 0.0, "num_input_tokens_seen": 10325168, "step": 36875 }, { "epoch": 409.77777777777777, "grad_norm": 1.289011379412841e-06, "learning_rate": 7.473115050254941e-07, "loss": 0.0, "num_input_tokens_seen": 10326608, "step": 36880 }, { "epoch": 409.8333333333333, "grad_norm": 2.2123245457805751e-07, "learning_rate": 7.449309147213173e-07, "loss": 0.0, "num_input_tokens_seen": 10327984, "step": 36885 }, { "epoch": 409.8888888888889, "grad_norm": 2.720621807839052e-07, "learning_rate": 7.425540648536067e-07, "loss": 0.0, "num_input_tokens_seen": 10329360, "step": 36890 }, { "epoch": 409.94444444444446, "grad_norm": 4.4266994336794596e-07, "learning_rate": 7.40180955788894e-07, "loss": 0.0, "num_input_tokens_seen": 10330752, "step": 36895 }, { "epoch": 410.0, "grad_norm": 3.8139378375490196e-06, "learning_rate": 7.378115878931474e-07, "loss": 0.0, "num_input_tokens_seen": 10332176, "step": 36900 }, { "epoch": 410.05555555555554, "grad_norm": 1.2033119674015325e-06, "learning_rate": 7.354459615317527e-07, "loss": 0.0, "num_input_tokens_seen": 10333600, "step": 36905 }, { "epoch": 410.1111111111111, "grad_norm": 2.693337535220053e-07, "learning_rate": 7.33084077069518e-07, "loss": 0.0, "num_input_tokens_seen": 10335024, "step": 36910 }, { "epoch": 410.1666666666667, "grad_norm": 1.6012940307064127e-07, "learning_rate": 7.307259348706768e-07, "loss": 0.0, "num_input_tokens_seen": 10336448, "step": 36915 }, { "epoch": 410.22222222222223, "grad_norm": 8.815512586579644e-08, "learning_rate": 7.283715352988801e-07, "loss": 0.0, "num_input_tokens_seen": 10337856, "step": 36920 }, { "epoch": 410.27777777777777, "grad_norm": 3.9279416341742035e-06, "learning_rate": 7.260208787172068e-07, "loss": 0.0, "num_input_tokens_seen": 10339200, "step": 36925 }, { "epoch": 410.3333333333333, "grad_norm": 1.4228958207240794e-06, "learning_rate": 7.23673965488167e-07, "loss": 0.0, "num_input_tokens_seen": 10340608, "step": 36930 }, { "epoch": 410.3888888888889, "grad_norm": 2.8280271635594545e-07, "learning_rate": 7.213307959736709e-07, "loss": 0.0, "num_input_tokens_seen": 10342000, "step": 36935 }, { "epoch": 410.44444444444446, "grad_norm": 3.097839282872883e-07, "learning_rate": 7.189913705350715e-07, "loss": 0.0, "num_input_tokens_seen": 10343424, "step": 36940 }, { "epoch": 410.5, "grad_norm": 5.213235567680385e-07, "learning_rate": 7.166556895331411e-07, "loss": 0.0, "num_input_tokens_seen": 10344784, "step": 36945 }, { "epoch": 410.55555555555554, "grad_norm": 2.989897893712623e-07, "learning_rate": 7.143237533280639e-07, "loss": 0.0, "num_input_tokens_seen": 10346128, "step": 36950 }, { "epoch": 410.6111111111111, "grad_norm": 3.055561137443874e-08, "learning_rate": 7.119955622794578e-07, "loss": 0.0, "num_input_tokens_seen": 10347488, "step": 36955 }, { "epoch": 410.6666666666667, "grad_norm": 2.8850857347606507e-07, "learning_rate": 7.096711167463577e-07, "loss": 0.0, "num_input_tokens_seen": 10348912, "step": 36960 }, { "epoch": 410.72222222222223, "grad_norm": 2.979727184992953e-07, "learning_rate": 7.073504170872213e-07, "loss": 0.0, "num_input_tokens_seen": 10350288, "step": 36965 }, { "epoch": 410.77777777777777, "grad_norm": 1.5916783979719185e-07, "learning_rate": 7.05033463659932e-07, "loss": 0.0, "num_input_tokens_seen": 10351712, "step": 36970 }, { "epoch": 410.8333333333333, "grad_norm": 5.540093752642861e-07, "learning_rate": 7.027202568217928e-07, "loss": 0.0, "num_input_tokens_seen": 10353136, "step": 36975 }, { "epoch": 410.8888888888889, "grad_norm": 2.564594012710586e-07, "learning_rate": 7.004107969295293e-07, "loss": 0.0, "num_input_tokens_seen": 10354576, "step": 36980 }, { "epoch": 410.94444444444446, "grad_norm": 2.0044279835929046e-07, "learning_rate": 6.9810508433929e-07, "loss": 0.0, "num_input_tokens_seen": 10356000, "step": 36985 }, { "epoch": 411.0, "grad_norm": 2.4621064653729263e-07, "learning_rate": 6.958031194066406e-07, "loss": 0.0, "num_input_tokens_seen": 10357408, "step": 36990 }, { "epoch": 411.05555555555554, "grad_norm": 4.453420743288916e-08, "learning_rate": 6.935049024865776e-07, "loss": 0.0, "num_input_tokens_seen": 10358784, "step": 36995 }, { "epoch": 411.1111111111111, "grad_norm": 1.1758069717870967e-07, "learning_rate": 6.912104339335118e-07, "loss": 0.0, "num_input_tokens_seen": 10360192, "step": 37000 }, { "epoch": 411.1111111111111, "eval_loss": 1.2468361854553223, "eval_runtime": 1.3919, "eval_samples_per_second": 28.739, "eval_steps_per_second": 14.369, "num_input_tokens_seen": 10360192, "step": 37000 }, { "epoch": 411.1666666666667, "grad_norm": 3.4548079952401167e-07, "learning_rate": 6.889197141012799e-07, "loss": 0.0, "num_input_tokens_seen": 10361616, "step": 37005 }, { "epoch": 411.22222222222223, "grad_norm": 5.030284810914054e-08, "learning_rate": 6.866327433431435e-07, "loss": 0.0, "num_input_tokens_seen": 10363040, "step": 37010 }, { "epoch": 411.27777777777777, "grad_norm": 1.0212124834652059e-06, "learning_rate": 6.843495220117735e-07, "loss": 0.0, "num_input_tokens_seen": 10364432, "step": 37015 }, { "epoch": 411.3333333333333, "grad_norm": 1.9227969971780112e-07, "learning_rate": 6.820700504592798e-07, "loss": 0.0, "num_input_tokens_seen": 10365808, "step": 37020 }, { "epoch": 411.3888888888889, "grad_norm": 1.0476905742962117e-07, "learning_rate": 6.797943290371839e-07, "loss": 0.0, "num_input_tokens_seen": 10367200, "step": 37025 }, { "epoch": 411.44444444444446, "grad_norm": 1.4961754857267806e-07, "learning_rate": 6.775223580964274e-07, "loss": 0.0, "num_input_tokens_seen": 10368592, "step": 37030 }, { "epoch": 411.5, "grad_norm": 3.4206805139547214e-06, "learning_rate": 6.7525413798738e-07, "loss": 0.0, "num_input_tokens_seen": 10370016, "step": 37035 }, { "epoch": 411.55555555555554, "grad_norm": 1.3032894230491365e-07, "learning_rate": 6.729896690598259e-07, "loss": 0.0, "num_input_tokens_seen": 10371424, "step": 37040 }, { "epoch": 411.6111111111111, "grad_norm": 1.7440821409309137e-07, "learning_rate": 6.707289516629772e-07, "loss": 0.0, "num_input_tokens_seen": 10372832, "step": 37045 }, { "epoch": 411.6666666666667, "grad_norm": 2.8898914479214e-06, "learning_rate": 6.684719861454692e-07, "loss": 0.0, "num_input_tokens_seen": 10374192, "step": 37050 }, { "epoch": 411.72222222222223, "grad_norm": 1.0073894429751817e-07, "learning_rate": 6.662187728553481e-07, "loss": 0.0, "num_input_tokens_seen": 10375600, "step": 37055 }, { "epoch": 411.77777777777777, "grad_norm": 1.401358258590335e-05, "learning_rate": 6.639693121400892e-07, "loss": 0.0, "num_input_tokens_seen": 10377008, "step": 37060 }, { "epoch": 411.8333333333333, "grad_norm": 1.3046820868112263e-07, "learning_rate": 6.617236043465868e-07, "loss": 0.0, "num_input_tokens_seen": 10378432, "step": 37065 }, { "epoch": 411.8888888888889, "grad_norm": 4.021312065560778e-07, "learning_rate": 6.594816498211587e-07, "loss": 0.0, "num_input_tokens_seen": 10379840, "step": 37070 }, { "epoch": 411.94444444444446, "grad_norm": 9.04250782696181e-07, "learning_rate": 6.572434489095447e-07, "loss": 0.0, "num_input_tokens_seen": 10381264, "step": 37075 }, { "epoch": 412.0, "grad_norm": 2.4051187210716307e-06, "learning_rate": 6.550090019568994e-07, "loss": 0.0, "num_input_tokens_seen": 10382640, "step": 37080 }, { "epoch": 412.05555555555554, "grad_norm": 8.008527174752089e-08, "learning_rate": 6.527783093078027e-07, "loss": 0.0, "num_input_tokens_seen": 10384016, "step": 37085 }, { "epoch": 412.1111111111111, "grad_norm": 2.821626878812822e-07, "learning_rate": 6.5055137130626e-07, "loss": 0.0, "num_input_tokens_seen": 10385440, "step": 37090 }, { "epoch": 412.1666666666667, "grad_norm": 1.0457937804631001e-07, "learning_rate": 6.483281882956854e-07, "loss": 0.0, "num_input_tokens_seen": 10386848, "step": 37095 }, { "epoch": 412.22222222222223, "grad_norm": 6.847336919690861e-08, "learning_rate": 6.461087606189298e-07, "loss": 0.0, "num_input_tokens_seen": 10388240, "step": 37100 }, { "epoch": 412.27777777777777, "grad_norm": 3.594427369080222e-07, "learning_rate": 6.438930886182554e-07, "loss": 0.0, "num_input_tokens_seen": 10389648, "step": 37105 }, { "epoch": 412.3333333333333, "grad_norm": 1.3730814316659234e-05, "learning_rate": 6.416811726353417e-07, "loss": 0.0, "num_input_tokens_seen": 10391072, "step": 37110 }, { "epoch": 412.3888888888889, "grad_norm": 1.667183653353277e-07, "learning_rate": 6.394730130112991e-07, "loss": 0.0, "num_input_tokens_seen": 10392544, "step": 37115 }, { "epoch": 412.44444444444446, "grad_norm": 1.3318325500222272e-06, "learning_rate": 6.372686100866471e-07, "loss": 0.0, "num_input_tokens_seen": 10393936, "step": 37120 }, { "epoch": 412.5, "grad_norm": 3.1344436024483e-07, "learning_rate": 6.350679642013413e-07, "loss": 0.0, "num_input_tokens_seen": 10395312, "step": 37125 }, { "epoch": 412.55555555555554, "grad_norm": 5.096715582908473e-08, "learning_rate": 6.328710756947437e-07, "loss": 0.0, "num_input_tokens_seen": 10396704, "step": 37130 }, { "epoch": 412.6111111111111, "grad_norm": 5.889054932595172e-07, "learning_rate": 6.306779449056416e-07, "loss": 0.0, "num_input_tokens_seen": 10398096, "step": 37135 }, { "epoch": 412.6666666666667, "grad_norm": 1.3076449079107988e-07, "learning_rate": 6.284885721722422e-07, "loss": 0.0, "num_input_tokens_seen": 10399520, "step": 37140 }, { "epoch": 412.72222222222223, "grad_norm": 6.685068001388572e-06, "learning_rate": 6.26302957832181e-07, "loss": 0.0, "num_input_tokens_seen": 10400944, "step": 37145 }, { "epoch": 412.77777777777777, "grad_norm": 3.2927162010309985e-06, "learning_rate": 6.241211022224997e-07, "loss": 0.0, "num_input_tokens_seen": 10402336, "step": 37150 }, { "epoch": 412.8333333333333, "grad_norm": 9.137133361036831e-08, "learning_rate": 6.219430056796732e-07, "loss": 0.0, "num_input_tokens_seen": 10403712, "step": 37155 }, { "epoch": 412.8888888888889, "grad_norm": 1.0386465874034911e-06, "learning_rate": 6.19768668539586e-07, "loss": 0.0, "num_input_tokens_seen": 10405104, "step": 37160 }, { "epoch": 412.94444444444446, "grad_norm": 1.2214291018608492e-06, "learning_rate": 6.175980911375528e-07, "loss": 0.0, "num_input_tokens_seen": 10406528, "step": 37165 }, { "epoch": 413.0, "grad_norm": 1.665440407805363e-07, "learning_rate": 6.154312738083034e-07, "loss": 0.0, "num_input_tokens_seen": 10407920, "step": 37170 }, { "epoch": 413.05555555555554, "grad_norm": 3.8198463698790874e-06, "learning_rate": 6.132682168859843e-07, "loss": 0.0, "num_input_tokens_seen": 10409312, "step": 37175 }, { "epoch": 413.1111111111111, "grad_norm": 4.320067148455564e-07, "learning_rate": 6.111089207041704e-07, "loss": 0.0, "num_input_tokens_seen": 10410640, "step": 37180 }, { "epoch": 413.1666666666667, "grad_norm": 7.974143159117375e-07, "learning_rate": 6.089533855958507e-07, "loss": 0.0, "num_input_tokens_seen": 10412016, "step": 37185 }, { "epoch": 413.22222222222223, "grad_norm": 3.6153386417936417e-07, "learning_rate": 6.068016118934372e-07, "loss": 0.0, "num_input_tokens_seen": 10413472, "step": 37190 }, { "epoch": 413.27777777777777, "grad_norm": 2.3512860991559137e-07, "learning_rate": 6.04653599928759e-07, "loss": 0.0, "num_input_tokens_seen": 10414848, "step": 37195 }, { "epoch": 413.3333333333333, "grad_norm": 2.1837817598679976e-07, "learning_rate": 6.025093500330675e-07, "loss": 0.0, "num_input_tokens_seen": 10416288, "step": 37200 }, { "epoch": 413.3333333333333, "eval_loss": 1.249448537826538, "eval_runtime": 1.3918, "eval_samples_per_second": 28.74, "eval_steps_per_second": 14.37, "num_input_tokens_seen": 10416288, "step": 37200 }, { "epoch": 413.3888888888889, "grad_norm": 8.091941339216646e-08, "learning_rate": 6.003688625370291e-07, "loss": 0.0, "num_input_tokens_seen": 10417696, "step": 37205 }, { "epoch": 413.44444444444446, "grad_norm": 2.393663862676476e-07, "learning_rate": 5.982321377707406e-07, "loss": 0.0, "num_input_tokens_seen": 10419136, "step": 37210 }, { "epoch": 413.5, "grad_norm": 5.775448812528339e-07, "learning_rate": 5.96099176063708e-07, "loss": 0.0, "num_input_tokens_seen": 10420560, "step": 37215 }, { "epoch": 413.55555555555554, "grad_norm": 6.275234909480787e-08, "learning_rate": 5.93969977744857e-07, "loss": 0.0, "num_input_tokens_seen": 10421984, "step": 37220 }, { "epoch": 413.6111111111111, "grad_norm": 1.3372547982726246e-05, "learning_rate": 5.918445431425445e-07, "loss": 0.0, "num_input_tokens_seen": 10423408, "step": 37225 }, { "epoch": 413.6666666666667, "grad_norm": 8.234996130340733e-06, "learning_rate": 5.897228725845333e-07, "loss": 0.0, "num_input_tokens_seen": 10424800, "step": 37230 }, { "epoch": 413.72222222222223, "grad_norm": 5.396516939981666e-07, "learning_rate": 5.876049663980171e-07, "loss": 0.0, "num_input_tokens_seen": 10426208, "step": 37235 }, { "epoch": 413.77777777777777, "grad_norm": 9.7721419933805e-07, "learning_rate": 5.854908249095959e-07, "loss": 0.0, "num_input_tokens_seen": 10427616, "step": 37240 }, { "epoch": 413.8333333333333, "grad_norm": 1.5930028212096659e-06, "learning_rate": 5.833804484453031e-07, "loss": 0.0, "num_input_tokens_seen": 10429008, "step": 37245 }, { "epoch": 413.8888888888889, "grad_norm": 2.4034966372710187e-07, "learning_rate": 5.81273837330587e-07, "loss": 0.0, "num_input_tokens_seen": 10430416, "step": 37250 }, { "epoch": 413.94444444444446, "grad_norm": 1.0540875194919863e-07, "learning_rate": 5.791709918903071e-07, "loss": 0.0, "num_input_tokens_seen": 10431824, "step": 37255 }, { "epoch": 414.0, "grad_norm": 5.7680167628859635e-06, "learning_rate": 5.770719124487483e-07, "loss": 0.0, "num_input_tokens_seen": 10433200, "step": 37260 }, { "epoch": 414.05555555555554, "grad_norm": 8.301631737595017e-08, "learning_rate": 5.749765993296241e-07, "loss": 0.0, "num_input_tokens_seen": 10434624, "step": 37265 }, { "epoch": 414.1111111111111, "grad_norm": 1.203099770918925e-07, "learning_rate": 5.728850528560509e-07, "loss": 0.0, "num_input_tokens_seen": 10436016, "step": 37270 }, { "epoch": 414.1666666666667, "grad_norm": 5.516790793080872e-07, "learning_rate": 5.707972733505707e-07, "loss": 0.0, "num_input_tokens_seen": 10437408, "step": 37275 }, { "epoch": 414.22222222222223, "grad_norm": 2.644037238042074e-07, "learning_rate": 5.687132611351509e-07, "loss": 0.0, "num_input_tokens_seen": 10438784, "step": 37280 }, { "epoch": 414.27777777777777, "grad_norm": 5.331127113095135e-07, "learning_rate": 5.666330165311651e-07, "loss": 0.0, "num_input_tokens_seen": 10440176, "step": 37285 }, { "epoch": 414.3333333333333, "grad_norm": 7.371275501100172e-07, "learning_rate": 5.645565398594204e-07, "loss": 0.0, "num_input_tokens_seen": 10441600, "step": 37290 }, { "epoch": 414.3888888888889, "grad_norm": 4.2196660388071905e-07, "learning_rate": 5.624838314401304e-07, "loss": 0.0, "num_input_tokens_seen": 10442976, "step": 37295 }, { "epoch": 414.44444444444446, "grad_norm": 1.4517000863634166e-06, "learning_rate": 5.604148915929336e-07, "loss": 0.0, "num_input_tokens_seen": 10444384, "step": 37300 }, { "epoch": 414.5, "grad_norm": 4.140429155086167e-06, "learning_rate": 5.583497206368887e-07, "loss": 0.0, "num_input_tokens_seen": 10445728, "step": 37305 }, { "epoch": 414.55555555555554, "grad_norm": 1.0452474725752836e-06, "learning_rate": 5.562883188904688e-07, "loss": 0.0, "num_input_tokens_seen": 10447168, "step": 37310 }, { "epoch": 414.6111111111111, "grad_norm": 1.0802939414134016e-06, "learning_rate": 5.542306866715724e-07, "loss": 0.0, "num_input_tokens_seen": 10448560, "step": 37315 }, { "epoch": 414.6666666666667, "grad_norm": 3.5470766306389123e-06, "learning_rate": 5.52176824297504e-07, "loss": 0.0, "num_input_tokens_seen": 10449968, "step": 37320 }, { "epoch": 414.72222222222223, "grad_norm": 2.7944659564127505e-07, "learning_rate": 5.501267320850018e-07, "loss": 0.0, "num_input_tokens_seen": 10451376, "step": 37325 }, { "epoch": 414.77777777777777, "grad_norm": 2.3454654183296952e-06, "learning_rate": 5.480804103502157e-07, "loss": 0.0, "num_input_tokens_seen": 10452800, "step": 37330 }, { "epoch": 414.8333333333333, "grad_norm": 7.147662586248771e-07, "learning_rate": 5.460378594087101e-07, "loss": 0.0, "num_input_tokens_seen": 10454192, "step": 37335 }, { "epoch": 414.8888888888889, "grad_norm": 6.432437515968559e-08, "learning_rate": 5.439990795754773e-07, "loss": 0.0, "num_input_tokens_seen": 10455600, "step": 37340 }, { "epoch": 414.94444444444446, "grad_norm": 3.784169564369222e-07, "learning_rate": 5.419640711649188e-07, "loss": 0.0, "num_input_tokens_seen": 10456992, "step": 37345 }, { "epoch": 415.0, "grad_norm": 6.720244272173659e-08, "learning_rate": 5.399328344908583e-07, "loss": 0.0, "num_input_tokens_seen": 10458416, "step": 37350 }, { "epoch": 415.05555555555554, "grad_norm": 2.8470580559769587e-07, "learning_rate": 5.379053698665399e-07, "loss": 0.0, "num_input_tokens_seen": 10459856, "step": 37355 }, { "epoch": 415.1111111111111, "grad_norm": 3.7410478626043187e-07, "learning_rate": 5.358816776046216e-07, "loss": 0.0, "num_input_tokens_seen": 10461248, "step": 37360 }, { "epoch": 415.1666666666667, "grad_norm": 5.159610338978382e-08, "learning_rate": 5.338617580171817e-07, "loss": 0.0, "num_input_tokens_seen": 10462608, "step": 37365 }, { "epoch": 415.22222222222223, "grad_norm": 2.806380337005976e-07, "learning_rate": 5.318456114157239e-07, "loss": 0.0, "num_input_tokens_seen": 10464016, "step": 37370 }, { "epoch": 415.27777777777777, "grad_norm": 2.227239122021274e-07, "learning_rate": 5.298332381111576e-07, "loss": 0.0, "num_input_tokens_seen": 10465392, "step": 37375 }, { "epoch": 415.3333333333333, "grad_norm": 4.3946240424475036e-08, "learning_rate": 5.27824638413818e-07, "loss": 0.0, "num_input_tokens_seen": 10466800, "step": 37380 }, { "epoch": 415.3888888888889, "grad_norm": 2.564491126122448e-07, "learning_rate": 5.258198126334546e-07, "loss": 0.0, "num_input_tokens_seen": 10468192, "step": 37385 }, { "epoch": 415.44444444444446, "grad_norm": 7.287280112677763e-08, "learning_rate": 5.238187610792367e-07, "loss": 0.0, "num_input_tokens_seen": 10469520, "step": 37390 }, { "epoch": 415.5, "grad_norm": 1.758374423843634e-08, "learning_rate": 5.218214840597563e-07, "loss": 0.0, "num_input_tokens_seen": 10470912, "step": 37395 }, { "epoch": 415.55555555555554, "grad_norm": 3.6817698401137022e-06, "learning_rate": 5.198279818830115e-07, "loss": 0.0, "num_input_tokens_seen": 10472368, "step": 37400 }, { "epoch": 415.55555555555554, "eval_loss": 1.2068923711776733, "eval_runtime": 1.3927, "eval_samples_per_second": 28.722, "eval_steps_per_second": 14.361, "num_input_tokens_seen": 10472368, "step": 37400 }, { "epoch": 415.6111111111111, "grad_norm": 1.7435746713090339e-06, "learning_rate": 5.178382548564287e-07, "loss": 0.0, "num_input_tokens_seen": 10473776, "step": 37405 }, { "epoch": 415.6666666666667, "grad_norm": 3.2518698844796745e-07, "learning_rate": 5.15852303286854e-07, "loss": 0.0, "num_input_tokens_seen": 10475216, "step": 37410 }, { "epoch": 415.72222222222223, "grad_norm": 2.0160773317456915e-07, "learning_rate": 5.138701274805396e-07, "loss": 0.0, "num_input_tokens_seen": 10476608, "step": 37415 }, { "epoch": 415.77777777777777, "grad_norm": 4.4884936301059497e-07, "learning_rate": 5.118917277431606e-07, "loss": 0.0, "num_input_tokens_seen": 10477984, "step": 37420 }, { "epoch": 415.8333333333333, "grad_norm": 1.1836757494165795e-06, "learning_rate": 5.099171043798145e-07, "loss": 0.0, "num_input_tokens_seen": 10479360, "step": 37425 }, { "epoch": 415.8888888888889, "grad_norm": 1.3991167179483455e-05, "learning_rate": 5.079462576950133e-07, "loss": 0.0, "num_input_tokens_seen": 10480752, "step": 37430 }, { "epoch": 415.94444444444446, "grad_norm": 2.0632901396311354e-06, "learning_rate": 5.059791879926862e-07, "loss": 0.0, "num_input_tokens_seen": 10482160, "step": 37435 }, { "epoch": 416.0, "grad_norm": 1.4257365421599388e-07, "learning_rate": 5.040158955761793e-07, "loss": 0.0, "num_input_tokens_seen": 10483600, "step": 37440 }, { "epoch": 416.05555555555554, "grad_norm": 4.5172842533247604e-07, "learning_rate": 5.020563807482559e-07, "loss": 0.0, "num_input_tokens_seen": 10485024, "step": 37445 }, { "epoch": 416.1111111111111, "grad_norm": 5.447013791126665e-07, "learning_rate": 5.001006438110995e-07, "loss": 0.0, "num_input_tokens_seen": 10486432, "step": 37450 }, { "epoch": 416.1666666666667, "grad_norm": 2.58229164273871e-07, "learning_rate": 4.981486850663075e-07, "loss": 0.0, "num_input_tokens_seen": 10487792, "step": 37455 }, { "epoch": 416.22222222222223, "grad_norm": 1.6178447026504728e-07, "learning_rate": 4.962005048149005e-07, "loss": 0.0, "num_input_tokens_seen": 10489136, "step": 37460 }, { "epoch": 416.27777777777777, "grad_norm": 3.839003966277232e-06, "learning_rate": 4.942561033573073e-07, "loss": 0.0, "num_input_tokens_seen": 10490528, "step": 37465 }, { "epoch": 416.3333333333333, "grad_norm": 7.837326876369843e-08, "learning_rate": 4.923154809933827e-07, "loss": 0.0, "num_input_tokens_seen": 10491920, "step": 37470 }, { "epoch": 416.3888888888889, "grad_norm": 2.2406145205877692e-07, "learning_rate": 4.903786380223957e-07, "loss": 0.0, "num_input_tokens_seen": 10493360, "step": 37475 }, { "epoch": 416.44444444444446, "grad_norm": 8.063418590609217e-07, "learning_rate": 4.884455747430266e-07, "loss": 0.0, "num_input_tokens_seen": 10494784, "step": 37480 }, { "epoch": 416.5, "grad_norm": 9.819626939133741e-06, "learning_rate": 4.865162914533816e-07, "loss": 0.0, "num_input_tokens_seen": 10496144, "step": 37485 }, { "epoch": 416.55555555555554, "grad_norm": 8.596642828706536e-07, "learning_rate": 4.845907884509809e-07, "loss": 0.0, "num_input_tokens_seen": 10497520, "step": 37490 }, { "epoch": 416.6111111111111, "grad_norm": 2.7556163217923313e-07, "learning_rate": 4.82669066032762e-07, "loss": 0.0, "num_input_tokens_seen": 10498928, "step": 37495 }, { "epoch": 416.6666666666667, "grad_norm": 8.625341251899954e-06, "learning_rate": 4.807511244950768e-07, "loss": 0.0, "num_input_tokens_seen": 10500336, "step": 37500 }, { "epoch": 416.72222222222223, "grad_norm": 6.41689723579475e-08, "learning_rate": 4.788369641336943e-07, "loss": 0.0, "num_input_tokens_seen": 10501696, "step": 37505 }, { "epoch": 416.77777777777777, "grad_norm": 4.1761550164665096e-06, "learning_rate": 4.769265852438032e-07, "loss": 0.0, "num_input_tokens_seen": 10503136, "step": 37510 }, { "epoch": 416.8333333333333, "grad_norm": 2.6356394755566726e-07, "learning_rate": 4.750199881200124e-07, "loss": 0.0, "num_input_tokens_seen": 10504544, "step": 37515 }, { "epoch": 416.8888888888889, "grad_norm": 1.9529099404280714e-07, "learning_rate": 4.7311717305633664e-07, "loss": 0.0, "num_input_tokens_seen": 10505984, "step": 37520 }, { "epoch": 416.94444444444446, "grad_norm": 4.904181949427766e-08, "learning_rate": 4.7121814034621623e-07, "loss": 0.0, "num_input_tokens_seen": 10507408, "step": 37525 }, { "epoch": 417.0, "grad_norm": 5.402582132774114e-07, "learning_rate": 4.693228902825114e-07, "loss": 0.0, "num_input_tokens_seen": 10508816, "step": 37530 }, { "epoch": 417.05555555555554, "grad_norm": 4.412868747749599e-06, "learning_rate": 4.6743142315748277e-07, "loss": 0.0, "num_input_tokens_seen": 10510176, "step": 37535 }, { "epoch": 417.1111111111111, "grad_norm": 3.3200017242052127e-06, "learning_rate": 4.655437392628276e-07, "loss": 0.0, "num_input_tokens_seen": 10511552, "step": 37540 }, { "epoch": 417.1666666666667, "grad_norm": 4.918458529346026e-08, "learning_rate": 4.636598388896463e-07, "loss": 0.0, "num_input_tokens_seen": 10512928, "step": 37545 }, { "epoch": 417.22222222222223, "grad_norm": 9.906506193146924e-08, "learning_rate": 4.6177972232845925e-07, "loss": 0.0, "num_input_tokens_seen": 10514304, "step": 37550 }, { "epoch": 417.27777777777777, "grad_norm": 6.786995498941906e-08, "learning_rate": 4.5990338986920953e-07, "loss": 0.0, "num_input_tokens_seen": 10515696, "step": 37555 }, { "epoch": 417.3333333333333, "grad_norm": 6.37440507489373e-07, "learning_rate": 4.5803084180124633e-07, "loss": 0.0, "num_input_tokens_seen": 10517072, "step": 37560 }, { "epoch": 417.3888888888889, "grad_norm": 3.877519816342101e-07, "learning_rate": 4.561620784133386e-07, "loss": 0.0, "num_input_tokens_seen": 10518432, "step": 37565 }, { "epoch": 417.44444444444446, "grad_norm": 6.629375661759696e-07, "learning_rate": 4.5429709999367796e-07, "loss": 0.0, "num_input_tokens_seen": 10519872, "step": 37570 }, { "epoch": 417.5, "grad_norm": 3.614489401115861e-07, "learning_rate": 4.5243590682986223e-07, "loss": 0.0, "num_input_tokens_seen": 10521296, "step": 37575 }, { "epoch": 417.55555555555554, "grad_norm": 1.574503869505861e-07, "learning_rate": 4.5057849920891735e-07, "loss": 0.0, "num_input_tokens_seen": 10522736, "step": 37580 }, { "epoch": 417.6111111111111, "grad_norm": 5.959317377346451e-07, "learning_rate": 4.487248774172698e-07, "loss": 0.0, "num_input_tokens_seen": 10524128, "step": 37585 }, { "epoch": 417.6666666666667, "grad_norm": 2.4461999714731064e-07, "learning_rate": 4.4687504174077965e-07, "loss": 0.0, "num_input_tokens_seen": 10525536, "step": 37590 }, { "epoch": 417.72222222222223, "grad_norm": 6.551881028826756e-07, "learning_rate": 4.450289924647133e-07, "loss": 0.0, "num_input_tokens_seen": 10526960, "step": 37595 }, { "epoch": 417.77777777777777, "grad_norm": 7.70711068298624e-08, "learning_rate": 4.431867298737513e-07, "loss": 0.0, "num_input_tokens_seen": 10528352, "step": 37600 }, { "epoch": 417.77777777777777, "eval_loss": 1.22310471534729, "eval_runtime": 1.3916, "eval_samples_per_second": 28.744, "eval_steps_per_second": 14.372, "num_input_tokens_seen": 10528352, "step": 37600 }, { "epoch": 417.8333333333333, "grad_norm": 5.23119410900108e-07, "learning_rate": 4.41348254251997e-07, "loss": 0.0, "num_input_tokens_seen": 10529792, "step": 37605 }, { "epoch": 417.8888888888889, "grad_norm": 3.1221426866068214e-07, "learning_rate": 4.395135658829652e-07, "loss": 0.0, "num_input_tokens_seen": 10531184, "step": 37610 }, { "epoch": 417.94444444444446, "grad_norm": 2.1142027151199727e-07, "learning_rate": 4.376826650495852e-07, "loss": 0.0, "num_input_tokens_seen": 10532624, "step": 37615 }, { "epoch": 418.0, "grad_norm": 1.4027878592060006e-07, "learning_rate": 4.358555520342117e-07, "loss": 0.0, "num_input_tokens_seen": 10534016, "step": 37620 }, { "epoch": 418.05555555555554, "grad_norm": 3.584374326237594e-07, "learning_rate": 4.3403222711860257e-07, "loss": 0.0, "num_input_tokens_seen": 10535456, "step": 37625 }, { "epoch": 418.1111111111111, "grad_norm": 4.1189085209225595e-07, "learning_rate": 4.3221269058394133e-07, "loss": 0.0, "num_input_tokens_seen": 10536848, "step": 37630 }, { "epoch": 418.1666666666667, "grad_norm": 1.1071062289147449e-07, "learning_rate": 4.303969427108173e-07, "loss": 0.0, "num_input_tokens_seen": 10538320, "step": 37635 }, { "epoch": 418.22222222222223, "grad_norm": 4.3745740185840987e-07, "learning_rate": 4.2858498377924825e-07, "loss": 0.0, "num_input_tokens_seen": 10539728, "step": 37640 }, { "epoch": 418.27777777777777, "grad_norm": 2.42506848735502e-06, "learning_rate": 4.267768140686579e-07, "loss": 0.0, "num_input_tokens_seen": 10541104, "step": 37645 }, { "epoch": 418.3333333333333, "grad_norm": 1.8150627738577896e-06, "learning_rate": 4.2497243385788975e-07, "loss": 0.0, "num_input_tokens_seen": 10542480, "step": 37650 }, { "epoch": 418.3888888888889, "grad_norm": 1.1403900401774081e-07, "learning_rate": 4.231718434251991e-07, "loss": 0.0, "num_input_tokens_seen": 10543824, "step": 37655 }, { "epoch": 418.44444444444446, "grad_norm": 2.3671882445341907e-06, "learning_rate": 4.213750430482666e-07, "loss": 0.0, "num_input_tokens_seen": 10545200, "step": 37660 }, { "epoch": 418.5, "grad_norm": 1.781166787395705e-07, "learning_rate": 4.1958203300417054e-07, "loss": 0.0, "num_input_tokens_seen": 10546656, "step": 37665 }, { "epoch": 418.55555555555554, "grad_norm": 4.333669494371861e-06, "learning_rate": 4.177928135694259e-07, "loss": 0.0, "num_input_tokens_seen": 10548048, "step": 37670 }, { "epoch": 418.6111111111111, "grad_norm": 1.524921913187427e-07, "learning_rate": 4.1600738501994807e-07, "loss": 0.0, "num_input_tokens_seen": 10549472, "step": 37675 }, { "epoch": 418.6666666666667, "grad_norm": 7.459291140321511e-08, "learning_rate": 4.1422574763107237e-07, "loss": 0.0, "num_input_tokens_seen": 10550848, "step": 37680 }, { "epoch": 418.72222222222223, "grad_norm": 8.110657176985114e-07, "learning_rate": 4.124479016775512e-07, "loss": 0.0, "num_input_tokens_seen": 10552224, "step": 37685 }, { "epoch": 418.77777777777777, "grad_norm": 2.6074093284478295e-07, "learning_rate": 4.106738474335514e-07, "loss": 0.0, "num_input_tokens_seen": 10553632, "step": 37690 }, { "epoch": 418.8333333333333, "grad_norm": 4.8850797185195916e-08, "learning_rate": 4.089035851726486e-07, "loss": 0.0, "num_input_tokens_seen": 10555008, "step": 37695 }, { "epoch": 418.8888888888889, "grad_norm": 3.319102290788578e-07, "learning_rate": 4.0713711516784937e-07, "loss": 0.0, "num_input_tokens_seen": 10556368, "step": 37700 }, { "epoch": 418.94444444444446, "grad_norm": 1.1058882165571049e-07, "learning_rate": 4.05374437691558e-07, "loss": 0.0, "num_input_tokens_seen": 10557776, "step": 37705 }, { "epoch": 419.0, "grad_norm": 9.864077583188191e-06, "learning_rate": 4.036155530156044e-07, "loss": 0.0, "num_input_tokens_seen": 10559152, "step": 37710 }, { "epoch": 419.05555555555554, "grad_norm": 1.3139093368863541e-07, "learning_rate": 4.018604614112298e-07, "loss": 0.0, "num_input_tokens_seen": 10560560, "step": 37715 }, { "epoch": 419.1111111111111, "grad_norm": 6.435143973249069e-07, "learning_rate": 4.0010916314908996e-07, "loss": 0.0, "num_input_tokens_seen": 10562016, "step": 37720 }, { "epoch": 419.1666666666667, "grad_norm": 2.3493461753787415e-07, "learning_rate": 3.983616584992578e-07, "loss": 0.0, "num_input_tokens_seen": 10563424, "step": 37725 }, { "epoch": 419.22222222222223, "grad_norm": 2.8076775038243795e-07, "learning_rate": 3.9661794773122595e-07, "loss": 0.0, "num_input_tokens_seen": 10564800, "step": 37730 }, { "epoch": 419.27777777777777, "grad_norm": 4.545112801679352e-07, "learning_rate": 3.9487803111388777e-07, "loss": 0.0, "num_input_tokens_seen": 10566192, "step": 37735 }, { "epoch": 419.3333333333333, "grad_norm": 1.1649248108369648e-06, "learning_rate": 3.9314190891556747e-07, "loss": 0.0, "num_input_tokens_seen": 10567600, "step": 37740 }, { "epoch": 419.3888888888889, "grad_norm": 2.0938706768447446e-07, "learning_rate": 3.914095814039925e-07, "loss": 0.0, "num_input_tokens_seen": 10568976, "step": 37745 }, { "epoch": 419.44444444444446, "grad_norm": 9.340359241605256e-08, "learning_rate": 3.896810488463104e-07, "loss": 0.0, "num_input_tokens_seen": 10570368, "step": 37750 }, { "epoch": 419.5, "grad_norm": 1.7207968028287723e-07, "learning_rate": 3.8795631150908565e-07, "loss": 0.0, "num_input_tokens_seen": 10571808, "step": 37755 }, { "epoch": 419.55555555555554, "grad_norm": 3.748466724573518e-06, "learning_rate": 3.862353696582888e-07, "loss": 0.0, "num_input_tokens_seen": 10573248, "step": 37760 }, { "epoch": 419.6111111111111, "grad_norm": 4.292368885217002e-06, "learning_rate": 3.8451822355931313e-07, "loss": 0.0, "num_input_tokens_seen": 10574624, "step": 37765 }, { "epoch": 419.6666666666667, "grad_norm": 1.7094818360874342e-07, "learning_rate": 3.82804873476969e-07, "loss": 0.0, "num_input_tokens_seen": 10576016, "step": 37770 }, { "epoch": 419.72222222222223, "grad_norm": 4.8339251179641e-07, "learning_rate": 3.810953196754702e-07, "loss": 0.0, "num_input_tokens_seen": 10577408, "step": 37775 }, { "epoch": 419.77777777777777, "grad_norm": 1.1454746129402338e-07, "learning_rate": 3.793895624184529e-07, "loss": 0.0, "num_input_tokens_seen": 10578768, "step": 37780 }, { "epoch": 419.8333333333333, "grad_norm": 7.637528142367955e-06, "learning_rate": 3.776876019689679e-07, "loss": 0.0, "num_input_tokens_seen": 10580128, "step": 37785 }, { "epoch": 419.8888888888889, "grad_norm": 1.87365799320105e-07, "learning_rate": 3.7598943858947743e-07, "loss": 0.0, "num_input_tokens_seen": 10581552, "step": 37790 }, { "epoch": 419.94444444444446, "grad_norm": 7.504511359002208e-07, "learning_rate": 3.742950725418637e-07, "loss": 0.0, "num_input_tokens_seen": 10582960, "step": 37795 }, { "epoch": 420.0, "grad_norm": 4.3342043909433414e-07, "learning_rate": 3.726045040874093e-07, "loss": 0.0, "num_input_tokens_seen": 10584384, "step": 37800 }, { "epoch": 420.0, "eval_loss": 1.2686867713928223, "eval_runtime": 1.385, "eval_samples_per_second": 28.88, "eval_steps_per_second": 14.44, "num_input_tokens_seen": 10584384, "step": 37800 }, { "epoch": 420.05555555555554, "grad_norm": 1.100404816156697e-07, "learning_rate": 3.709177334868308e-07, "loss": 0.0, "num_input_tokens_seen": 10585808, "step": 37805 }, { "epoch": 420.1111111111111, "grad_norm": 3.7791025988553884e-06, "learning_rate": 3.692347610002478e-07, "loss": 0.0, "num_input_tokens_seen": 10587216, "step": 37810 }, { "epoch": 420.1666666666667, "grad_norm": 1.9048385979658633e-07, "learning_rate": 3.675555868871916e-07, "loss": 0.0, "num_input_tokens_seen": 10588656, "step": 37815 }, { "epoch": 420.22222222222223, "grad_norm": 9.434963743615299e-08, "learning_rate": 3.658802114066162e-07, "loss": 0.0, "num_input_tokens_seen": 10590032, "step": 37820 }, { "epoch": 420.27777777777777, "grad_norm": 1.1330930504982462e-07, "learning_rate": 3.6420863481688437e-07, "loss": 0.0, "num_input_tokens_seen": 10591424, "step": 37825 }, { "epoch": 420.3333333333333, "grad_norm": 3.106540304997907e-07, "learning_rate": 3.625408573757705e-07, "loss": 0.0, "num_input_tokens_seen": 10592816, "step": 37830 }, { "epoch": 420.3888888888889, "grad_norm": 2.984920115522982e-07, "learning_rate": 3.608768793404743e-07, "loss": 0.0, "num_input_tokens_seen": 10594272, "step": 37835 }, { "epoch": 420.44444444444446, "grad_norm": 1.2683635759458411e-05, "learning_rate": 3.592167009675934e-07, "loss": 0.0, "num_input_tokens_seen": 10595648, "step": 37840 }, { "epoch": 420.5, "grad_norm": 4.4838435542260413e-07, "learning_rate": 3.575603225131563e-07, "loss": 0.0, "num_input_tokens_seen": 10597056, "step": 37845 }, { "epoch": 420.55555555555554, "grad_norm": 6.529581241920823e-06, "learning_rate": 3.55907744232592e-07, "loss": 0.0, "num_input_tokens_seen": 10598480, "step": 37850 }, { "epoch": 420.6111111111111, "grad_norm": 7.044739049888449e-06, "learning_rate": 3.5425896638075217e-07, "loss": 0.0, "num_input_tokens_seen": 10599856, "step": 37855 }, { "epoch": 420.6666666666667, "grad_norm": 3.445116476541443e-07, "learning_rate": 3.5261398921189736e-07, "loss": 0.0, "num_input_tokens_seen": 10601248, "step": 37860 }, { "epoch": 420.72222222222223, "grad_norm": 1.0299923047796256e-07, "learning_rate": 3.509728129797024e-07, "loss": 0.0, "num_input_tokens_seen": 10602672, "step": 37865 }, { "epoch": 420.77777777777777, "grad_norm": 6.150834792606474e-07, "learning_rate": 3.4933543793725656e-07, "loss": 0.0, "num_input_tokens_seen": 10604064, "step": 37870 }, { "epoch": 420.8333333333333, "grad_norm": 2.306109365690645e-07, "learning_rate": 3.4770186433707163e-07, "loss": 0.0, "num_input_tokens_seen": 10605440, "step": 37875 }, { "epoch": 420.8888888888889, "grad_norm": 3.8292301951514673e-07, "learning_rate": 3.4607209243105453e-07, "loss": 0.0, "num_input_tokens_seen": 10606848, "step": 37880 }, { "epoch": 420.94444444444446, "grad_norm": 3.5485812986735255e-05, "learning_rate": 3.444461224705431e-07, "loss": 0.0, "num_input_tokens_seen": 10608224, "step": 37885 }, { "epoch": 421.0, "grad_norm": 7.403904191960464e-07, "learning_rate": 3.4282395470628116e-07, "loss": 0.0, "num_input_tokens_seen": 10609616, "step": 37890 }, { "epoch": 421.05555555555554, "grad_norm": 2.964916347991675e-07, "learning_rate": 3.4120558938842417e-07, "loss": 0.0, "num_input_tokens_seen": 10611008, "step": 37895 }, { "epoch": 421.1111111111111, "grad_norm": 6.544521511386847e-07, "learning_rate": 3.395910267665503e-07, "loss": 0.0, "num_input_tokens_seen": 10612400, "step": 37900 }, { "epoch": 421.1666666666667, "grad_norm": 5.1095748858642764e-06, "learning_rate": 3.3798026708964094e-07, "loss": 0.0, "num_input_tokens_seen": 10613792, "step": 37905 }, { "epoch": 421.22222222222223, "grad_norm": 1.2411183547555993e-07, "learning_rate": 3.3637331060609456e-07, "loss": 0.0, "num_input_tokens_seen": 10615200, "step": 37910 }, { "epoch": 421.27777777777777, "grad_norm": 1.0530307008593809e-06, "learning_rate": 3.3477015756372966e-07, "loss": 0.0, "num_input_tokens_seen": 10616576, "step": 37915 }, { "epoch": 421.3333333333333, "grad_norm": 5.390313617681386e-07, "learning_rate": 3.3317080820976785e-07, "loss": 0.0, "num_input_tokens_seen": 10618000, "step": 37920 }, { "epoch": 421.3888888888889, "grad_norm": 3.764285111174104e-06, "learning_rate": 3.315752627908508e-07, "loss": 0.0, "num_input_tokens_seen": 10619408, "step": 37925 }, { "epoch": 421.44444444444446, "grad_norm": 7.707799909439927e-07, "learning_rate": 3.299835215530317e-07, "loss": 0.0, "num_input_tokens_seen": 10620800, "step": 37930 }, { "epoch": 421.5, "grad_norm": 1.7842249633304164e-07, "learning_rate": 3.2839558474177245e-07, "loss": 0.0, "num_input_tokens_seen": 10622160, "step": 37935 }, { "epoch": 421.55555555555554, "grad_norm": 1.0676676538423635e-06, "learning_rate": 3.2681145260196056e-07, "loss": 0.0, "num_input_tokens_seen": 10623520, "step": 37940 }, { "epoch": 421.6111111111111, "grad_norm": 3.682980320718343e-07, "learning_rate": 3.252311253778839e-07, "loss": 0.0, "num_input_tokens_seen": 10624928, "step": 37945 }, { "epoch": 421.6666666666667, "grad_norm": 1.6547532766253426e-07, "learning_rate": 3.2365460331325034e-07, "loss": 0.0, "num_input_tokens_seen": 10626384, "step": 37950 }, { "epoch": 421.72222222222223, "grad_norm": 7.276661051491828e-08, "learning_rate": 3.2208188665117934e-07, "loss": 0.0, "num_input_tokens_seen": 10627776, "step": 37955 }, { "epoch": 421.77777777777777, "grad_norm": 4.347570836671366e-07, "learning_rate": 3.205129756342018e-07, "loss": 0.0, "num_input_tokens_seen": 10629216, "step": 37960 }, { "epoch": 421.8333333333333, "grad_norm": 3.7313382108550286e-06, "learning_rate": 3.189478705042659e-07, "loss": 0.0, "num_input_tokens_seen": 10630608, "step": 37965 }, { "epoch": 421.8888888888889, "grad_norm": 2.0373265385842387e-07, "learning_rate": 3.173865715027341e-07, "loss": 0.0, "num_input_tokens_seen": 10632000, "step": 37970 }, { "epoch": 421.94444444444446, "grad_norm": 7.767605580966119e-08, "learning_rate": 3.158290788703694e-07, "loss": 0.0, "num_input_tokens_seen": 10633440, "step": 37975 }, { "epoch": 422.0, "grad_norm": 2.230563723060186e-06, "learning_rate": 3.1427539284736297e-07, "loss": 0.0, "num_input_tokens_seen": 10634864, "step": 37980 }, { "epoch": 422.05555555555554, "grad_norm": 1.2101786239782086e-07, "learning_rate": 3.127255136733093e-07, "loss": 0.0, "num_input_tokens_seen": 10636256, "step": 37985 }, { "epoch": 422.1111111111111, "grad_norm": 4.5969389361744106e-07, "learning_rate": 3.1117944158722544e-07, "loss": 0.0, "num_input_tokens_seen": 10637680, "step": 37990 }, { "epoch": 422.1666666666667, "grad_norm": 1.38763834911515e-06, "learning_rate": 3.0963717682752635e-07, "loss": 0.0, "num_input_tokens_seen": 10639104, "step": 37995 }, { "epoch": 422.22222222222223, "grad_norm": 8.38643870793021e-07, "learning_rate": 3.080987196320578e-07, "loss": 0.0, "num_input_tokens_seen": 10640496, "step": 38000 }, { "epoch": 422.22222222222223, "eval_loss": 1.215099573135376, "eval_runtime": 1.3937, "eval_samples_per_second": 28.701, "eval_steps_per_second": 14.351, "num_input_tokens_seen": 10640496, "step": 38000 }, { "epoch": 422.27777777777777, "grad_norm": 4.0277859625348356e-07, "learning_rate": 3.065640702380607e-07, "loss": 0.0, "num_input_tokens_seen": 10641936, "step": 38005 }, { "epoch": 422.3333333333333, "grad_norm": 9.102583931053232e-08, "learning_rate": 3.050332288822011e-07, "loss": 0.0, "num_input_tokens_seen": 10643344, "step": 38010 }, { "epoch": 422.3888888888889, "grad_norm": 2.0958850655006245e-06, "learning_rate": 3.035061958005542e-07, "loss": 0.0, "num_input_tokens_seen": 10644768, "step": 38015 }, { "epoch": 422.44444444444446, "grad_norm": 2.391813040958368e-07, "learning_rate": 3.019829712286093e-07, "loss": 0.0, "num_input_tokens_seen": 10646144, "step": 38020 }, { "epoch": 422.5, "grad_norm": 7.731889581918949e-07, "learning_rate": 3.004635554012647e-07, "loss": 0.0, "num_input_tokens_seen": 10647536, "step": 38025 }, { "epoch": 422.55555555555554, "grad_norm": 8.770337700525488e-08, "learning_rate": 2.9894794855283017e-07, "loss": 0.0, "num_input_tokens_seen": 10648896, "step": 38030 }, { "epoch": 422.6111111111111, "grad_norm": 6.660790177193121e-07, "learning_rate": 2.9743615091703816e-07, "loss": 0.0, "num_input_tokens_seen": 10650320, "step": 38035 }, { "epoch": 422.6666666666667, "grad_norm": 2.432412031794229e-07, "learning_rate": 2.959281627270216e-07, "loss": 0.0, "num_input_tokens_seen": 10651760, "step": 38040 }, { "epoch": 422.72222222222223, "grad_norm": 4.856341817571774e-08, "learning_rate": 2.944239842153362e-07, "loss": 0.0, "num_input_tokens_seen": 10653168, "step": 38045 }, { "epoch": 422.77777777777777, "grad_norm": 1.2866379961451457e-07, "learning_rate": 2.929236156139381e-07, "loss": 0.0, "num_input_tokens_seen": 10654592, "step": 38050 }, { "epoch": 422.8333333333333, "grad_norm": 4.3095835167150653e-07, "learning_rate": 2.9142705715420883e-07, "loss": 0.0, "num_input_tokens_seen": 10655968, "step": 38055 }, { "epoch": 422.8888888888889, "grad_norm": 2.0574610459789255e-07, "learning_rate": 2.8993430906693595e-07, "loss": 0.0, "num_input_tokens_seen": 10657344, "step": 38060 }, { "epoch": 422.94444444444446, "grad_norm": 1.9530514805410348e-07, "learning_rate": 2.88445371582316e-07, "loss": 0.0, "num_input_tokens_seen": 10658688, "step": 38065 }, { "epoch": 423.0, "grad_norm": 4.902679211227223e-05, "learning_rate": 2.8696024492996796e-07, "loss": 0.0, "num_input_tokens_seen": 10660080, "step": 38070 }, { "epoch": 423.05555555555554, "grad_norm": 7.711977332292008e-07, "learning_rate": 2.854789293389115e-07, "loss": 0.0, "num_input_tokens_seen": 10661472, "step": 38075 }, { "epoch": 423.1111111111111, "grad_norm": 1.2269070737147558e-07, "learning_rate": 2.8400142503758606e-07, "loss": 0.0, "num_input_tokens_seen": 10662848, "step": 38080 }, { "epoch": 423.1666666666667, "grad_norm": 2.577411351012415e-07, "learning_rate": 2.8252773225384276e-07, "loss": 0.0, "num_input_tokens_seen": 10664256, "step": 38085 }, { "epoch": 423.22222222222223, "grad_norm": 9.953779454008327e-07, "learning_rate": 2.8105785121494143e-07, "loss": 0.0, "num_input_tokens_seen": 10665728, "step": 38090 }, { "epoch": 423.27777777777777, "grad_norm": 4.066016145998219e-08, "learning_rate": 2.795917821475563e-07, "loss": 0.0, "num_input_tokens_seen": 10667120, "step": 38095 }, { "epoch": 423.3333333333333, "grad_norm": 4.318430001148954e-06, "learning_rate": 2.78129525277776e-07, "loss": 0.0, "num_input_tokens_seen": 10668512, "step": 38100 }, { "epoch": 423.3888888888889, "grad_norm": 4.8521268780632454e-08, "learning_rate": 2.766710808310952e-07, "loss": 0.0, "num_input_tokens_seen": 10669904, "step": 38105 }, { "epoch": 423.44444444444446, "grad_norm": 5.2653877702368845e-08, "learning_rate": 2.7521644903242827e-07, "loss": 0.0, "num_input_tokens_seen": 10671328, "step": 38110 }, { "epoch": 423.5, "grad_norm": 7.300922106878716e-07, "learning_rate": 2.7376563010609593e-07, "loss": 0.0, "num_input_tokens_seen": 10672752, "step": 38115 }, { "epoch": 423.55555555555554, "grad_norm": 3.1222978691403114e-07, "learning_rate": 2.72318624275833e-07, "loss": 0.0, "num_input_tokens_seen": 10674144, "step": 38120 }, { "epoch": 423.6111111111111, "grad_norm": 7.658663037091173e-08, "learning_rate": 2.7087543176478324e-07, "loss": 0.0, "num_input_tokens_seen": 10675520, "step": 38125 }, { "epoch": 423.6666666666667, "grad_norm": 4.765363428305136e-06, "learning_rate": 2.694360527955103e-07, "loss": 0.0, "num_input_tokens_seen": 10676912, "step": 38130 }, { "epoch": 423.72222222222223, "grad_norm": 8.036644771891588e-08, "learning_rate": 2.680004875899811e-07, "loss": 0.0, "num_input_tokens_seen": 10678320, "step": 38135 }, { "epoch": 423.77777777777777, "grad_norm": 1.5223039895317925e-07, "learning_rate": 2.665687363695768e-07, "loss": 0.0, "num_input_tokens_seen": 10679744, "step": 38140 }, { "epoch": 423.8333333333333, "grad_norm": 2.0400830180733465e-06, "learning_rate": 2.6514079935509584e-07, "loss": 0.0, "num_input_tokens_seen": 10681168, "step": 38145 }, { "epoch": 423.8888888888889, "grad_norm": 2.229245268381419e-07, "learning_rate": 2.6371667676673983e-07, "loss": 0.0, "num_input_tokens_seen": 10682576, "step": 38150 }, { "epoch": 423.94444444444446, "grad_norm": 1.991872068174416e-06, "learning_rate": 2.6229636882412755e-07, "loss": 0.0, "num_input_tokens_seen": 10683952, "step": 38155 }, { "epoch": 424.0, "grad_norm": 5.537623906093359e-07, "learning_rate": 2.6087987574628935e-07, "loss": 0.0, "num_input_tokens_seen": 10685344, "step": 38160 }, { "epoch": 424.05555555555554, "grad_norm": 4.579979773211562e-08, "learning_rate": 2.5946719775166437e-07, "loss": 0.0, "num_input_tokens_seen": 10686720, "step": 38165 }, { "epoch": 424.1111111111111, "grad_norm": 2.480244063463033e-07, "learning_rate": 2.5805833505810616e-07, "loss": 0.0, "num_input_tokens_seen": 10688112, "step": 38170 }, { "epoch": 424.1666666666667, "grad_norm": 3.5719253332899825e-07, "learning_rate": 2.566532878828798e-07, "loss": 0.0, "num_input_tokens_seen": 10689552, "step": 38175 }, { "epoch": 424.22222222222223, "grad_norm": 1.1443083991480307e-07, "learning_rate": 2.552520564426619e-07, "loss": 0.0, "num_input_tokens_seen": 10690928, "step": 38180 }, { "epoch": 424.27777777777777, "grad_norm": 3.2153769780052244e-07, "learning_rate": 2.5385464095353803e-07, "loss": 0.0, "num_input_tokens_seen": 10692336, "step": 38185 }, { "epoch": 424.3333333333333, "grad_norm": 2.82427805586849e-07, "learning_rate": 2.5246104163100804e-07, "loss": 0.0, "num_input_tokens_seen": 10693728, "step": 38190 }, { "epoch": 424.3888888888889, "grad_norm": 5.245883016868902e-07, "learning_rate": 2.510712586899833e-07, "loss": 0.0, "num_input_tokens_seen": 10695104, "step": 38195 }, { "epoch": 424.44444444444446, "grad_norm": 2.422443503746763e-05, "learning_rate": 2.4968529234478124e-07, "loss": 0.0, "num_input_tokens_seen": 10696528, "step": 38200 }, { "epoch": 424.44444444444446, "eval_loss": 1.2404636144638062, "eval_runtime": 1.3924, "eval_samples_per_second": 28.728, "eval_steps_per_second": 14.364, "num_input_tokens_seen": 10696528, "step": 38200 }, { "epoch": 424.5, "grad_norm": 2.899915898524341e-07, "learning_rate": 2.483031428091448e-07, "loss": 0.0, "num_input_tokens_seen": 10697920, "step": 38205 }, { "epoch": 424.55555555555554, "grad_norm": 7.895798717072466e-07, "learning_rate": 2.469248102962091e-07, "loss": 0.0, "num_input_tokens_seen": 10699392, "step": 38210 }, { "epoch": 424.6111111111111, "grad_norm": 2.4791955866021453e-07, "learning_rate": 2.4555029501853455e-07, "loss": 0.0, "num_input_tokens_seen": 10700768, "step": 38215 }, { "epoch": 424.6666666666667, "grad_norm": 4.34221510658972e-06, "learning_rate": 2.441795971880906e-07, "loss": 0.0, "num_input_tokens_seen": 10702176, "step": 38220 }, { "epoch": 424.72222222222223, "grad_norm": 9.873456008335779e-08, "learning_rate": 2.4281271701625255e-07, "loss": 0.0, "num_input_tokens_seen": 10703536, "step": 38225 }, { "epoch": 424.77777777777777, "grad_norm": 7.085635189696404e-08, "learning_rate": 2.4144965471381007e-07, "loss": 0.0, "num_input_tokens_seen": 10704928, "step": 38230 }, { "epoch": 424.8333333333333, "grad_norm": 2.7819555725727696e-07, "learning_rate": 2.400904104909674e-07, "loss": 0.0, "num_input_tokens_seen": 10706416, "step": 38235 }, { "epoch": 424.8888888888889, "grad_norm": 1.9442657048784895e-06, "learning_rate": 2.3873498455733725e-07, "loss": 0.0, "num_input_tokens_seen": 10707808, "step": 38240 }, { "epoch": 424.94444444444446, "grad_norm": 4.2619416262823506e-07, "learning_rate": 2.3738337712194137e-07, "loss": 0.0, "num_input_tokens_seen": 10709200, "step": 38245 }, { "epoch": 425.0, "grad_norm": 3.7569114397228986e-07, "learning_rate": 2.3603558839321305e-07, "loss": 0.0, "num_input_tokens_seen": 10710608, "step": 38250 }, { "epoch": 425.05555555555554, "grad_norm": 7.57438044729497e-07, "learning_rate": 2.3469161857900267e-07, "loss": 0.0, "num_input_tokens_seen": 10712016, "step": 38255 }, { "epoch": 425.1111111111111, "grad_norm": 6.915874450896808e-08, "learning_rate": 2.3335146788656393e-07, "loss": 0.0, "num_input_tokens_seen": 10713424, "step": 38260 }, { "epoch": 425.1666666666667, "grad_norm": 3.018059032910969e-06, "learning_rate": 2.3201513652256757e-07, "loss": 0.0, "num_input_tokens_seen": 10714832, "step": 38265 }, { "epoch": 425.22222222222223, "grad_norm": 3.722054771060357e-06, "learning_rate": 2.3068262469308766e-07, "loss": 0.0, "num_input_tokens_seen": 10716224, "step": 38270 }, { "epoch": 425.27777777777777, "grad_norm": 1.420505526539273e-07, "learning_rate": 2.2935393260362093e-07, "loss": 0.0, "num_input_tokens_seen": 10717632, "step": 38275 }, { "epoch": 425.3333333333333, "grad_norm": 6.444241762437741e-07, "learning_rate": 2.2802906045906458e-07, "loss": 0.0, "num_input_tokens_seen": 10718992, "step": 38280 }, { "epoch": 425.3888888888889, "grad_norm": 3.926863882952603e-06, "learning_rate": 2.2670800846373018e-07, "loss": 0.0, "num_input_tokens_seen": 10720416, "step": 38285 }, { "epoch": 425.44444444444446, "grad_norm": 4.523952057411407e-08, "learning_rate": 2.2539077682134367e-07, "loss": 0.0, "num_input_tokens_seen": 10721808, "step": 38290 }, { "epoch": 425.5, "grad_norm": 7.231107446159513e-08, "learning_rate": 2.2407736573503423e-07, "loss": 0.0, "num_input_tokens_seen": 10723248, "step": 38295 }, { "epoch": 425.55555555555554, "grad_norm": 2.7202091246181226e-07, "learning_rate": 2.2276777540735093e-07, "loss": 0.0, "num_input_tokens_seen": 10724656, "step": 38300 }, { "epoch": 425.6111111111111, "grad_norm": 6.702686050630291e-08, "learning_rate": 2.2146200604024613e-07, "loss": 0.0, "num_input_tokens_seen": 10726096, "step": 38305 }, { "epoch": 425.6666666666667, "grad_norm": 9.997265806305222e-07, "learning_rate": 2.2016005783508375e-07, "loss": 0.0, "num_input_tokens_seen": 10727536, "step": 38310 }, { "epoch": 425.72222222222223, "grad_norm": 8.595907843300665e-07, "learning_rate": 2.1886193099264763e-07, "loss": 0.0, "num_input_tokens_seen": 10728912, "step": 38315 }, { "epoch": 425.77777777777777, "grad_norm": 1.7308074973243492e-07, "learning_rate": 2.175676257131165e-07, "loss": 0.0, "num_input_tokens_seen": 10730256, "step": 38320 }, { "epoch": 425.8333333333333, "grad_norm": 3.106272004060884e-07, "learning_rate": 2.162771421960974e-07, "loss": 0.0, "num_input_tokens_seen": 10731648, "step": 38325 }, { "epoch": 425.8888888888889, "grad_norm": 4.547095784346311e-07, "learning_rate": 2.1499048064059224e-07, "loss": 0.0, "num_input_tokens_seen": 10733024, "step": 38330 }, { "epoch": 425.94444444444446, "grad_norm": 8.366617976207635e-08, "learning_rate": 2.1370764124502285e-07, "loss": 0.0, "num_input_tokens_seen": 10734416, "step": 38335 }, { "epoch": 426.0, "grad_norm": 1.6343376501026796e-06, "learning_rate": 2.1242862420721988e-07, "loss": 0.0, "num_input_tokens_seen": 10735840, "step": 38340 }, { "epoch": 426.05555555555554, "grad_norm": 7.111207196430769e-07, "learning_rate": 2.1115342972442276e-07, "loss": 0.0, "num_input_tokens_seen": 10737200, "step": 38345 }, { "epoch": 426.1111111111111, "grad_norm": 1.6473840958042274e-07, "learning_rate": 2.0988205799328252e-07, "loss": 0.0, "num_input_tokens_seen": 10738608, "step": 38350 }, { "epoch": 426.1666666666667, "grad_norm": 6.041288997948868e-07, "learning_rate": 2.0861450920986182e-07, "loss": 0.0, "num_input_tokens_seen": 10740032, "step": 38355 }, { "epoch": 426.22222222222223, "grad_norm": 1.9312986410113808e-07, "learning_rate": 2.07350783569632e-07, "loss": 0.0, "num_input_tokens_seen": 10741408, "step": 38360 }, { "epoch": 426.27777777777777, "grad_norm": 7.637684262817857e-08, "learning_rate": 2.060908812674761e-07, "loss": 0.0, "num_input_tokens_seen": 10742864, "step": 38365 }, { "epoch": 426.3333333333333, "grad_norm": 1.46002790302191e-07, "learning_rate": 2.0483480249768317e-07, "loss": 0.0, "num_input_tokens_seen": 10744224, "step": 38370 }, { "epoch": 426.3888888888889, "grad_norm": 7.609094154759077e-07, "learning_rate": 2.035825474539621e-07, "loss": 0.0, "num_input_tokens_seen": 10745616, "step": 38375 }, { "epoch": 426.44444444444446, "grad_norm": 1.5172885241554468e-06, "learning_rate": 2.0233411632942235e-07, "loss": 0.0, "num_input_tokens_seen": 10747056, "step": 38380 }, { "epoch": 426.5, "grad_norm": 1.2857663023169152e-07, "learning_rate": 2.0108950931658764e-07, "loss": 0.0, "num_input_tokens_seen": 10748416, "step": 38385 }, { "epoch": 426.55555555555554, "grad_norm": 2.8177873900858685e-07, "learning_rate": 1.998487266073934e-07, "loss": 0.0, "num_input_tokens_seen": 10749824, "step": 38390 }, { "epoch": 426.6111111111111, "grad_norm": 8.937409035070232e-08, "learning_rate": 1.986117683931865e-07, "loss": 0.0, "num_input_tokens_seen": 10751248, "step": 38395 }, { "epoch": 426.6666666666667, "grad_norm": 7.624934283967377e-08, "learning_rate": 1.9737863486471442e-07, "loss": 0.0, "num_input_tokens_seen": 10752640, "step": 38400 }, { "epoch": 426.6666666666667, "eval_loss": 1.2779897451400757, "eval_runtime": 1.3981, "eval_samples_per_second": 28.609, "eval_steps_per_second": 14.305, "num_input_tokens_seen": 10752640, "step": 38400 }, { "epoch": 426.72222222222223, "grad_norm": 1.4342103327180666e-07, "learning_rate": 1.9614932621215e-07, "loss": 0.0, "num_input_tokens_seen": 10754096, "step": 38405 }, { "epoch": 426.77777777777777, "grad_norm": 1.786996364216975e-07, "learning_rate": 1.9492384262506102e-07, "loss": 0.0, "num_input_tokens_seen": 10755456, "step": 38410 }, { "epoch": 426.8333333333333, "grad_norm": 2.387936035574967e-07, "learning_rate": 1.9370218429243524e-07, "loss": 0.0, "num_input_tokens_seen": 10756880, "step": 38415 }, { "epoch": 426.8888888888889, "grad_norm": 1.0339791600699755e-07, "learning_rate": 1.9248435140267197e-07, "loss": 0.0, "num_input_tokens_seen": 10758288, "step": 38420 }, { "epoch": 426.94444444444446, "grad_norm": 7.85766047783909e-08, "learning_rate": 1.9127034414356814e-07, "loss": 0.0, "num_input_tokens_seen": 10759696, "step": 38425 }, { "epoch": 427.0, "grad_norm": 2.581651585842337e-07, "learning_rate": 1.9006016270234627e-07, "loss": 0.0, "num_input_tokens_seen": 10761104, "step": 38430 }, { "epoch": 427.05555555555554, "grad_norm": 1.887032254899168e-07, "learning_rate": 1.888538072656293e-07, "loss": 0.0, "num_input_tokens_seen": 10762496, "step": 38435 }, { "epoch": 427.1111111111111, "grad_norm": 1.4949935689401173e-07, "learning_rate": 1.8765127801944893e-07, "loss": 0.0, "num_input_tokens_seen": 10763936, "step": 38440 }, { "epoch": 427.1666666666667, "grad_norm": 5.378412311074499e-07, "learning_rate": 1.8645257514925406e-07, "loss": 0.0, "num_input_tokens_seen": 10765360, "step": 38445 }, { "epoch": 427.22222222222223, "grad_norm": 6.237464731384534e-07, "learning_rate": 1.8525769883989685e-07, "loss": 0.0, "num_input_tokens_seen": 10766752, "step": 38450 }, { "epoch": 427.27777777777777, "grad_norm": 9.142852945842606e-07, "learning_rate": 1.8406664927564654e-07, "loss": 0.0, "num_input_tokens_seen": 10768176, "step": 38455 }, { "epoch": 427.3333333333333, "grad_norm": 2.3970966367414803e-07, "learning_rate": 1.8287942664017566e-07, "loss": 0.0, "num_input_tokens_seen": 10769584, "step": 38460 }, { "epoch": 427.3888888888889, "grad_norm": 3.1249723519977124e-07, "learning_rate": 1.8169603111656552e-07, "loss": 0.0, "num_input_tokens_seen": 10770976, "step": 38465 }, { "epoch": 427.44444444444446, "grad_norm": 9.839874337558285e-08, "learning_rate": 1.805164628873146e-07, "loss": 0.0, "num_input_tokens_seen": 10772336, "step": 38470 }, { "epoch": 427.5, "grad_norm": 1.108451570530633e-07, "learning_rate": 1.793407221343274e-07, "loss": 0.0, "num_input_tokens_seen": 10773744, "step": 38475 }, { "epoch": 427.55555555555554, "grad_norm": 1.0010851383412955e-06, "learning_rate": 1.781688090389172e-07, "loss": 0.0, "num_input_tokens_seen": 10775136, "step": 38480 }, { "epoch": 427.6111111111111, "grad_norm": 1.746267059843376e-07, "learning_rate": 1.770007237818061e-07, "loss": 0.0, "num_input_tokens_seen": 10776512, "step": 38485 }, { "epoch": 427.6666666666667, "grad_norm": 4.625207111530472e-06, "learning_rate": 1.7583646654313059e-07, "loss": 0.0, "num_input_tokens_seen": 10777888, "step": 38490 }, { "epoch": 427.72222222222223, "grad_norm": 1.1468001730463584e-06, "learning_rate": 1.7467603750242757e-07, "loss": 0.0, "num_input_tokens_seen": 10779264, "step": 38495 }, { "epoch": 427.77777777777777, "grad_norm": 9.905212209559977e-06, "learning_rate": 1.7351943683865944e-07, "loss": 0.0, "num_input_tokens_seen": 10780688, "step": 38500 }, { "epoch": 427.8333333333333, "grad_norm": 9.99483091845832e-08, "learning_rate": 1.723666647301808e-07, "loss": 0.0, "num_input_tokens_seen": 10782112, "step": 38505 }, { "epoch": 427.8888888888889, "grad_norm": 1.9747228634514613e-06, "learning_rate": 1.712177213547661e-07, "loss": 0.0, "num_input_tokens_seen": 10783472, "step": 38510 }, { "epoch": 427.94444444444446, "grad_norm": 1.8069285090405174e-07, "learning_rate": 1.7007260688959581e-07, "loss": 0.0, "num_input_tokens_seen": 10784896, "step": 38515 }, { "epoch": 428.0, "grad_norm": 3.338434737543139e-07, "learning_rate": 1.68931321511262e-07, "loss": 0.0, "num_input_tokens_seen": 10786256, "step": 38520 }, { "epoch": 428.05555555555554, "grad_norm": 1.905698212567586e-07, "learning_rate": 1.6779386539576835e-07, "loss": 0.0, "num_input_tokens_seen": 10787632, "step": 38525 }, { "epoch": 428.1111111111111, "grad_norm": 1.1806669419911486e-07, "learning_rate": 1.666602387185162e-07, "loss": 0.0, "num_input_tokens_seen": 10789072, "step": 38530 }, { "epoch": 428.1666666666667, "grad_norm": 9.972931991342193e-08, "learning_rate": 1.655304416543352e-07, "loss": 0.0, "num_input_tokens_seen": 10790432, "step": 38535 }, { "epoch": 428.22222222222223, "grad_norm": 2.3621822720087948e-07, "learning_rate": 1.6440447437744698e-07, "loss": 0.0, "num_input_tokens_seen": 10791824, "step": 38540 }, { "epoch": 428.27777777777777, "grad_norm": 1.046657871484058e-05, "learning_rate": 1.6328233706149332e-07, "loss": 0.0, "num_input_tokens_seen": 10793248, "step": 38545 }, { "epoch": 428.3333333333333, "grad_norm": 1.9063996603563282e-07, "learning_rate": 1.6216402987951906e-07, "loss": 0.0, "num_input_tokens_seen": 10794624, "step": 38550 }, { "epoch": 428.3888888888889, "grad_norm": 2.7679770937538706e-05, "learning_rate": 1.6104955300398627e-07, "loss": 0.0, "num_input_tokens_seen": 10796048, "step": 38555 }, { "epoch": 428.44444444444446, "grad_norm": 1.9183617041562684e-06, "learning_rate": 1.5993890660675748e-07, "loss": 0.0, "num_input_tokens_seen": 10797424, "step": 38560 }, { "epoch": 428.5, "grad_norm": 2.7532507829164388e-06, "learning_rate": 1.5883209085910678e-07, "loss": 0.0, "num_input_tokens_seen": 10798816, "step": 38565 }, { "epoch": 428.55555555555554, "grad_norm": 8.16899614619615e-08, "learning_rate": 1.5772910593172264e-07, "loss": 0.0, "num_input_tokens_seen": 10800224, "step": 38570 }, { "epoch": 428.6111111111111, "grad_norm": 2.1348128598219773e-08, "learning_rate": 1.5662995199469954e-07, "loss": 0.0, "num_input_tokens_seen": 10801616, "step": 38575 }, { "epoch": 428.6666666666667, "grad_norm": 4.2709606873359007e-07, "learning_rate": 1.5553462921753802e-07, "loss": 0.0, "num_input_tokens_seen": 10803024, "step": 38580 }, { "epoch": 428.72222222222223, "grad_norm": 4.4238472582947e-06, "learning_rate": 1.544431377691502e-07, "loss": 0.0, "num_input_tokens_seen": 10804432, "step": 38585 }, { "epoch": 428.77777777777777, "grad_norm": 2.174056817239034e-06, "learning_rate": 1.5335547781785975e-07, "loss": 0.0, "num_input_tokens_seen": 10805840, "step": 38590 }, { "epoch": 428.8333333333333, "grad_norm": 4.753993835038273e-06, "learning_rate": 1.5227164953139917e-07, "loss": 0.0, "num_input_tokens_seen": 10807264, "step": 38595 }, { "epoch": 428.8888888888889, "grad_norm": 1.841673764602092e-07, "learning_rate": 1.511916530769042e-07, "loss": 0.0, "num_input_tokens_seen": 10808672, "step": 38600 }, { "epoch": 428.8888888888889, "eval_loss": 1.2368600368499756, "eval_runtime": 1.4012, "eval_samples_per_second": 28.547, "eval_steps_per_second": 14.274, "num_input_tokens_seen": 10808672, "step": 38600 }, { "epoch": 428.94444444444446, "grad_norm": 1.7628268267344538e-07, "learning_rate": 1.5011548862092773e-07, "loss": 0.0, "num_input_tokens_seen": 10810048, "step": 38605 }, { "epoch": 429.0, "grad_norm": 3.173588254412607e-07, "learning_rate": 1.490431563294231e-07, "loss": 0.0, "num_input_tokens_seen": 10811408, "step": 38610 }, { "epoch": 429.05555555555554, "grad_norm": 2.091713668050943e-06, "learning_rate": 1.4797465636776365e-07, "loss": 0.0, "num_input_tokens_seen": 10812800, "step": 38615 }, { "epoch": 429.1111111111111, "grad_norm": 5.519793333519374e-08, "learning_rate": 1.4690998890072027e-07, "loss": 0.0, "num_input_tokens_seen": 10814160, "step": 38620 }, { "epoch": 429.1666666666667, "grad_norm": 6.465485853368591e-07, "learning_rate": 1.4584915409248112e-07, "loss": 0.0, "num_input_tokens_seen": 10815568, "step": 38625 }, { "epoch": 429.22222222222223, "grad_norm": 6.2602903199149296e-06, "learning_rate": 1.4479215210663754e-07, "loss": 0.0, "num_input_tokens_seen": 10816976, "step": 38630 }, { "epoch": 429.27777777777777, "grad_norm": 6.545886321873695e-08, "learning_rate": 1.4373898310619528e-07, "loss": 0.0, "num_input_tokens_seen": 10818352, "step": 38635 }, { "epoch": 429.3333333333333, "grad_norm": 7.464090145958835e-08, "learning_rate": 1.4268964725356604e-07, "loss": 0.0, "num_input_tokens_seen": 10819712, "step": 38640 }, { "epoch": 429.3888888888889, "grad_norm": 7.42512753504343e-08, "learning_rate": 1.4164414471056764e-07, "loss": 0.0, "num_input_tokens_seen": 10821168, "step": 38645 }, { "epoch": 429.44444444444446, "grad_norm": 4.816308418753579e-08, "learning_rate": 1.4060247563843497e-07, "loss": 0.0, "num_input_tokens_seen": 10822560, "step": 38650 }, { "epoch": 429.5, "grad_norm": 2.3819785610612598e-07, "learning_rate": 1.3956464019780068e-07, "loss": 0.0, "num_input_tokens_seen": 10823952, "step": 38655 }, { "epoch": 429.55555555555554, "grad_norm": 3.655230784715968e-07, "learning_rate": 1.385306385487145e-07, "loss": 0.0, "num_input_tokens_seen": 10825376, "step": 38660 }, { "epoch": 429.6111111111111, "grad_norm": 1.2192479914574506e-07, "learning_rate": 1.3750047085063222e-07, "loss": 0.0, "num_input_tokens_seen": 10826784, "step": 38665 }, { "epoch": 429.6666666666667, "grad_norm": 2.6376858386356616e-07, "learning_rate": 1.3647413726242119e-07, "loss": 0.0, "num_input_tokens_seen": 10828208, "step": 38670 }, { "epoch": 429.72222222222223, "grad_norm": 9.316987643614993e-07, "learning_rate": 1.3545163794235205e-07, "loss": 0.0, "num_input_tokens_seen": 10829616, "step": 38675 }, { "epoch": 429.77777777777777, "grad_norm": 6.544249572471017e-06, "learning_rate": 1.3443297304810698e-07, "loss": 0.0, "num_input_tokens_seen": 10830976, "step": 38680 }, { "epoch": 429.8333333333333, "grad_norm": 3.664202665731864e-07, "learning_rate": 1.3341814273677977e-07, "loss": 0.0, "num_input_tokens_seen": 10832384, "step": 38685 }, { "epoch": 429.8888888888889, "grad_norm": 2.566208365806233e-07, "learning_rate": 1.324071471648647e-07, "loss": 0.0, "num_input_tokens_seen": 10833808, "step": 38690 }, { "epoch": 429.94444444444446, "grad_norm": 2.742021081303392e-07, "learning_rate": 1.3139998648827312e-07, "loss": 0.0, "num_input_tokens_seen": 10835168, "step": 38695 }, { "epoch": 430.0, "grad_norm": 4.629794858601599e-08, "learning_rate": 1.3039666086232526e-07, "loss": 0.0, "num_input_tokens_seen": 10836592, "step": 38700 }, { "epoch": 430.05555555555554, "grad_norm": 4.527328201220371e-05, "learning_rate": 1.2939717044174183e-07, "loss": 0.0, "num_input_tokens_seen": 10838032, "step": 38705 }, { "epoch": 430.1111111111111, "grad_norm": 4.949579306412488e-07, "learning_rate": 1.284015153806578e-07, "loss": 0.0, "num_input_tokens_seen": 10839408, "step": 38710 }, { "epoch": 430.1666666666667, "grad_norm": 2.7428407634033647e-07, "learning_rate": 1.274096958326171e-07, "loss": 0.0, "num_input_tokens_seen": 10840768, "step": 38715 }, { "epoch": 430.22222222222223, "grad_norm": 1.589359754916586e-07, "learning_rate": 1.2642171195056952e-07, "loss": 0.0, "num_input_tokens_seen": 10842176, "step": 38720 }, { "epoch": 430.27777777777777, "grad_norm": 4.643295028472494e-07, "learning_rate": 1.2543756388687377e-07, "loss": 0.0, "num_input_tokens_seen": 10843552, "step": 38725 }, { "epoch": 430.3333333333333, "grad_norm": 3.883087273948149e-08, "learning_rate": 1.2445725179330014e-07, "loss": 0.0, "num_input_tokens_seen": 10844976, "step": 38730 }, { "epoch": 430.3888888888889, "grad_norm": 4.5848361196476617e-07, "learning_rate": 1.2348077582102212e-07, "loss": 0.0, "num_input_tokens_seen": 10846400, "step": 38735 }, { "epoch": 430.44444444444446, "grad_norm": 2.228785405122835e-07, "learning_rate": 1.2250813612062762e-07, "loss": 0.0, "num_input_tokens_seen": 10847744, "step": 38740 }, { "epoch": 430.5, "grad_norm": 7.349572683779115e-08, "learning_rate": 1.215393328421105e-07, "loss": 0.0, "num_input_tokens_seen": 10849136, "step": 38745 }, { "epoch": 430.55555555555554, "grad_norm": 5.780574952041206e-07, "learning_rate": 1.2057436613486796e-07, "loss": 0.0, "num_input_tokens_seen": 10850496, "step": 38750 }, { "epoch": 430.6111111111111, "grad_norm": 1.7764659787644632e-07, "learning_rate": 1.1961323614771424e-07, "loss": 0.0, "num_input_tokens_seen": 10851920, "step": 38755 }, { "epoch": 430.6666666666667, "grad_norm": 5.680395133822458e-06, "learning_rate": 1.1865594302886418e-07, "loss": 0.0, "num_input_tokens_seen": 10853328, "step": 38760 }, { "epoch": 430.72222222222223, "grad_norm": 4.736488534717864e-08, "learning_rate": 1.1770248692594687e-07, "loss": 0.0, "num_input_tokens_seen": 10854688, "step": 38765 }, { "epoch": 430.77777777777777, "grad_norm": 1.226277504429163e-06, "learning_rate": 1.167528679859975e-07, "loss": 0.0, "num_input_tokens_seen": 10856112, "step": 38770 }, { "epoch": 430.8333333333333, "grad_norm": 8.20295724679454e-07, "learning_rate": 1.1580708635545446e-07, "loss": 0.0, "num_input_tokens_seen": 10857552, "step": 38775 }, { "epoch": 430.8888888888889, "grad_norm": 1.1847848924162463e-07, "learning_rate": 1.1486514218017885e-07, "loss": 0.0, "num_input_tokens_seen": 10858976, "step": 38780 }, { "epoch": 430.94444444444446, "grad_norm": 9.096829955979047e-08, "learning_rate": 1.1392703560542117e-07, "loss": 0.0, "num_input_tokens_seen": 10860352, "step": 38785 }, { "epoch": 431.0, "grad_norm": 2.2727991222382116e-07, "learning_rate": 1.129927667758518e-07, "loss": 0.0, "num_input_tokens_seen": 10861760, "step": 38790 }, { "epoch": 431.05555555555554, "grad_norm": 1.379024752168334e-07, "learning_rate": 1.1206233583554992e-07, "loss": 0.0, "num_input_tokens_seen": 10863152, "step": 38795 }, { "epoch": 431.1111111111111, "grad_norm": 3.854003693959385e-08, "learning_rate": 1.1113574292799523e-07, "loss": 0.0, "num_input_tokens_seen": 10864512, "step": 38800 }, { "epoch": 431.1111111111111, "eval_loss": 1.2550748586654663, "eval_runtime": 1.3908, "eval_samples_per_second": 28.76, "eval_steps_per_second": 14.38, "num_input_tokens_seen": 10864512, "step": 38800 }, { "epoch": 431.1666666666667, "grad_norm": 1.202544552825202e-07, "learning_rate": 1.1021298819608449e-07, "loss": 0.0, "num_input_tokens_seen": 10865968, "step": 38805 }, { "epoch": 431.22222222222223, "grad_norm": 4.1175775322699337e-07, "learning_rate": 1.0929407178211226e-07, "loss": 0.0, "num_input_tokens_seen": 10867392, "step": 38810 }, { "epoch": 431.27777777777777, "grad_norm": 8.403858942074294e-08, "learning_rate": 1.0837899382779293e-07, "loss": 0.0, "num_input_tokens_seen": 10868768, "step": 38815 }, { "epoch": 431.3333333333333, "grad_norm": 4.151670225382986e-07, "learning_rate": 1.0746775447423862e-07, "loss": 0.0, "num_input_tokens_seen": 10870176, "step": 38820 }, { "epoch": 431.3888888888889, "grad_norm": 4.0586098748462973e-07, "learning_rate": 1.0656035386197583e-07, "loss": 0.0, "num_input_tokens_seen": 10871536, "step": 38825 }, { "epoch": 431.44444444444446, "grad_norm": 1.1252618747903398e-07, "learning_rate": 1.0565679213093982e-07, "loss": 0.0, "num_input_tokens_seen": 10872880, "step": 38830 }, { "epoch": 431.5, "grad_norm": 3.4032148050755495e-06, "learning_rate": 1.0475706942046638e-07, "loss": 0.0, "num_input_tokens_seen": 10874208, "step": 38835 }, { "epoch": 431.55555555555554, "grad_norm": 4.0589830518911185e-07, "learning_rate": 1.0386118586930282e-07, "loss": 0.0, "num_input_tokens_seen": 10875632, "step": 38840 }, { "epoch": 431.6111111111111, "grad_norm": 5.17771695740521e-07, "learning_rate": 1.0296914161561367e-07, "loss": 0.0, "num_input_tokens_seen": 10877072, "step": 38845 }, { "epoch": 431.6666666666667, "grad_norm": 3.611616875787149e-06, "learning_rate": 1.0208093679695552e-07, "loss": 0.0, "num_input_tokens_seen": 10878432, "step": 38850 }, { "epoch": 431.72222222222223, "grad_norm": 8.433127618445724e-07, "learning_rate": 1.0119657155030493e-07, "loss": 0.0, "num_input_tokens_seen": 10879776, "step": 38855 }, { "epoch": 431.77777777777777, "grad_norm": 5.4659629711295565e-08, "learning_rate": 1.003160460120417e-07, "loss": 0.0, "num_input_tokens_seen": 10881200, "step": 38860 }, { "epoch": 431.8333333333333, "grad_norm": 8.234826509578852e-08, "learning_rate": 9.943936031795165e-08, "loss": 0.0, "num_input_tokens_seen": 10882624, "step": 38865 }, { "epoch": 431.8888888888889, "grad_norm": 7.907125620931765e-08, "learning_rate": 9.856651460323219e-08, "loss": 0.0, "num_input_tokens_seen": 10884080, "step": 38870 }, { "epoch": 431.94444444444446, "grad_norm": 1.6918062328841188e-07, "learning_rate": 9.769750900248953e-08, "loss": 0.0, "num_input_tokens_seen": 10885504, "step": 38875 }, { "epoch": 432.0, "grad_norm": 1.2935855693285703e-06, "learning_rate": 9.683234364973038e-08, "loss": 0.0, "num_input_tokens_seen": 10886912, "step": 38880 }, { "epoch": 432.05555555555554, "grad_norm": 1.4703714157349168e-07, "learning_rate": 9.597101867837854e-08, "loss": 0.0, "num_input_tokens_seen": 10888304, "step": 38885 }, { "epoch": 432.1111111111111, "grad_norm": 1.6594746909959213e-07, "learning_rate": 9.511353422125835e-08, "loss": 0.0, "num_input_tokens_seen": 10889696, "step": 38890 }, { "epoch": 432.1666666666667, "grad_norm": 2.451050988838688e-07, "learning_rate": 9.42598904106029e-08, "loss": 0.0, "num_input_tokens_seen": 10891072, "step": 38895 }, { "epoch": 432.22222222222223, "grad_norm": 1.0804252781326795e-07, "learning_rate": 9.341008737806245e-08, "loss": 0.0, "num_input_tokens_seen": 10892480, "step": 38900 }, { "epoch": 432.27777777777777, "grad_norm": 6.833209198475743e-08, "learning_rate": 9.256412525467661e-08, "loss": 0.0, "num_input_tokens_seen": 10893888, "step": 38905 }, { "epoch": 432.3333333333333, "grad_norm": 2.9157699827919714e-06, "learning_rate": 9.172200417091326e-08, "loss": 0.0, "num_input_tokens_seen": 10895328, "step": 38910 }, { "epoch": 432.3888888888889, "grad_norm": 5.407063596862827e-08, "learning_rate": 9.088372425663239e-08, "loss": 0.0, "num_input_tokens_seen": 10896736, "step": 38915 }, { "epoch": 432.44444444444446, "grad_norm": 1.3592193681688514e-07, "learning_rate": 9.004928564110837e-08, "loss": 0.0, "num_input_tokens_seen": 10898096, "step": 38920 }, { "epoch": 432.5, "grad_norm": 4.58867610575453e-08, "learning_rate": 8.92186884530244e-08, "loss": 0.0, "num_input_tokens_seen": 10899472, "step": 38925 }, { "epoch": 432.55555555555554, "grad_norm": 1.6185518347811012e-07, "learning_rate": 8.83919328204641e-08, "loss": 0.0, "num_input_tokens_seen": 10900864, "step": 38930 }, { "epoch": 432.6111111111111, "grad_norm": 2.613857361666305e-07, "learning_rate": 8.756901887093105e-08, "loss": 0.0, "num_input_tokens_seen": 10902224, "step": 38935 }, { "epoch": 432.6666666666667, "grad_norm": 3.533996277838014e-06, "learning_rate": 8.674994673132098e-08, "loss": 0.0, "num_input_tokens_seen": 10903616, "step": 38940 }, { "epoch": 432.72222222222223, "grad_norm": 6.722587841068162e-06, "learning_rate": 8.593471652794949e-08, "loss": 0.0, "num_input_tokens_seen": 10905024, "step": 38945 }, { "epoch": 432.77777777777777, "grad_norm": 4.397073098516557e-06, "learning_rate": 8.512332838653548e-08, "loss": 0.0, "num_input_tokens_seen": 10906464, "step": 38950 }, { "epoch": 432.8333333333333, "grad_norm": 2.449159239859e-07, "learning_rate": 8.431578243220106e-08, "loss": 0.0, "num_input_tokens_seen": 10907856, "step": 38955 }, { "epoch": 432.8888888888889, "grad_norm": 2.015349593875726e-07, "learning_rate": 8.351207878948552e-08, "loss": 0.0, "num_input_tokens_seen": 10909280, "step": 38960 }, { "epoch": 432.94444444444446, "grad_norm": 2.199154067739073e-07, "learning_rate": 8.271221758232583e-08, "loss": 0.0, "num_input_tokens_seen": 10910688, "step": 38965 }, { "epoch": 433.0, "grad_norm": 1.6484901266267116e-07, "learning_rate": 8.191619893407332e-08, "loss": 0.0, "num_input_tokens_seen": 10912096, "step": 38970 }, { "epoch": 433.05555555555554, "grad_norm": 3.90359957691544e-08, "learning_rate": 8.112402296748534e-08, "loss": 0.0, "num_input_tokens_seen": 10913536, "step": 38975 }, { "epoch": 433.1111111111111, "grad_norm": 7.448470000781526e-07, "learning_rate": 8.033568980471973e-08, "loss": 0.0, "num_input_tokens_seen": 10914960, "step": 38980 }, { "epoch": 433.1666666666667, "grad_norm": 9.97614961306681e-07, "learning_rate": 7.955119956735146e-08, "loss": 0.0, "num_input_tokens_seen": 10916352, "step": 38985 }, { "epoch": 433.22222222222223, "grad_norm": 6.622908443887354e-08, "learning_rate": 7.877055237636155e-08, "loss": 0.0, "num_input_tokens_seen": 10917776, "step": 38990 }, { "epoch": 433.27777777777777, "grad_norm": 4.139239138112316e-07, "learning_rate": 7.79937483521287e-08, "loss": 0.0, "num_input_tokens_seen": 10919200, "step": 38995 }, { "epoch": 433.3333333333333, "grad_norm": 1.034143792821851e-06, "learning_rate": 7.722078761444873e-08, "loss": 0.0, "num_input_tokens_seen": 10920608, "step": 39000 }, { "epoch": 433.3333333333333, "eval_loss": 1.2631747722625732, "eval_runtime": 1.3907, "eval_samples_per_second": 28.762, "eval_steps_per_second": 14.381, "num_input_tokens_seen": 10920608, "step": 39000 }, { "epoch": 433.3888888888889, "grad_norm": 5.639682285618619e-07, "learning_rate": 7.645167028252631e-08, "loss": 0.0, "num_input_tokens_seen": 10921968, "step": 39005 }, { "epoch": 433.44444444444446, "grad_norm": 2.514684922516608e-07, "learning_rate": 7.568639647496379e-08, "loss": 0.0, "num_input_tokens_seen": 10923360, "step": 39010 }, { "epoch": 433.5, "grad_norm": 4.025465841550613e-06, "learning_rate": 7.492496630977508e-08, "loss": 0.0, "num_input_tokens_seen": 10924768, "step": 39015 }, { "epoch": 433.55555555555554, "grad_norm": 3.548695531208068e-06, "learning_rate": 7.416737990438571e-08, "loss": 0.0, "num_input_tokens_seen": 10926160, "step": 39020 }, { "epoch": 433.6111111111111, "grad_norm": 4.041394277010113e-07, "learning_rate": 7.341363737562445e-08, "loss": 0.0, "num_input_tokens_seen": 10927600, "step": 39025 }, { "epoch": 433.6666666666667, "grad_norm": 4.295770850148983e-08, "learning_rate": 7.266373883972887e-08, "loss": 0.0, "num_input_tokens_seen": 10928992, "step": 39030 }, { "epoch": 433.72222222222223, "grad_norm": 8.191955771508219e-07, "learning_rate": 7.191768441233981e-08, "loss": 0.0, "num_input_tokens_seen": 10930400, "step": 39035 }, { "epoch": 433.77777777777777, "grad_norm": 7.384974196611438e-07, "learning_rate": 7.11754742085069e-08, "loss": 0.0, "num_input_tokens_seen": 10931824, "step": 39040 }, { "epoch": 433.8333333333333, "grad_norm": 2.471984430485463e-07, "learning_rate": 7.043710834269413e-08, "loss": 0.0, "num_input_tokens_seen": 10933232, "step": 39045 }, { "epoch": 433.8888888888889, "grad_norm": 7.073562642290199e-07, "learning_rate": 6.970258692876319e-08, "loss": 0.0, "num_input_tokens_seen": 10934608, "step": 39050 }, { "epoch": 433.94444444444446, "grad_norm": 7.618161390610112e-08, "learning_rate": 6.897191007998738e-08, "loss": 0.0, "num_input_tokens_seen": 10935984, "step": 39055 }, { "epoch": 434.0, "grad_norm": 3.2120530590873386e-07, "learning_rate": 6.824507790904599e-08, "loss": 0.0, "num_input_tokens_seen": 10937392, "step": 39060 }, { "epoch": 434.05555555555554, "grad_norm": 1.3272602927827393e-06, "learning_rate": 6.752209052802439e-08, "loss": 0.0, "num_input_tokens_seen": 10938800, "step": 39065 }, { "epoch": 434.1111111111111, "grad_norm": 6.533256282637012e-07, "learning_rate": 6.680294804841946e-08, "loss": 0.0, "num_input_tokens_seen": 10940160, "step": 39070 }, { "epoch": 434.1666666666667, "grad_norm": 8.727123201879294e-08, "learning_rate": 6.608765058112865e-08, "loss": 0.0, "num_input_tokens_seen": 10941552, "step": 39075 }, { "epoch": 434.22222222222223, "grad_norm": 1.2249267911101924e-07, "learning_rate": 6.537619823646368e-08, "loss": 0.0, "num_input_tokens_seen": 10942960, "step": 39080 }, { "epoch": 434.27777777777777, "grad_norm": 1.911831759571214e-06, "learning_rate": 6.466859112413404e-08, "loss": 0.0, "num_input_tokens_seen": 10944416, "step": 39085 }, { "epoch": 434.3333333333333, "grad_norm": 1.6984225226224225e-07, "learning_rate": 6.39648293532663e-08, "loss": 0.0, "num_input_tokens_seen": 10945808, "step": 39090 }, { "epoch": 434.3888888888889, "grad_norm": 4.050940205502229e-08, "learning_rate": 6.32649130323848e-08, "loss": 0.0, "num_input_tokens_seen": 10947216, "step": 39095 }, { "epoch": 434.44444444444446, "grad_norm": 3.324977342344937e-07, "learning_rate": 6.256884226943094e-08, "loss": 0.0, "num_input_tokens_seen": 10948608, "step": 39100 }, { "epoch": 434.5, "grad_norm": 5.645721330438391e-07, "learning_rate": 6.187661717174386e-08, "loss": 0.0, "num_input_tokens_seen": 10950016, "step": 39105 }, { "epoch": 434.55555555555554, "grad_norm": 7.843968319320993e-08, "learning_rate": 6.118823784607708e-08, "loss": 0.0, "num_input_tokens_seen": 10951360, "step": 39110 }, { "epoch": 434.6111111111111, "grad_norm": 3.464551127763116e-06, "learning_rate": 6.050370439858178e-08, "loss": 0.0, "num_input_tokens_seen": 10952752, "step": 39115 }, { "epoch": 434.6666666666667, "grad_norm": 2.1968914154513186e-07, "learning_rate": 5.98230169348235e-08, "loss": 0.0, "num_input_tokens_seen": 10954144, "step": 39120 }, { "epoch": 434.72222222222223, "grad_norm": 5.635743249854386e-08, "learning_rate": 5.914617555977664e-08, "loss": 0.0, "num_input_tokens_seen": 10955520, "step": 39125 }, { "epoch": 434.77777777777777, "grad_norm": 1.004509613267146e-07, "learning_rate": 5.8473180377816017e-08, "loss": 0.0, "num_input_tokens_seen": 10956912, "step": 39130 }, { "epoch": 434.8333333333333, "grad_norm": 3.5829438616019615e-07, "learning_rate": 5.780403149272251e-08, "loss": 0.0, "num_input_tokens_seen": 10958352, "step": 39135 }, { "epoch": 434.8888888888889, "grad_norm": 1.864933523165746e-07, "learning_rate": 5.7138729007694126e-08, "loss": 0.0, "num_input_tokens_seen": 10959744, "step": 39140 }, { "epoch": 434.94444444444446, "grad_norm": 7.965180884639267e-06, "learning_rate": 5.64772730253238e-08, "loss": 0.0, "num_input_tokens_seen": 10961200, "step": 39145 }, { "epoch": 435.0, "grad_norm": 1.3355990802210727e-07, "learning_rate": 5.5819663647618814e-08, "loss": 0.0, "num_input_tokens_seen": 10962608, "step": 39150 }, { "epoch": 435.05555555555554, "grad_norm": 2.2816224998223333e-07, "learning_rate": 5.5165900975989723e-08, "loss": 0.0, "num_input_tokens_seen": 10964000, "step": 39155 }, { "epoch": 435.1111111111111, "grad_norm": 1.1370764951834644e-07, "learning_rate": 5.451598511125311e-08, "loss": 0.0, "num_input_tokens_seen": 10965456, "step": 39160 }, { "epoch": 435.1666666666667, "grad_norm": 5.3115189757591e-07, "learning_rate": 5.3869916153637124e-08, "loss": 0.0, "num_input_tokens_seen": 10966864, "step": 39165 }, { "epoch": 435.22222222222223, "grad_norm": 1.8451197547619813e-07, "learning_rate": 5.322769420277318e-08, "loss": 0.0, "num_input_tokens_seen": 10968272, "step": 39170 }, { "epoch": 435.27777777777777, "grad_norm": 5.343838438420789e-07, "learning_rate": 5.258931935769873e-08, "loss": 0.0, "num_input_tokens_seen": 10969648, "step": 39175 }, { "epoch": 435.3333333333333, "grad_norm": 1.583009456851414e-08, "learning_rate": 5.19547917168628e-08, "loss": 0.0, "num_input_tokens_seen": 10970992, "step": 39180 }, { "epoch": 435.3888888888889, "grad_norm": 2.581681428637239e-07, "learning_rate": 5.13241113781121e-08, "loss": 0.0, "num_input_tokens_seen": 10972400, "step": 39185 }, { "epoch": 435.44444444444446, "grad_norm": 2.1291425511549278e-08, "learning_rate": 5.0697278438707755e-08, "loss": 0.0, "num_input_tokens_seen": 10973824, "step": 39190 }, { "epoch": 435.5, "grad_norm": 2.9463834039233916e-08, "learning_rate": 5.0074292995316854e-08, "loss": 0.0, "num_input_tokens_seen": 10975216, "step": 39195 }, { "epoch": 435.55555555555554, "grad_norm": 9.018649507197551e-07, "learning_rate": 4.945515514400978e-08, "loss": 0.0, "num_input_tokens_seen": 10976624, "step": 39200 }, { "epoch": 435.55555555555554, "eval_loss": 1.2148511409759521, "eval_runtime": 1.4192, "eval_samples_per_second": 28.186, "eval_steps_per_second": 14.093, "num_input_tokens_seen": 10976624, "step": 39200 }, { "epoch": 435.6111111111111, "grad_norm": 3.526355385474744e-08, "learning_rate": 4.883986498026571e-08, "loss": 0.0, "num_input_tokens_seen": 10977936, "step": 39205 }, { "epoch": 435.6666666666667, "grad_norm": 3.8793655221525114e-07, "learning_rate": 4.822842259896987e-08, "loss": 0.0, "num_input_tokens_seen": 10979312, "step": 39210 }, { "epoch": 435.72222222222223, "grad_norm": 1.8239045118662034e-07, "learning_rate": 4.762082809441626e-08, "loss": 0.0, "num_input_tokens_seen": 10980704, "step": 39215 }, { "epoch": 435.77777777777777, "grad_norm": 3.2124777504805024e-08, "learning_rate": 4.7017081560302156e-08, "loss": 0.0, "num_input_tokens_seen": 10982064, "step": 39220 }, { "epoch": 435.8333333333333, "grad_norm": 3.782262410823023e-06, "learning_rate": 4.6417183089730866e-08, "loss": 0.0, "num_input_tokens_seen": 10983520, "step": 39225 }, { "epoch": 435.8888888888889, "grad_norm": 4.615148441189376e-08, "learning_rate": 4.5821132775217265e-08, "loss": 0.0, "num_input_tokens_seen": 10984960, "step": 39230 }, { "epoch": 435.94444444444446, "grad_norm": 2.687503695142368e-07, "learning_rate": 4.5228930708679504e-08, "loss": 0.0, "num_input_tokens_seen": 10986368, "step": 39235 }, { "epoch": 436.0, "grad_norm": 8.490266623084608e-08, "learning_rate": 4.464057698144175e-08, "loss": 0.0, "num_input_tokens_seen": 10987808, "step": 39240 }, { "epoch": 436.05555555555554, "grad_norm": 7.495272598134761e-07, "learning_rate": 4.4056071684236974e-08, "loss": 0.0, "num_input_tokens_seen": 10989216, "step": 39245 }, { "epoch": 436.1111111111111, "grad_norm": 8.504675719223087e-08, "learning_rate": 4.347541490719864e-08, "loss": 0.0, "num_input_tokens_seen": 10990608, "step": 39250 }, { "epoch": 436.1666666666667, "grad_norm": 1.0379599757470714e-07, "learning_rate": 4.2898606739877336e-08, "loss": 0.0, "num_input_tokens_seen": 10992048, "step": 39255 }, { "epoch": 436.22222222222223, "grad_norm": 6.663051408395404e-07, "learning_rate": 4.232564727122135e-08, "loss": 0.0, "num_input_tokens_seen": 10993424, "step": 39260 }, { "epoch": 436.27777777777777, "grad_norm": 4.492829646096652e-07, "learning_rate": 4.1756536589585004e-08, "loss": 0.0, "num_input_tokens_seen": 10994800, "step": 39265 }, { "epoch": 436.3333333333333, "grad_norm": 1.7932882201421307e-07, "learning_rate": 4.119127478273976e-08, "loss": 0.0, "num_input_tokens_seen": 10996240, "step": 39270 }, { "epoch": 436.3888888888889, "grad_norm": 9.015339941242928e-08, "learning_rate": 4.062986193784923e-08, "loss": 0.0, "num_input_tokens_seen": 10997616, "step": 39275 }, { "epoch": 436.44444444444446, "grad_norm": 3.490532662908663e-06, "learning_rate": 4.007229814149416e-08, "loss": 0.0, "num_input_tokens_seen": 10998960, "step": 39280 }, { "epoch": 436.5, "grad_norm": 3.8017310544091742e-06, "learning_rate": 3.951858347965576e-08, "loss": 0.0, "num_input_tokens_seen": 11000336, "step": 39285 }, { "epoch": 436.55555555555554, "grad_norm": 7.569764761683473e-07, "learning_rate": 3.896871803772684e-08, "loss": 0.0, "num_input_tokens_seen": 11001744, "step": 39290 }, { "epoch": 436.6111111111111, "grad_norm": 7.573440257147013e-07, "learning_rate": 3.842270190050068e-08, "loss": 0.0, "num_input_tokens_seen": 11003136, "step": 39295 }, { "epoch": 436.6666666666667, "grad_norm": 4.61844450683202e-07, "learning_rate": 3.7880535152179376e-08, "loss": 0.0, "num_input_tokens_seen": 11004512, "step": 39300 }, { "epoch": 436.72222222222223, "grad_norm": 1.2496823273977498e-06, "learning_rate": 3.734221787637382e-08, "loss": 0.0, "num_input_tokens_seen": 11005888, "step": 39305 }, { "epoch": 436.77777777777777, "grad_norm": 5.614128326669743e-07, "learning_rate": 3.680775015609817e-08, "loss": 0.0, "num_input_tokens_seen": 11007280, "step": 39310 }, { "epoch": 436.8333333333333, "grad_norm": 1.0192240551987197e-05, "learning_rate": 3.627713207377537e-08, "loss": 0.0, "num_input_tokens_seen": 11008704, "step": 39315 }, { "epoch": 436.8888888888889, "grad_norm": 9.332883905699418e-07, "learning_rate": 3.575036371123164e-08, "loss": 0.0, "num_input_tokens_seen": 11010112, "step": 39320 }, { "epoch": 436.94444444444446, "grad_norm": 1.0294046433045878e-06, "learning_rate": 3.5227445149704776e-08, "loss": 0.0, "num_input_tokens_seen": 11011552, "step": 39325 }, { "epoch": 437.0, "grad_norm": 3.383219393526815e-07, "learning_rate": 3.470837646983027e-08, "loss": 0.0, "num_input_tokens_seen": 11012960, "step": 39330 }, { "epoch": 437.05555555555554, "grad_norm": 7.493098337363335e-07, "learning_rate": 3.419315775165799e-08, "loss": 0.0, "num_input_tokens_seen": 11014352, "step": 39335 }, { "epoch": 437.1111111111111, "grad_norm": 4.949350085325932e-08, "learning_rate": 3.368178907464103e-08, "loss": 0.0, "num_input_tokens_seen": 11015728, "step": 39340 }, { "epoch": 437.1666666666667, "grad_norm": 1.0432240742375143e-05, "learning_rate": 3.317427051763855e-08, "loss": 0.0, "num_input_tokens_seen": 11017184, "step": 39345 }, { "epoch": 437.22222222222223, "grad_norm": 6.592748604816734e-07, "learning_rate": 3.267060215891571e-08, "loss": 0.0, "num_input_tokens_seen": 11018544, "step": 39350 }, { "epoch": 437.27777777777777, "grad_norm": 3.7236517869132513e-07, "learning_rate": 3.217078407614649e-08, "loss": 0.0, "num_input_tokens_seen": 11019936, "step": 39355 }, { "epoch": 437.3333333333333, "grad_norm": 3.753096279979218e-06, "learning_rate": 3.1674816346405345e-08, "loss": 0.0, "num_input_tokens_seen": 11021376, "step": 39360 }, { "epoch": 437.3888888888889, "grad_norm": 2.0641572007207287e-07, "learning_rate": 3.11826990461811e-08, "loss": 0.0, "num_input_tokens_seen": 11022720, "step": 39365 }, { "epoch": 437.44444444444446, "grad_norm": 4.829772137782129e-07, "learning_rate": 3.069443225136304e-08, "loss": 0.0, "num_input_tokens_seen": 11024128, "step": 39370 }, { "epoch": 437.5, "grad_norm": 1.3172029866836965e-07, "learning_rate": 3.021001603724372e-08, "loss": 0.0, "num_input_tokens_seen": 11025504, "step": 39375 }, { "epoch": 437.55555555555554, "grad_norm": 4.260633374997269e-07, "learning_rate": 2.9729450478532818e-08, "loss": 0.0, "num_input_tokens_seen": 11026928, "step": 39380 }, { "epoch": 437.6111111111111, "grad_norm": 2.3596594473929144e-06, "learning_rate": 2.9252735649337726e-08, "loss": 0.0, "num_input_tokens_seen": 11028368, "step": 39385 }, { "epoch": 437.6666666666667, "grad_norm": 7.761490792290715e-07, "learning_rate": 2.8779871623171863e-08, "loss": 0.0, "num_input_tokens_seen": 11029776, "step": 39390 }, { "epoch": 437.72222222222223, "grad_norm": 4.499524663970078e-08, "learning_rate": 2.8310858472957448e-08, "loss": 0.0, "num_input_tokens_seen": 11031168, "step": 39395 }, { "epoch": 437.77777777777777, "grad_norm": 4.934405524181784e-07, "learning_rate": 2.784569627101996e-08, "loss": 0.0, "num_input_tokens_seen": 11032608, "step": 39400 }, { "epoch": 437.77777777777777, "eval_loss": 1.2244216203689575, "eval_runtime": 1.3969, "eval_samples_per_second": 28.636, "eval_steps_per_second": 14.318, "num_input_tokens_seen": 11032608, "step": 39400 }, { "epoch": 437.8333333333333, "grad_norm": 8.998927114589605e-07, "learning_rate": 2.738438508909924e-08, "loss": 0.0, "num_input_tokens_seen": 11034032, "step": 39405 }, { "epoch": 437.8888888888889, "grad_norm": 1.1193137794407448e-07, "learning_rate": 2.692692499833005e-08, "loss": 0.0, "num_input_tokens_seen": 11035424, "step": 39410 }, { "epoch": 437.94444444444446, "grad_norm": 3.285652212525747e-07, "learning_rate": 2.647331606926151e-08, "loss": 0.0, "num_input_tokens_seen": 11036832, "step": 39415 }, { "epoch": 438.0, "grad_norm": 2.3632179591004387e-07, "learning_rate": 2.6023558371843225e-08, "loss": 0.0, "num_input_tokens_seen": 11038224, "step": 39420 }, { "epoch": 438.05555555555554, "grad_norm": 9.46601574014494e-07, "learning_rate": 2.557765197543638e-08, "loss": 0.0, "num_input_tokens_seen": 11039664, "step": 39425 }, { "epoch": 438.1111111111111, "grad_norm": 1.1607252758949471e-07, "learning_rate": 2.513559694880263e-08, "loss": 0.0, "num_input_tokens_seen": 11041104, "step": 39430 }, { "epoch": 438.1666666666667, "grad_norm": 1.015244379232172e-06, "learning_rate": 2.469739336011523e-08, "loss": 0.0, "num_input_tokens_seen": 11042512, "step": 39435 }, { "epoch": 438.22222222222223, "grad_norm": 6.741188371961471e-07, "learning_rate": 2.4263041276947894e-08, "loss": 0.0, "num_input_tokens_seen": 11043920, "step": 39440 }, { "epoch": 438.27777777777777, "grad_norm": 5.844580641678476e-07, "learning_rate": 2.3832540766283164e-08, "loss": 0.0, "num_input_tokens_seen": 11045344, "step": 39445 }, { "epoch": 438.3333333333333, "grad_norm": 2.724747787397064e-07, "learning_rate": 2.3405891894512366e-08, "loss": 0.0, "num_input_tokens_seen": 11046784, "step": 39450 }, { "epoch": 438.3888888888889, "grad_norm": 7.797053314106961e-08, "learning_rate": 2.29830947274301e-08, "loss": 0.0, "num_input_tokens_seen": 11048208, "step": 39455 }, { "epoch": 438.44444444444446, "grad_norm": 1.7103335494539351e-06, "learning_rate": 2.2564149330231432e-08, "loss": 0.0, "num_input_tokens_seen": 11049584, "step": 39460 }, { "epoch": 438.5, "grad_norm": 6.564294352529032e-08, "learning_rate": 2.2149055767528572e-08, "loss": 0.0, "num_input_tokens_seen": 11050992, "step": 39465 }, { "epoch": 438.55555555555554, "grad_norm": 9.954350389307365e-06, "learning_rate": 2.1737814103334197e-08, "loss": 0.0, "num_input_tokens_seen": 11052368, "step": 39470 }, { "epoch": 438.6111111111111, "grad_norm": 7.391416545488028e-08, "learning_rate": 2.1330424401064253e-08, "loss": 0.0, "num_input_tokens_seen": 11053760, "step": 39475 }, { "epoch": 438.6666666666667, "grad_norm": 2.0509838805082836e-07, "learning_rate": 2.092688672354348e-08, "loss": 0.0, "num_input_tokens_seen": 11055152, "step": 39480 }, { "epoch": 438.72222222222223, "grad_norm": 1.3171553803204006e-07, "learning_rate": 2.0527201133005435e-08, "loss": 0.0, "num_input_tokens_seen": 11056528, "step": 39485 }, { "epoch": 438.77777777777777, "grad_norm": 8.340261814510086e-08, "learning_rate": 2.0131367691084148e-08, "loss": 0.0, "num_input_tokens_seen": 11057920, "step": 39490 }, { "epoch": 438.8333333333333, "grad_norm": 5.632055177784423e-08, "learning_rate": 1.9739386458819675e-08, "loss": 0.0, "num_input_tokens_seen": 11059344, "step": 39495 }, { "epoch": 438.8888888888889, "grad_norm": 1.975634518203151e-07, "learning_rate": 1.9351257496666442e-08, "loss": 0.0, "num_input_tokens_seen": 11060736, "step": 39500 }, { "epoch": 438.94444444444446, "grad_norm": 1.2583825537149096e-07, "learning_rate": 1.896698086447657e-08, "loss": 0.0, "num_input_tokens_seen": 11062128, "step": 39505 }, { "epoch": 439.0, "grad_norm": 1.8202842966275057e-07, "learning_rate": 1.8586556621505436e-08, "loss": 0.0, "num_input_tokens_seen": 11063488, "step": 39510 }, { "epoch": 439.05555555555554, "grad_norm": 6.259837448396866e-08, "learning_rate": 1.820998482642833e-08, "loss": 0.0, "num_input_tokens_seen": 11064864, "step": 39515 }, { "epoch": 439.1111111111111, "grad_norm": 2.1718827269978647e-07, "learning_rate": 1.7837265537309912e-08, "loss": 0.0, "num_input_tokens_seen": 11066288, "step": 39520 }, { "epoch": 439.1666666666667, "grad_norm": 2.912879324412643e-07, "learning_rate": 1.7468398811629206e-08, "loss": 0.0, "num_input_tokens_seen": 11067664, "step": 39525 }, { "epoch": 439.22222222222223, "grad_norm": 3.244373374400311e-07, "learning_rate": 1.710338470627404e-08, "loss": 0.0, "num_input_tokens_seen": 11069056, "step": 39530 }, { "epoch": 439.27777777777777, "grad_norm": 1.4560119154793938e-07, "learning_rate": 1.6742223277529945e-08, "loss": 0.0, "num_input_tokens_seen": 11070480, "step": 39535 }, { "epoch": 439.3333333333333, "grad_norm": 8.620109781531937e-08, "learning_rate": 1.6384914581094036e-08, "loss": 0.0, "num_input_tokens_seen": 11071824, "step": 39540 }, { "epoch": 439.3888888888889, "grad_norm": 3.399937043013779e-07, "learning_rate": 1.6031458672069455e-08, "loss": 0.0, "num_input_tokens_seen": 11073200, "step": 39545 }, { "epoch": 439.44444444444446, "grad_norm": 2.0062270777998492e-07, "learning_rate": 1.5681855604962602e-08, "loss": 0.0, "num_input_tokens_seen": 11074576, "step": 39550 }, { "epoch": 439.5, "grad_norm": 7.71569716562226e-07, "learning_rate": 1.5336105433683135e-08, "loss": 0.0, "num_input_tokens_seen": 11076032, "step": 39555 }, { "epoch": 439.55555555555554, "grad_norm": 2.67430220191045e-08, "learning_rate": 1.499420821155506e-08, "loss": 0.0, "num_input_tokens_seen": 11077408, "step": 39560 }, { "epoch": 439.6111111111111, "grad_norm": 1.9879078081430634e-06, "learning_rate": 1.4656163991302874e-08, "loss": 0.0, "num_input_tokens_seen": 11078800, "step": 39565 }, { "epoch": 439.6666666666667, "grad_norm": 1.024310094521752e-07, "learning_rate": 1.4321972825051544e-08, "loss": 0.0, "num_input_tokens_seen": 11080224, "step": 39570 }, { "epoch": 439.72222222222223, "grad_norm": 1.4734935405158467e-07, "learning_rate": 1.3991634764345951e-08, "loss": 0.0, "num_input_tokens_seen": 11081664, "step": 39575 }, { "epoch": 439.77777777777777, "grad_norm": 1.258461225006613e-06, "learning_rate": 1.3665149860120352e-08, "loss": 0.0, "num_input_tokens_seen": 11083104, "step": 39580 }, { "epoch": 439.8333333333333, "grad_norm": 1.9377703210921027e-06, "learning_rate": 1.3342518162728912e-08, "loss": 0.0, "num_input_tokens_seen": 11084512, "step": 39585 }, { "epoch": 439.8888888888889, "grad_norm": 3.0284235208455357e-07, "learning_rate": 1.30237397219235e-08, "loss": 0.0, "num_input_tokens_seen": 11085888, "step": 39590 }, { "epoch": 439.94444444444446, "grad_norm": 9.946237469193875e-07, "learning_rate": 1.2708814586862016e-08, "loss": 0.0, "num_input_tokens_seen": 11087296, "step": 39595 }, { "epoch": 440.0, "grad_norm": 1.0208773346676026e-06, "learning_rate": 1.2397742806111168e-08, "loss": 0.0, "num_input_tokens_seen": 11088720, "step": 39600 }, { "epoch": 440.0, "eval_loss": 1.233507752418518, "eval_runtime": 1.3865, "eval_samples_per_second": 28.849, "eval_steps_per_second": 14.425, "num_input_tokens_seen": 11088720, "step": 39600 }, { "epoch": 440.05555555555554, "grad_norm": 1.626520855779745e-07, "learning_rate": 1.209052442764369e-08, "loss": 0.0, "num_input_tokens_seen": 11090144, "step": 39605 }, { "epoch": 440.1111111111111, "grad_norm": 4.007997347343917e-07, "learning_rate": 1.17871594988328e-08, "loss": 0.0, "num_input_tokens_seen": 11091552, "step": 39610 }, { "epoch": 440.1666666666667, "grad_norm": 2.900929132465535e-07, "learning_rate": 1.1487648066466072e-08, "loss": 0.0, "num_input_tokens_seen": 11092912, "step": 39615 }, { "epoch": 440.22222222222223, "grad_norm": 2.1600295951884618e-07, "learning_rate": 1.1191990176728784e-08, "loss": 0.0, "num_input_tokens_seen": 11094320, "step": 39620 }, { "epoch": 440.27777777777777, "grad_norm": 3.353999318278511e-07, "learning_rate": 1.0900185875215018e-08, "loss": 0.0, "num_input_tokens_seen": 11095696, "step": 39625 }, { "epoch": 440.3333333333333, "grad_norm": 1.0735790283433744e-06, "learning_rate": 1.0612235206924891e-08, "loss": 0.0, "num_input_tokens_seen": 11097088, "step": 39630 }, { "epoch": 440.3888888888889, "grad_norm": 1.5039121308291215e-06, "learning_rate": 1.0328138216264549e-08, "loss": 0.0, "num_input_tokens_seen": 11098544, "step": 39635 }, { "epoch": 440.44444444444446, "grad_norm": 5.258871169644408e-07, "learning_rate": 1.004789494704339e-08, "loss": 0.0, "num_input_tokens_seen": 11099952, "step": 39640 }, { "epoch": 440.5, "grad_norm": 1.3456072338158265e-05, "learning_rate": 9.771505442482397e-09, "loss": 0.0, "num_input_tokens_seen": 11101376, "step": 39645 }, { "epoch": 440.55555555555554, "grad_norm": 4.0276850654663576e-07, "learning_rate": 9.498969745200259e-09, "loss": 0.0, "num_input_tokens_seen": 11102752, "step": 39650 }, { "epoch": 440.6111111111111, "grad_norm": 1.3907445861605083e-07, "learning_rate": 9.230287897230017e-09, "loss": 0.0, "num_input_tokens_seen": 11104128, "step": 39655 }, { "epoch": 440.6666666666667, "grad_norm": 3.9195747376652434e-06, "learning_rate": 8.965459940002419e-09, "loss": 0.0, "num_input_tokens_seen": 11105504, "step": 39660 }, { "epoch": 440.72222222222223, "grad_norm": 5.364996695789159e-07, "learning_rate": 8.704485914357019e-09, "loss": 0.0, "num_input_tokens_seen": 11106928, "step": 39665 }, { "epoch": 440.77777777777777, "grad_norm": 3.9320818245869305e-07, "learning_rate": 8.447365860539402e-09, "loss": 0.0, "num_input_tokens_seen": 11108304, "step": 39670 }, { "epoch": 440.8333333333333, "grad_norm": 1.4013030522619374e-06, "learning_rate": 8.194099818201184e-09, "loss": 0.0, "num_input_tokens_seen": 11109696, "step": 39675 }, { "epoch": 440.8888888888889, "grad_norm": 5.543582304312622e-08, "learning_rate": 7.944687826400011e-09, "loss": 0.0, "num_input_tokens_seen": 11111104, "step": 39680 }, { "epoch": 440.94444444444446, "grad_norm": 1.0345985401727376e-06, "learning_rate": 7.699129923599557e-09, "loss": 0.0, "num_input_tokens_seen": 11112496, "step": 39685 }, { "epoch": 441.0, "grad_norm": 3.818285676970845e-06, "learning_rate": 7.457426147663982e-09, "loss": 0.0, "num_input_tokens_seen": 11113888, "step": 39690 }, { "epoch": 441.05555555555554, "grad_norm": 2.268211119371699e-06, "learning_rate": 7.219576535871797e-09, "loss": 0.0, "num_input_tokens_seen": 11115264, "step": 39695 }, { "epoch": 441.1111111111111, "grad_norm": 7.20788406738393e-08, "learning_rate": 6.985581124896445e-09, "loss": 0.0, "num_input_tokens_seen": 11116672, "step": 39700 }, { "epoch": 441.1666666666667, "grad_norm": 9.146414754468424e-07, "learning_rate": 6.755439950828501e-09, "loss": 0.0, "num_input_tokens_seen": 11118096, "step": 39705 }, { "epoch": 441.22222222222223, "grad_norm": 2.1416668971596664e-07, "learning_rate": 6.5291530491562444e-09, "loss": 0.0, "num_input_tokens_seen": 11119520, "step": 39710 }, { "epoch": 441.27777777777777, "grad_norm": 3.942921580346592e-07, "learning_rate": 6.3067204547739845e-09, "loss": 0.0, "num_input_tokens_seen": 11120912, "step": 39715 }, { "epoch": 441.3333333333333, "grad_norm": 5.246170644568338e-07, "learning_rate": 6.088142201987612e-09, "loss": 0.0, "num_input_tokens_seen": 11122320, "step": 39720 }, { "epoch": 441.3888888888889, "grad_norm": 1.7012530406645965e-07, "learning_rate": 5.873418324503499e-09, "loss": 0.0, "num_input_tokens_seen": 11123728, "step": 39725 }, { "epoch": 441.44444444444446, "grad_norm": 3.4482144428693573e-07, "learning_rate": 5.6625488554340465e-09, "loss": 0.0, "num_input_tokens_seen": 11125168, "step": 39730 }, { "epoch": 441.5, "grad_norm": 3.730700086634897e-07, "learning_rate": 5.455533827297688e-09, "loss": 0.0, "num_input_tokens_seen": 11126576, "step": 39735 }, { "epoch": 441.55555555555554, "grad_norm": 6.913685979270667e-08, "learning_rate": 5.252373272018885e-09, "loss": 0.0, "num_input_tokens_seen": 11127952, "step": 39740 }, { "epoch": 441.6111111111111, "grad_norm": 3.0574827292184636e-07, "learning_rate": 5.053067220925356e-09, "loss": 0.0, "num_input_tokens_seen": 11129360, "step": 39745 }, { "epoch": 441.6666666666667, "grad_norm": 1.603163894969839e-07, "learning_rate": 4.857615704759177e-09, "loss": 0.0, "num_input_tokens_seen": 11130736, "step": 39750 }, { "epoch": 441.72222222222223, "grad_norm": 3.676893456372454e-08, "learning_rate": 4.666018753654577e-09, "loss": 0.0, "num_input_tokens_seen": 11132144, "step": 39755 }, { "epoch": 441.77777777777777, "grad_norm": 4.925417670165189e-06, "learning_rate": 4.478276397162917e-09, "loss": 0.0, "num_input_tokens_seen": 11133552, "step": 39760 }, { "epoch": 441.8333333333333, "grad_norm": 5.3810886413430126e-08, "learning_rate": 4.294388664233262e-09, "loss": 0.0, "num_input_tokens_seen": 11134912, "step": 39765 }, { "epoch": 441.8888888888889, "grad_norm": 2.901164464219619e-07, "learning_rate": 4.114355583223484e-09, "loss": 0.0, "num_input_tokens_seen": 11136336, "step": 39770 }, { "epoch": 441.94444444444446, "grad_norm": 9.018673154059798e-06, "learning_rate": 3.9381771818974845e-09, "loss": 0.0, "num_input_tokens_seen": 11137744, "step": 39775 }, { "epoch": 442.0, "grad_norm": 4.234654227275314e-07, "learning_rate": 3.765853487427973e-09, "loss": 0.0, "num_input_tokens_seen": 11139120, "step": 39780 }, { "epoch": 442.05555555555554, "grad_norm": 6.401437531167176e-07, "learning_rate": 3.5973845263825857e-09, "loss": 0.0, "num_input_tokens_seen": 11140544, "step": 39785 }, { "epoch": 442.1111111111111, "grad_norm": 1.7735497692683566e-07, "learning_rate": 3.4327703247488684e-09, "loss": 0.0, "num_input_tokens_seen": 11141920, "step": 39790 }, { "epoch": 442.1666666666667, "grad_norm": 1.6711129546820302e-07, "learning_rate": 3.2720109079037443e-09, "loss": 0.0, "num_input_tokens_seen": 11143312, "step": 39795 }, { "epoch": 442.22222222222223, "grad_norm": 5.394645654632768e-07, "learning_rate": 3.1151063006468193e-09, "loss": 0.0, "num_input_tokens_seen": 11144688, "step": 39800 }, { "epoch": 442.22222222222223, "eval_loss": 1.2725274562835693, "eval_runtime": 1.3989, "eval_samples_per_second": 28.594, "eval_steps_per_second": 14.297, "num_input_tokens_seen": 11144688, "step": 39800 }, { "epoch": 442.27777777777777, "grad_norm": 5.315893645274627e-07, "learning_rate": 2.962056527169854e-09, "loss": 0.0, "num_input_tokens_seen": 11146048, "step": 39805 }, { "epoch": 442.3333333333333, "grad_norm": 2.430999472835538e-07, "learning_rate": 2.8128616110761898e-09, "loss": 0.0, "num_input_tokens_seen": 11147488, "step": 39810 }, { "epoch": 442.3888888888889, "grad_norm": 4.1986126575466187e-07, "learning_rate": 2.6675215753724223e-09, "loss": 0.0, "num_input_tokens_seen": 11148912, "step": 39815 }, { "epoch": 442.44444444444446, "grad_norm": 1.984032496693544e-05, "learning_rate": 2.5260364424739557e-09, "loss": 0.0, "num_input_tokens_seen": 11150288, "step": 39820 }, { "epoch": 442.5, "grad_norm": 3.9466601720050676e-07, "learning_rate": 2.3884062341994475e-09, "loss": 0.0, "num_input_tokens_seen": 11151664, "step": 39825 }, { "epoch": 442.55555555555554, "grad_norm": 3.5023975897274795e-07, "learning_rate": 2.25463097177081e-09, "loss": 0.0, "num_input_tokens_seen": 11152992, "step": 39830 }, { "epoch": 442.6111111111111, "grad_norm": 3.087361903908459e-07, "learning_rate": 2.1247106758215397e-09, "loss": 0.0, "num_input_tokens_seen": 11154416, "step": 39835 }, { "epoch": 442.6666666666667, "grad_norm": 2.0708748138531519e-07, "learning_rate": 1.998645366382834e-09, "loss": 0.0, "num_input_tokens_seen": 11155824, "step": 39840 }, { "epoch": 442.72222222222223, "grad_norm": 2.1363473479141248e-07, "learning_rate": 1.876435062897475e-09, "loss": 0.0, "num_input_tokens_seen": 11157216, "step": 39845 }, { "epoch": 442.77777777777777, "grad_norm": 1.1499054153318866e-07, "learning_rate": 1.758079784211497e-09, "loss": 0.0, "num_input_tokens_seen": 11158624, "step": 39850 }, { "epoch": 442.8333333333333, "grad_norm": 1.6751438636219973e-07, "learning_rate": 1.6435795485797434e-09, "loss": 0.0, "num_input_tokens_seen": 11160000, "step": 39855 }, { "epoch": 442.8888888888889, "grad_norm": 8.924629355533398e-07, "learning_rate": 1.5329343736547596e-09, "loss": 0.0, "num_input_tokens_seen": 11161456, "step": 39860 }, { "epoch": 442.94444444444446, "grad_norm": 4.828223609365523e-06, "learning_rate": 1.4261442765006739e-09, "loss": 0.0, "num_input_tokens_seen": 11162896, "step": 39865 }, { "epoch": 443.0, "grad_norm": 3.763068789908175e-08, "learning_rate": 1.3232092735876445e-09, "loss": 0.0, "num_input_tokens_seen": 11164336, "step": 39870 }, { "epoch": 443.05555555555554, "grad_norm": 4.5509332835536043e-07, "learning_rate": 1.2241293807918607e-09, "loss": 0.0, "num_input_tokens_seen": 11165712, "step": 39875 }, { "epoch": 443.1111111111111, "grad_norm": 1.930721111875755e-07, "learning_rate": 1.128904613387216e-09, "loss": 0.0, "num_input_tokens_seen": 11167104, "step": 39880 }, { "epoch": 443.1666666666667, "grad_norm": 5.436601213659742e-07, "learning_rate": 1.0375349860591853e-09, "loss": 0.0, "num_input_tokens_seen": 11168560, "step": 39885 }, { "epoch": 443.22222222222223, "grad_norm": 5.080055842654474e-08, "learning_rate": 9.5002051290205e-10, "loss": 0.0, "num_input_tokens_seen": 11169968, "step": 39890 }, { "epoch": 443.27777777777777, "grad_norm": 7.397937679343158e-06, "learning_rate": 8.663612074077954e-10, "loss": 0.0, "num_input_tokens_seen": 11171376, "step": 39895 }, { "epoch": 443.3333333333333, "grad_norm": 8.443445409511696e-08, "learning_rate": 7.865570824799884e-10, "loss": 0.0, "num_input_tokens_seen": 11172768, "step": 39900 }, { "epoch": 443.3888888888889, "grad_norm": 2.9905481824243907e-07, "learning_rate": 7.106081504254514e-10, "loss": 0.0, "num_input_tokens_seen": 11174160, "step": 39905 }, { "epoch": 443.44444444444446, "grad_norm": 5.178172841624473e-07, "learning_rate": 6.385144229570372e-10, "loss": 0.0, "num_input_tokens_seen": 11175584, "step": 39910 }, { "epoch": 443.5, "grad_norm": 4.456080660020234e-08, "learning_rate": 5.70275911190854e-10, "loss": 0.0, "num_input_tokens_seen": 11176928, "step": 39915 }, { "epoch": 443.55555555555554, "grad_norm": 8.255192796013944e-08, "learning_rate": 5.058926256490403e-10, "loss": 0.0, "num_input_tokens_seen": 11178304, "step": 39920 }, { "epoch": 443.6111111111111, "grad_norm": 1.1714479342117556e-07, "learning_rate": 4.4536457626254134e-10, "loss": 0.0, "num_input_tokens_seen": 11179728, "step": 39925 }, { "epoch": 443.6666666666667, "grad_norm": 2.480014131833741e-07, "learning_rate": 3.88691772365557e-10, "loss": 0.0, "num_input_tokens_seen": 11181104, "step": 39930 }, { "epoch": 443.72222222222223, "grad_norm": 9.049216487255762e-08, "learning_rate": 3.358742226955425e-10, "loss": 0.0, "num_input_tokens_seen": 11182512, "step": 39935 }, { "epoch": 443.77777777777777, "grad_norm": 1.8009421864917385e-07, "learning_rate": 2.8691193539875925e-10, "loss": 0.0, "num_input_tokens_seen": 11183952, "step": 39940 }, { "epoch": 443.8333333333333, "grad_norm": 6.059113388801052e-08, "learning_rate": 2.418049180274995e-10, "loss": 0.0, "num_input_tokens_seen": 11185344, "step": 39945 }, { "epoch": 443.8888888888889, "grad_norm": 9.574938530931831e-08, "learning_rate": 2.005531775373104e-10, "loss": 0.0, "num_input_tokens_seen": 11186752, "step": 39950 }, { "epoch": 443.94444444444446, "grad_norm": 3.4814790978998644e-06, "learning_rate": 1.6315672028699435e-10, "loss": 0.0, "num_input_tokens_seen": 11188176, "step": 39955 }, { "epoch": 444.0, "grad_norm": 6.309709021934395e-08, "learning_rate": 1.2961555204693555e-10, "loss": 0.0, "num_input_tokens_seen": 11189552, "step": 39960 }, { "epoch": 444.05555555555554, "grad_norm": 2.8488329917308874e-07, "learning_rate": 9.992967798799768e-11, "loss": 0.0, "num_input_tokens_seen": 11190960, "step": 39965 }, { "epoch": 444.1111111111111, "grad_norm": 4.843337464421893e-08, "learning_rate": 7.409910268707521e-11, "loss": 0.0, "num_input_tokens_seen": 11192368, "step": 39970 }, { "epoch": 444.1666666666667, "grad_norm": 7.982529837136099e-07, "learning_rate": 5.212383012986877e-11, "loss": 0.0, "num_input_tokens_seen": 11193776, "step": 39975 }, { "epoch": 444.22222222222223, "grad_norm": 4.797757924279722e-07, "learning_rate": 3.400386370533415e-11, "loss": 0.0, "num_input_tokens_seen": 11195216, "step": 39980 }, { "epoch": 444.27777777777777, "grad_norm": 2.8305370847192535e-07, "learning_rate": 1.9739206205682258e-11, "loss": 0.0, "num_input_tokens_seen": 11196624, "step": 39985 }, { "epoch": 444.3333333333333, "grad_norm": 2.2482335282347776e-07, "learning_rate": 9.329859829154685e-12, "loss": 0.0, "num_input_tokens_seen": 11198032, "step": 39990 }, { "epoch": 444.3888888888889, "grad_norm": 1.7705808375012566e-07, "learning_rate": 2.7758261855748148e-12, "loss": 0.0, "num_input_tokens_seen": 11199424, "step": 39995 }, { "epoch": 444.44444444444446, "grad_norm": 3.799516079538989e-08, "learning_rate": 7.710628524559838e-14, "loss": 0.0, "num_input_tokens_seen": 11200800, "step": 40000 }, { "epoch": 444.44444444444446, "eval_loss": 1.2688204050064087, "eval_runtime": 1.4014, "eval_samples_per_second": 28.542, "eval_steps_per_second": 14.271, "num_input_tokens_seen": 11200800, "step": 40000 }, { "epoch": 444.44444444444446, "num_input_tokens_seen": 11200800, "step": 40000, "total_flos": 4.72061397030912e+16, "train_loss": 0.004416686975649121, "train_runtime": 17416.8724, "train_samples_per_second": 9.186, "train_steps_per_second": 2.297 } ], "logging_steps": 5, "max_steps": 40000, "num_input_tokens_seen": 11200800, "num_train_epochs": 445, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.72061397030912e+16, "train_batch_size": 2, "trial_name": null, "trial_params": null }