{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 6.036524590163935, "eval_steps": 500, "global_step": 24000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.000546448087431694, "grad_norm": 5.3125, "learning_rate": 6.124999999999999e-06, "loss": 9.6445, "step": 50 }, { "epoch": 0.001092896174863388, "grad_norm": 5.40625, "learning_rate": 1.2375e-05, "loss": 8.1001, "step": 100 }, { "epoch": 0.001639344262295082, "grad_norm": 4.0, "learning_rate": 1.8625e-05, "loss": 7.5007, "step": 150 }, { "epoch": 0.002185792349726776, "grad_norm": 6.125, "learning_rate": 2.4874999999999998e-05, "loss": 7.0884, "step": 200 }, { "epoch": 0.00273224043715847, "grad_norm": 3.703125, "learning_rate": 3.1125e-05, "loss": 6.6637, "step": 250 }, { "epoch": 0.003278688524590164, "grad_norm": 3.25, "learning_rate": 3.7375e-05, "loss": 6.3275, "step": 300 }, { "epoch": 0.003825136612021858, "grad_norm": 2.578125, "learning_rate": 4.3624999999999997e-05, "loss": 6.0834, "step": 350 }, { "epoch": 0.004371584699453552, "grad_norm": 2.640625, "learning_rate": 4.9875e-05, "loss": 5.8539, "step": 400 }, { "epoch": 0.004918032786885246, "grad_norm": 2.578125, "learning_rate": 5.6124999999999995e-05, "loss": 5.7353, "step": 450 }, { "epoch": 0.00546448087431694, "grad_norm": 2.75, "learning_rate": 6.2375e-05, "loss": 5.6411, "step": 500 }, { "epoch": 0.006010928961748634, "grad_norm": 1.8359375, "learning_rate": 6.8625e-05, "loss": 5.488, "step": 550 }, { "epoch": 0.006557377049180328, "grad_norm": 2.609375, "learning_rate": 7.487499999999999e-05, "loss": 5.3441, "step": 600 }, { "epoch": 0.007103825136612022, "grad_norm": 1.765625, "learning_rate": 8.112500000000001e-05, "loss": 5.2922, "step": 650 }, { "epoch": 0.007650273224043716, "grad_norm": 1.7265625, "learning_rate": 8.7375e-05, "loss": 5.1776, "step": 700 }, { "epoch": 0.00819672131147541, "grad_norm": 2.28125, "learning_rate": 9.362499999999999e-05, "loss": 5.0849, "step": 750 }, { "epoch": 0.008743169398907104, "grad_norm": 1.171875, "learning_rate": 9.9875e-05, "loss": 4.9607, "step": 800 }, { "epoch": 0.009289617486338797, "grad_norm": 1.5234375, "learning_rate": 0.00010612499999999999, "loss": 4.8847, "step": 850 }, { "epoch": 0.009836065573770493, "grad_norm": 1.15625, "learning_rate": 0.000112375, "loss": 4.8233, "step": 900 }, { "epoch": 0.010382513661202186, "grad_norm": 1.3359375, "learning_rate": 0.000118625, "loss": 4.7077, "step": 950 }, { "epoch": 0.01092896174863388, "grad_norm": 0.9921875, "learning_rate": 0.00012487499999999999, "loss": 4.6274, "step": 1000 }, { "epoch": 0.011475409836065573, "grad_norm": 1.15625, "learning_rate": 0.00013112499999999998, "loss": 4.5506, "step": 1050 }, { "epoch": 0.012021857923497269, "grad_norm": 1.03125, "learning_rate": 0.000137375, "loss": 4.4801, "step": 1100 }, { "epoch": 0.012568306010928962, "grad_norm": 1.3046875, "learning_rate": 0.00014362499999999998, "loss": 4.425, "step": 1150 }, { "epoch": 0.013114754098360656, "grad_norm": 1.046875, "learning_rate": 0.000149875, "loss": 4.3204, "step": 1200 }, { "epoch": 0.01366120218579235, "grad_norm": 7.5, "learning_rate": 0.000156125, "loss": 4.2244, "step": 1250 }, { "epoch": 0.014207650273224045, "grad_norm": 0.99609375, "learning_rate": 0.00016237499999999998, "loss": 4.1705, "step": 1300 }, { "epoch": 0.014754098360655738, "grad_norm": 0.83984375, "learning_rate": 0.000168625, "loss": 4.1256, "step": 1350 }, { "epoch": 0.015300546448087432, "grad_norm": 0.93359375, "learning_rate": 0.00017487499999999998, "loss": 4.0456, "step": 1400 }, { "epoch": 0.015846994535519125, "grad_norm": 0.90234375, "learning_rate": 0.000181125, "loss": 3.9429, "step": 1450 }, { "epoch": 0.01639344262295082, "grad_norm": 0.8671875, "learning_rate": 0.000187375, "loss": 3.8733, "step": 1500 }, { "epoch": 0.016939890710382512, "grad_norm": 0.9140625, "learning_rate": 0.00019362499999999998, "loss": 3.8094, "step": 1550 }, { "epoch": 0.017486338797814208, "grad_norm": 0.890625, "learning_rate": 0.000199875, "loss": 3.8117, "step": 1600 }, { "epoch": 0.018032786885245903, "grad_norm": 1.0390625, "learning_rate": 0.00020612499999999998, "loss": 3.6919, "step": 1650 }, { "epoch": 0.018579234972677595, "grad_norm": 0.94921875, "learning_rate": 0.00021237499999999997, "loss": 3.6148, "step": 1700 }, { "epoch": 0.01912568306010929, "grad_norm": 1.0546875, "learning_rate": 0.000218625, "loss": 3.5881, "step": 1750 }, { "epoch": 0.019672131147540985, "grad_norm": 0.953125, "learning_rate": 0.000224875, "loss": 3.46, "step": 1800 }, { "epoch": 0.020218579234972677, "grad_norm": 0.9765625, "learning_rate": 0.00023112499999999997, "loss": 3.4731, "step": 1850 }, { "epoch": 0.020765027322404372, "grad_norm": 0.984375, "learning_rate": 0.00023737499999999998, "loss": 3.3781, "step": 1900 }, { "epoch": 0.021311475409836064, "grad_norm": 1.0078125, "learning_rate": 0.000243625, "loss": 3.3166, "step": 1950 }, { "epoch": 0.02185792349726776, "grad_norm": 1.0859375, "learning_rate": 0.000249875, "loss": 3.3105, "step": 2000 }, { "epoch": 0.022404371584699455, "grad_norm": 0.86328125, "learning_rate": 0.000256125, "loss": 3.25, "step": 2050 }, { "epoch": 0.022950819672131147, "grad_norm": 0.9375, "learning_rate": 0.00026237499999999997, "loss": 3.1414, "step": 2100 }, { "epoch": 0.023497267759562842, "grad_norm": 0.86328125, "learning_rate": 0.000268625, "loss": 3.1565, "step": 2150 }, { "epoch": 0.024043715846994537, "grad_norm": 0.80859375, "learning_rate": 0.000274875, "loss": 3.1131, "step": 2200 }, { "epoch": 0.02459016393442623, "grad_norm": 0.91015625, "learning_rate": 0.00028112499999999996, "loss": 3.0784, "step": 2250 }, { "epoch": 0.025136612021857924, "grad_norm": 0.80859375, "learning_rate": 0.000287375, "loss": 3.0332, "step": 2300 }, { "epoch": 0.025683060109289616, "grad_norm": 0.85546875, "learning_rate": 0.000293625, "loss": 3.0955, "step": 2350 }, { "epoch": 0.02622950819672131, "grad_norm": 0.734375, "learning_rate": 0.000299875, "loss": 3.045, "step": 2400 }, { "epoch": 0.026775956284153007, "grad_norm": 0.8359375, "learning_rate": 0.0002999997761290961, "loss": 2.995, "step": 2450 }, { "epoch": 0.0273224043715847, "grad_norm": 0.75390625, "learning_rate": 0.0002999990861486685, "loss": 2.9428, "step": 2500 }, { "epoch": 0.027868852459016394, "grad_norm": 0.71484375, "learning_rate": 0.00029999792996762107, "loss": 2.9131, "step": 2550 }, { "epoch": 0.02841530054644809, "grad_norm": 0.88671875, "learning_rate": 0.00029999630758954706, "loss": 2.896, "step": 2600 }, { "epoch": 0.02896174863387978, "grad_norm": 0.73828125, "learning_rate": 0.000299994219019489, "loss": 2.8605, "step": 2650 }, { "epoch": 0.029508196721311476, "grad_norm": 0.66796875, "learning_rate": 0.0002999916642639382, "loss": 2.8407, "step": 2700 }, { "epoch": 0.030054644808743168, "grad_norm": 0.68359375, "learning_rate": 0.0002999886433308348, "loss": 2.8313, "step": 2750 }, { "epoch": 0.030601092896174863, "grad_norm": 0.69140625, "learning_rate": 0.00029998515622956803, "loss": 2.8194, "step": 2800 }, { "epoch": 0.03114754098360656, "grad_norm": 0.61328125, "learning_rate": 0.00029998120297097586, "loss": 2.7874, "step": 2850 }, { "epoch": 0.03169398907103825, "grad_norm": 0.67578125, "learning_rate": 0.00029997678356734504, "loss": 2.7631, "step": 2900 }, { "epoch": 0.03224043715846994, "grad_norm": 0.6875, "learning_rate": 0.0002999718980324113, "loss": 2.7603, "step": 2950 }, { "epoch": 0.03278688524590164, "grad_norm": 0.62890625, "learning_rate": 0.0002999665463813589, "loss": 2.7229, "step": 3000 }, { "epoch": 0.03333333333333333, "grad_norm": 0.671875, "learning_rate": 0.00029996072863082093, "loss": 2.7895, "step": 3050 }, { "epoch": 0.033879781420765025, "grad_norm": 0.91796875, "learning_rate": 0.0002999544447988791, "loss": 2.6505, "step": 3100 }, { "epoch": 0.03442622950819672, "grad_norm": 0.60546875, "learning_rate": 0.0002999476949050637, "loss": 2.6744, "step": 3150 }, { "epoch": 0.034972677595628415, "grad_norm": 0.59375, "learning_rate": 0.0002999404789703535, "loss": 2.6869, "step": 3200 }, { "epoch": 0.03551912568306011, "grad_norm": 0.76953125, "learning_rate": 0.0002999327970171759, "loss": 2.6726, "step": 3250 }, { "epoch": 0.036065573770491806, "grad_norm": 0.66015625, "learning_rate": 0.0002999246490694065, "loss": 2.6444, "step": 3300 }, { "epoch": 0.0366120218579235, "grad_norm": 0.69921875, "learning_rate": 0.0002999160351523693, "loss": 2.6568, "step": 3350 }, { "epoch": 0.03715846994535519, "grad_norm": 0.625, "learning_rate": 0.00029990695529283665, "loss": 2.6436, "step": 3400 }, { "epoch": 1.0000765027322405, "grad_norm": 0.5390625, "learning_rate": 0.00029989740951902885, "loss": 2.6468, "step": 3450 }, { "epoch": 1.0006229508196722, "grad_norm": 0.578125, "learning_rate": 0.0002998873978606145, "loss": 2.5703, "step": 3500 }, { "epoch": 1.0011693989071038, "grad_norm": 0.6171875, "learning_rate": 0.0002998769203487099, "loss": 2.6321, "step": 3550 }, { "epoch": 1.0017158469945355, "grad_norm": 0.62109375, "learning_rate": 0.0002998659770158796, "loss": 2.5518, "step": 3600 }, { "epoch": 1.0022622950819673, "grad_norm": 0.6171875, "learning_rate": 0.0002998545678961356, "loss": 2.5255, "step": 3650 }, { "epoch": 1.0028087431693988, "grad_norm": 0.66015625, "learning_rate": 0.00029984269302493776, "loss": 2.4976, "step": 3700 }, { "epoch": 1.0033551912568306, "grad_norm": 0.58984375, "learning_rate": 0.0002998303524391934, "loss": 2.532, "step": 3750 }, { "epoch": 1.0039016393442624, "grad_norm": 0.6484375, "learning_rate": 0.00029981754617725747, "loss": 2.5321, "step": 3800 }, { "epoch": 1.004448087431694, "grad_norm": 0.53515625, "learning_rate": 0.0002998042742789319, "loss": 2.4924, "step": 3850 }, { "epoch": 1.0049945355191257, "grad_norm": 0.6015625, "learning_rate": 0.0002997905367854663, "loss": 2.492, "step": 3900 }, { "epoch": 1.0055409836065574, "grad_norm": 0.61328125, "learning_rate": 0.00029977633373955696, "loss": 2.5266, "step": 3950 }, { "epoch": 1.0060874316939892, "grad_norm": 0.58984375, "learning_rate": 0.00029976166518534735, "loss": 2.4739, "step": 4000 }, { "epoch": 1.0066338797814207, "grad_norm": 0.5546875, "learning_rate": 0.00029974653116842764, "loss": 2.4487, "step": 4050 }, { "epoch": 1.0071803278688525, "grad_norm": 0.60546875, "learning_rate": 0.0002997309317358347, "loss": 2.4674, "step": 4100 }, { "epoch": 1.0077267759562842, "grad_norm": 0.67578125, "learning_rate": 0.0002997148669360519, "loss": 2.4814, "step": 4150 }, { "epoch": 1.0082732240437158, "grad_norm": 0.53515625, "learning_rate": 0.00029969833681900914, "loss": 2.448, "step": 4200 }, { "epoch": 1.0088196721311475, "grad_norm": 0.68359375, "learning_rate": 0.0002996813414360822, "loss": 2.4299, "step": 4250 }, { "epoch": 1.0093661202185793, "grad_norm": 0.57421875, "learning_rate": 0.00029966388084009334, "loss": 2.4271, "step": 4300 }, { "epoch": 1.0099125683060108, "grad_norm": 0.515625, "learning_rate": 0.00029964595508531034, "loss": 2.4848, "step": 4350 }, { "epoch": 1.0104590163934426, "grad_norm": 0.53125, "learning_rate": 0.00029962756422744695, "loss": 2.414, "step": 4400 }, { "epoch": 1.0110054644808744, "grad_norm": 0.51171875, "learning_rate": 0.00029960870832366224, "loss": 2.3993, "step": 4450 }, { "epoch": 1.0115519125683061, "grad_norm": 0.54296875, "learning_rate": 0.000299589387432561, "loss": 2.4171, "step": 4500 }, { "epoch": 1.0120983606557377, "grad_norm": 0.490234375, "learning_rate": 0.00029956960161419283, "loss": 2.4038, "step": 4550 }, { "epoch": 1.0126448087431694, "grad_norm": 0.498046875, "learning_rate": 0.0002995493509300526, "loss": 2.4128, "step": 4600 }, { "epoch": 1.0131912568306012, "grad_norm": 0.51171875, "learning_rate": 0.0002995286354430799, "loss": 2.3721, "step": 4650 }, { "epoch": 1.0137377049180327, "grad_norm": 0.51171875, "learning_rate": 0.0002995074552176589, "loss": 2.3734, "step": 4700 }, { "epoch": 1.0142841530054645, "grad_norm": 0.515625, "learning_rate": 0.00029948581031961826, "loss": 2.3805, "step": 4750 }, { "epoch": 1.0148306010928962, "grad_norm": 0.515625, "learning_rate": 0.0002994637008162308, "loss": 2.3819, "step": 4800 }, { "epoch": 1.0153770491803278, "grad_norm": 0.53515625, "learning_rate": 0.00029944112677621345, "loss": 2.3839, "step": 4850 }, { "epoch": 1.0159234972677595, "grad_norm": 0.4921875, "learning_rate": 0.00029941808826972673, "loss": 2.336, "step": 4900 }, { "epoch": 1.0164699453551913, "grad_norm": 0.515625, "learning_rate": 0.0002993945853683749, "loss": 2.3126, "step": 4950 }, { "epoch": 1.0170163934426228, "grad_norm": 0.53515625, "learning_rate": 0.00029937061814520546, "loss": 2.3271, "step": 5000 }, { "epoch": 1.0175628415300546, "grad_norm": 0.53125, "learning_rate": 0.00029934618667470925, "loss": 2.3275, "step": 5050 }, { "epoch": 1.0181092896174864, "grad_norm": 0.51953125, "learning_rate": 0.0002993212910328197, "loss": 2.2837, "step": 5100 }, { "epoch": 1.0186557377049181, "grad_norm": 0.56640625, "learning_rate": 0.00029929593129691305, "loss": 2.2964, "step": 5150 }, { "epoch": 1.0192021857923497, "grad_norm": 0.50390625, "learning_rate": 0.000299270107545808, "loss": 2.3155, "step": 5200 }, { "epoch": 1.0197486338797814, "grad_norm": 0.55078125, "learning_rate": 0.00029924381985976534, "loss": 2.2722, "step": 5250 }, { "epoch": 1.0202950819672132, "grad_norm": 0.49609375, "learning_rate": 0.00029921706832048784, "loss": 2.3175, "step": 5300 }, { "epoch": 1.0208415300546447, "grad_norm": 0.48828125, "learning_rate": 0.00029918985301111985, "loss": 2.2834, "step": 5350 }, { "epoch": 1.0213879781420765, "grad_norm": 0.734375, "learning_rate": 0.00029916217401624716, "loss": 2.2522, "step": 5400 }, { "epoch": 1.0219344262295083, "grad_norm": 0.46484375, "learning_rate": 0.00029913403142189677, "loss": 2.2872, "step": 5450 }, { "epoch": 1.0224808743169398, "grad_norm": 0.51953125, "learning_rate": 0.00029910542531553656, "loss": 2.2793, "step": 5500 }, { "epoch": 1.0230273224043716, "grad_norm": 0.474609375, "learning_rate": 0.00029907635578607487, "loss": 2.218, "step": 5550 }, { "epoch": 1.0235737704918033, "grad_norm": 0.49609375, "learning_rate": 0.00029904682292386053, "loss": 2.2309, "step": 5600 }, { "epoch": 1.024120218579235, "grad_norm": 0.52734375, "learning_rate": 0.0002990168268206823, "loss": 2.2285, "step": 5650 }, { "epoch": 1.0246666666666666, "grad_norm": 0.48828125, "learning_rate": 0.00029898636756976884, "loss": 2.2338, "step": 5700 }, { "epoch": 1.0252131147540984, "grad_norm": 0.462890625, "learning_rate": 0.0002989554452657881, "loss": 2.2048, "step": 5750 }, { "epoch": 1.0257595628415301, "grad_norm": 0.62109375, "learning_rate": 0.0002989240600048475, "loss": 2.2716, "step": 5800 }, { "epoch": 1.0263060109289617, "grad_norm": 0.6015625, "learning_rate": 0.00029889221188449295, "loss": 2.2618, "step": 5850 }, { "epoch": 1.0268524590163934, "grad_norm": 0.47265625, "learning_rate": 0.0002988599010037092, "loss": 2.2181, "step": 5900 }, { "epoch": 1.0273989071038252, "grad_norm": 0.5234375, "learning_rate": 0.0002988271274629192, "loss": 2.2005, "step": 5950 }, { "epoch": 1.0279453551912567, "grad_norm": 0.515625, "learning_rate": 0.00029879389136398403, "loss": 2.1958, "step": 6000 }, { "epoch": 1.0284918032786885, "grad_norm": 0.52734375, "learning_rate": 0.00029876019281020207, "loss": 2.1853, "step": 6050 }, { "epoch": 1.0290382513661203, "grad_norm": 0.52734375, "learning_rate": 0.00029872603190630927, "loss": 2.1753, "step": 6100 }, { "epoch": 1.029584699453552, "grad_norm": 0.5546875, "learning_rate": 0.00029869140875847847, "loss": 2.1931, "step": 6150 }, { "epoch": 1.0301311475409836, "grad_norm": 0.53125, "learning_rate": 0.0002986563234743193, "loss": 2.1846, "step": 6200 }, { "epoch": 1.0306775956284153, "grad_norm": 0.51953125, "learning_rate": 0.0002986207761628775, "loss": 2.1928, "step": 6250 }, { "epoch": 1.031224043715847, "grad_norm": 0.6171875, "learning_rate": 0.00029858476693463506, "loss": 2.1942, "step": 6300 }, { "epoch": 1.0317704918032786, "grad_norm": 0.48828125, "learning_rate": 0.0002985482959015094, "loss": 2.1653, "step": 6350 }, { "epoch": 1.0323169398907104, "grad_norm": 0.5234375, "learning_rate": 0.00029851136317685345, "loss": 2.1659, "step": 6400 }, { "epoch": 1.0328633879781421, "grad_norm": 0.5234375, "learning_rate": 0.00029847396887545485, "loss": 2.1829, "step": 6450 }, { "epoch": 1.0334098360655737, "grad_norm": 0.54296875, "learning_rate": 0.00029843611311353597, "loss": 2.1911, "step": 6500 }, { "epoch": 1.0339562841530054, "grad_norm": 0.51953125, "learning_rate": 0.00029839779600875343, "loss": 2.1041, "step": 6550 }, { "epoch": 1.0345027322404372, "grad_norm": 0.53125, "learning_rate": 0.00029835901768019763, "loss": 2.1634, "step": 6600 }, { "epoch": 1.0350491803278687, "grad_norm": 0.515625, "learning_rate": 0.0002983197782483926, "loss": 2.1642, "step": 6650 }, { "epoch": 1.0355956284153005, "grad_norm": 0.58203125, "learning_rate": 0.00029828007783529533, "loss": 2.1621, "step": 6700 }, { "epoch": 1.0361420765027323, "grad_norm": 0.5078125, "learning_rate": 0.0002982399165642956, "loss": 2.1553, "step": 6750 }, { "epoch": 1.036688524590164, "grad_norm": 0.51171875, "learning_rate": 0.00029819929456021565, "loss": 2.1592, "step": 6800 }, { "epoch": 1.0372349726775956, "grad_norm": 0.51953125, "learning_rate": 0.0002981582119493095, "loss": 2.1527, "step": 6850 }, { "epoch": 2.000153005464481, "grad_norm": 0.5, "learning_rate": 0.0002981166688592629, "loss": 2.1674, "step": 6900 }, { "epoch": 2.0006994535519125, "grad_norm": 0.5390625, "learning_rate": 0.00029807466541919273, "loss": 2.1168, "step": 6950 }, { "epoch": 2.0012459016393445, "grad_norm": 0.53515625, "learning_rate": 0.00029803220175964675, "loss": 2.1439, "step": 7000 }, { "epoch": 2.001792349726776, "grad_norm": 0.5, "learning_rate": 0.0002979892780126028, "loss": 2.0952, "step": 7050 }, { "epoch": 2.0023387978142075, "grad_norm": 0.490234375, "learning_rate": 0.00029794589431146904, "loss": 2.0817, "step": 7100 }, { "epoch": 2.0028852459016395, "grad_norm": 0.462890625, "learning_rate": 0.00029790205079108294, "loss": 2.0643, "step": 7150 }, { "epoch": 2.003431693989071, "grad_norm": 0.55078125, "learning_rate": 0.00029785774758771114, "loss": 2.0993, "step": 7200 }, { "epoch": 2.0039781420765026, "grad_norm": 0.51953125, "learning_rate": 0.00029781298483904907, "loss": 2.1085, "step": 7250 }, { "epoch": 2.0045245901639346, "grad_norm": 0.4921875, "learning_rate": 0.0002977677626842204, "loss": 2.0645, "step": 7300 }, { "epoch": 2.005071038251366, "grad_norm": 0.50390625, "learning_rate": 0.0002977220812637766, "loss": 2.0929, "step": 7350 }, { "epoch": 2.0056174863387977, "grad_norm": 0.53515625, "learning_rate": 0.0002976759407196966, "loss": 2.0919, "step": 7400 }, { "epoch": 2.0061639344262296, "grad_norm": 0.55078125, "learning_rate": 0.00029762934119538623, "loss": 2.0903, "step": 7450 }, { "epoch": 2.006710382513661, "grad_norm": 0.5625, "learning_rate": 0.00029758228283567796, "loss": 2.0481, "step": 7500 }, { "epoch": 2.0072568306010927, "grad_norm": 0.5703125, "learning_rate": 0.00029753476578683023, "loss": 2.0737, "step": 7550 }, { "epoch": 2.0078032786885247, "grad_norm": 0.49609375, "learning_rate": 0.00029748679019652704, "loss": 2.1032, "step": 7600 }, { "epoch": 2.0083497267759562, "grad_norm": 0.5546875, "learning_rate": 0.00029743835621387775, "loss": 2.0722, "step": 7650 }, { "epoch": 2.008896174863388, "grad_norm": 0.5234375, "learning_rate": 0.00029738946398941623, "loss": 2.057, "step": 7700 }, { "epoch": 2.0094426229508198, "grad_norm": 0.490234375, "learning_rate": 0.0002973401136751007, "loss": 2.0802, "step": 7750 }, { "epoch": 2.0099890710382513, "grad_norm": 0.5, "learning_rate": 0.0002972903054243129, "loss": 2.1094, "step": 7800 }, { "epoch": 2.010535519125683, "grad_norm": 0.515625, "learning_rate": 0.0002972400393918583, "loss": 2.0409, "step": 7850 }, { "epoch": 2.011081967213115, "grad_norm": 0.578125, "learning_rate": 0.0002971893157339647, "loss": 2.0543, "step": 7900 }, { "epoch": 2.0116284153005464, "grad_norm": 0.5234375, "learning_rate": 0.0002971381346082824, "loss": 2.0776, "step": 7950 }, { "epoch": 2.0121748633879784, "grad_norm": 0.55078125, "learning_rate": 0.00029708649617388356, "loss": 2.0629, "step": 8000 }, { "epoch": 2.01272131147541, "grad_norm": 0.515625, "learning_rate": 0.0002970344005912617, "loss": 2.0588, "step": 8050 }, { "epoch": 2.0132677595628414, "grad_norm": 0.4765625, "learning_rate": 0.000296981848022331, "loss": 2.0373, "step": 8100 }, { "epoch": 2.0138142076502734, "grad_norm": 0.44921875, "learning_rate": 0.000296928838630426, "loss": 2.0348, "step": 8150 }, { "epoch": 2.014360655737705, "grad_norm": 0.490234375, "learning_rate": 0.0002968753725803013, "loss": 2.0784, "step": 8200 }, { "epoch": 2.0149071038251365, "grad_norm": 0.5390625, "learning_rate": 0.0002968214500381304, "loss": 2.0531, "step": 8250 }, { "epoch": 2.0154535519125685, "grad_norm": 0.546875, "learning_rate": 0.000296767071171506, "loss": 2.0482, "step": 8300 }, { "epoch": 2.016, "grad_norm": 0.484375, "learning_rate": 0.00029671223614943874, "loss": 2.0193, "step": 8350 }, { "epoch": 2.0165464480874316, "grad_norm": 0.4765625, "learning_rate": 0.0002966569451423572, "loss": 2.007, "step": 8400 }, { "epoch": 2.0170928961748635, "grad_norm": 0.48828125, "learning_rate": 0.000296601198322107, "loss": 2.0325, "step": 8450 }, { "epoch": 2.017639344262295, "grad_norm": 0.4921875, "learning_rate": 0.0002965449958619508, "loss": 2.0173, "step": 8500 }, { "epoch": 2.0181857923497266, "grad_norm": 0.51953125, "learning_rate": 0.0002964883379365668, "loss": 1.9927, "step": 8550 }, { "epoch": 2.0187322404371586, "grad_norm": 0.546875, "learning_rate": 0.00029643122472204934, "loss": 2.0149, "step": 8600 }, { "epoch": 2.01927868852459, "grad_norm": 0.49609375, "learning_rate": 0.00029637365639590763, "loss": 2.0077, "step": 8650 }, { "epoch": 2.0198251366120217, "grad_norm": 0.5390625, "learning_rate": 0.00029631563313706525, "loss": 1.9926, "step": 8700 }, { "epoch": 2.0203715846994537, "grad_norm": 0.494140625, "learning_rate": 0.0002962571551258599, "loss": 2.0248, "step": 8750 }, { "epoch": 2.020918032786885, "grad_norm": 0.51953125, "learning_rate": 0.00029619822254404256, "loss": 1.998, "step": 8800 }, { "epoch": 2.0214644808743167, "grad_norm": 0.478515625, "learning_rate": 0.00029613883557477706, "loss": 1.9957, "step": 8850 }, { "epoch": 2.0220109289617487, "grad_norm": 0.49609375, "learning_rate": 0.00029607899440263946, "loss": 2.0132, "step": 8900 }, { "epoch": 2.0225573770491803, "grad_norm": 0.5390625, "learning_rate": 0.00029601869921361756, "loss": 2.0038, "step": 8950 }, { "epoch": 2.0231038251366122, "grad_norm": 0.458984375, "learning_rate": 0.00029595795019511005, "loss": 1.9447, "step": 9000 }, { "epoch": 2.023650273224044, "grad_norm": 0.5078125, "learning_rate": 0.00029589674753592647, "loss": 1.9806, "step": 9050 }, { "epoch": 2.0241967213114753, "grad_norm": 0.5546875, "learning_rate": 0.000295835091426286, "loss": 1.9738, "step": 9100 }, { "epoch": 2.0247431693989073, "grad_norm": 0.5078125, "learning_rate": 0.00029577298205781726, "loss": 1.9568, "step": 9150 }, { "epoch": 2.025289617486339, "grad_norm": 0.5078125, "learning_rate": 0.00029571041962355755, "loss": 1.9778, "step": 9200 }, { "epoch": 2.0258360655737704, "grad_norm": 0.5078125, "learning_rate": 0.0002956474043179525, "loss": 2.0212, "step": 9250 }, { "epoch": 2.0263825136612024, "grad_norm": 0.51953125, "learning_rate": 0.0002955839363368549, "loss": 1.9981, "step": 9300 }, { "epoch": 2.026928961748634, "grad_norm": 0.5, "learning_rate": 0.00029552001587752495, "loss": 1.9763, "step": 9350 }, { "epoch": 2.0274754098360654, "grad_norm": 0.515625, "learning_rate": 0.0002954556431386288, "loss": 1.9687, "step": 9400 }, { "epoch": 2.0280218579234974, "grad_norm": 0.5390625, "learning_rate": 0.00029539081832023837, "loss": 1.9391, "step": 9450 }, { "epoch": 2.028568306010929, "grad_norm": 0.56640625, "learning_rate": 0.0002953255416238308, "loss": 1.9614, "step": 9500 }, { "epoch": 2.0291147540983605, "grad_norm": 0.490234375, "learning_rate": 0.0002952598132522874, "loss": 1.9405, "step": 9550 }, { "epoch": 2.0296612021857925, "grad_norm": 0.546875, "learning_rate": 0.00029519363340989367, "loss": 1.9653, "step": 9600 }, { "epoch": 2.030207650273224, "grad_norm": 0.50390625, "learning_rate": 0.0002951270023023379, "loss": 1.9704, "step": 9650 }, { "epoch": 2.0307540983606556, "grad_norm": 0.53515625, "learning_rate": 0.00029505992013671126, "loss": 1.9592, "step": 9700 }, { "epoch": 2.0313005464480876, "grad_norm": 0.474609375, "learning_rate": 0.0002949923871215065, "loss": 1.9679, "step": 9750 }, { "epoch": 2.031846994535519, "grad_norm": 0.5546875, "learning_rate": 0.000294924403466618, "loss": 1.9398, "step": 9800 }, { "epoch": 2.0323934426229506, "grad_norm": 0.5625, "learning_rate": 0.00029485596938334037, "loss": 1.9469, "step": 9850 }, { "epoch": 2.0329398907103826, "grad_norm": 0.51953125, "learning_rate": 0.00029478708508436834, "loss": 1.9742, "step": 9900 }, { "epoch": 2.033486338797814, "grad_norm": 0.482421875, "learning_rate": 0.000294717750783796, "loss": 1.9619, "step": 9950 }, { "epoch": 2.0340327868852457, "grad_norm": 0.6015625, "learning_rate": 0.0002946479666971158, "loss": 1.8817, "step": 10000 }, { "epoch": 2.0345792349726777, "grad_norm": 0.50390625, "learning_rate": 0.0002945777330412184, "loss": 1.9465, "step": 10050 }, { "epoch": 2.035125683060109, "grad_norm": 0.5703125, "learning_rate": 0.00029450705003439156, "loss": 1.9628, "step": 10100 }, { "epoch": 2.035672131147541, "grad_norm": 0.52734375, "learning_rate": 0.0002944359178963198, "loss": 1.9421, "step": 10150 }, { "epoch": 2.0362185792349727, "grad_norm": 0.51171875, "learning_rate": 0.00029436433684808336, "loss": 1.953, "step": 10200 }, { "epoch": 2.0367650273224043, "grad_norm": 0.5390625, "learning_rate": 0.0002942923071121578, "loss": 1.9499, "step": 10250 }, { "epoch": 2.0373114754098363, "grad_norm": 0.53515625, "learning_rate": 0.0002942198289124132, "loss": 1.9522, "step": 10300 }, { "epoch": 3.000229508196721, "grad_norm": 0.5078125, "learning_rate": 0.00029414690247411346, "loss": 1.9537, "step": 10350 }, { "epoch": 3.000775956284153, "grad_norm": 0.5390625, "learning_rate": 0.0002940735280239157, "loss": 1.9224, "step": 10400 }, { "epoch": 3.0013224043715847, "grad_norm": 0.5234375, "learning_rate": 0.0002939997057898693, "loss": 1.9165, "step": 10450 }, { "epoch": 3.0018688524590162, "grad_norm": 0.490234375, "learning_rate": 0.0002939254360014156, "loss": 1.8814, "step": 10500 }, { "epoch": 3.0024153005464482, "grad_norm": 0.52734375, "learning_rate": 0.0002938507188893867, "loss": 1.8547, "step": 10550 }, { "epoch": 3.0029617486338798, "grad_norm": 0.8125, "learning_rate": 0.00029377555468600516, "loss": 1.9014, "step": 10600 }, { "epoch": 3.0035081967213113, "grad_norm": 0.8125, "learning_rate": 0.00029369994362488306, "loss": 1.8837, "step": 10650 }, { "epoch": 3.0040546448087433, "grad_norm": 0.51953125, "learning_rate": 0.0002936238859410213, "loss": 1.9095, "step": 10700 }, { "epoch": 3.004601092896175, "grad_norm": 0.52734375, "learning_rate": 0.0002935473818708089, "loss": 1.8654, "step": 10750 }, { "epoch": 3.0051475409836064, "grad_norm": 0.58203125, "learning_rate": 0.00029347043165202233, "loss": 1.9018, "step": 10800 }, { "epoch": 3.0056939890710384, "grad_norm": 0.51171875, "learning_rate": 0.0002933930355238246, "loss": 1.895, "step": 10850 }, { "epoch": 3.00624043715847, "grad_norm": 0.50390625, "learning_rate": 0.0002933151937267647, "loss": 1.8872, "step": 10900 }, { "epoch": 3.0067868852459014, "grad_norm": 0.53125, "learning_rate": 0.0002932369065027767, "loss": 1.8532, "step": 10950 }, { "epoch": 3.0073333333333334, "grad_norm": 0.546875, "learning_rate": 0.0002931581740951791, "loss": 1.8935, "step": 11000 }, { "epoch": 3.007879781420765, "grad_norm": 0.5234375, "learning_rate": 0.00029307899674867405, "loss": 1.8991, "step": 11050 }, { "epoch": 3.008426229508197, "grad_norm": 0.53515625, "learning_rate": 0.00029299937470934656, "loss": 1.8784, "step": 11100 }, { "epoch": 3.0089726775956285, "grad_norm": 0.52734375, "learning_rate": 0.00029291930822466383, "loss": 1.8775, "step": 11150 }, { "epoch": 3.00951912568306, "grad_norm": 0.56640625, "learning_rate": 0.0002928387975434742, "loss": 1.8874, "step": 11200 }, { "epoch": 3.010065573770492, "grad_norm": 0.59765625, "learning_rate": 0.00029275784291600684, "loss": 1.9137, "step": 11250 }, { "epoch": 3.0106120218579235, "grad_norm": 0.53125, "learning_rate": 0.0002926764445938705, "loss": 1.8568, "step": 11300 }, { "epoch": 3.011158469945355, "grad_norm": 0.53125, "learning_rate": 0.0002925946028300532, "loss": 1.8578, "step": 11350 }, { "epoch": 3.011704918032787, "grad_norm": 0.53515625, "learning_rate": 0.0002925123178789209, "loss": 1.9092, "step": 11400 }, { "epoch": 3.0122513661202186, "grad_norm": 0.515625, "learning_rate": 0.00029242958999621717, "loss": 1.8663, "step": 11450 }, { "epoch": 3.01279781420765, "grad_norm": 0.54296875, "learning_rate": 0.00029234641943906223, "loss": 1.862, "step": 11500 }, { "epoch": 3.013344262295082, "grad_norm": 0.56640625, "learning_rate": 0.0002922628064659519, "loss": 1.8594, "step": 11550 }, { "epoch": 3.0138907103825137, "grad_norm": 0.546875, "learning_rate": 0.0002921787513367575, "loss": 1.8633, "step": 11600 }, { "epoch": 3.014437158469945, "grad_norm": 0.51953125, "learning_rate": 0.0002920942543127241, "loss": 1.8929, "step": 11650 }, { "epoch": 3.014983606557377, "grad_norm": 0.54296875, "learning_rate": 0.0002920093156564705, "loss": 1.8794, "step": 11700 }, { "epoch": 3.0155300546448087, "grad_norm": 0.5234375, "learning_rate": 0.0002919239356319879, "loss": 1.8691, "step": 11750 }, { "epoch": 3.0160765027322403, "grad_norm": 0.53515625, "learning_rate": 0.00029183811450463954, "loss": 1.8429, "step": 11800 }, { "epoch": 3.0166229508196722, "grad_norm": 0.5, "learning_rate": 0.00029175185254115934, "loss": 1.8325, "step": 11850 }, { "epoch": 3.017169398907104, "grad_norm": 0.53125, "learning_rate": 0.00029166515000965154, "loss": 1.8598, "step": 11900 }, { "epoch": 3.0177158469945353, "grad_norm": 0.55078125, "learning_rate": 0.0002915780071795896, "loss": 1.8376, "step": 11950 }, { "epoch": 3.0182622950819673, "grad_norm": 0.57421875, "learning_rate": 0.0002914904243218154, "loss": 1.8142, "step": 12000 }, { "epoch": 3.018808743169399, "grad_norm": 0.55078125, "learning_rate": 0.00029140240170853857, "loss": 1.8505, "step": 12050 }, { "epoch": 3.0193551912568304, "grad_norm": 0.5546875, "learning_rate": 0.0002913139396133353, "loss": 1.8315, "step": 12100 }, { "epoch": 3.0199016393442624, "grad_norm": 0.53515625, "learning_rate": 0.0002912250383111479, "loss": 1.8337, "step": 12150 }, { "epoch": 3.020448087431694, "grad_norm": 0.5234375, "learning_rate": 0.0002911356980782837, "loss": 1.8647, "step": 12200 }, { "epoch": 3.020994535519126, "grad_norm": 0.498046875, "learning_rate": 0.0002910459191924141, "loss": 1.8303, "step": 12250 }, { "epoch": 3.0215409836065574, "grad_norm": 0.515625, "learning_rate": 0.00029095570193257405, "loss": 1.8347, "step": 12300 }, { "epoch": 3.022087431693989, "grad_norm": 0.515625, "learning_rate": 0.0002908650465791608, "loss": 1.842, "step": 12350 }, { "epoch": 3.022633879781421, "grad_norm": 0.48046875, "learning_rate": 0.00029077395341393334, "loss": 1.8282, "step": 12400 }, { "epoch": 3.0231803278688525, "grad_norm": 0.546875, "learning_rate": 0.00029068242272001135, "loss": 1.7943, "step": 12450 }, { "epoch": 3.023726775956284, "grad_norm": 0.5078125, "learning_rate": 0.00029059045478187424, "loss": 1.8147, "step": 12500 }, { "epoch": 3.024273224043716, "grad_norm": 0.51953125, "learning_rate": 0.00029049804988536053, "loss": 1.8135, "step": 12550 }, { "epoch": 3.0248196721311476, "grad_norm": 0.52734375, "learning_rate": 0.00029040520831766676, "loss": 1.8067, "step": 12600 }, { "epoch": 3.025366120218579, "grad_norm": 0.515625, "learning_rate": 0.00029031193036734666, "loss": 1.8333, "step": 12650 }, { "epoch": 3.025912568306011, "grad_norm": 0.5546875, "learning_rate": 0.0002902182163243103, "loss": 1.8624, "step": 12700 }, { "epoch": 3.0264590163934426, "grad_norm": 0.57421875, "learning_rate": 0.00029012406647982306, "loss": 1.8277, "step": 12750 }, { "epoch": 3.027005464480874, "grad_norm": 0.51953125, "learning_rate": 0.0002900294811265048, "loss": 1.8209, "step": 12800 }, { "epoch": 3.027551912568306, "grad_norm": 0.5078125, "learning_rate": 0.0002899344605583291, "loss": 1.8295, "step": 12850 }, { "epoch": 3.0280983606557377, "grad_norm": 0.4921875, "learning_rate": 0.0002898390050706219, "loss": 1.7926, "step": 12900 }, { "epoch": 3.028644808743169, "grad_norm": 0.515625, "learning_rate": 0.0002897431149600612, "loss": 1.8064, "step": 12950 }, { "epoch": 3.029191256830601, "grad_norm": 0.51953125, "learning_rate": 0.0002896467905246755, "loss": 1.7923, "step": 13000 }, { "epoch": 3.0297377049180327, "grad_norm": 0.5625, "learning_rate": 0.00028955003206384357, "loss": 1.8346, "step": 13050 }, { "epoch": 3.0302841530054643, "grad_norm": 0.5078125, "learning_rate": 0.0002894528398782929, "loss": 1.8187, "step": 13100 }, { "epoch": 3.0308306010928963, "grad_norm": 0.55078125, "learning_rate": 0.0002893552142700989, "loss": 1.8035, "step": 13150 }, { "epoch": 3.031377049180328, "grad_norm": 0.55078125, "learning_rate": 0.0002892571555426843, "loss": 1.8248, "step": 13200 }, { "epoch": 3.0319234972677593, "grad_norm": 0.51953125, "learning_rate": 0.00028915866400081795, "loss": 1.8066, "step": 13250 }, { "epoch": 3.0324699453551913, "grad_norm": 0.50390625, "learning_rate": 0.00028905973995061373, "loss": 1.8087, "step": 13300 }, { "epoch": 3.033016393442623, "grad_norm": 0.53515625, "learning_rate": 0.00028896038369953, "loss": 1.8208, "step": 13350 }, { "epoch": 3.033562841530055, "grad_norm": 0.53125, "learning_rate": 0.00028886059555636816, "loss": 1.8282, "step": 13400 }, { "epoch": 3.0341092896174864, "grad_norm": 0.5546875, "learning_rate": 0.00028876037583127213, "loss": 1.7288, "step": 13450 }, { "epoch": 3.034655737704918, "grad_norm": 0.58203125, "learning_rate": 0.000288659724835727, "loss": 1.841, "step": 13500 }, { "epoch": 3.03520218579235, "grad_norm": 0.51171875, "learning_rate": 0.00028855864288255856, "loss": 1.8044, "step": 13550 }, { "epoch": 3.0357486338797814, "grad_norm": 0.52734375, "learning_rate": 0.00028845713028593183, "loss": 1.8101, "step": 13600 }, { "epoch": 3.036295081967213, "grad_norm": 0.53515625, "learning_rate": 0.00028835518736135013, "loss": 1.8193, "step": 13650 }, { "epoch": 3.036841530054645, "grad_norm": 0.5625, "learning_rate": 0.0002882528144256546, "loss": 1.8219, "step": 13700 }, { "epoch": 3.0373879781420765, "grad_norm": 0.53125, "learning_rate": 0.00028815001179702265, "loss": 1.8044, "step": 13750 }, { "epoch": 4.000306010928962, "grad_norm": 0.546875, "learning_rate": 0.0002880467797949671, "loss": 1.8068, "step": 13800 }, { "epoch": 4.000852459016394, "grad_norm": 0.57421875, "learning_rate": 0.00028794311874033563, "loss": 1.7919, "step": 13850 }, { "epoch": 4.001398907103825, "grad_norm": 0.53515625, "learning_rate": 0.00028783902895530893, "loss": 1.7501, "step": 13900 }, { "epoch": 4.001945355191257, "grad_norm": 0.6171875, "learning_rate": 0.00028773451076340064, "loss": 1.7494, "step": 13950 }, { "epoch": 4.002491803278689, "grad_norm": 0.546875, "learning_rate": 0.00028762956448945563, "loss": 1.6976, "step": 14000 }, { "epoch": 4.00303825136612, "grad_norm": 0.53125, "learning_rate": 0.00028752419045964935, "loss": 1.7673, "step": 14050 }, { "epoch": 4.003584699453552, "grad_norm": 0.6328125, "learning_rate": 0.0002874183890014867, "loss": 1.7385, "step": 14100 }, { "epoch": 4.004131147540984, "grad_norm": 0.5703125, "learning_rate": 0.0002873121604438011, "loss": 1.7526, "step": 14150 }, { "epoch": 4.004677595628415, "grad_norm": 0.53125, "learning_rate": 0.0002872055051167533, "loss": 1.7323, "step": 14200 }, { "epoch": 4.005224043715847, "grad_norm": 0.578125, "learning_rate": 0.0002870984233518306, "loss": 1.7666, "step": 14250 }, { "epoch": 4.005770491803279, "grad_norm": 0.5546875, "learning_rate": 0.0002869909154818455, "loss": 1.7382, "step": 14300 }, { "epoch": 4.00631693989071, "grad_norm": 0.55859375, "learning_rate": 0.00028688298184093497, "loss": 1.7366, "step": 14350 }, { "epoch": 4.006863387978142, "grad_norm": 0.5703125, "learning_rate": 0.0002867746227645593, "loss": 1.7206, "step": 14400 }, { "epoch": 4.007409836065574, "grad_norm": 0.5703125, "learning_rate": 0.000286665838589501, "loss": 1.7655, "step": 14450 }, { "epoch": 4.007956284153005, "grad_norm": 0.55859375, "learning_rate": 0.0002865566296538637, "loss": 1.7599, "step": 14500 }, { "epoch": 4.008502732240437, "grad_norm": 0.62890625, "learning_rate": 0.00028644699629707136, "loss": 1.7298, "step": 14550 }, { "epoch": 4.009049180327869, "grad_norm": 0.671875, "learning_rate": 0.00028633693885986696, "loss": 1.7392, "step": 14600 }, { "epoch": 4.0095956284153, "grad_norm": 0.5703125, "learning_rate": 0.0002862264576843116, "loss": 1.7556, "step": 14650 }, { "epoch": 4.010142076502732, "grad_norm": 0.57421875, "learning_rate": 0.0002861155531137833, "loss": 1.7677, "step": 14700 }, { "epoch": 4.010688524590164, "grad_norm": 0.62890625, "learning_rate": 0.00028600422549297604, "loss": 1.7283, "step": 14750 }, { "epoch": 4.011234972677595, "grad_norm": 0.51953125, "learning_rate": 0.00028589247516789856, "loss": 1.7389, "step": 14800 }, { "epoch": 4.011781420765027, "grad_norm": 0.5703125, "learning_rate": 0.0002857803024858735, "loss": 1.7769, "step": 14850 }, { "epoch": 4.012327868852459, "grad_norm": 0.5234375, "learning_rate": 0.00028566770779553613, "loss": 1.7306, "step": 14900 }, { "epoch": 4.01287431693989, "grad_norm": 0.54296875, "learning_rate": 0.00028555469144683337, "loss": 1.7341, "step": 14950 }, { "epoch": 4.013420765027322, "grad_norm": 0.53515625, "learning_rate": 0.00028544125379102264, "loss": 1.7364, "step": 15000 }, { "epoch": 4.013967213114754, "grad_norm": 0.52734375, "learning_rate": 0.0002853273951806708, "loss": 1.7203, "step": 15050 }, { "epoch": 4.0145136612021854, "grad_norm": 0.57421875, "learning_rate": 0.00028521311596965297, "loss": 1.7735, "step": 15100 }, { "epoch": 4.015060109289617, "grad_norm": 0.5546875, "learning_rate": 0.00028509841651315156, "loss": 1.7457, "step": 15150 }, { "epoch": 4.015606557377049, "grad_norm": 0.53125, "learning_rate": 0.0002849832971676553, "loss": 1.7317, "step": 15200 }, { "epoch": 4.0161530054644805, "grad_norm": 0.5625, "learning_rate": 0.0002848677582909576, "loss": 1.7168, "step": 15250 }, { "epoch": 4.0166994535519125, "grad_norm": 0.5625, "learning_rate": 0.000284751800242156, "loss": 1.714, "step": 15300 }, { "epoch": 4.0172459016393445, "grad_norm": 0.53125, "learning_rate": 0.0002846354233816508, "loss": 1.7342, "step": 15350 }, { "epoch": 4.017792349726776, "grad_norm": 0.58203125, "learning_rate": 0.00028451862807114396, "loss": 1.7201, "step": 15400 }, { "epoch": 4.0183387978142076, "grad_norm": 0.55078125, "learning_rate": 0.00028440141467363803, "loss": 1.6881, "step": 15450 }, { "epoch": 4.0188852459016395, "grad_norm": 0.54296875, "learning_rate": 0.00028428378355343495, "loss": 1.7184, "step": 15500 }, { "epoch": 4.019431693989071, "grad_norm": 0.54296875, "learning_rate": 0.00028416573507613485, "loss": 1.7055, "step": 15550 }, { "epoch": 4.019978142076503, "grad_norm": 0.58203125, "learning_rate": 0.0002840472696086353, "loss": 1.72, "step": 15600 }, { "epoch": 4.020524590163935, "grad_norm": 0.53125, "learning_rate": 0.0002839283875191295, "loss": 1.7308, "step": 15650 }, { "epoch": 4.021071038251366, "grad_norm": 0.5625, "learning_rate": 0.0002838090891771059, "loss": 1.702, "step": 15700 }, { "epoch": 4.021617486338798, "grad_norm": 0.59375, "learning_rate": 0.0002836893749533465, "loss": 1.7191, "step": 15750 }, { "epoch": 4.02216393442623, "grad_norm": 0.54296875, "learning_rate": 0.0002835692452199257, "loss": 1.7215, "step": 15800 }, { "epoch": 4.022710382513662, "grad_norm": 0.54296875, "learning_rate": 0.00028344870035020963, "loss": 1.6984, "step": 15850 }, { "epoch": 4.023256830601093, "grad_norm": 0.53125, "learning_rate": 0.0002833277407188545, "loss": 1.6847, "step": 15900 }, { "epoch": 4.023803278688525, "grad_norm": 0.53125, "learning_rate": 0.00028320636670180557, "loss": 1.6739, "step": 15950 }, { "epoch": 4.024349726775957, "grad_norm": 0.52734375, "learning_rate": 0.0002830845786762962, "loss": 1.7125, "step": 16000 }, { "epoch": 4.024896174863388, "grad_norm": 0.6796875, "learning_rate": 0.0002829623770208463, "loss": 1.6678, "step": 16050 }, { "epoch": 4.02544262295082, "grad_norm": 0.55859375, "learning_rate": 0.00028283976211526137, "loss": 1.7396, "step": 16100 }, { "epoch": 4.025989071038252, "grad_norm": 0.578125, "learning_rate": 0.0002827167343406315, "loss": 1.752, "step": 16150 }, { "epoch": 4.026535519125683, "grad_norm": 0.56640625, "learning_rate": 0.0002825932940793298, "loss": 1.6994, "step": 16200 }, { "epoch": 4.027081967213115, "grad_norm": 0.6015625, "learning_rate": 0.00028246944171501145, "loss": 1.7189, "step": 16250 }, { "epoch": 4.027628415300547, "grad_norm": 0.5625, "learning_rate": 0.00028234517763261243, "loss": 1.7158, "step": 16300 }, { "epoch": 4.028174863387978, "grad_norm": 0.5625, "learning_rate": 0.00028222050221834847, "loss": 1.6885, "step": 16350 }, { "epoch": 4.02872131147541, "grad_norm": 0.5078125, "learning_rate": 0.0002820954158597134, "loss": 1.6865, "step": 16400 }, { "epoch": 4.029267759562842, "grad_norm": 0.52734375, "learning_rate": 0.0002819699189454788, "loss": 1.6946, "step": 16450 }, { "epoch": 4.029814207650273, "grad_norm": 0.5703125, "learning_rate": 0.0002818440118656918, "loss": 1.7194, "step": 16500 }, { "epoch": 4.030360655737705, "grad_norm": 0.52734375, "learning_rate": 0.0002817176950116746, "loss": 1.7022, "step": 16550 }, { "epoch": 4.030907103825137, "grad_norm": 0.61328125, "learning_rate": 0.00028159096877602275, "loss": 1.7159, "step": 16600 }, { "epoch": 4.031453551912568, "grad_norm": 0.5625, "learning_rate": 0.00028146383355260446, "loss": 1.7063, "step": 16650 }, { "epoch": 4.032, "grad_norm": 0.58984375, "learning_rate": 0.00028133628973655894, "loss": 1.7064, "step": 16700 }, { "epoch": 4.032546448087432, "grad_norm": 0.55859375, "learning_rate": 0.00028120833772429517, "loss": 1.6992, "step": 16750 }, { "epoch": 4.033092896174863, "grad_norm": 0.5546875, "learning_rate": 0.0002810799779134911, "loss": 1.7132, "step": 16800 }, { "epoch": 4.033639344262295, "grad_norm": 0.5390625, "learning_rate": 0.0002809512107030919, "loss": 1.7125, "step": 16850 }, { "epoch": 4.034185792349727, "grad_norm": 0.57421875, "learning_rate": 0.0002808220364933091, "loss": 1.6373, "step": 16900 }, { "epoch": 4.034732240437158, "grad_norm": 0.5234375, "learning_rate": 0.00028069245568561904, "loss": 1.7379, "step": 16950 }, { "epoch": 4.03527868852459, "grad_norm": 0.6015625, "learning_rate": 0.00028056246868276186, "loss": 1.699, "step": 17000 }, { "epoch": 4.035825136612022, "grad_norm": 0.5546875, "learning_rate": 0.0002804320758887403, "loss": 1.6939, "step": 17050 }, { "epoch": 4.036371584699453, "grad_norm": 0.498046875, "learning_rate": 0.000280301277708818, "loss": 1.7146, "step": 17100 }, { "epoch": 4.036918032786885, "grad_norm": 0.5859375, "learning_rate": 0.00028017007454951884, "loss": 1.7363, "step": 17150 }, { "epoch": 4.037464480874317, "grad_norm": 0.55078125, "learning_rate": 0.00028003846681862524, "loss": 1.6965, "step": 17200 }, { "epoch": 5.000382513661203, "grad_norm": 0.5703125, "learning_rate": 0.00027990645492517697, "loss": 1.6887, "step": 17250 }, { "epoch": 5.000928961748634, "grad_norm": 0.62109375, "learning_rate": 0.0002797740392794702, "loss": 1.6829, "step": 17300 }, { "epoch": 5.001475409836066, "grad_norm": 0.58984375, "learning_rate": 0.0002796412202930557, "loss": 1.6235, "step": 17350 }, { "epoch": 5.002021857923498, "grad_norm": 0.59375, "learning_rate": 0.00027950799837873794, "loss": 1.634, "step": 17400 }, { "epoch": 5.002568306010929, "grad_norm": 0.55859375, "learning_rate": 0.0002793743739505738, "loss": 1.5863, "step": 17450 }, { "epoch": 5.003114754098361, "grad_norm": 0.5546875, "learning_rate": 0.0002792403474238709, "loss": 1.6513, "step": 17500 }, { "epoch": 5.003661202185793, "grad_norm": 0.59375, "learning_rate": 0.000279105919215187, "loss": 1.6388, "step": 17550 }, { "epoch": 5.004207650273224, "grad_norm": 0.59375, "learning_rate": 0.00027897108974232797, "loss": 1.6302, "step": 17600 }, { "epoch": 5.004754098360656, "grad_norm": 1.1875, "learning_rate": 0.0002788358594243469, "loss": 1.6336, "step": 17650 }, { "epoch": 5.005300546448088, "grad_norm": 0.58984375, "learning_rate": 0.0002787002286815428, "loss": 1.6513, "step": 17700 }, { "epoch": 5.005846994535519, "grad_norm": 0.6015625, "learning_rate": 0.000278564197935459, "loss": 1.6185, "step": 17750 }, { "epoch": 5.006393442622951, "grad_norm": 0.58984375, "learning_rate": 0.00027842776760888236, "loss": 1.6366, "step": 17800 }, { "epoch": 5.006939890710383, "grad_norm": 0.578125, "learning_rate": 0.00027829093812584143, "loss": 1.6162, "step": 17850 }, { "epoch": 5.007486338797814, "grad_norm": 0.56640625, "learning_rate": 0.0002781537099116054, "loss": 1.6665, "step": 17900 }, { "epoch": 5.008032786885246, "grad_norm": 0.55859375, "learning_rate": 0.00027801608339268275, "loss": 1.6505, "step": 17950 }, { "epoch": 5.008579234972678, "grad_norm": 0.578125, "learning_rate": 0.00027787805899681976, "loss": 1.6079, "step": 18000 }, { "epoch": 5.009125683060109, "grad_norm": 0.640625, "learning_rate": 0.00027773963715299957, "loss": 1.6429, "step": 18050 }, { "epoch": 5.009672131147541, "grad_norm": 0.640625, "learning_rate": 0.00027760081829144044, "loss": 1.6671, "step": 18100 }, { "epoch": 5.010218579234973, "grad_norm": 0.6484375, "learning_rate": 0.0002774616028435946, "loss": 1.6619, "step": 18150 }, { "epoch": 5.010765027322404, "grad_norm": 0.546875, "learning_rate": 0.00027732199124214676, "loss": 1.6221, "step": 18200 }, { "epoch": 5.011311475409836, "grad_norm": 0.63671875, "learning_rate": 0.0002771819839210131, "loss": 1.6311, "step": 18250 }, { "epoch": 5.011857923497268, "grad_norm": 0.54296875, "learning_rate": 0.0002770415813153396, "loss": 1.669, "step": 18300 }, { "epoch": 5.012404371584699, "grad_norm": 0.609375, "learning_rate": 0.00027690078386150084, "loss": 1.6177, "step": 18350 }, { "epoch": 5.012950819672131, "grad_norm": 0.59375, "learning_rate": 0.0002767595919970984, "loss": 1.622, "step": 18400 }, { "epoch": 5.013497267759563, "grad_norm": 0.609375, "learning_rate": 0.00027661800616096, "loss": 1.6329, "step": 18450 }, { "epoch": 5.014043715846994, "grad_norm": 0.58203125, "learning_rate": 0.00027647602679313764, "loss": 1.6369, "step": 18500 }, { "epoch": 5.014590163934426, "grad_norm": 0.55078125, "learning_rate": 0.0002763336543349065, "loss": 1.6624, "step": 18550 }, { "epoch": 5.015136612021858, "grad_norm": 0.5859375, "learning_rate": 0.0002761908892287633, "loss": 1.6422, "step": 18600 }, { "epoch": 5.015683060109289, "grad_norm": 0.5703125, "learning_rate": 0.0002760477319184255, "loss": 1.6101, "step": 18650 }, { "epoch": 5.016229508196721, "grad_norm": 0.57421875, "learning_rate": 0.0002759041828488292, "loss": 1.6355, "step": 18700 }, { "epoch": 5.016775956284153, "grad_norm": 0.58984375, "learning_rate": 0.0002757602424661283, "loss": 1.6114, "step": 18750 }, { "epoch": 5.017322404371584, "grad_norm": 0.59375, "learning_rate": 0.00027561591121769277, "loss": 1.6274, "step": 18800 }, { "epoch": 5.017868852459016, "grad_norm": 0.57421875, "learning_rate": 0.00027547118955210747, "loss": 1.6139, "step": 18850 }, { "epoch": 5.018415300546448, "grad_norm": 0.55859375, "learning_rate": 0.0002753260779191706, "loss": 1.6027, "step": 18900 }, { "epoch": 5.018961748633879, "grad_norm": 0.62109375, "learning_rate": 0.0002751805767698927, "loss": 1.6036, "step": 18950 }, { "epoch": 5.019508196721311, "grad_norm": 0.59375, "learning_rate": 0.0002750346865564944, "loss": 1.6073, "step": 19000 }, { "epoch": 5.020054644808743, "grad_norm": 0.54296875, "learning_rate": 0.0002748884077324061, "loss": 1.6285, "step": 19050 }, { "epoch": 5.020601092896175, "grad_norm": 0.59765625, "learning_rate": 0.0002747417407522656, "loss": 1.6254, "step": 19100 }, { "epoch": 5.021147540983606, "grad_norm": 0.5546875, "learning_rate": 0.00027459468607191723, "loss": 1.6058, "step": 19150 }, { "epoch": 5.021693989071038, "grad_norm": 0.60546875, "learning_rate": 0.00027444724414841046, "loss": 1.623, "step": 19200 }, { "epoch": 5.02224043715847, "grad_norm": 0.59765625, "learning_rate": 0.00027429941543999814, "loss": 1.623, "step": 19250 }, { "epoch": 5.0227868852459014, "grad_norm": 0.5390625, "learning_rate": 0.0002741512004061353, "loss": 1.601, "step": 19300 }, { "epoch": 5.023333333333333, "grad_norm": 0.56640625, "learning_rate": 0.0002740025995074777, "loss": 1.5935, "step": 19350 }, { "epoch": 5.023879781420765, "grad_norm": 0.578125, "learning_rate": 0.00027385361320588034, "loss": 1.5831, "step": 19400 }, { "epoch": 5.0244262295081965, "grad_norm": 0.59765625, "learning_rate": 0.0002737042419643961, "loss": 1.5974, "step": 19450 }, { "epoch": 5.0249726775956285, "grad_norm": 0.61328125, "learning_rate": 0.0002735544862472742, "loss": 1.5629, "step": 19500 }, { "epoch": 5.0255191256830605, "grad_norm": 0.640625, "learning_rate": 0.00027340434651995887, "loss": 1.6589, "step": 19550 }, { "epoch": 5.026065573770492, "grad_norm": 0.59375, "learning_rate": 0.0002732538232490879, "loss": 1.6576, "step": 19600 }, { "epoch": 5.0266120218579236, "grad_norm": 0.5859375, "learning_rate": 0.0002731029169024911, "loss": 1.6112, "step": 19650 }, { "epoch": 5.0271584699453555, "grad_norm": 0.609375, "learning_rate": 0.00027295162794918875, "loss": 1.6232, "step": 19700 }, { "epoch": 5.027704918032787, "grad_norm": 0.6328125, "learning_rate": 0.00027279995685939055, "loss": 1.6242, "step": 19750 }, { "epoch": 5.028251366120219, "grad_norm": 0.546875, "learning_rate": 0.00027264790410449363, "loss": 1.6029, "step": 19800 }, { "epoch": 5.028797814207651, "grad_norm": 0.6875, "learning_rate": 0.0002724954701570816, "loss": 1.5935, "step": 19850 }, { "epoch": 5.029344262295082, "grad_norm": 0.6640625, "learning_rate": 0.00027234265549092257, "loss": 1.6099, "step": 19900 }, { "epoch": 5.029890710382514, "grad_norm": 0.6015625, "learning_rate": 0.00027218946058096805, "loss": 1.6366, "step": 19950 }, { "epoch": 5.030437158469946, "grad_norm": 0.57421875, "learning_rate": 0.0002720358859033514, "loss": 1.6249, "step": 20000 }, { "epoch": 5.030983606557377, "grad_norm": 0.59375, "learning_rate": 0.00027188193193538625, "loss": 1.6114, "step": 20050 }, { "epoch": 5.031530054644809, "grad_norm": 0.671875, "learning_rate": 0.00027172759915556504, "loss": 1.6081, "step": 20100 }, { "epoch": 5.032076502732241, "grad_norm": 0.62890625, "learning_rate": 0.0002715728880435577, "loss": 1.627, "step": 20150 }, { "epoch": 5.032622950819672, "grad_norm": 0.62890625, "learning_rate": 0.00027141779908020986, "loss": 1.5912, "step": 20200 }, { "epoch": 5.033169398907104, "grad_norm": 0.65234375, "learning_rate": 0.00027126233274754163, "loss": 1.6476, "step": 20250 }, { "epoch": 5.033715846994536, "grad_norm": 0.59375, "learning_rate": 0.00027110648952874595, "loss": 1.6051, "step": 20300 }, { "epoch": 5.034262295081967, "grad_norm": 0.65234375, "learning_rate": 0.0002709502699081871, "loss": 1.5705, "step": 20350 }, { "epoch": 5.034808743169399, "grad_norm": 0.578125, "learning_rate": 0.00027079367437139935, "loss": 1.6533, "step": 20400 }, { "epoch": 5.035355191256831, "grad_norm": 0.59765625, "learning_rate": 0.00027063670340508514, "loss": 1.6099, "step": 20450 }, { "epoch": 5.035901639344262, "grad_norm": 0.5703125, "learning_rate": 0.00027047935749711395, "loss": 1.6018, "step": 20500 }, { "epoch": 5.036448087431694, "grad_norm": 0.59375, "learning_rate": 0.0002703216371365204, "loss": 1.637, "step": 20550 }, { "epoch": 5.036994535519126, "grad_norm": 0.56640625, "learning_rate": 0.00027016354281350315, "loss": 1.6394, "step": 20600 }, { "epoch": 5.037540983606557, "grad_norm": 0.6015625, "learning_rate": 0.00027000507501942283, "loss": 1.6253, "step": 20650 }, { "epoch": 6.000459016393442, "grad_norm": 0.57421875, "learning_rate": 0.0002698462342468011, "loss": 1.5897, "step": 20700 }, { "epoch": 6.001005464480874, "grad_norm": 0.56640625, "learning_rate": 0.0002696870209893187, "loss": 1.569, "step": 20750 }, { "epoch": 6.001551912568306, "grad_norm": 0.59765625, "learning_rate": 0.00026952743574181414, "loss": 1.5482, "step": 20800 }, { "epoch": 6.002098360655737, "grad_norm": 0.59765625, "learning_rate": 0.00026936747900028205, "loss": 1.5458, "step": 20850 }, { "epoch": 6.002644808743169, "grad_norm": 0.734375, "learning_rate": 0.00026920715126187167, "loss": 1.4787, "step": 20900 }, { "epoch": 6.003191256830601, "grad_norm": 0.7109375, "learning_rate": 0.0002690464530248853, "loss": 1.5565, "step": 20950 }, { "epoch": 6.0037377049180325, "grad_norm": 0.64453125, "learning_rate": 0.00026888538478877675, "loss": 1.5588, "step": 21000 }, { "epoch": 6.0042841530054645, "grad_norm": 0.6953125, "learning_rate": 0.0002687239470541498, "loss": 1.5347, "step": 21050 }, { "epoch": 6.0048306010928965, "grad_norm": 0.58984375, "learning_rate": 0.00026856214032275675, "loss": 1.5341, "step": 21100 }, { "epoch": 6.0053770491803276, "grad_norm": 0.73046875, "learning_rate": 0.00026839996509749655, "loss": 1.5441, "step": 21150 }, { "epoch": 6.0059234972677595, "grad_norm": 0.58984375, "learning_rate": 0.00026823742188241366, "loss": 1.5405, "step": 21200 }, { "epoch": 6.0064699453551915, "grad_norm": 0.64453125, "learning_rate": 0.000268074511182696, "loss": 1.5327, "step": 21250 }, { "epoch": 6.007016393442623, "grad_norm": 0.61328125, "learning_rate": 0.00026791123350467384, "loss": 1.5338, "step": 21300 }, { "epoch": 6.007562841530055, "grad_norm": 0.58984375, "learning_rate": 0.000267747589355818, "loss": 1.5663, "step": 21350 }, { "epoch": 6.008109289617487, "grad_norm": 0.67578125, "learning_rate": 0.0002675835792447382, "loss": 1.5519, "step": 21400 }, { "epoch": 6.008655737704918, "grad_norm": 0.66015625, "learning_rate": 0.0002674192036811818, "loss": 1.5256, "step": 21450 }, { "epoch": 6.00920218579235, "grad_norm": 0.578125, "learning_rate": 0.0002672544631760317, "loss": 1.5428, "step": 21500 }, { "epoch": 6.009748633879782, "grad_norm": 0.75, "learning_rate": 0.00026708935824130514, "loss": 1.5806, "step": 21550 }, { "epoch": 6.010295081967213, "grad_norm": 0.57421875, "learning_rate": 0.00026692388939015226, "loss": 1.5628, "step": 21600 }, { "epoch": 6.010841530054645, "grad_norm": 0.61328125, "learning_rate": 0.00026675805713685387, "loss": 1.5275, "step": 21650 }, { "epoch": 6.011387978142077, "grad_norm": 0.61328125, "learning_rate": 0.0002665918619968206, "loss": 1.5547, "step": 21700 }, { "epoch": 6.011934426229508, "grad_norm": 0.57421875, "learning_rate": 0.0002664253044865907, "loss": 1.575, "step": 21750 }, { "epoch": 6.01248087431694, "grad_norm": 0.59765625, "learning_rate": 0.0002662583851238287, "loss": 1.5386, "step": 21800 }, { "epoch": 6.013027322404372, "grad_norm": 0.6484375, "learning_rate": 0.000266091104427324, "loss": 1.5107, "step": 21850 }, { "epoch": 6.013573770491803, "grad_norm": 0.625, "learning_rate": 0.00026592346291698864, "loss": 1.5516, "step": 21900 }, { "epoch": 6.014120218579235, "grad_norm": 0.5703125, "learning_rate": 0.00026575546111385647, "loss": 1.5431, "step": 21950 }, { "epoch": 6.014666666666667, "grad_norm": 0.6015625, "learning_rate": 0.00026558709954008095, "loss": 1.566, "step": 22000 }, { "epoch": 6.015213114754098, "grad_norm": 0.62890625, "learning_rate": 0.00026541837871893367, "loss": 1.5595, "step": 22050 }, { "epoch": 6.01575956284153, "grad_norm": 0.5625, "learning_rate": 0.0002652492991748029, "loss": 1.5206, "step": 22100 }, { "epoch": 6.016306010928962, "grad_norm": 0.640625, "learning_rate": 0.00026507986143319164, "loss": 1.5374, "step": 22150 }, { "epoch": 6.016852459016394, "grad_norm": 0.703125, "learning_rate": 0.0002649100660207164, "loss": 1.522, "step": 22200 }, { "epoch": 6.017398907103825, "grad_norm": 0.65625, "learning_rate": 0.0002647399134651053, "loss": 1.5532, "step": 22250 }, { "epoch": 6.017945355191257, "grad_norm": 0.63671875, "learning_rate": 0.0002645694042951963, "loss": 1.5274, "step": 22300 }, { "epoch": 6.018491803278689, "grad_norm": 0.65625, "learning_rate": 0.00026439853904093586, "loss": 1.517, "step": 22350 }, { "epoch": 6.01903825136612, "grad_norm": 0.6875, "learning_rate": 0.00026422731823337717, "loss": 1.5197, "step": 22400 }, { "epoch": 6.019584699453552, "grad_norm": 0.59765625, "learning_rate": 0.0002640557424046784, "loss": 1.5081, "step": 22450 }, { "epoch": 6.020131147540984, "grad_norm": 0.65234375, "learning_rate": 0.0002638838120881012, "loss": 1.5526, "step": 22500 }, { "epoch": 6.020677595628415, "grad_norm": 0.62109375, "learning_rate": 0.000263711527818009, "loss": 1.534, "step": 22550 }, { "epoch": 6.021224043715847, "grad_norm": 0.60546875, "learning_rate": 0.0002635388901298652, "loss": 1.5156, "step": 22600 }, { "epoch": 6.021770491803279, "grad_norm": 0.71875, "learning_rate": 0.0002633658995602318, "loss": 1.5402, "step": 22650 }, { "epoch": 6.02231693989071, "grad_norm": 0.68359375, "learning_rate": 0.0002631925566467674, "loss": 1.5367, "step": 22700 }, { "epoch": 6.022863387978142, "grad_norm": 0.62890625, "learning_rate": 0.00026301886192822585, "loss": 1.5126, "step": 22750 }, { "epoch": 6.023409836065574, "grad_norm": 0.625, "learning_rate": 0.00026284481594445434, "loss": 1.5097, "step": 22800 }, { "epoch": 6.023956284153005, "grad_norm": 0.60546875, "learning_rate": 0.00026267041923639175, "loss": 1.5058, "step": 22850 }, { "epoch": 6.024502732240437, "grad_norm": 0.703125, "learning_rate": 0.00026249567234606707, "loss": 1.5004, "step": 22900 }, { "epoch": 6.025049180327869, "grad_norm": 0.60546875, "learning_rate": 0.00026232057581659777, "loss": 1.4884, "step": 22950 }, { "epoch": 6.0255956284153, "grad_norm": 0.65234375, "learning_rate": 0.0002621451301921878, "loss": 1.5884, "step": 23000 }, { "epoch": 6.026142076502732, "grad_norm": 0.6484375, "learning_rate": 0.00026196933601812616, "loss": 1.565, "step": 23050 }, { "epoch": 6.026688524590164, "grad_norm": 0.69140625, "learning_rate": 0.00026179319384078535, "loss": 1.5399, "step": 23100 }, { "epoch": 6.027234972677595, "grad_norm": 0.57421875, "learning_rate": 0.0002616167042076192, "loss": 1.5319, "step": 23150 }, { "epoch": 6.027781420765027, "grad_norm": 0.62890625, "learning_rate": 0.0002614398676671616, "loss": 1.5379, "step": 23200 }, { "epoch": 6.028327868852459, "grad_norm": 0.69921875, "learning_rate": 0.0002612626847690247, "loss": 1.5344, "step": 23250 }, { "epoch": 6.02887431693989, "grad_norm": 0.59765625, "learning_rate": 0.0002610851560638968, "loss": 1.5054, "step": 23300 }, { "epoch": 6.029420765027322, "grad_norm": 0.6484375, "learning_rate": 0.0002609072821035415, "loss": 1.5421, "step": 23350 }, { "epoch": 6.029967213114754, "grad_norm": 0.69921875, "learning_rate": 0.00026072906344079484, "loss": 1.5625, "step": 23400 }, { "epoch": 6.0305136612021855, "grad_norm": 0.59765625, "learning_rate": 0.0002605505006295648, "loss": 1.5476, "step": 23450 }, { "epoch": 6.031060109289617, "grad_norm": 0.609375, "learning_rate": 0.00026037159422482865, "loss": 1.537, "step": 23500 }, { "epoch": 6.031606557377049, "grad_norm": 0.60546875, "learning_rate": 0.00026019234478263155, "loss": 1.5204, "step": 23550 }, { "epoch": 6.0321530054644805, "grad_norm": 0.69921875, "learning_rate": 0.000260012752860085, "loss": 1.5479, "step": 23600 }, { "epoch": 6.0326994535519125, "grad_norm": 0.6171875, "learning_rate": 0.00025983281901536474, "loss": 1.5304, "step": 23650 }, { "epoch": 6.0332459016393445, "grad_norm": 0.6171875, "learning_rate": 0.00025965254380770945, "loss": 1.5738, "step": 23700 }, { "epoch": 6.033792349726776, "grad_norm": 0.62109375, "learning_rate": 0.0002594719277974185, "loss": 1.5168, "step": 23750 }, { "epoch": 6.034338797814208, "grad_norm": 0.6328125, "learning_rate": 0.0002592909715458506, "loss": 1.4984, "step": 23800 }, { "epoch": 6.0348852459016395, "grad_norm": 0.57421875, "learning_rate": 0.0002591096756154221, "loss": 1.5721, "step": 23850 }, { "epoch": 6.035431693989071, "grad_norm": 0.6015625, "learning_rate": 0.0002589280405696048, "loss": 1.5369, "step": 23900 }, { "epoch": 6.035978142076503, "grad_norm": 0.61328125, "learning_rate": 0.00025874606697292473, "loss": 1.5236, "step": 23950 }, { "epoch": 6.036524590163935, "grad_norm": 0.6953125, "learning_rate": 0.00025856375539095986, "loss": 1.577, "step": 24000 } ], "logging_steps": 50, "max_steps": 91500, "num_input_tokens_seen": 0, "num_train_epochs": 9223372036854775807, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.2834709186604433e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }