| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 6.036524590163935, |
| "eval_steps": 500, |
| "global_step": 24000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.000546448087431694, |
| "grad_norm": 5.3125, |
| "learning_rate": 6.124999999999999e-06, |
| "loss": 9.6445, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.001092896174863388, |
| "grad_norm": 5.40625, |
| "learning_rate": 1.2375e-05, |
| "loss": 8.1001, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.001639344262295082, |
| "grad_norm": 4.0, |
| "learning_rate": 1.8625e-05, |
| "loss": 7.5007, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.002185792349726776, |
| "grad_norm": 6.125, |
| "learning_rate": 2.4874999999999998e-05, |
| "loss": 7.0884, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.00273224043715847, |
| "grad_norm": 3.703125, |
| "learning_rate": 3.1125e-05, |
| "loss": 6.6637, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.003278688524590164, |
| "grad_norm": 3.25, |
| "learning_rate": 3.7375e-05, |
| "loss": 6.3275, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.003825136612021858, |
| "grad_norm": 2.578125, |
| "learning_rate": 4.3624999999999997e-05, |
| "loss": 6.0834, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.004371584699453552, |
| "grad_norm": 2.640625, |
| "learning_rate": 4.9875e-05, |
| "loss": 5.8539, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.004918032786885246, |
| "grad_norm": 2.578125, |
| "learning_rate": 5.6124999999999995e-05, |
| "loss": 5.7353, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.00546448087431694, |
| "grad_norm": 2.75, |
| "learning_rate": 6.2375e-05, |
| "loss": 5.6411, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.006010928961748634, |
| "grad_norm": 1.8359375, |
| "learning_rate": 6.8625e-05, |
| "loss": 5.488, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.006557377049180328, |
| "grad_norm": 2.609375, |
| "learning_rate": 7.487499999999999e-05, |
| "loss": 5.3441, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.007103825136612022, |
| "grad_norm": 1.765625, |
| "learning_rate": 8.112500000000001e-05, |
| "loss": 5.2922, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.007650273224043716, |
| "grad_norm": 1.7265625, |
| "learning_rate": 8.7375e-05, |
| "loss": 5.1776, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.00819672131147541, |
| "grad_norm": 2.28125, |
| "learning_rate": 9.362499999999999e-05, |
| "loss": 5.0849, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.008743169398907104, |
| "grad_norm": 1.171875, |
| "learning_rate": 9.9875e-05, |
| "loss": 4.9607, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.009289617486338797, |
| "grad_norm": 1.5234375, |
| "learning_rate": 0.00010612499999999999, |
| "loss": 4.8847, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.009836065573770493, |
| "grad_norm": 1.15625, |
| "learning_rate": 0.000112375, |
| "loss": 4.8233, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.010382513661202186, |
| "grad_norm": 1.3359375, |
| "learning_rate": 0.000118625, |
| "loss": 4.7077, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.01092896174863388, |
| "grad_norm": 0.9921875, |
| "learning_rate": 0.00012487499999999999, |
| "loss": 4.6274, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.011475409836065573, |
| "grad_norm": 1.15625, |
| "learning_rate": 0.00013112499999999998, |
| "loss": 4.5506, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.012021857923497269, |
| "grad_norm": 1.03125, |
| "learning_rate": 0.000137375, |
| "loss": 4.4801, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.012568306010928962, |
| "grad_norm": 1.3046875, |
| "learning_rate": 0.00014362499999999998, |
| "loss": 4.425, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.013114754098360656, |
| "grad_norm": 1.046875, |
| "learning_rate": 0.000149875, |
| "loss": 4.3204, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.01366120218579235, |
| "grad_norm": 7.5, |
| "learning_rate": 0.000156125, |
| "loss": 4.2244, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.014207650273224045, |
| "grad_norm": 0.99609375, |
| "learning_rate": 0.00016237499999999998, |
| "loss": 4.1705, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.014754098360655738, |
| "grad_norm": 0.83984375, |
| "learning_rate": 0.000168625, |
| "loss": 4.1256, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.015300546448087432, |
| "grad_norm": 0.93359375, |
| "learning_rate": 0.00017487499999999998, |
| "loss": 4.0456, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.015846994535519125, |
| "grad_norm": 0.90234375, |
| "learning_rate": 0.000181125, |
| "loss": 3.9429, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.01639344262295082, |
| "grad_norm": 0.8671875, |
| "learning_rate": 0.000187375, |
| "loss": 3.8733, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.016939890710382512, |
| "grad_norm": 0.9140625, |
| "learning_rate": 0.00019362499999999998, |
| "loss": 3.8094, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.017486338797814208, |
| "grad_norm": 0.890625, |
| "learning_rate": 0.000199875, |
| "loss": 3.8117, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.018032786885245903, |
| "grad_norm": 1.0390625, |
| "learning_rate": 0.00020612499999999998, |
| "loss": 3.6919, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.018579234972677595, |
| "grad_norm": 0.94921875, |
| "learning_rate": 0.00021237499999999997, |
| "loss": 3.6148, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.01912568306010929, |
| "grad_norm": 1.0546875, |
| "learning_rate": 0.000218625, |
| "loss": 3.5881, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.019672131147540985, |
| "grad_norm": 0.953125, |
| "learning_rate": 0.000224875, |
| "loss": 3.46, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.020218579234972677, |
| "grad_norm": 0.9765625, |
| "learning_rate": 0.00023112499999999997, |
| "loss": 3.4731, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.020765027322404372, |
| "grad_norm": 0.984375, |
| "learning_rate": 0.00023737499999999998, |
| "loss": 3.3781, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.021311475409836064, |
| "grad_norm": 1.0078125, |
| "learning_rate": 0.000243625, |
| "loss": 3.3166, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.02185792349726776, |
| "grad_norm": 1.0859375, |
| "learning_rate": 0.000249875, |
| "loss": 3.3105, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.022404371584699455, |
| "grad_norm": 0.86328125, |
| "learning_rate": 0.000256125, |
| "loss": 3.25, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.022950819672131147, |
| "grad_norm": 0.9375, |
| "learning_rate": 0.00026237499999999997, |
| "loss": 3.1414, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.023497267759562842, |
| "grad_norm": 0.86328125, |
| "learning_rate": 0.000268625, |
| "loss": 3.1565, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.024043715846994537, |
| "grad_norm": 0.80859375, |
| "learning_rate": 0.000274875, |
| "loss": 3.1131, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.02459016393442623, |
| "grad_norm": 0.91015625, |
| "learning_rate": 0.00028112499999999996, |
| "loss": 3.0784, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.025136612021857924, |
| "grad_norm": 0.80859375, |
| "learning_rate": 0.000287375, |
| "loss": 3.0332, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.025683060109289616, |
| "grad_norm": 0.85546875, |
| "learning_rate": 0.000293625, |
| "loss": 3.0955, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.02622950819672131, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.000299875, |
| "loss": 3.045, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.026775956284153007, |
| "grad_norm": 0.8359375, |
| "learning_rate": 0.0002999997761290961, |
| "loss": 2.995, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.0273224043715847, |
| "grad_norm": 0.75390625, |
| "learning_rate": 0.0002999990861486685, |
| "loss": 2.9428, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.027868852459016394, |
| "grad_norm": 0.71484375, |
| "learning_rate": 0.00029999792996762107, |
| "loss": 2.9131, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.02841530054644809, |
| "grad_norm": 0.88671875, |
| "learning_rate": 0.00029999630758954706, |
| "loss": 2.896, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.02896174863387978, |
| "grad_norm": 0.73828125, |
| "learning_rate": 0.000299994219019489, |
| "loss": 2.8605, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.029508196721311476, |
| "grad_norm": 0.66796875, |
| "learning_rate": 0.0002999916642639382, |
| "loss": 2.8407, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.030054644808743168, |
| "grad_norm": 0.68359375, |
| "learning_rate": 0.0002999886433308348, |
| "loss": 2.8313, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.030601092896174863, |
| "grad_norm": 0.69140625, |
| "learning_rate": 0.00029998515622956803, |
| "loss": 2.8194, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.03114754098360656, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.00029998120297097586, |
| "loss": 2.7874, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.03169398907103825, |
| "grad_norm": 0.67578125, |
| "learning_rate": 0.00029997678356734504, |
| "loss": 2.7631, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.03224043715846994, |
| "grad_norm": 0.6875, |
| "learning_rate": 0.0002999718980324113, |
| "loss": 2.7603, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.03278688524590164, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.0002999665463813589, |
| "loss": 2.7229, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.03333333333333333, |
| "grad_norm": 0.671875, |
| "learning_rate": 0.00029996072863082093, |
| "loss": 2.7895, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.033879781420765025, |
| "grad_norm": 0.91796875, |
| "learning_rate": 0.0002999544447988791, |
| "loss": 2.6505, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.03442622950819672, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.0002999476949050637, |
| "loss": 2.6744, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.034972677595628415, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.0002999404789703535, |
| "loss": 2.6869, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.03551912568306011, |
| "grad_norm": 0.76953125, |
| "learning_rate": 0.0002999327970171759, |
| "loss": 2.6726, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.036065573770491806, |
| "grad_norm": 0.66015625, |
| "learning_rate": 0.0002999246490694065, |
| "loss": 2.6444, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.0366120218579235, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.0002999160351523693, |
| "loss": 2.6568, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.03715846994535519, |
| "grad_norm": 0.625, |
| "learning_rate": 0.00029990695529283665, |
| "loss": 2.6436, |
| "step": 3400 |
| }, |
| { |
| "epoch": 1.0000765027322405, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.00029989740951902885, |
| "loss": 2.6468, |
| "step": 3450 |
| }, |
| { |
| "epoch": 1.0006229508196722, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.0002998873978606145, |
| "loss": 2.5703, |
| "step": 3500 |
| }, |
| { |
| "epoch": 1.0011693989071038, |
| "grad_norm": 0.6171875, |
| "learning_rate": 0.0002998769203487099, |
| "loss": 2.6321, |
| "step": 3550 |
| }, |
| { |
| "epoch": 1.0017158469945355, |
| "grad_norm": 0.62109375, |
| "learning_rate": 0.0002998659770158796, |
| "loss": 2.5518, |
| "step": 3600 |
| }, |
| { |
| "epoch": 1.0022622950819673, |
| "grad_norm": 0.6171875, |
| "learning_rate": 0.0002998545678961356, |
| "loss": 2.5255, |
| "step": 3650 |
| }, |
| { |
| "epoch": 1.0028087431693988, |
| "grad_norm": 0.66015625, |
| "learning_rate": 0.00029984269302493776, |
| "loss": 2.4976, |
| "step": 3700 |
| }, |
| { |
| "epoch": 1.0033551912568306, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.0002998303524391934, |
| "loss": 2.532, |
| "step": 3750 |
| }, |
| { |
| "epoch": 1.0039016393442624, |
| "grad_norm": 0.6484375, |
| "learning_rate": 0.00029981754617725747, |
| "loss": 2.5321, |
| "step": 3800 |
| }, |
| { |
| "epoch": 1.004448087431694, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.0002998042742789319, |
| "loss": 2.4924, |
| "step": 3850 |
| }, |
| { |
| "epoch": 1.0049945355191257, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.0002997905367854663, |
| "loss": 2.492, |
| "step": 3900 |
| }, |
| { |
| "epoch": 1.0055409836065574, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.00029977633373955696, |
| "loss": 2.5266, |
| "step": 3950 |
| }, |
| { |
| "epoch": 1.0060874316939892, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.00029976166518534735, |
| "loss": 2.4739, |
| "step": 4000 |
| }, |
| { |
| "epoch": 1.0066338797814207, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.00029974653116842764, |
| "loss": 2.4487, |
| "step": 4050 |
| }, |
| { |
| "epoch": 1.0071803278688525, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.0002997309317358347, |
| "loss": 2.4674, |
| "step": 4100 |
| }, |
| { |
| "epoch": 1.0077267759562842, |
| "grad_norm": 0.67578125, |
| "learning_rate": 0.0002997148669360519, |
| "loss": 2.4814, |
| "step": 4150 |
| }, |
| { |
| "epoch": 1.0082732240437158, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029969833681900914, |
| "loss": 2.448, |
| "step": 4200 |
| }, |
| { |
| "epoch": 1.0088196721311475, |
| "grad_norm": 0.68359375, |
| "learning_rate": 0.0002996813414360822, |
| "loss": 2.4299, |
| "step": 4250 |
| }, |
| { |
| "epoch": 1.0093661202185793, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.00029966388084009334, |
| "loss": 2.4271, |
| "step": 4300 |
| }, |
| { |
| "epoch": 1.0099125683060108, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.00029964595508531034, |
| "loss": 2.4848, |
| "step": 4350 |
| }, |
| { |
| "epoch": 1.0104590163934426, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00029962756422744695, |
| "loss": 2.414, |
| "step": 4400 |
| }, |
| { |
| "epoch": 1.0110054644808744, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.00029960870832366224, |
| "loss": 2.3993, |
| "step": 4450 |
| }, |
| { |
| "epoch": 1.0115519125683061, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.000299589387432561, |
| "loss": 2.4171, |
| "step": 4500 |
| }, |
| { |
| "epoch": 1.0120983606557377, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.00029956960161419283, |
| "loss": 2.4038, |
| "step": 4550 |
| }, |
| { |
| "epoch": 1.0126448087431694, |
| "grad_norm": 0.498046875, |
| "learning_rate": 0.0002995493509300526, |
| "loss": 2.4128, |
| "step": 4600 |
| }, |
| { |
| "epoch": 1.0131912568306012, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.0002995286354430799, |
| "loss": 2.3721, |
| "step": 4650 |
| }, |
| { |
| "epoch": 1.0137377049180327, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.0002995074552176589, |
| "loss": 2.3734, |
| "step": 4700 |
| }, |
| { |
| "epoch": 1.0142841530054645, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.00029948581031961826, |
| "loss": 2.3805, |
| "step": 4750 |
| }, |
| { |
| "epoch": 1.0148306010928962, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002994637008162308, |
| "loss": 2.3819, |
| "step": 4800 |
| }, |
| { |
| "epoch": 1.0153770491803278, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029944112677621345, |
| "loss": 2.3839, |
| "step": 4850 |
| }, |
| { |
| "epoch": 1.0159234972677595, |
| "grad_norm": 0.4921875, |
| "learning_rate": 0.00029941808826972673, |
| "loss": 2.336, |
| "step": 4900 |
| }, |
| { |
| "epoch": 1.0164699453551913, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002993945853683749, |
| "loss": 2.3126, |
| "step": 4950 |
| }, |
| { |
| "epoch": 1.0170163934426228, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029937061814520546, |
| "loss": 2.3271, |
| "step": 5000 |
| }, |
| { |
| "epoch": 1.0175628415300546, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00029934618667470925, |
| "loss": 2.3275, |
| "step": 5050 |
| }, |
| { |
| "epoch": 1.0181092896174864, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002993212910328197, |
| "loss": 2.2837, |
| "step": 5100 |
| }, |
| { |
| "epoch": 1.0186557377049181, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.00029929593129691305, |
| "loss": 2.2964, |
| "step": 5150 |
| }, |
| { |
| "epoch": 1.0192021857923497, |
| "grad_norm": 0.50390625, |
| "learning_rate": 0.000299270107545808, |
| "loss": 2.3155, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.0197486338797814, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00029924381985976534, |
| "loss": 2.2722, |
| "step": 5250 |
| }, |
| { |
| "epoch": 1.0202950819672132, |
| "grad_norm": 0.49609375, |
| "learning_rate": 0.00029921706832048784, |
| "loss": 2.3175, |
| "step": 5300 |
| }, |
| { |
| "epoch": 1.0208415300546447, |
| "grad_norm": 0.48828125, |
| "learning_rate": 0.00029918985301111985, |
| "loss": 2.2834, |
| "step": 5350 |
| }, |
| { |
| "epoch": 1.0213879781420765, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.00029916217401624716, |
| "loss": 2.2522, |
| "step": 5400 |
| }, |
| { |
| "epoch": 1.0219344262295083, |
| "grad_norm": 0.46484375, |
| "learning_rate": 0.00029913403142189677, |
| "loss": 2.2872, |
| "step": 5450 |
| }, |
| { |
| "epoch": 1.0224808743169398, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00029910542531553656, |
| "loss": 2.2793, |
| "step": 5500 |
| }, |
| { |
| "epoch": 1.0230273224043716, |
| "grad_norm": 0.474609375, |
| "learning_rate": 0.00029907635578607487, |
| "loss": 2.218, |
| "step": 5550 |
| }, |
| { |
| "epoch": 1.0235737704918033, |
| "grad_norm": 0.49609375, |
| "learning_rate": 0.00029904682292386053, |
| "loss": 2.2309, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.024120218579235, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002990168268206823, |
| "loss": 2.2285, |
| "step": 5650 |
| }, |
| { |
| "epoch": 1.0246666666666666, |
| "grad_norm": 0.48828125, |
| "learning_rate": 0.00029898636756976884, |
| "loss": 2.2338, |
| "step": 5700 |
| }, |
| { |
| "epoch": 1.0252131147540984, |
| "grad_norm": 0.462890625, |
| "learning_rate": 0.0002989554452657881, |
| "loss": 2.2048, |
| "step": 5750 |
| }, |
| { |
| "epoch": 1.0257595628415301, |
| "grad_norm": 0.62109375, |
| "learning_rate": 0.0002989240600048475, |
| "loss": 2.2716, |
| "step": 5800 |
| }, |
| { |
| "epoch": 1.0263060109289617, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.00029889221188449295, |
| "loss": 2.2618, |
| "step": 5850 |
| }, |
| { |
| "epoch": 1.0268524590163934, |
| "grad_norm": 0.47265625, |
| "learning_rate": 0.0002988599010037092, |
| "loss": 2.2181, |
| "step": 5900 |
| }, |
| { |
| "epoch": 1.0273989071038252, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.0002988271274629192, |
| "loss": 2.2005, |
| "step": 5950 |
| }, |
| { |
| "epoch": 1.0279453551912567, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.00029879389136398403, |
| "loss": 2.1958, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.0284918032786885, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.00029876019281020207, |
| "loss": 2.1853, |
| "step": 6050 |
| }, |
| { |
| "epoch": 1.0290382513661203, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.00029872603190630927, |
| "loss": 2.1753, |
| "step": 6100 |
| }, |
| { |
| "epoch": 1.029584699453552, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.00029869140875847847, |
| "loss": 2.1931, |
| "step": 6150 |
| }, |
| { |
| "epoch": 1.0301311475409836, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002986563234743193, |
| "loss": 2.1846, |
| "step": 6200 |
| }, |
| { |
| "epoch": 1.0306775956284153, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002986207761628775, |
| "loss": 2.1928, |
| "step": 6250 |
| }, |
| { |
| "epoch": 1.031224043715847, |
| "grad_norm": 0.6171875, |
| "learning_rate": 0.00029858476693463506, |
| "loss": 2.1942, |
| "step": 6300 |
| }, |
| { |
| "epoch": 1.0317704918032786, |
| "grad_norm": 0.48828125, |
| "learning_rate": 0.0002985482959015094, |
| "loss": 2.1653, |
| "step": 6350 |
| }, |
| { |
| "epoch": 1.0323169398907104, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.00029851136317685345, |
| "loss": 2.1659, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.0328633879781421, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.00029847396887545485, |
| "loss": 2.1829, |
| "step": 6450 |
| }, |
| { |
| "epoch": 1.0334098360655737, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.00029843611311353597, |
| "loss": 2.1911, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.0339562841530054, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00029839779600875343, |
| "loss": 2.1041, |
| "step": 6550 |
| }, |
| { |
| "epoch": 1.0345027322404372, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00029835901768019763, |
| "loss": 2.1634, |
| "step": 6600 |
| }, |
| { |
| "epoch": 1.0350491803278687, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002983197782483926, |
| "loss": 2.1642, |
| "step": 6650 |
| }, |
| { |
| "epoch": 1.0355956284153005, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.00029828007783529533, |
| "loss": 2.1621, |
| "step": 6700 |
| }, |
| { |
| "epoch": 1.0361420765027323, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.0002982399165642956, |
| "loss": 2.1553, |
| "step": 6750 |
| }, |
| { |
| "epoch": 1.036688524590164, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.00029819929456021565, |
| "loss": 2.1592, |
| "step": 6800 |
| }, |
| { |
| "epoch": 1.0372349726775956, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002981582119493095, |
| "loss": 2.1527, |
| "step": 6850 |
| }, |
| { |
| "epoch": 2.000153005464481, |
| "grad_norm": 0.5, |
| "learning_rate": 0.0002981166688592629, |
| "loss": 2.1674, |
| "step": 6900 |
| }, |
| { |
| "epoch": 2.0006994535519125, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.00029807466541919273, |
| "loss": 2.1168, |
| "step": 6950 |
| }, |
| { |
| "epoch": 2.0012459016393445, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029803220175964675, |
| "loss": 2.1439, |
| "step": 7000 |
| }, |
| { |
| "epoch": 2.001792349726776, |
| "grad_norm": 0.5, |
| "learning_rate": 0.0002979892780126028, |
| "loss": 2.0952, |
| "step": 7050 |
| }, |
| { |
| "epoch": 2.0023387978142075, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.00029794589431146904, |
| "loss": 2.0817, |
| "step": 7100 |
| }, |
| { |
| "epoch": 2.0028852459016395, |
| "grad_norm": 0.462890625, |
| "learning_rate": 0.00029790205079108294, |
| "loss": 2.0643, |
| "step": 7150 |
| }, |
| { |
| "epoch": 2.003431693989071, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00029785774758771114, |
| "loss": 2.0993, |
| "step": 7200 |
| }, |
| { |
| "epoch": 2.0039781420765026, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00029781298483904907, |
| "loss": 2.1085, |
| "step": 7250 |
| }, |
| { |
| "epoch": 2.0045245901639346, |
| "grad_norm": 0.4921875, |
| "learning_rate": 0.0002977677626842204, |
| "loss": 2.0645, |
| "step": 7300 |
| }, |
| { |
| "epoch": 2.005071038251366, |
| "grad_norm": 0.50390625, |
| "learning_rate": 0.0002977220812637766, |
| "loss": 2.0929, |
| "step": 7350 |
| }, |
| { |
| "epoch": 2.0056174863387977, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.0002976759407196966, |
| "loss": 2.0919, |
| "step": 7400 |
| }, |
| { |
| "epoch": 2.0061639344262296, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00029762934119538623, |
| "loss": 2.0903, |
| "step": 7450 |
| }, |
| { |
| "epoch": 2.006710382513661, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.00029758228283567796, |
| "loss": 2.0481, |
| "step": 7500 |
| }, |
| { |
| "epoch": 2.0072568306010927, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.00029753476578683023, |
| "loss": 2.0737, |
| "step": 7550 |
| }, |
| { |
| "epoch": 2.0078032786885247, |
| "grad_norm": 0.49609375, |
| "learning_rate": 0.00029748679019652704, |
| "loss": 2.1032, |
| "step": 7600 |
| }, |
| { |
| "epoch": 2.0083497267759562, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.00029743835621387775, |
| "loss": 2.0722, |
| "step": 7650 |
| }, |
| { |
| "epoch": 2.008896174863388, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.00029738946398941623, |
| "loss": 2.057, |
| "step": 7700 |
| }, |
| { |
| "epoch": 2.0094426229508198, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.0002973401136751007, |
| "loss": 2.0802, |
| "step": 7750 |
| }, |
| { |
| "epoch": 2.0099890710382513, |
| "grad_norm": 0.5, |
| "learning_rate": 0.0002972903054243129, |
| "loss": 2.1094, |
| "step": 7800 |
| }, |
| { |
| "epoch": 2.010535519125683, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002972400393918583, |
| "loss": 2.0409, |
| "step": 7850 |
| }, |
| { |
| "epoch": 2.011081967213115, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.0002971893157339647, |
| "loss": 2.0543, |
| "step": 7900 |
| }, |
| { |
| "epoch": 2.0116284153005464, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.0002971381346082824, |
| "loss": 2.0776, |
| "step": 7950 |
| }, |
| { |
| "epoch": 2.0121748633879784, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00029708649617388356, |
| "loss": 2.0629, |
| "step": 8000 |
| }, |
| { |
| "epoch": 2.01272131147541, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002970344005912617, |
| "loss": 2.0588, |
| "step": 8050 |
| }, |
| { |
| "epoch": 2.0132677595628414, |
| "grad_norm": 0.4765625, |
| "learning_rate": 0.000296981848022331, |
| "loss": 2.0373, |
| "step": 8100 |
| }, |
| { |
| "epoch": 2.0138142076502734, |
| "grad_norm": 0.44921875, |
| "learning_rate": 0.000296928838630426, |
| "loss": 2.0348, |
| "step": 8150 |
| }, |
| { |
| "epoch": 2.014360655737705, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.0002968753725803013, |
| "loss": 2.0784, |
| "step": 8200 |
| }, |
| { |
| "epoch": 2.0149071038251365, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.0002968214500381304, |
| "loss": 2.0531, |
| "step": 8250 |
| }, |
| { |
| "epoch": 2.0154535519125685, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.000296767071171506, |
| "loss": 2.0482, |
| "step": 8300 |
| }, |
| { |
| "epoch": 2.016, |
| "grad_norm": 0.484375, |
| "learning_rate": 0.00029671223614943874, |
| "loss": 2.0193, |
| "step": 8350 |
| }, |
| { |
| "epoch": 2.0165464480874316, |
| "grad_norm": 0.4765625, |
| "learning_rate": 0.0002966569451423572, |
| "loss": 2.007, |
| "step": 8400 |
| }, |
| { |
| "epoch": 2.0170928961748635, |
| "grad_norm": 0.48828125, |
| "learning_rate": 0.000296601198322107, |
| "loss": 2.0325, |
| "step": 8450 |
| }, |
| { |
| "epoch": 2.017639344262295, |
| "grad_norm": 0.4921875, |
| "learning_rate": 0.0002965449958619508, |
| "loss": 2.0173, |
| "step": 8500 |
| }, |
| { |
| "epoch": 2.0181857923497266, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002964883379365668, |
| "loss": 1.9927, |
| "step": 8550 |
| }, |
| { |
| "epoch": 2.0187322404371586, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.00029643122472204934, |
| "loss": 2.0149, |
| "step": 8600 |
| }, |
| { |
| "epoch": 2.01927868852459, |
| "grad_norm": 0.49609375, |
| "learning_rate": 0.00029637365639590763, |
| "loss": 2.0077, |
| "step": 8650 |
| }, |
| { |
| "epoch": 2.0198251366120217, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.00029631563313706525, |
| "loss": 1.9926, |
| "step": 8700 |
| }, |
| { |
| "epoch": 2.0203715846994537, |
| "grad_norm": 0.494140625, |
| "learning_rate": 0.0002962571551258599, |
| "loss": 2.0248, |
| "step": 8750 |
| }, |
| { |
| "epoch": 2.020918032786885, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00029619822254404256, |
| "loss": 1.998, |
| "step": 8800 |
| }, |
| { |
| "epoch": 2.0214644808743167, |
| "grad_norm": 0.478515625, |
| "learning_rate": 0.00029613883557477706, |
| "loss": 1.9957, |
| "step": 8850 |
| }, |
| { |
| "epoch": 2.0220109289617487, |
| "grad_norm": 0.49609375, |
| "learning_rate": 0.00029607899440263946, |
| "loss": 2.0132, |
| "step": 8900 |
| }, |
| { |
| "epoch": 2.0225573770491803, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.00029601869921361756, |
| "loss": 2.0038, |
| "step": 8950 |
| }, |
| { |
| "epoch": 2.0231038251366122, |
| "grad_norm": 0.458984375, |
| "learning_rate": 0.00029595795019511005, |
| "loss": 1.9447, |
| "step": 9000 |
| }, |
| { |
| "epoch": 2.023650273224044, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.00029589674753592647, |
| "loss": 1.9806, |
| "step": 9050 |
| }, |
| { |
| "epoch": 2.0241967213114753, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.000295835091426286, |
| "loss": 1.9738, |
| "step": 9100 |
| }, |
| { |
| "epoch": 2.0247431693989073, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.00029577298205781726, |
| "loss": 1.9568, |
| "step": 9150 |
| }, |
| { |
| "epoch": 2.025289617486339, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.00029571041962355755, |
| "loss": 1.9778, |
| "step": 9200 |
| }, |
| { |
| "epoch": 2.0258360655737704, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.0002956474043179525, |
| "loss": 2.0212, |
| "step": 9250 |
| }, |
| { |
| "epoch": 2.0263825136612024, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002955839363368549, |
| "loss": 1.9981, |
| "step": 9300 |
| }, |
| { |
| "epoch": 2.026928961748634, |
| "grad_norm": 0.5, |
| "learning_rate": 0.00029552001587752495, |
| "loss": 1.9763, |
| "step": 9350 |
| }, |
| { |
| "epoch": 2.0274754098360654, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002954556431386288, |
| "loss": 1.9687, |
| "step": 9400 |
| }, |
| { |
| "epoch": 2.0280218579234974, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.00029539081832023837, |
| "loss": 1.9391, |
| "step": 9450 |
| }, |
| { |
| "epoch": 2.028568306010929, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002953255416238308, |
| "loss": 1.9614, |
| "step": 9500 |
| }, |
| { |
| "epoch": 2.0291147540983605, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.0002952598132522874, |
| "loss": 1.9405, |
| "step": 9550 |
| }, |
| { |
| "epoch": 2.0296612021857925, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.00029519363340989367, |
| "loss": 1.9653, |
| "step": 9600 |
| }, |
| { |
| "epoch": 2.030207650273224, |
| "grad_norm": 0.50390625, |
| "learning_rate": 0.0002951270023023379, |
| "loss": 1.9704, |
| "step": 9650 |
| }, |
| { |
| "epoch": 2.0307540983606556, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029505992013671126, |
| "loss": 1.9592, |
| "step": 9700 |
| }, |
| { |
| "epoch": 2.0313005464480876, |
| "grad_norm": 0.474609375, |
| "learning_rate": 0.0002949923871215065, |
| "loss": 1.9679, |
| "step": 9750 |
| }, |
| { |
| "epoch": 2.031846994535519, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.000294924403466618, |
| "loss": 1.9398, |
| "step": 9800 |
| }, |
| { |
| "epoch": 2.0323934426229506, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.00029485596938334037, |
| "loss": 1.9469, |
| "step": 9850 |
| }, |
| { |
| "epoch": 2.0329398907103826, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00029478708508436834, |
| "loss": 1.9742, |
| "step": 9900 |
| }, |
| { |
| "epoch": 2.033486338797814, |
| "grad_norm": 0.482421875, |
| "learning_rate": 0.000294717750783796, |
| "loss": 1.9619, |
| "step": 9950 |
| }, |
| { |
| "epoch": 2.0340327868852457, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.0002946479666971158, |
| "loss": 1.8817, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.0345792349726777, |
| "grad_norm": 0.50390625, |
| "learning_rate": 0.0002945777330412184, |
| "loss": 1.9465, |
| "step": 10050 |
| }, |
| { |
| "epoch": 2.035125683060109, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.00029450705003439156, |
| "loss": 1.9628, |
| "step": 10100 |
| }, |
| { |
| "epoch": 2.035672131147541, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002944359178963198, |
| "loss": 1.9421, |
| "step": 10150 |
| }, |
| { |
| "epoch": 2.0362185792349727, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.00029436433684808336, |
| "loss": 1.953, |
| "step": 10200 |
| }, |
| { |
| "epoch": 2.0367650273224043, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.0002942923071121578, |
| "loss": 1.9499, |
| "step": 10250 |
| }, |
| { |
| "epoch": 2.0373114754098363, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.0002942198289124132, |
| "loss": 1.9522, |
| "step": 10300 |
| }, |
| { |
| "epoch": 3.000229508196721, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.00029414690247411346, |
| "loss": 1.9537, |
| "step": 10350 |
| }, |
| { |
| "epoch": 3.000775956284153, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.0002940735280239157, |
| "loss": 1.9224, |
| "step": 10400 |
| }, |
| { |
| "epoch": 3.0013224043715847, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.0002939997057898693, |
| "loss": 1.9165, |
| "step": 10450 |
| }, |
| { |
| "epoch": 3.0018688524590162, |
| "grad_norm": 0.490234375, |
| "learning_rate": 0.0002939254360014156, |
| "loss": 1.8814, |
| "step": 10500 |
| }, |
| { |
| "epoch": 3.0024153005464482, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002938507188893867, |
| "loss": 1.8547, |
| "step": 10550 |
| }, |
| { |
| "epoch": 3.0029617486338798, |
| "grad_norm": 0.8125, |
| "learning_rate": 0.00029377555468600516, |
| "loss": 1.9014, |
| "step": 10600 |
| }, |
| { |
| "epoch": 3.0035081967213113, |
| "grad_norm": 0.8125, |
| "learning_rate": 0.00029369994362488306, |
| "loss": 1.8837, |
| "step": 10650 |
| }, |
| { |
| "epoch": 3.0040546448087433, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002936238859410213, |
| "loss": 1.9095, |
| "step": 10700 |
| }, |
| { |
| "epoch": 3.004601092896175, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002935473818708089, |
| "loss": 1.8654, |
| "step": 10750 |
| }, |
| { |
| "epoch": 3.0051475409836064, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.00029347043165202233, |
| "loss": 1.9018, |
| "step": 10800 |
| }, |
| { |
| "epoch": 3.0056939890710384, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.0002933930355238246, |
| "loss": 1.895, |
| "step": 10850 |
| }, |
| { |
| "epoch": 3.00624043715847, |
| "grad_norm": 0.50390625, |
| "learning_rate": 0.0002933151937267647, |
| "loss": 1.8872, |
| "step": 10900 |
| }, |
| { |
| "epoch": 3.0067868852459014, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002932369065027767, |
| "loss": 1.8532, |
| "step": 10950 |
| }, |
| { |
| "epoch": 3.0073333333333334, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.0002931581740951791, |
| "loss": 1.8935, |
| "step": 11000 |
| }, |
| { |
| "epoch": 3.007879781420765, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.00029307899674867405, |
| "loss": 1.8991, |
| "step": 11050 |
| }, |
| { |
| "epoch": 3.008426229508197, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029299937470934656, |
| "loss": 1.8784, |
| "step": 11100 |
| }, |
| { |
| "epoch": 3.0089726775956285, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.00029291930822466383, |
| "loss": 1.8775, |
| "step": 11150 |
| }, |
| { |
| "epoch": 3.00951912568306, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002928387975434742, |
| "loss": 1.8874, |
| "step": 11200 |
| }, |
| { |
| "epoch": 3.010065573770492, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.00029275784291600684, |
| "loss": 1.9137, |
| "step": 11250 |
| }, |
| { |
| "epoch": 3.0106120218579235, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002926764445938705, |
| "loss": 1.8568, |
| "step": 11300 |
| }, |
| { |
| "epoch": 3.011158469945355, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002925946028300532, |
| "loss": 1.8578, |
| "step": 11350 |
| }, |
| { |
| "epoch": 3.011704918032787, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.0002925123178789209, |
| "loss": 1.9092, |
| "step": 11400 |
| }, |
| { |
| "epoch": 3.0122513661202186, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.00029242958999621717, |
| "loss": 1.8663, |
| "step": 11450 |
| }, |
| { |
| "epoch": 3.01279781420765, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.00029234641943906223, |
| "loss": 1.862, |
| "step": 11500 |
| }, |
| { |
| "epoch": 3.013344262295082, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002922628064659519, |
| "loss": 1.8594, |
| "step": 11550 |
| }, |
| { |
| "epoch": 3.0138907103825137, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.0002921787513367575, |
| "loss": 1.8633, |
| "step": 11600 |
| }, |
| { |
| "epoch": 3.014437158469945, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002920942543127241, |
| "loss": 1.8929, |
| "step": 11650 |
| }, |
| { |
| "epoch": 3.014983606557377, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.0002920093156564705, |
| "loss": 1.8794, |
| "step": 11700 |
| }, |
| { |
| "epoch": 3.0155300546448087, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.0002919239356319879, |
| "loss": 1.8691, |
| "step": 11750 |
| }, |
| { |
| "epoch": 3.0160765027322403, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00029183811450463954, |
| "loss": 1.8429, |
| "step": 11800 |
| }, |
| { |
| "epoch": 3.0166229508196722, |
| "grad_norm": 0.5, |
| "learning_rate": 0.00029175185254115934, |
| "loss": 1.8325, |
| "step": 11850 |
| }, |
| { |
| "epoch": 3.017169398907104, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00029166515000965154, |
| "loss": 1.8598, |
| "step": 11900 |
| }, |
| { |
| "epoch": 3.0177158469945353, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.0002915780071795896, |
| "loss": 1.8376, |
| "step": 11950 |
| }, |
| { |
| "epoch": 3.0182622950819673, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002914904243218154, |
| "loss": 1.8142, |
| "step": 12000 |
| }, |
| { |
| "epoch": 3.018808743169399, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00029140240170853857, |
| "loss": 1.8505, |
| "step": 12050 |
| }, |
| { |
| "epoch": 3.0193551912568304, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.0002913139396133353, |
| "loss": 1.8315, |
| "step": 12100 |
| }, |
| { |
| "epoch": 3.0199016393442624, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.0002912250383111479, |
| "loss": 1.8337, |
| "step": 12150 |
| }, |
| { |
| "epoch": 3.020448087431694, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.0002911356980782837, |
| "loss": 1.8647, |
| "step": 12200 |
| }, |
| { |
| "epoch": 3.020994535519126, |
| "grad_norm": 0.498046875, |
| "learning_rate": 0.0002910459191924141, |
| "loss": 1.8303, |
| "step": 12250 |
| }, |
| { |
| "epoch": 3.0215409836065574, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.00029095570193257405, |
| "loss": 1.8347, |
| "step": 12300 |
| }, |
| { |
| "epoch": 3.022087431693989, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002908650465791608, |
| "loss": 1.842, |
| "step": 12350 |
| }, |
| { |
| "epoch": 3.022633879781421, |
| "grad_norm": 0.48046875, |
| "learning_rate": 0.00029077395341393334, |
| "loss": 1.8282, |
| "step": 12400 |
| }, |
| { |
| "epoch": 3.0231803278688525, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.00029068242272001135, |
| "loss": 1.7943, |
| "step": 12450 |
| }, |
| { |
| "epoch": 3.023726775956284, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.00029059045478187424, |
| "loss": 1.8147, |
| "step": 12500 |
| }, |
| { |
| "epoch": 3.024273224043716, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00029049804988536053, |
| "loss": 1.8135, |
| "step": 12550 |
| }, |
| { |
| "epoch": 3.0248196721311476, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.00029040520831766676, |
| "loss": 1.8067, |
| "step": 12600 |
| }, |
| { |
| "epoch": 3.025366120218579, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.00029031193036734666, |
| "loss": 1.8333, |
| "step": 12650 |
| }, |
| { |
| "epoch": 3.025912568306011, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.0002902182163243103, |
| "loss": 1.8624, |
| "step": 12700 |
| }, |
| { |
| "epoch": 3.0264590163934426, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.00029012406647982306, |
| "loss": 1.8277, |
| "step": 12750 |
| }, |
| { |
| "epoch": 3.027005464480874, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002900294811265048, |
| "loss": 1.8209, |
| "step": 12800 |
| }, |
| { |
| "epoch": 3.027551912568306, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.0002899344605583291, |
| "loss": 1.8295, |
| "step": 12850 |
| }, |
| { |
| "epoch": 3.0280983606557377, |
| "grad_norm": 0.4921875, |
| "learning_rate": 0.0002898390050706219, |
| "loss": 1.7926, |
| "step": 12900 |
| }, |
| { |
| "epoch": 3.028644808743169, |
| "grad_norm": 0.515625, |
| "learning_rate": 0.0002897431149600612, |
| "loss": 1.8064, |
| "step": 12950 |
| }, |
| { |
| "epoch": 3.029191256830601, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.0002896467905246755, |
| "loss": 1.7923, |
| "step": 13000 |
| }, |
| { |
| "epoch": 3.0297377049180327, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.00028955003206384357, |
| "loss": 1.8346, |
| "step": 13050 |
| }, |
| { |
| "epoch": 3.0302841530054643, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.0002894528398782929, |
| "loss": 1.8187, |
| "step": 13100 |
| }, |
| { |
| "epoch": 3.0308306010928963, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.0002893552142700989, |
| "loss": 1.8035, |
| "step": 13150 |
| }, |
| { |
| "epoch": 3.031377049180328, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.0002892571555426843, |
| "loss": 1.8248, |
| "step": 13200 |
| }, |
| { |
| "epoch": 3.0319234972677593, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00028915866400081795, |
| "loss": 1.8066, |
| "step": 13250 |
| }, |
| { |
| "epoch": 3.0324699453551913, |
| "grad_norm": 0.50390625, |
| "learning_rate": 0.00028905973995061373, |
| "loss": 1.8087, |
| "step": 13300 |
| }, |
| { |
| "epoch": 3.033016393442623, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00028896038369953, |
| "loss": 1.8208, |
| "step": 13350 |
| }, |
| { |
| "epoch": 3.033562841530055, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00028886059555636816, |
| "loss": 1.8282, |
| "step": 13400 |
| }, |
| { |
| "epoch": 3.0341092896174864, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.00028876037583127213, |
| "loss": 1.7288, |
| "step": 13450 |
| }, |
| { |
| "epoch": 3.034655737704918, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.000288659724835727, |
| "loss": 1.841, |
| "step": 13500 |
| }, |
| { |
| "epoch": 3.03520218579235, |
| "grad_norm": 0.51171875, |
| "learning_rate": 0.00028855864288255856, |
| "loss": 1.8044, |
| "step": 13550 |
| }, |
| { |
| "epoch": 3.0357486338797814, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.00028845713028593183, |
| "loss": 1.8101, |
| "step": 13600 |
| }, |
| { |
| "epoch": 3.036295081967213, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00028835518736135013, |
| "loss": 1.8193, |
| "step": 13650 |
| }, |
| { |
| "epoch": 3.036841530054645, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.0002882528144256546, |
| "loss": 1.8219, |
| "step": 13700 |
| }, |
| { |
| "epoch": 3.0373879781420765, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00028815001179702265, |
| "loss": 1.8044, |
| "step": 13750 |
| }, |
| { |
| "epoch": 4.000306010928962, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.0002880467797949671, |
| "loss": 1.8068, |
| "step": 13800 |
| }, |
| { |
| "epoch": 4.000852459016394, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.00028794311874033563, |
| "loss": 1.7919, |
| "step": 13850 |
| }, |
| { |
| "epoch": 4.001398907103825, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00028783902895530893, |
| "loss": 1.7501, |
| "step": 13900 |
| }, |
| { |
| "epoch": 4.001945355191257, |
| "grad_norm": 0.6171875, |
| "learning_rate": 0.00028773451076340064, |
| "loss": 1.7494, |
| "step": 13950 |
| }, |
| { |
| "epoch": 4.002491803278689, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.00028762956448945563, |
| "loss": 1.6976, |
| "step": 14000 |
| }, |
| { |
| "epoch": 4.00303825136612, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00028752419045964935, |
| "loss": 1.7673, |
| "step": 14050 |
| }, |
| { |
| "epoch": 4.003584699453552, |
| "grad_norm": 0.6328125, |
| "learning_rate": 0.0002874183890014867, |
| "loss": 1.7385, |
| "step": 14100 |
| }, |
| { |
| "epoch": 4.004131147540984, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.0002873121604438011, |
| "loss": 1.7526, |
| "step": 14150 |
| }, |
| { |
| "epoch": 4.004677595628415, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002872055051167533, |
| "loss": 1.7323, |
| "step": 14200 |
| }, |
| { |
| "epoch": 4.005224043715847, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.0002870984233518306, |
| "loss": 1.7666, |
| "step": 14250 |
| }, |
| { |
| "epoch": 4.005770491803279, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.0002869909154818455, |
| "loss": 1.7382, |
| "step": 14300 |
| }, |
| { |
| "epoch": 4.00631693989071, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.00028688298184093497, |
| "loss": 1.7366, |
| "step": 14350 |
| }, |
| { |
| "epoch": 4.006863387978142, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.0002867746227645593, |
| "loss": 1.7206, |
| "step": 14400 |
| }, |
| { |
| "epoch": 4.007409836065574, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.000286665838589501, |
| "loss": 1.7655, |
| "step": 14450 |
| }, |
| { |
| "epoch": 4.007956284153005, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.0002865566296538637, |
| "loss": 1.7599, |
| "step": 14500 |
| }, |
| { |
| "epoch": 4.008502732240437, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.00028644699629707136, |
| "loss": 1.7298, |
| "step": 14550 |
| }, |
| { |
| "epoch": 4.009049180327869, |
| "grad_norm": 0.671875, |
| "learning_rate": 0.00028633693885986696, |
| "loss": 1.7392, |
| "step": 14600 |
| }, |
| { |
| "epoch": 4.0095956284153, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.0002862264576843116, |
| "loss": 1.7556, |
| "step": 14650 |
| }, |
| { |
| "epoch": 4.010142076502732, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002861155531137833, |
| "loss": 1.7677, |
| "step": 14700 |
| }, |
| { |
| "epoch": 4.010688524590164, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.00028600422549297604, |
| "loss": 1.7283, |
| "step": 14750 |
| }, |
| { |
| "epoch": 4.011234972677595, |
| "grad_norm": 0.51953125, |
| "learning_rate": 0.00028589247516789856, |
| "loss": 1.7389, |
| "step": 14800 |
| }, |
| { |
| "epoch": 4.011781420765027, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.0002857803024858735, |
| "loss": 1.7769, |
| "step": 14850 |
| }, |
| { |
| "epoch": 4.012327868852459, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.00028566770779553613, |
| "loss": 1.7306, |
| "step": 14900 |
| }, |
| { |
| "epoch": 4.01287431693989, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.00028555469144683337, |
| "loss": 1.7341, |
| "step": 14950 |
| }, |
| { |
| "epoch": 4.013420765027322, |
| "grad_norm": 0.53515625, |
| "learning_rate": 0.00028544125379102264, |
| "loss": 1.7364, |
| "step": 15000 |
| }, |
| { |
| "epoch": 4.013967213114754, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002853273951806708, |
| "loss": 1.7203, |
| "step": 15050 |
| }, |
| { |
| "epoch": 4.0145136612021854, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.00028521311596965297, |
| "loss": 1.7735, |
| "step": 15100 |
| }, |
| { |
| "epoch": 4.015060109289617, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.00028509841651315156, |
| "loss": 1.7457, |
| "step": 15150 |
| }, |
| { |
| "epoch": 4.015606557377049, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002849832971676553, |
| "loss": 1.7317, |
| "step": 15200 |
| }, |
| { |
| "epoch": 4.0161530054644805, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.0002848677582909576, |
| "loss": 1.7168, |
| "step": 15250 |
| }, |
| { |
| "epoch": 4.0166994535519125, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.000284751800242156, |
| "loss": 1.714, |
| "step": 15300 |
| }, |
| { |
| "epoch": 4.0172459016393445, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002846354233816508, |
| "loss": 1.7342, |
| "step": 15350 |
| }, |
| { |
| "epoch": 4.017792349726776, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.00028451862807114396, |
| "loss": 1.7201, |
| "step": 15400 |
| }, |
| { |
| "epoch": 4.0183387978142076, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00028440141467363803, |
| "loss": 1.6881, |
| "step": 15450 |
| }, |
| { |
| "epoch": 4.0188852459016395, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.00028428378355343495, |
| "loss": 1.7184, |
| "step": 15500 |
| }, |
| { |
| "epoch": 4.019431693989071, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.00028416573507613485, |
| "loss": 1.7055, |
| "step": 15550 |
| }, |
| { |
| "epoch": 4.019978142076503, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.0002840472696086353, |
| "loss": 1.72, |
| "step": 15600 |
| }, |
| { |
| "epoch": 4.020524590163935, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002839283875191295, |
| "loss": 1.7308, |
| "step": 15650 |
| }, |
| { |
| "epoch": 4.021071038251366, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.0002838090891771059, |
| "loss": 1.702, |
| "step": 15700 |
| }, |
| { |
| "epoch": 4.021617486338798, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.0002836893749533465, |
| "loss": 1.7191, |
| "step": 15750 |
| }, |
| { |
| "epoch": 4.02216393442623, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.0002835692452199257, |
| "loss": 1.7215, |
| "step": 15800 |
| }, |
| { |
| "epoch": 4.022710382513662, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.00028344870035020963, |
| "loss": 1.6984, |
| "step": 15850 |
| }, |
| { |
| "epoch": 4.023256830601093, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.0002833277407188545, |
| "loss": 1.6847, |
| "step": 15900 |
| }, |
| { |
| "epoch": 4.023803278688525, |
| "grad_norm": 0.53125, |
| "learning_rate": 0.00028320636670180557, |
| "loss": 1.6739, |
| "step": 15950 |
| }, |
| { |
| "epoch": 4.024349726775957, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002830845786762962, |
| "loss": 1.7125, |
| "step": 16000 |
| }, |
| { |
| "epoch": 4.024896174863388, |
| "grad_norm": 0.6796875, |
| "learning_rate": 0.0002829623770208463, |
| "loss": 1.6678, |
| "step": 16050 |
| }, |
| { |
| "epoch": 4.02544262295082, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.00028283976211526137, |
| "loss": 1.7396, |
| "step": 16100 |
| }, |
| { |
| "epoch": 4.025989071038252, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.0002827167343406315, |
| "loss": 1.752, |
| "step": 16150 |
| }, |
| { |
| "epoch": 4.026535519125683, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002825932940793298, |
| "loss": 1.6994, |
| "step": 16200 |
| }, |
| { |
| "epoch": 4.027081967213115, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.00028246944171501145, |
| "loss": 1.7189, |
| "step": 16250 |
| }, |
| { |
| "epoch": 4.027628415300547, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.00028234517763261243, |
| "loss": 1.7158, |
| "step": 16300 |
| }, |
| { |
| "epoch": 4.028174863387978, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.00028222050221834847, |
| "loss": 1.6885, |
| "step": 16350 |
| }, |
| { |
| "epoch": 4.02872131147541, |
| "grad_norm": 0.5078125, |
| "learning_rate": 0.0002820954158597134, |
| "loss": 1.6865, |
| "step": 16400 |
| }, |
| { |
| "epoch": 4.029267759562842, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002819699189454788, |
| "loss": 1.6946, |
| "step": 16450 |
| }, |
| { |
| "epoch": 4.029814207650273, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.0002818440118656918, |
| "loss": 1.7194, |
| "step": 16500 |
| }, |
| { |
| "epoch": 4.030360655737705, |
| "grad_norm": 0.52734375, |
| "learning_rate": 0.0002817176950116746, |
| "loss": 1.7022, |
| "step": 16550 |
| }, |
| { |
| "epoch": 4.030907103825137, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.00028159096877602275, |
| "loss": 1.7159, |
| "step": 16600 |
| }, |
| { |
| "epoch": 4.031453551912568, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.00028146383355260446, |
| "loss": 1.7063, |
| "step": 16650 |
| }, |
| { |
| "epoch": 4.032, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.00028133628973655894, |
| "loss": 1.7064, |
| "step": 16700 |
| }, |
| { |
| "epoch": 4.032546448087432, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.00028120833772429517, |
| "loss": 1.6992, |
| "step": 16750 |
| }, |
| { |
| "epoch": 4.033092896174863, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.0002810799779134911, |
| "loss": 1.7132, |
| "step": 16800 |
| }, |
| { |
| "epoch": 4.033639344262295, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.0002809512107030919, |
| "loss": 1.7125, |
| "step": 16850 |
| }, |
| { |
| "epoch": 4.034185792349727, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002808220364933091, |
| "loss": 1.6373, |
| "step": 16900 |
| }, |
| { |
| "epoch": 4.034732240437158, |
| "grad_norm": 0.5234375, |
| "learning_rate": 0.00028069245568561904, |
| "loss": 1.7379, |
| "step": 16950 |
| }, |
| { |
| "epoch": 4.03527868852459, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.00028056246868276186, |
| "loss": 1.699, |
| "step": 17000 |
| }, |
| { |
| "epoch": 4.035825136612022, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.0002804320758887403, |
| "loss": 1.6939, |
| "step": 17050 |
| }, |
| { |
| "epoch": 4.036371584699453, |
| "grad_norm": 0.498046875, |
| "learning_rate": 0.000280301277708818, |
| "loss": 1.7146, |
| "step": 17100 |
| }, |
| { |
| "epoch": 4.036918032786885, |
| "grad_norm": 0.5859375, |
| "learning_rate": 0.00028017007454951884, |
| "loss": 1.7363, |
| "step": 17150 |
| }, |
| { |
| "epoch": 4.037464480874317, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.00028003846681862524, |
| "loss": 1.6965, |
| "step": 17200 |
| }, |
| { |
| "epoch": 5.000382513661203, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.00027990645492517697, |
| "loss": 1.6887, |
| "step": 17250 |
| }, |
| { |
| "epoch": 5.000928961748634, |
| "grad_norm": 0.62109375, |
| "learning_rate": 0.0002797740392794702, |
| "loss": 1.6829, |
| "step": 17300 |
| }, |
| { |
| "epoch": 5.001475409836066, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.0002796412202930557, |
| "loss": 1.6235, |
| "step": 17350 |
| }, |
| { |
| "epoch": 5.002021857923498, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.00027950799837873794, |
| "loss": 1.634, |
| "step": 17400 |
| }, |
| { |
| "epoch": 5.002568306010929, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.0002793743739505738, |
| "loss": 1.5863, |
| "step": 17450 |
| }, |
| { |
| "epoch": 5.003114754098361, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.0002792403474238709, |
| "loss": 1.6513, |
| "step": 17500 |
| }, |
| { |
| "epoch": 5.003661202185793, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.000279105919215187, |
| "loss": 1.6388, |
| "step": 17550 |
| }, |
| { |
| "epoch": 5.004207650273224, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.00027897108974232797, |
| "loss": 1.6302, |
| "step": 17600 |
| }, |
| { |
| "epoch": 5.004754098360656, |
| "grad_norm": 1.1875, |
| "learning_rate": 0.0002788358594243469, |
| "loss": 1.6336, |
| "step": 17650 |
| }, |
| { |
| "epoch": 5.005300546448088, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.0002787002286815428, |
| "loss": 1.6513, |
| "step": 17700 |
| }, |
| { |
| "epoch": 5.005846994535519, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.000278564197935459, |
| "loss": 1.6185, |
| "step": 17750 |
| }, |
| { |
| "epoch": 5.006393442622951, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.00027842776760888236, |
| "loss": 1.6366, |
| "step": 17800 |
| }, |
| { |
| "epoch": 5.006939890710383, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.00027829093812584143, |
| "loss": 1.6162, |
| "step": 17850 |
| }, |
| { |
| "epoch": 5.007486338797814, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002781537099116054, |
| "loss": 1.6665, |
| "step": 17900 |
| }, |
| { |
| "epoch": 5.008032786885246, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.00027801608339268275, |
| "loss": 1.6505, |
| "step": 17950 |
| }, |
| { |
| "epoch": 5.008579234972678, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.00027787805899681976, |
| "loss": 1.6079, |
| "step": 18000 |
| }, |
| { |
| "epoch": 5.009125683060109, |
| "grad_norm": 0.640625, |
| "learning_rate": 0.00027773963715299957, |
| "loss": 1.6429, |
| "step": 18050 |
| }, |
| { |
| "epoch": 5.009672131147541, |
| "grad_norm": 0.640625, |
| "learning_rate": 0.00027760081829144044, |
| "loss": 1.6671, |
| "step": 18100 |
| }, |
| { |
| "epoch": 5.010218579234973, |
| "grad_norm": 0.6484375, |
| "learning_rate": 0.0002774616028435946, |
| "loss": 1.6619, |
| "step": 18150 |
| }, |
| { |
| "epoch": 5.010765027322404, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.00027732199124214676, |
| "loss": 1.6221, |
| "step": 18200 |
| }, |
| { |
| "epoch": 5.011311475409836, |
| "grad_norm": 0.63671875, |
| "learning_rate": 0.0002771819839210131, |
| "loss": 1.6311, |
| "step": 18250 |
| }, |
| { |
| "epoch": 5.011857923497268, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.0002770415813153396, |
| "loss": 1.669, |
| "step": 18300 |
| }, |
| { |
| "epoch": 5.012404371584699, |
| "grad_norm": 0.609375, |
| "learning_rate": 0.00027690078386150084, |
| "loss": 1.6177, |
| "step": 18350 |
| }, |
| { |
| "epoch": 5.012950819672131, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.0002767595919970984, |
| "loss": 1.622, |
| "step": 18400 |
| }, |
| { |
| "epoch": 5.013497267759563, |
| "grad_norm": 0.609375, |
| "learning_rate": 0.00027661800616096, |
| "loss": 1.6329, |
| "step": 18450 |
| }, |
| { |
| "epoch": 5.014043715846994, |
| "grad_norm": 0.58203125, |
| "learning_rate": 0.00027647602679313764, |
| "loss": 1.6369, |
| "step": 18500 |
| }, |
| { |
| "epoch": 5.014590163934426, |
| "grad_norm": 0.55078125, |
| "learning_rate": 0.0002763336543349065, |
| "loss": 1.6624, |
| "step": 18550 |
| }, |
| { |
| "epoch": 5.015136612021858, |
| "grad_norm": 0.5859375, |
| "learning_rate": 0.0002761908892287633, |
| "loss": 1.6422, |
| "step": 18600 |
| }, |
| { |
| "epoch": 5.015683060109289, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.0002760477319184255, |
| "loss": 1.6101, |
| "step": 18650 |
| }, |
| { |
| "epoch": 5.016229508196721, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002759041828488292, |
| "loss": 1.6355, |
| "step": 18700 |
| }, |
| { |
| "epoch": 5.016775956284153, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.0002757602424661283, |
| "loss": 1.6114, |
| "step": 18750 |
| }, |
| { |
| "epoch": 5.017322404371584, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.00027561591121769277, |
| "loss": 1.6274, |
| "step": 18800 |
| }, |
| { |
| "epoch": 5.017868852459016, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.00027547118955210747, |
| "loss": 1.6139, |
| "step": 18850 |
| }, |
| { |
| "epoch": 5.018415300546448, |
| "grad_norm": 0.55859375, |
| "learning_rate": 0.0002753260779191706, |
| "loss": 1.6027, |
| "step": 18900 |
| }, |
| { |
| "epoch": 5.018961748633879, |
| "grad_norm": 0.62109375, |
| "learning_rate": 0.0002751805767698927, |
| "loss": 1.6036, |
| "step": 18950 |
| }, |
| { |
| "epoch": 5.019508196721311, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.0002750346865564944, |
| "loss": 1.6073, |
| "step": 19000 |
| }, |
| { |
| "epoch": 5.020054644808743, |
| "grad_norm": 0.54296875, |
| "learning_rate": 0.0002748884077324061, |
| "loss": 1.6285, |
| "step": 19050 |
| }, |
| { |
| "epoch": 5.020601092896175, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.0002747417407522656, |
| "loss": 1.6254, |
| "step": 19100 |
| }, |
| { |
| "epoch": 5.021147540983606, |
| "grad_norm": 0.5546875, |
| "learning_rate": 0.00027459468607191723, |
| "loss": 1.6058, |
| "step": 19150 |
| }, |
| { |
| "epoch": 5.021693989071038, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.00027444724414841046, |
| "loss": 1.623, |
| "step": 19200 |
| }, |
| { |
| "epoch": 5.02224043715847, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.00027429941543999814, |
| "loss": 1.623, |
| "step": 19250 |
| }, |
| { |
| "epoch": 5.0227868852459014, |
| "grad_norm": 0.5390625, |
| "learning_rate": 0.0002741512004061353, |
| "loss": 1.601, |
| "step": 19300 |
| }, |
| { |
| "epoch": 5.023333333333333, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002740025995074777, |
| "loss": 1.5935, |
| "step": 19350 |
| }, |
| { |
| "epoch": 5.023879781420765, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.00027385361320588034, |
| "loss": 1.5831, |
| "step": 19400 |
| }, |
| { |
| "epoch": 5.0244262295081965, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.0002737042419643961, |
| "loss": 1.5974, |
| "step": 19450 |
| }, |
| { |
| "epoch": 5.0249726775956285, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.0002735544862472742, |
| "loss": 1.5629, |
| "step": 19500 |
| }, |
| { |
| "epoch": 5.0255191256830605, |
| "grad_norm": 0.640625, |
| "learning_rate": 0.00027340434651995887, |
| "loss": 1.6589, |
| "step": 19550 |
| }, |
| { |
| "epoch": 5.026065573770492, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.0002732538232490879, |
| "loss": 1.6576, |
| "step": 19600 |
| }, |
| { |
| "epoch": 5.0266120218579236, |
| "grad_norm": 0.5859375, |
| "learning_rate": 0.0002731029169024911, |
| "loss": 1.6112, |
| "step": 19650 |
| }, |
| { |
| "epoch": 5.0271584699453555, |
| "grad_norm": 0.609375, |
| "learning_rate": 0.00027295162794918875, |
| "loss": 1.6232, |
| "step": 19700 |
| }, |
| { |
| "epoch": 5.027704918032787, |
| "grad_norm": 0.6328125, |
| "learning_rate": 0.00027279995685939055, |
| "loss": 1.6242, |
| "step": 19750 |
| }, |
| { |
| "epoch": 5.028251366120219, |
| "grad_norm": 0.546875, |
| "learning_rate": 0.00027264790410449363, |
| "loss": 1.6029, |
| "step": 19800 |
| }, |
| { |
| "epoch": 5.028797814207651, |
| "grad_norm": 0.6875, |
| "learning_rate": 0.0002724954701570816, |
| "loss": 1.5935, |
| "step": 19850 |
| }, |
| { |
| "epoch": 5.029344262295082, |
| "grad_norm": 0.6640625, |
| "learning_rate": 0.00027234265549092257, |
| "loss": 1.6099, |
| "step": 19900 |
| }, |
| { |
| "epoch": 5.029890710382514, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.00027218946058096805, |
| "loss": 1.6366, |
| "step": 19950 |
| }, |
| { |
| "epoch": 5.030437158469946, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002720358859033514, |
| "loss": 1.6249, |
| "step": 20000 |
| }, |
| { |
| "epoch": 5.030983606557377, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.00027188193193538625, |
| "loss": 1.6114, |
| "step": 20050 |
| }, |
| { |
| "epoch": 5.031530054644809, |
| "grad_norm": 0.671875, |
| "learning_rate": 0.00027172759915556504, |
| "loss": 1.6081, |
| "step": 20100 |
| }, |
| { |
| "epoch": 5.032076502732241, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.0002715728880435577, |
| "loss": 1.627, |
| "step": 20150 |
| }, |
| { |
| "epoch": 5.032622950819672, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.00027141779908020986, |
| "loss": 1.5912, |
| "step": 20200 |
| }, |
| { |
| "epoch": 5.033169398907104, |
| "grad_norm": 0.65234375, |
| "learning_rate": 0.00027126233274754163, |
| "loss": 1.6476, |
| "step": 20250 |
| }, |
| { |
| "epoch": 5.033715846994536, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.00027110648952874595, |
| "loss": 1.6051, |
| "step": 20300 |
| }, |
| { |
| "epoch": 5.034262295081967, |
| "grad_norm": 0.65234375, |
| "learning_rate": 0.0002709502699081871, |
| "loss": 1.5705, |
| "step": 20350 |
| }, |
| { |
| "epoch": 5.034808743169399, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.00027079367437139935, |
| "loss": 1.6533, |
| "step": 20400 |
| }, |
| { |
| "epoch": 5.035355191256831, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.00027063670340508514, |
| "loss": 1.6099, |
| "step": 20450 |
| }, |
| { |
| "epoch": 5.035901639344262, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.00027047935749711395, |
| "loss": 1.6018, |
| "step": 20500 |
| }, |
| { |
| "epoch": 5.036448087431694, |
| "grad_norm": 0.59375, |
| "learning_rate": 0.0002703216371365204, |
| "loss": 1.637, |
| "step": 20550 |
| }, |
| { |
| "epoch": 5.036994535519126, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.00027016354281350315, |
| "loss": 1.6394, |
| "step": 20600 |
| }, |
| { |
| "epoch": 5.037540983606557, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.00027000507501942283, |
| "loss": 1.6253, |
| "step": 20650 |
| }, |
| { |
| "epoch": 6.000459016393442, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002698462342468011, |
| "loss": 1.5897, |
| "step": 20700 |
| }, |
| { |
| "epoch": 6.001005464480874, |
| "grad_norm": 0.56640625, |
| "learning_rate": 0.0002696870209893187, |
| "loss": 1.569, |
| "step": 20750 |
| }, |
| { |
| "epoch": 6.001551912568306, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.00026952743574181414, |
| "loss": 1.5482, |
| "step": 20800 |
| }, |
| { |
| "epoch": 6.002098360655737, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.00026936747900028205, |
| "loss": 1.5458, |
| "step": 20850 |
| }, |
| { |
| "epoch": 6.002644808743169, |
| "grad_norm": 0.734375, |
| "learning_rate": 0.00026920715126187167, |
| "loss": 1.4787, |
| "step": 20900 |
| }, |
| { |
| "epoch": 6.003191256830601, |
| "grad_norm": 0.7109375, |
| "learning_rate": 0.0002690464530248853, |
| "loss": 1.5565, |
| "step": 20950 |
| }, |
| { |
| "epoch": 6.0037377049180325, |
| "grad_norm": 0.64453125, |
| "learning_rate": 0.00026888538478877675, |
| "loss": 1.5588, |
| "step": 21000 |
| }, |
| { |
| "epoch": 6.0042841530054645, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.0002687239470541498, |
| "loss": 1.5347, |
| "step": 21050 |
| }, |
| { |
| "epoch": 6.0048306010928965, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.00026856214032275675, |
| "loss": 1.5341, |
| "step": 21100 |
| }, |
| { |
| "epoch": 6.0053770491803276, |
| "grad_norm": 0.73046875, |
| "learning_rate": 0.00026839996509749655, |
| "loss": 1.5441, |
| "step": 21150 |
| }, |
| { |
| "epoch": 6.0059234972677595, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.00026823742188241366, |
| "loss": 1.5405, |
| "step": 21200 |
| }, |
| { |
| "epoch": 6.0064699453551915, |
| "grad_norm": 0.64453125, |
| "learning_rate": 0.000268074511182696, |
| "loss": 1.5327, |
| "step": 21250 |
| }, |
| { |
| "epoch": 6.007016393442623, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.00026791123350467384, |
| "loss": 1.5338, |
| "step": 21300 |
| }, |
| { |
| "epoch": 6.007562841530055, |
| "grad_norm": 0.58984375, |
| "learning_rate": 0.000267747589355818, |
| "loss": 1.5663, |
| "step": 21350 |
| }, |
| { |
| "epoch": 6.008109289617487, |
| "grad_norm": 0.67578125, |
| "learning_rate": 0.0002675835792447382, |
| "loss": 1.5519, |
| "step": 21400 |
| }, |
| { |
| "epoch": 6.008655737704918, |
| "grad_norm": 0.66015625, |
| "learning_rate": 0.0002674192036811818, |
| "loss": 1.5256, |
| "step": 21450 |
| }, |
| { |
| "epoch": 6.00920218579235, |
| "grad_norm": 0.578125, |
| "learning_rate": 0.0002672544631760317, |
| "loss": 1.5428, |
| "step": 21500 |
| }, |
| { |
| "epoch": 6.009748633879782, |
| "grad_norm": 0.75, |
| "learning_rate": 0.00026708935824130514, |
| "loss": 1.5806, |
| "step": 21550 |
| }, |
| { |
| "epoch": 6.010295081967213, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.00026692388939015226, |
| "loss": 1.5628, |
| "step": 21600 |
| }, |
| { |
| "epoch": 6.010841530054645, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.00026675805713685387, |
| "loss": 1.5275, |
| "step": 21650 |
| }, |
| { |
| "epoch": 6.011387978142077, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.0002665918619968206, |
| "loss": 1.5547, |
| "step": 21700 |
| }, |
| { |
| "epoch": 6.011934426229508, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002664253044865907, |
| "loss": 1.575, |
| "step": 21750 |
| }, |
| { |
| "epoch": 6.01248087431694, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.0002662583851238287, |
| "loss": 1.5386, |
| "step": 21800 |
| }, |
| { |
| "epoch": 6.013027322404372, |
| "grad_norm": 0.6484375, |
| "learning_rate": 0.000266091104427324, |
| "loss": 1.5107, |
| "step": 21850 |
| }, |
| { |
| "epoch": 6.013573770491803, |
| "grad_norm": 0.625, |
| "learning_rate": 0.00026592346291698864, |
| "loss": 1.5516, |
| "step": 21900 |
| }, |
| { |
| "epoch": 6.014120218579235, |
| "grad_norm": 0.5703125, |
| "learning_rate": 0.00026575546111385647, |
| "loss": 1.5431, |
| "step": 21950 |
| }, |
| { |
| "epoch": 6.014666666666667, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.00026558709954008095, |
| "loss": 1.566, |
| "step": 22000 |
| }, |
| { |
| "epoch": 6.015213114754098, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.00026541837871893367, |
| "loss": 1.5595, |
| "step": 22050 |
| }, |
| { |
| "epoch": 6.01575956284153, |
| "grad_norm": 0.5625, |
| "learning_rate": 0.0002652492991748029, |
| "loss": 1.5206, |
| "step": 22100 |
| }, |
| { |
| "epoch": 6.016306010928962, |
| "grad_norm": 0.640625, |
| "learning_rate": 0.00026507986143319164, |
| "loss": 1.5374, |
| "step": 22150 |
| }, |
| { |
| "epoch": 6.016852459016394, |
| "grad_norm": 0.703125, |
| "learning_rate": 0.0002649100660207164, |
| "loss": 1.522, |
| "step": 22200 |
| }, |
| { |
| "epoch": 6.017398907103825, |
| "grad_norm": 0.65625, |
| "learning_rate": 0.0002647399134651053, |
| "loss": 1.5532, |
| "step": 22250 |
| }, |
| { |
| "epoch": 6.017945355191257, |
| "grad_norm": 0.63671875, |
| "learning_rate": 0.0002645694042951963, |
| "loss": 1.5274, |
| "step": 22300 |
| }, |
| { |
| "epoch": 6.018491803278689, |
| "grad_norm": 0.65625, |
| "learning_rate": 0.00026439853904093586, |
| "loss": 1.517, |
| "step": 22350 |
| }, |
| { |
| "epoch": 6.01903825136612, |
| "grad_norm": 0.6875, |
| "learning_rate": 0.00026422731823337717, |
| "loss": 1.5197, |
| "step": 22400 |
| }, |
| { |
| "epoch": 6.019584699453552, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.0002640557424046784, |
| "loss": 1.5081, |
| "step": 22450 |
| }, |
| { |
| "epoch": 6.020131147540984, |
| "grad_norm": 0.65234375, |
| "learning_rate": 0.0002638838120881012, |
| "loss": 1.5526, |
| "step": 22500 |
| }, |
| { |
| "epoch": 6.020677595628415, |
| "grad_norm": 0.62109375, |
| "learning_rate": 0.000263711527818009, |
| "loss": 1.534, |
| "step": 22550 |
| }, |
| { |
| "epoch": 6.021224043715847, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.0002635388901298652, |
| "loss": 1.5156, |
| "step": 22600 |
| }, |
| { |
| "epoch": 6.021770491803279, |
| "grad_norm": 0.71875, |
| "learning_rate": 0.0002633658995602318, |
| "loss": 1.5402, |
| "step": 22650 |
| }, |
| { |
| "epoch": 6.02231693989071, |
| "grad_norm": 0.68359375, |
| "learning_rate": 0.0002631925566467674, |
| "loss": 1.5367, |
| "step": 22700 |
| }, |
| { |
| "epoch": 6.022863387978142, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.00026301886192822585, |
| "loss": 1.5126, |
| "step": 22750 |
| }, |
| { |
| "epoch": 6.023409836065574, |
| "grad_norm": 0.625, |
| "learning_rate": 0.00026284481594445434, |
| "loss": 1.5097, |
| "step": 22800 |
| }, |
| { |
| "epoch": 6.023956284153005, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.00026267041923639175, |
| "loss": 1.5058, |
| "step": 22850 |
| }, |
| { |
| "epoch": 6.024502732240437, |
| "grad_norm": 0.703125, |
| "learning_rate": 0.00026249567234606707, |
| "loss": 1.5004, |
| "step": 22900 |
| }, |
| { |
| "epoch": 6.025049180327869, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.00026232057581659777, |
| "loss": 1.4884, |
| "step": 22950 |
| }, |
| { |
| "epoch": 6.0255956284153, |
| "grad_norm": 0.65234375, |
| "learning_rate": 0.0002621451301921878, |
| "loss": 1.5884, |
| "step": 23000 |
| }, |
| { |
| "epoch": 6.026142076502732, |
| "grad_norm": 0.6484375, |
| "learning_rate": 0.00026196933601812616, |
| "loss": 1.565, |
| "step": 23050 |
| }, |
| { |
| "epoch": 6.026688524590164, |
| "grad_norm": 0.69140625, |
| "learning_rate": 0.00026179319384078535, |
| "loss": 1.5399, |
| "step": 23100 |
| }, |
| { |
| "epoch": 6.027234972677595, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002616167042076192, |
| "loss": 1.5319, |
| "step": 23150 |
| }, |
| { |
| "epoch": 6.027781420765027, |
| "grad_norm": 0.62890625, |
| "learning_rate": 0.0002614398676671616, |
| "loss": 1.5379, |
| "step": 23200 |
| }, |
| { |
| "epoch": 6.028327868852459, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.0002612626847690247, |
| "loss": 1.5344, |
| "step": 23250 |
| }, |
| { |
| "epoch": 6.02887431693989, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.0002610851560638968, |
| "loss": 1.5054, |
| "step": 23300 |
| }, |
| { |
| "epoch": 6.029420765027322, |
| "grad_norm": 0.6484375, |
| "learning_rate": 0.0002609072821035415, |
| "loss": 1.5421, |
| "step": 23350 |
| }, |
| { |
| "epoch": 6.029967213114754, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.00026072906344079484, |
| "loss": 1.5625, |
| "step": 23400 |
| }, |
| { |
| "epoch": 6.0305136612021855, |
| "grad_norm": 0.59765625, |
| "learning_rate": 0.0002605505006295648, |
| "loss": 1.5476, |
| "step": 23450 |
| }, |
| { |
| "epoch": 6.031060109289617, |
| "grad_norm": 0.609375, |
| "learning_rate": 0.00026037159422482865, |
| "loss": 1.537, |
| "step": 23500 |
| }, |
| { |
| "epoch": 6.031606557377049, |
| "grad_norm": 0.60546875, |
| "learning_rate": 0.00026019234478263155, |
| "loss": 1.5204, |
| "step": 23550 |
| }, |
| { |
| "epoch": 6.0321530054644805, |
| "grad_norm": 0.69921875, |
| "learning_rate": 0.000260012752860085, |
| "loss": 1.5479, |
| "step": 23600 |
| }, |
| { |
| "epoch": 6.0326994535519125, |
| "grad_norm": 0.6171875, |
| "learning_rate": 0.00025983281901536474, |
| "loss": 1.5304, |
| "step": 23650 |
| }, |
| { |
| "epoch": 6.0332459016393445, |
| "grad_norm": 0.6171875, |
| "learning_rate": 0.00025965254380770945, |
| "loss": 1.5738, |
| "step": 23700 |
| }, |
| { |
| "epoch": 6.033792349726776, |
| "grad_norm": 0.62109375, |
| "learning_rate": 0.0002594719277974185, |
| "loss": 1.5168, |
| "step": 23750 |
| }, |
| { |
| "epoch": 6.034338797814208, |
| "grad_norm": 0.6328125, |
| "learning_rate": 0.0002592909715458506, |
| "loss": 1.4984, |
| "step": 23800 |
| }, |
| { |
| "epoch": 6.0348852459016395, |
| "grad_norm": 0.57421875, |
| "learning_rate": 0.0002591096756154221, |
| "loss": 1.5721, |
| "step": 23850 |
| }, |
| { |
| "epoch": 6.035431693989071, |
| "grad_norm": 0.6015625, |
| "learning_rate": 0.0002589280405696048, |
| "loss": 1.5369, |
| "step": 23900 |
| }, |
| { |
| "epoch": 6.035978142076503, |
| "grad_norm": 0.61328125, |
| "learning_rate": 0.00025874606697292473, |
| "loss": 1.5236, |
| "step": 23950 |
| }, |
| { |
| "epoch": 6.036524590163935, |
| "grad_norm": 0.6953125, |
| "learning_rate": 0.00025856375539095986, |
| "loss": 1.577, |
| "step": 24000 |
| } |
| ], |
| "logging_steps": 50, |
| "max_steps": 91500, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 9223372036854775807, |
| "save_steps": 2000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.2834709186604433e+19, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|