{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 4091, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0002444390124663896, "grad_norm": 20.966943740844727, "learning_rate": 1.6260162601626018e-07, "loss": 3.4835, "step": 1 }, { "epoch": 0.0004888780249327792, "grad_norm": 18.250547409057617, "learning_rate": 3.2520325203252037e-07, "loss": 3.3028, "step": 2 }, { "epoch": 0.0007333170373991689, "grad_norm": 18.719200134277344, "learning_rate": 4.878048780487805e-07, "loss": 3.249, "step": 3 }, { "epoch": 0.0009777560498655585, "grad_norm": 17.085054397583008, "learning_rate": 6.504065040650407e-07, "loss": 3.4029, "step": 4 }, { "epoch": 0.0012221950623319481, "grad_norm": 18.758522033691406, "learning_rate": 8.130081300813009e-07, "loss": 3.3331, "step": 5 }, { "epoch": 0.0014666340747983377, "grad_norm": 15.967318534851074, "learning_rate": 9.75609756097561e-07, "loss": 3.176, "step": 6 }, { "epoch": 0.0017110730872647274, "grad_norm": 16.959604263305664, "learning_rate": 1.1382113821138213e-06, "loss": 3.2587, "step": 7 }, { "epoch": 0.001955512099731117, "grad_norm": 16.36026954650879, "learning_rate": 1.3008130081300815e-06, "loss": 3.1667, "step": 8 }, { "epoch": 0.002199951112197507, "grad_norm": 15.372851371765137, "learning_rate": 1.4634146341463414e-06, "loss": 3.1927, "step": 9 }, { "epoch": 0.0024443901246638962, "grad_norm": 14.929744720458984, "learning_rate": 1.6260162601626018e-06, "loss": 3.0885, "step": 10 }, { "epoch": 0.002688829137130286, "grad_norm": 12.868640899658203, "learning_rate": 1.788617886178862e-06, "loss": 2.7862, "step": 11 }, { "epoch": 0.0029332681495966755, "grad_norm": 11.283382415771484, "learning_rate": 1.951219512195122e-06, "loss": 2.5843, "step": 12 }, { "epoch": 0.0031777071620630653, "grad_norm": 14.134821891784668, "learning_rate": 2.1138211382113824e-06, "loss": 2.6906, "step": 13 }, { "epoch": 0.0034221461745294547, "grad_norm": 9.865968704223633, "learning_rate": 2.2764227642276426e-06, "loss": 2.6543, "step": 14 }, { "epoch": 0.0036665851869958446, "grad_norm": 6.860537052154541, "learning_rate": 2.4390243902439027e-06, "loss": 2.4197, "step": 15 }, { "epoch": 0.003911024199462234, "grad_norm": 6.645073890686035, "learning_rate": 2.601626016260163e-06, "loss": 2.2959, "step": 16 }, { "epoch": 0.004155463211928624, "grad_norm": 26.320695877075195, "learning_rate": 2.764227642276423e-06, "loss": 2.1923, "step": 17 }, { "epoch": 0.004399902224395014, "grad_norm": 8.401859283447266, "learning_rate": 2.926829268292683e-06, "loss": 2.4102, "step": 18 }, { "epoch": 0.0046443412368614035, "grad_norm": 6.93703556060791, "learning_rate": 3.0894308943089435e-06, "loss": 2.2069, "step": 19 }, { "epoch": 0.0048887802493277925, "grad_norm": 8.31473445892334, "learning_rate": 3.2520325203252037e-06, "loss": 2.0198, "step": 20 }, { "epoch": 0.005133219261794182, "grad_norm": 5.992040634155273, "learning_rate": 3.414634146341464e-06, "loss": 2.1501, "step": 21 }, { "epoch": 0.005377658274260572, "grad_norm": 5.812566757202148, "learning_rate": 3.577235772357724e-06, "loss": 2.2029, "step": 22 }, { "epoch": 0.005622097286726962, "grad_norm": 6.239164352416992, "learning_rate": 3.7398373983739838e-06, "loss": 2.0681, "step": 23 }, { "epoch": 0.005866536299193351, "grad_norm": 4.529574871063232, "learning_rate": 3.902439024390244e-06, "loss": 2.2424, "step": 24 }, { "epoch": 0.006110975311659741, "grad_norm": 5.873737335205078, "learning_rate": 4.0650406504065046e-06, "loss": 1.997, "step": 25 }, { "epoch": 0.006355414324126131, "grad_norm": 5.107316017150879, "learning_rate": 4.227642276422765e-06, "loss": 1.9068, "step": 26 }, { "epoch": 0.0065998533365925205, "grad_norm": 5.841495513916016, "learning_rate": 4.390243902439025e-06, "loss": 1.9078, "step": 27 }, { "epoch": 0.0068442923490589095, "grad_norm": 6.587307929992676, "learning_rate": 4.552845528455285e-06, "loss": 1.9157, "step": 28 }, { "epoch": 0.007088731361525299, "grad_norm": 7.674979209899902, "learning_rate": 4.715447154471545e-06, "loss": 1.9183, "step": 29 }, { "epoch": 0.007333170373991689, "grad_norm": 4.5966691970825195, "learning_rate": 4.8780487804878055e-06, "loss": 2.2402, "step": 30 }, { "epoch": 0.007577609386458079, "grad_norm": 6.864888668060303, "learning_rate": 5.040650406504065e-06, "loss": 1.7592, "step": 31 }, { "epoch": 0.007822048398924468, "grad_norm": 5.788065433502197, "learning_rate": 5.203252032520326e-06, "loss": 1.9031, "step": 32 }, { "epoch": 0.008066487411390858, "grad_norm": 5.776228904724121, "learning_rate": 5.365853658536586e-06, "loss": 1.8899, "step": 33 }, { "epoch": 0.008310926423857248, "grad_norm": 5.260758399963379, "learning_rate": 5.528455284552846e-06, "loss": 1.6202, "step": 34 }, { "epoch": 0.008555365436323637, "grad_norm": 5.431663990020752, "learning_rate": 5.691056910569106e-06, "loss": 1.871, "step": 35 }, { "epoch": 0.008799804448790027, "grad_norm": 5.790746688842773, "learning_rate": 5.853658536585366e-06, "loss": 1.6639, "step": 36 }, { "epoch": 0.009044243461256417, "grad_norm": 5.473887920379639, "learning_rate": 6.016260162601627e-06, "loss": 1.8165, "step": 37 }, { "epoch": 0.009288682473722807, "grad_norm": 5.433403968811035, "learning_rate": 6.178861788617887e-06, "loss": 1.6229, "step": 38 }, { "epoch": 0.009533121486189195, "grad_norm": 11.307058334350586, "learning_rate": 6.341463414634147e-06, "loss": 1.8198, "step": 39 }, { "epoch": 0.009777560498655585, "grad_norm": 5.7803568840026855, "learning_rate": 6.504065040650407e-06, "loss": 1.7236, "step": 40 }, { "epoch": 0.010021999511121975, "grad_norm": 8.382798194885254, "learning_rate": 6.666666666666667e-06, "loss": 1.6812, "step": 41 }, { "epoch": 0.010266438523588365, "grad_norm": 6.12075138092041, "learning_rate": 6.829268292682928e-06, "loss": 1.6711, "step": 42 }, { "epoch": 0.010510877536054754, "grad_norm": 6.378519535064697, "learning_rate": 6.991869918699188e-06, "loss": 1.5586, "step": 43 }, { "epoch": 0.010755316548521144, "grad_norm": 8.325606346130371, "learning_rate": 7.154471544715448e-06, "loss": 1.7467, "step": 44 }, { "epoch": 0.010999755560987534, "grad_norm": 5.427236557006836, "learning_rate": 7.317073170731707e-06, "loss": 1.561, "step": 45 }, { "epoch": 0.011244194573453924, "grad_norm": 5.533506870269775, "learning_rate": 7.4796747967479676e-06, "loss": 1.6548, "step": 46 }, { "epoch": 0.011488633585920312, "grad_norm": 11.618112564086914, "learning_rate": 7.64227642276423e-06, "loss": 1.5789, "step": 47 }, { "epoch": 0.011733072598386702, "grad_norm": 6.130812644958496, "learning_rate": 7.804878048780489e-06, "loss": 1.518, "step": 48 }, { "epoch": 0.011977511610853092, "grad_norm": 6.173574447631836, "learning_rate": 7.967479674796748e-06, "loss": 1.7447, "step": 49 }, { "epoch": 0.012221950623319482, "grad_norm": 5.371044158935547, "learning_rate": 8.130081300813009e-06, "loss": 1.7225, "step": 50 }, { "epoch": 0.012466389635785871, "grad_norm": 6.262181282043457, "learning_rate": 8.292682926829268e-06, "loss": 1.5532, "step": 51 }, { "epoch": 0.012710828648252261, "grad_norm": 6.0705671310424805, "learning_rate": 8.45528455284553e-06, "loss": 1.5654, "step": 52 }, { "epoch": 0.012955267660718651, "grad_norm": 7.2067694664001465, "learning_rate": 8.617886178861789e-06, "loss": 1.4073, "step": 53 }, { "epoch": 0.013199706673185041, "grad_norm": 5.209422588348389, "learning_rate": 8.78048780487805e-06, "loss": 1.6653, "step": 54 }, { "epoch": 0.01344414568565143, "grad_norm": 14.538374900817871, "learning_rate": 8.94308943089431e-06, "loss": 1.547, "step": 55 }, { "epoch": 0.013688584698117819, "grad_norm": 6.08756685256958, "learning_rate": 9.10569105691057e-06, "loss": 1.5763, "step": 56 }, { "epoch": 0.013933023710584209, "grad_norm": 5.936624050140381, "learning_rate": 9.268292682926831e-06, "loss": 1.6264, "step": 57 }, { "epoch": 0.014177462723050599, "grad_norm": 4.346157550811768, "learning_rate": 9.43089430894309e-06, "loss": 1.4903, "step": 58 }, { "epoch": 0.014421901735516988, "grad_norm": 5.19631814956665, "learning_rate": 9.59349593495935e-06, "loss": 1.4802, "step": 59 }, { "epoch": 0.014666340747983378, "grad_norm": 6.647718906402588, "learning_rate": 9.756097560975611e-06, "loss": 2.2965, "step": 60 }, { "epoch": 0.014910779760449768, "grad_norm": 4.844165802001953, "learning_rate": 9.91869918699187e-06, "loss": 1.4015, "step": 61 }, { "epoch": 0.015155218772916158, "grad_norm": 5.066519737243652, "learning_rate": 1.008130081300813e-05, "loss": 1.5471, "step": 62 }, { "epoch": 0.015399657785382548, "grad_norm": 6.175212383270264, "learning_rate": 1.024390243902439e-05, "loss": 1.5756, "step": 63 }, { "epoch": 0.015644096797848936, "grad_norm": 5.052133083343506, "learning_rate": 1.0406504065040652e-05, "loss": 1.4716, "step": 64 }, { "epoch": 0.015888535810315327, "grad_norm": 6.0629353523254395, "learning_rate": 1.0569105691056911e-05, "loss": 1.668, "step": 65 }, { "epoch": 0.016132974822781716, "grad_norm": 6.251193046569824, "learning_rate": 1.0731707317073172e-05, "loss": 1.4667, "step": 66 }, { "epoch": 0.016377413835248107, "grad_norm": 7.878139495849609, "learning_rate": 1.0894308943089431e-05, "loss": 1.5591, "step": 67 }, { "epoch": 0.016621852847714495, "grad_norm": 5.429769039154053, "learning_rate": 1.1056910569105692e-05, "loss": 1.4326, "step": 68 }, { "epoch": 0.016866291860180883, "grad_norm": 6.907913684844971, "learning_rate": 1.1219512195121953e-05, "loss": 1.5594, "step": 69 }, { "epoch": 0.017110730872647275, "grad_norm": 5.1678290367126465, "learning_rate": 1.1382113821138213e-05, "loss": 1.5511, "step": 70 }, { "epoch": 0.017355169885113663, "grad_norm": 5.462667465209961, "learning_rate": 1.1544715447154474e-05, "loss": 1.3895, "step": 71 }, { "epoch": 0.017599608897580055, "grad_norm": 6.608217239379883, "learning_rate": 1.1707317073170731e-05, "loss": 1.487, "step": 72 }, { "epoch": 0.017844047910046443, "grad_norm": 5.475932598114014, "learning_rate": 1.1869918699186992e-05, "loss": 1.5315, "step": 73 }, { "epoch": 0.018088486922512834, "grad_norm": 7.1279802322387695, "learning_rate": 1.2032520325203254e-05, "loss": 1.3711, "step": 74 }, { "epoch": 0.018332925934979222, "grad_norm": 6.620932579040527, "learning_rate": 1.2195121951219513e-05, "loss": 1.4346, "step": 75 }, { "epoch": 0.018577364947445614, "grad_norm": 6.356578826904297, "learning_rate": 1.2357723577235774e-05, "loss": 1.4521, "step": 76 }, { "epoch": 0.018821803959912002, "grad_norm": 5.530941486358643, "learning_rate": 1.2520325203252033e-05, "loss": 1.472, "step": 77 }, { "epoch": 0.01906624297237839, "grad_norm": 7.7910261154174805, "learning_rate": 1.2682926829268294e-05, "loss": 1.4296, "step": 78 }, { "epoch": 0.019310681984844782, "grad_norm": 7.19402551651001, "learning_rate": 1.2845528455284555e-05, "loss": 1.4176, "step": 79 }, { "epoch": 0.01955512099731117, "grad_norm": 6.1136040687561035, "learning_rate": 1.3008130081300815e-05, "loss": 1.4165, "step": 80 }, { "epoch": 0.01979956000977756, "grad_norm": 5.990948677062988, "learning_rate": 1.3170731707317076e-05, "loss": 1.4218, "step": 81 }, { "epoch": 0.02004399902224395, "grad_norm": 8.119118690490723, "learning_rate": 1.3333333333333333e-05, "loss": 1.4373, "step": 82 }, { "epoch": 0.02028843803471034, "grad_norm": 6.3870954513549805, "learning_rate": 1.3495934959349594e-05, "loss": 1.3869, "step": 83 }, { "epoch": 0.02053287704717673, "grad_norm": 8.236477851867676, "learning_rate": 1.3658536585365855e-05, "loss": 1.4277, "step": 84 }, { "epoch": 0.020777316059643117, "grad_norm": 17.77801513671875, "learning_rate": 1.3821138211382115e-05, "loss": 1.308, "step": 85 }, { "epoch": 0.02102175507210951, "grad_norm": 6.244637489318848, "learning_rate": 1.3983739837398376e-05, "loss": 1.3706, "step": 86 }, { "epoch": 0.021266194084575897, "grad_norm": 6.249431610107422, "learning_rate": 1.4146341463414635e-05, "loss": 1.3358, "step": 87 }, { "epoch": 0.02151063309704229, "grad_norm": 4.701333999633789, "learning_rate": 1.4308943089430896e-05, "loss": 2.2497, "step": 88 }, { "epoch": 0.021755072109508677, "grad_norm": 5.575697898864746, "learning_rate": 1.4471544715447157e-05, "loss": 1.4671, "step": 89 }, { "epoch": 0.02199951112197507, "grad_norm": 5.8239922523498535, "learning_rate": 1.4634146341463415e-05, "loss": 1.4833, "step": 90 }, { "epoch": 0.022243950134441456, "grad_norm": 6.4747185707092285, "learning_rate": 1.4796747967479676e-05, "loss": 1.5262, "step": 91 }, { "epoch": 0.022488389146907848, "grad_norm": 6.724368095397949, "learning_rate": 1.4959349593495935e-05, "loss": 1.422, "step": 92 }, { "epoch": 0.022732828159374236, "grad_norm": 5.154726982116699, "learning_rate": 1.5121951219512196e-05, "loss": 1.3532, "step": 93 }, { "epoch": 0.022977267171840624, "grad_norm": 4.542200565338135, "learning_rate": 1.528455284552846e-05, "loss": 2.1428, "step": 94 }, { "epoch": 0.023221706184307016, "grad_norm": 5.880817890167236, "learning_rate": 1.5447154471544717e-05, "loss": 1.5996, "step": 95 }, { "epoch": 0.023466145196773404, "grad_norm": 6.723649501800537, "learning_rate": 1.5609756097560978e-05, "loss": 1.5054, "step": 96 }, { "epoch": 0.023710584209239795, "grad_norm": 5.2248382568359375, "learning_rate": 1.5772357723577235e-05, "loss": 1.3787, "step": 97 }, { "epoch": 0.023955023221706184, "grad_norm": 5.563326835632324, "learning_rate": 1.5934959349593496e-05, "loss": 1.3925, "step": 98 }, { "epoch": 0.024199462234172575, "grad_norm": 6.171076774597168, "learning_rate": 1.6097560975609757e-05, "loss": 1.502, "step": 99 }, { "epoch": 0.024443901246638963, "grad_norm": 11.536863327026367, "learning_rate": 1.6260162601626018e-05, "loss": 1.4156, "step": 100 }, { "epoch": 0.024688340259105355, "grad_norm": 11.269830703735352, "learning_rate": 1.642276422764228e-05, "loss": 1.3481, "step": 101 }, { "epoch": 0.024932779271571743, "grad_norm": 7.356262683868408, "learning_rate": 1.6585365853658537e-05, "loss": 1.4403, "step": 102 }, { "epoch": 0.02517721828403813, "grad_norm": 10.883296012878418, "learning_rate": 1.6747967479674798e-05, "loss": 2.2085, "step": 103 }, { "epoch": 0.025421657296504523, "grad_norm": 6.563252925872803, "learning_rate": 1.691056910569106e-05, "loss": 1.4681, "step": 104 }, { "epoch": 0.02566609630897091, "grad_norm": 7.629971027374268, "learning_rate": 1.7073170731707317e-05, "loss": 1.4373, "step": 105 }, { "epoch": 0.025910535321437302, "grad_norm": 12.355507850646973, "learning_rate": 1.7235772357723578e-05, "loss": 1.4394, "step": 106 }, { "epoch": 0.02615497433390369, "grad_norm": 6.847367286682129, "learning_rate": 1.739837398373984e-05, "loss": 1.4561, "step": 107 }, { "epoch": 0.026399413346370082, "grad_norm": 7.732123374938965, "learning_rate": 1.75609756097561e-05, "loss": 1.4949, "step": 108 }, { "epoch": 0.02664385235883647, "grad_norm": 7.845799922943115, "learning_rate": 1.772357723577236e-05, "loss": 1.5389, "step": 109 }, { "epoch": 0.02688829137130286, "grad_norm": 5.964730739593506, "learning_rate": 1.788617886178862e-05, "loss": 1.4011, "step": 110 }, { "epoch": 0.02713273038376925, "grad_norm": 6.563551425933838, "learning_rate": 1.804878048780488e-05, "loss": 1.3924, "step": 111 }, { "epoch": 0.027377169396235638, "grad_norm": 7.457088947296143, "learning_rate": 1.821138211382114e-05, "loss": 1.5348, "step": 112 }, { "epoch": 0.02762160840870203, "grad_norm": 5.761719703674316, "learning_rate": 1.83739837398374e-05, "loss": 1.3819, "step": 113 }, { "epoch": 0.027866047421168418, "grad_norm": 10.146748542785645, "learning_rate": 1.8536585365853663e-05, "loss": 1.417, "step": 114 }, { "epoch": 0.02811048643363481, "grad_norm": 5.558394432067871, "learning_rate": 1.869918699186992e-05, "loss": 1.4842, "step": 115 }, { "epoch": 0.028354925446101197, "grad_norm": 8.84795093536377, "learning_rate": 1.886178861788618e-05, "loss": 1.3574, "step": 116 }, { "epoch": 0.02859936445856759, "grad_norm": 5.490208148956299, "learning_rate": 1.902439024390244e-05, "loss": 1.357, "step": 117 }, { "epoch": 0.028843803471033977, "grad_norm": 7.372227191925049, "learning_rate": 1.91869918699187e-05, "loss": 1.3379, "step": 118 }, { "epoch": 0.029088242483500365, "grad_norm": 11.52920913696289, "learning_rate": 1.934959349593496e-05, "loss": 1.4243, "step": 119 }, { "epoch": 0.029332681495966757, "grad_norm": 6.755467414855957, "learning_rate": 1.9512195121951222e-05, "loss": 1.5187, "step": 120 }, { "epoch": 0.029577120508433145, "grad_norm": 5.93511438369751, "learning_rate": 1.9674796747967483e-05, "loss": 1.3194, "step": 121 }, { "epoch": 0.029821559520899536, "grad_norm": 4.53206729888916, "learning_rate": 1.983739837398374e-05, "loss": 1.3059, "step": 122 }, { "epoch": 0.030065998533365924, "grad_norm": 6.463418483734131, "learning_rate": 2e-05, "loss": 1.2321, "step": 123 }, { "epoch": 0.030310437545832316, "grad_norm": 5.359195232391357, "learning_rate": 1.9999996865802212e-05, "loss": 1.3775, "step": 124 }, { "epoch": 0.030554876558298704, "grad_norm": 7.653247356414795, "learning_rate": 1.9999987463210805e-05, "loss": 1.4078, "step": 125 }, { "epoch": 0.030799315570765096, "grad_norm": 5.883333683013916, "learning_rate": 1.9999971792231676e-05, "loss": 1.4596, "step": 126 }, { "epoch": 0.031043754583231484, "grad_norm": 6.219335079193115, "learning_rate": 1.999994985287465e-05, "loss": 1.3468, "step": 127 }, { "epoch": 0.03128819359569787, "grad_norm": 9.816319465637207, "learning_rate": 1.999992164515348e-05, "loss": 1.2856, "step": 128 }, { "epoch": 0.03153263260816426, "grad_norm": 7.612930774688721, "learning_rate": 1.999988716908584e-05, "loss": 1.3644, "step": 129 }, { "epoch": 0.031777071620630655, "grad_norm": 7.602503299713135, "learning_rate": 1.999984642469335e-05, "loss": 1.4115, "step": 130 }, { "epoch": 0.03202151063309704, "grad_norm": 5.727510452270508, "learning_rate": 1.9999799412001547e-05, "loss": 1.228, "step": 131 }, { "epoch": 0.03226594964556343, "grad_norm": 5.721832752227783, "learning_rate": 1.99997461310399e-05, "loss": 1.335, "step": 132 }, { "epoch": 0.03251038865802982, "grad_norm": 6.187972068786621, "learning_rate": 1.9999686581841805e-05, "loss": 1.339, "step": 133 }, { "epoch": 0.032754827670496214, "grad_norm": 5.254082202911377, "learning_rate": 1.9999620764444594e-05, "loss": 1.2741, "step": 134 }, { "epoch": 0.0329992666829626, "grad_norm": 6.423061370849609, "learning_rate": 1.9999548678889523e-05, "loss": 1.3004, "step": 135 }, { "epoch": 0.03324370569542899, "grad_norm": 4.9685959815979, "learning_rate": 1.9999470325221778e-05, "loss": 1.3251, "step": 136 }, { "epoch": 0.03348814470789538, "grad_norm": 5.46399450302124, "learning_rate": 1.999938570349047e-05, "loss": 1.3795, "step": 137 }, { "epoch": 0.03373258372036177, "grad_norm": 4.732756614685059, "learning_rate": 1.999929481374865e-05, "loss": 1.2958, "step": 138 }, { "epoch": 0.03397702273282816, "grad_norm": 4.804397106170654, "learning_rate": 1.9999197656053288e-05, "loss": 1.2627, "step": 139 }, { "epoch": 0.03422146174529455, "grad_norm": 4.775611400604248, "learning_rate": 1.9999094230465285e-05, "loss": 1.4391, "step": 140 }, { "epoch": 0.03446590075776094, "grad_norm": 4.6918721199035645, "learning_rate": 1.9998984537049476e-05, "loss": 1.4239, "step": 141 }, { "epoch": 0.034710339770227326, "grad_norm": 4.3264899253845215, "learning_rate": 1.9998868575874616e-05, "loss": 1.1658, "step": 142 }, { "epoch": 0.03495477878269372, "grad_norm": 4.351634502410889, "learning_rate": 1.99987463470134e-05, "loss": 1.3883, "step": 143 }, { "epoch": 0.03519921779516011, "grad_norm": 5.136422634124756, "learning_rate": 1.999861785054244e-05, "loss": 1.438, "step": 144 }, { "epoch": 0.0354436568076265, "grad_norm": 5.345577716827393, "learning_rate": 1.999848308654229e-05, "loss": 1.4144, "step": 145 }, { "epoch": 0.035688095820092886, "grad_norm": 5.832770824432373, "learning_rate": 1.9998342055097416e-05, "loss": 1.299, "step": 146 }, { "epoch": 0.035932534832559274, "grad_norm": 7.734096527099609, "learning_rate": 1.999819475629623e-05, "loss": 1.4003, "step": 147 }, { "epoch": 0.03617697384502567, "grad_norm": 5.463601589202881, "learning_rate": 1.9998041190231063e-05, "loss": 1.5057, "step": 148 }, { "epoch": 0.03642141285749206, "grad_norm": 8.983975410461426, "learning_rate": 1.9997881356998172e-05, "loss": 1.3619, "step": 149 }, { "epoch": 0.036665851869958445, "grad_norm": 5.394047737121582, "learning_rate": 1.999771525669775e-05, "loss": 1.3032, "step": 150 }, { "epoch": 0.03691029088242483, "grad_norm": 4.557994842529297, "learning_rate": 1.9997542889433917e-05, "loss": 1.2587, "step": 151 }, { "epoch": 0.03715472989489123, "grad_norm": 4.662408351898193, "learning_rate": 1.9997364255314715e-05, "loss": 1.3185, "step": 152 }, { "epoch": 0.037399168907357616, "grad_norm": 7.458238124847412, "learning_rate": 1.9997179354452126e-05, "loss": 1.2708, "step": 153 }, { "epoch": 0.037643607919824004, "grad_norm": 4.436248302459717, "learning_rate": 1.9996988186962044e-05, "loss": 1.3901, "step": 154 }, { "epoch": 0.03788804693229039, "grad_norm": 5.528069972991943, "learning_rate": 1.9996790752964305e-05, "loss": 1.4166, "step": 155 }, { "epoch": 0.03813248594475678, "grad_norm": 4.3356733322143555, "learning_rate": 1.9996587052582672e-05, "loss": 1.152, "step": 156 }, { "epoch": 0.038376924957223175, "grad_norm": 5.925254821777344, "learning_rate": 1.9996377085944826e-05, "loss": 1.5081, "step": 157 }, { "epoch": 0.038621363969689564, "grad_norm": 4.844107151031494, "learning_rate": 1.999616085318239e-05, "loss": 1.3824, "step": 158 }, { "epoch": 0.03886580298215595, "grad_norm": 6.188292026519775, "learning_rate": 1.9995938354430898e-05, "loss": 1.3108, "step": 159 }, { "epoch": 0.03911024199462234, "grad_norm": 5.222842216491699, "learning_rate": 1.9995709589829828e-05, "loss": 1.433, "step": 160 }, { "epoch": 0.03935468100708873, "grad_norm": 6.492764949798584, "learning_rate": 1.9995474559522576e-05, "loss": 1.3901, "step": 161 }, { "epoch": 0.03959912001955512, "grad_norm": 6.479674339294434, "learning_rate": 1.999523326365647e-05, "loss": 1.3002, "step": 162 }, { "epoch": 0.03984355903202151, "grad_norm": 4.043074607849121, "learning_rate": 1.999498570238276e-05, "loss": 2.238, "step": 163 }, { "epoch": 0.0400879980444879, "grad_norm": 8.806992530822754, "learning_rate": 1.999473187585663e-05, "loss": 1.3225, "step": 164 }, { "epoch": 0.04033243705695429, "grad_norm": 5.6869964599609375, "learning_rate": 1.9994471784237188e-05, "loss": 1.3458, "step": 165 }, { "epoch": 0.04057687606942068, "grad_norm": 7.028317451477051, "learning_rate": 1.9994205427687473e-05, "loss": 1.2942, "step": 166 }, { "epoch": 0.04082131508188707, "grad_norm": 5.3759684562683105, "learning_rate": 1.999393280637444e-05, "loss": 1.3008, "step": 167 }, { "epoch": 0.04106575409435346, "grad_norm": 4.7559943199157715, "learning_rate": 1.999365392046899e-05, "loss": 1.3133, "step": 168 }, { "epoch": 0.04131019310681985, "grad_norm": 4.474522590637207, "learning_rate": 1.9993368770145927e-05, "loss": 1.3373, "step": 169 }, { "epoch": 0.041554632119286235, "grad_norm": 6.905185699462891, "learning_rate": 1.9993077355584e-05, "loss": 1.4465, "step": 170 }, { "epoch": 0.04179907113175263, "grad_norm": 5.875603675842285, "learning_rate": 1.9992779676965884e-05, "loss": 1.2736, "step": 171 }, { "epoch": 0.04204351014421902, "grad_norm": 3.587334156036377, "learning_rate": 1.999247573447817e-05, "loss": 2.1556, "step": 172 }, { "epoch": 0.042287949156685406, "grad_norm": 6.450631141662598, "learning_rate": 1.9992165528311385e-05, "loss": 1.33, "step": 173 }, { "epoch": 0.042532388169151794, "grad_norm": 6.471398830413818, "learning_rate": 1.9991849058659977e-05, "loss": 1.2372, "step": 174 }, { "epoch": 0.04277682718161819, "grad_norm": 6.2006425857543945, "learning_rate": 1.9991526325722316e-05, "loss": 1.2998, "step": 175 }, { "epoch": 0.04302126619408458, "grad_norm": 4.926456928253174, "learning_rate": 1.999119732970071e-05, "loss": 1.3577, "step": 176 }, { "epoch": 0.043265705206550965, "grad_norm": 5.1743292808532715, "learning_rate": 1.9990862070801388e-05, "loss": 1.3454, "step": 177 }, { "epoch": 0.04351014421901735, "grad_norm": 5.5031890869140625, "learning_rate": 1.9990520549234502e-05, "loss": 1.3274, "step": 178 }, { "epoch": 0.04375458323148374, "grad_norm": 16.7782039642334, "learning_rate": 1.999017276521413e-05, "loss": 1.3566, "step": 179 }, { "epoch": 0.04399902224395014, "grad_norm": 5.242146015167236, "learning_rate": 1.9989818718958275e-05, "loss": 1.3043, "step": 180 }, { "epoch": 0.044243461256416525, "grad_norm": 5.954908847808838, "learning_rate": 1.9989458410688865e-05, "loss": 1.2932, "step": 181 }, { "epoch": 0.04448790026888291, "grad_norm": 4.562159061431885, "learning_rate": 1.9989091840631767e-05, "loss": 1.269, "step": 182 }, { "epoch": 0.0447323392813493, "grad_norm": 5.205554485321045, "learning_rate": 1.9988719009016753e-05, "loss": 1.2356, "step": 183 }, { "epoch": 0.044976778293815696, "grad_norm": 6.116245746612549, "learning_rate": 1.998833991607753e-05, "loss": 1.3568, "step": 184 }, { "epoch": 0.045221217306282084, "grad_norm": 3.484712600708008, "learning_rate": 1.9987954562051724e-05, "loss": 2.1914, "step": 185 }, { "epoch": 0.04546565631874847, "grad_norm": 4.402046203613281, "learning_rate": 1.99875629471809e-05, "loss": 1.33, "step": 186 }, { "epoch": 0.04571009533121486, "grad_norm": 6.1173415184021, "learning_rate": 1.998716507171053e-05, "loss": 1.3433, "step": 187 }, { "epoch": 0.04595453434368125, "grad_norm": 4.56729793548584, "learning_rate": 1.998676093589002e-05, "loss": 1.2448, "step": 188 }, { "epoch": 0.04619897335614764, "grad_norm": 2.8044798374176025, "learning_rate": 1.99863505399727e-05, "loss": 2.1563, "step": 189 }, { "epoch": 0.04644341236861403, "grad_norm": 5.482027530670166, "learning_rate": 1.998593388421582e-05, "loss": 1.3307, "step": 190 }, { "epoch": 0.04668785138108042, "grad_norm": 6.116431713104248, "learning_rate": 1.9985510968880555e-05, "loss": 1.3009, "step": 191 }, { "epoch": 0.04693229039354681, "grad_norm": 5.399773120880127, "learning_rate": 1.9985081794232014e-05, "loss": 1.3656, "step": 192 }, { "epoch": 0.0471767294060132, "grad_norm": 4.578831195831299, "learning_rate": 1.998464636053921e-05, "loss": 1.3241, "step": 193 }, { "epoch": 0.04742116841847959, "grad_norm": 5.7253618240356445, "learning_rate": 1.9984204668075094e-05, "loss": 1.1836, "step": 194 }, { "epoch": 0.04766560743094598, "grad_norm": 5.640674114227295, "learning_rate": 1.9983756717116534e-05, "loss": 1.3452, "step": 195 }, { "epoch": 0.04791004644341237, "grad_norm": 5.7181291580200195, "learning_rate": 1.998330250794433e-05, "loss": 1.2925, "step": 196 }, { "epoch": 0.048154485455878755, "grad_norm": 5.187880039215088, "learning_rate": 1.9982842040843193e-05, "loss": 1.3884, "step": 197 }, { "epoch": 0.04839892446834515, "grad_norm": 5.154764175415039, "learning_rate": 1.9982375316101764e-05, "loss": 1.3073, "step": 198 }, { "epoch": 0.04864336348081154, "grad_norm": 3.99664568901062, "learning_rate": 1.99819023340126e-05, "loss": 2.2408, "step": 199 }, { "epoch": 0.048887802493277926, "grad_norm": 4.218594074249268, "learning_rate": 1.998142309487219e-05, "loss": 1.319, "step": 200 }, { "epoch": 0.049132241505744315, "grad_norm": 4.567066669464111, "learning_rate": 1.9980937598980943e-05, "loss": 1.2632, "step": 201 }, { "epoch": 0.04937668051821071, "grad_norm": 4.567723751068115, "learning_rate": 1.9980445846643176e-05, "loss": 1.1766, "step": 202 }, { "epoch": 0.0496211195306771, "grad_norm": 4.89946985244751, "learning_rate": 1.9979947838167152e-05, "loss": 1.3318, "step": 203 }, { "epoch": 0.049865558543143486, "grad_norm": 7.729551792144775, "learning_rate": 1.9979443573865033e-05, "loss": 1.3179, "step": 204 }, { "epoch": 0.050109997555609874, "grad_norm": 7.349613189697266, "learning_rate": 1.9978933054052913e-05, "loss": 1.2955, "step": 205 }, { "epoch": 0.05035443656807626, "grad_norm": 4.26453161239624, "learning_rate": 1.997841627905081e-05, "loss": 1.3304, "step": 206 }, { "epoch": 0.05059887558054266, "grad_norm": 7.018398284912109, "learning_rate": 1.9977893249182654e-05, "loss": 1.3471, "step": 207 }, { "epoch": 0.050843314593009045, "grad_norm": 6.137728691101074, "learning_rate": 1.9977363964776305e-05, "loss": 1.3536, "step": 208 }, { "epoch": 0.05108775360547543, "grad_norm": 5.061336994171143, "learning_rate": 1.9976828426163538e-05, "loss": 1.3652, "step": 209 }, { "epoch": 0.05133219261794182, "grad_norm": 5.005093574523926, "learning_rate": 1.9976286633680048e-05, "loss": 1.3201, "step": 210 }, { "epoch": 0.051576631630408216, "grad_norm": 3.8842194080352783, "learning_rate": 1.9975738587665455e-05, "loss": 1.3596, "step": 211 }, { "epoch": 0.051821070642874605, "grad_norm": 5.012754440307617, "learning_rate": 1.9975184288463295e-05, "loss": 1.2887, "step": 212 }, { "epoch": 0.05206550965534099, "grad_norm": 4.832705497741699, "learning_rate": 1.9974623736421025e-05, "loss": 1.3391, "step": 213 }, { "epoch": 0.05230994866780738, "grad_norm": 5.078959941864014, "learning_rate": 1.9974056931890018e-05, "loss": 1.2961, "step": 214 }, { "epoch": 0.05255438768027377, "grad_norm": 4.353578090667725, "learning_rate": 1.9973483875225573e-05, "loss": 1.2887, "step": 215 }, { "epoch": 0.052798826692740164, "grad_norm": 7.96591329574585, "learning_rate": 1.9972904566786903e-05, "loss": 1.3251, "step": 216 }, { "epoch": 0.05304326570520655, "grad_norm": 5.235666275024414, "learning_rate": 1.9972319006937143e-05, "loss": 1.2725, "step": 217 }, { "epoch": 0.05328770471767294, "grad_norm": 4.645522594451904, "learning_rate": 1.9971727196043344e-05, "loss": 1.3218, "step": 218 }, { "epoch": 0.05353214373013933, "grad_norm": 7.124841690063477, "learning_rate": 1.9971129134476474e-05, "loss": 1.3267, "step": 219 }, { "epoch": 0.05377658274260572, "grad_norm": 6.27811861038208, "learning_rate": 1.997052482261143e-05, "loss": 1.2694, "step": 220 }, { "epoch": 0.05402102175507211, "grad_norm": 6.613264560699463, "learning_rate": 1.996991426082701e-05, "loss": 1.2389, "step": 221 }, { "epoch": 0.0542654607675385, "grad_norm": 5.72642183303833, "learning_rate": 1.996929744950594e-05, "loss": 1.2542, "step": 222 }, { "epoch": 0.05450989978000489, "grad_norm": 6.274509906768799, "learning_rate": 1.996867438903486e-05, "loss": 1.3864, "step": 223 }, { "epoch": 0.054754338792471276, "grad_norm": 4.827954292297363, "learning_rate": 1.9968045079804333e-05, "loss": 1.311, "step": 224 }, { "epoch": 0.05499877780493767, "grad_norm": 4.8781280517578125, "learning_rate": 1.9967409522208835e-05, "loss": 1.2287, "step": 225 }, { "epoch": 0.05524321681740406, "grad_norm": 4.974234104156494, "learning_rate": 1.996676771664676e-05, "loss": 1.3661, "step": 226 }, { "epoch": 0.05548765582987045, "grad_norm": 4.70881986618042, "learning_rate": 1.996611966352041e-05, "loss": 1.2792, "step": 227 }, { "epoch": 0.055732094842336835, "grad_norm": 4.771082401275635, "learning_rate": 1.996546536323602e-05, "loss": 1.3459, "step": 228 }, { "epoch": 0.05597653385480322, "grad_norm": 4.687952995300293, "learning_rate": 1.9964804816203723e-05, "loss": 1.2545, "step": 229 }, { "epoch": 0.05622097286726962, "grad_norm": 3.7861790657043457, "learning_rate": 1.9964138022837575e-05, "loss": 2.1981, "step": 230 }, { "epoch": 0.056465411879736006, "grad_norm": 6.665339946746826, "learning_rate": 1.9963464983555557e-05, "loss": 1.4111, "step": 231 }, { "epoch": 0.056709850892202394, "grad_norm": 5.215484142303467, "learning_rate": 1.996278569877955e-05, "loss": 1.3479, "step": 232 }, { "epoch": 0.05695428990466878, "grad_norm": 3.6600303649902344, "learning_rate": 1.996210016893536e-05, "loss": 1.3194, "step": 233 }, { "epoch": 0.05719872891713518, "grad_norm": 6.554864883422852, "learning_rate": 1.9961408394452703e-05, "loss": 1.2998, "step": 234 }, { "epoch": 0.057443167929601566, "grad_norm": 6.014383316040039, "learning_rate": 1.9960710375765212e-05, "loss": 1.2345, "step": 235 }, { "epoch": 0.057687606942067954, "grad_norm": 6.356429576873779, "learning_rate": 1.9960006113310428e-05, "loss": 1.2493, "step": 236 }, { "epoch": 0.05793204595453434, "grad_norm": 10.219972610473633, "learning_rate": 1.9959295607529818e-05, "loss": 1.2908, "step": 237 }, { "epoch": 0.05817648496700073, "grad_norm": 4.4042558670043945, "learning_rate": 1.9958578858868747e-05, "loss": 1.3068, "step": 238 }, { "epoch": 0.058420923979467125, "grad_norm": 4.574854373931885, "learning_rate": 1.9957855867776507e-05, "loss": 1.3013, "step": 239 }, { "epoch": 0.05866536299193351, "grad_norm": 4.1353960037231445, "learning_rate": 1.9957126634706294e-05, "loss": 1.2734, "step": 240 }, { "epoch": 0.0589098020043999, "grad_norm": 4.964991092681885, "learning_rate": 1.9956391160115224e-05, "loss": 1.1964, "step": 241 }, { "epoch": 0.05915424101686629, "grad_norm": 6.201467514038086, "learning_rate": 1.995564944446432e-05, "loss": 1.2947, "step": 242 }, { "epoch": 0.059398680029332684, "grad_norm": 5.684882164001465, "learning_rate": 1.9954901488218515e-05, "loss": 1.3692, "step": 243 }, { "epoch": 0.05964311904179907, "grad_norm": 4.244503974914551, "learning_rate": 1.9954147291846663e-05, "loss": 2.2702, "step": 244 }, { "epoch": 0.05988755805426546, "grad_norm": 6.967065334320068, "learning_rate": 1.9953386855821523e-05, "loss": 1.2285, "step": 245 }, { "epoch": 0.06013199706673185, "grad_norm": 5.459507942199707, "learning_rate": 1.9952620180619765e-05, "loss": 1.266, "step": 246 }, { "epoch": 0.06037643607919824, "grad_norm": 7.973972320556641, "learning_rate": 1.995184726672197e-05, "loss": 1.3239, "step": 247 }, { "epoch": 0.06062087509166463, "grad_norm": 4.386531829833984, "learning_rate": 1.9951068114612637e-05, "loss": 1.3246, "step": 248 }, { "epoch": 0.06086531410413102, "grad_norm": 4.626567840576172, "learning_rate": 1.995028272478016e-05, "loss": 1.3018, "step": 249 }, { "epoch": 0.06110975311659741, "grad_norm": 4.793629169464111, "learning_rate": 1.9949491097716856e-05, "loss": 1.3344, "step": 250 }, { "epoch": 0.061354192129063796, "grad_norm": 6.8952412605285645, "learning_rate": 1.994869323391895e-05, "loss": 1.3457, "step": 251 }, { "epoch": 0.06159863114153019, "grad_norm": 6.089010238647461, "learning_rate": 1.994788913388658e-05, "loss": 1.3226, "step": 252 }, { "epoch": 0.06184307015399658, "grad_norm": 6.344317436218262, "learning_rate": 1.9947078798123778e-05, "loss": 1.2385, "step": 253 }, { "epoch": 0.06208750916646297, "grad_norm": 12.935145378112793, "learning_rate": 1.99462622271385e-05, "loss": 1.3256, "step": 254 }, { "epoch": 0.062331948178929356, "grad_norm": 4.718111991882324, "learning_rate": 1.9945439421442602e-05, "loss": 1.2985, "step": 255 }, { "epoch": 0.06257638719139574, "grad_norm": 5.535119533538818, "learning_rate": 1.9944610381551852e-05, "loss": 1.2133, "step": 256 }, { "epoch": 0.06282082620386213, "grad_norm": 4.996159076690674, "learning_rate": 1.9943775107985925e-05, "loss": 1.3943, "step": 257 }, { "epoch": 0.06306526521632852, "grad_norm": 5.818681240081787, "learning_rate": 1.9942933601268403e-05, "loss": 1.2941, "step": 258 }, { "epoch": 0.06330970422879492, "grad_norm": 5.313794136047363, "learning_rate": 1.994208586192678e-05, "loss": 1.225, "step": 259 }, { "epoch": 0.06355414324126131, "grad_norm": 5.420829772949219, "learning_rate": 1.994123189049245e-05, "loss": 1.4065, "step": 260 }, { "epoch": 0.0637985822537277, "grad_norm": 7.252939224243164, "learning_rate": 1.9940371687500713e-05, "loss": 1.2559, "step": 261 }, { "epoch": 0.06404302126619409, "grad_norm": 7.47566556930542, "learning_rate": 1.993950525349078e-05, "loss": 1.2258, "step": 262 }, { "epoch": 0.06428746027866047, "grad_norm": 6.010402679443359, "learning_rate": 1.993863258900577e-05, "loss": 1.2764, "step": 263 }, { "epoch": 0.06453189929112686, "grad_norm": 5.5107808113098145, "learning_rate": 1.9937753694592698e-05, "loss": 1.3388, "step": 264 }, { "epoch": 0.06477633830359325, "grad_norm": 4.4688334465026855, "learning_rate": 1.993686857080249e-05, "loss": 1.2992, "step": 265 }, { "epoch": 0.06502077731605964, "grad_norm": 4.5571818351745605, "learning_rate": 1.9935977218189982e-05, "loss": 1.127, "step": 266 }, { "epoch": 0.06526521632852603, "grad_norm": 4.901233673095703, "learning_rate": 1.9935079637313906e-05, "loss": 1.213, "step": 267 }, { "epoch": 0.06550965534099243, "grad_norm": 6.362276077270508, "learning_rate": 1.9934175828736898e-05, "loss": 1.3096, "step": 268 }, { "epoch": 0.06575409435345882, "grad_norm": 4.998250484466553, "learning_rate": 1.993326579302551e-05, "loss": 1.2048, "step": 269 }, { "epoch": 0.0659985333659252, "grad_norm": 3.9181408882141113, "learning_rate": 1.9932349530750173e-05, "loss": 1.1552, "step": 270 }, { "epoch": 0.06624297237839159, "grad_norm": 10.106731414794922, "learning_rate": 1.9931427042485252e-05, "loss": 1.1628, "step": 271 }, { "epoch": 0.06648741139085798, "grad_norm": 4.483226299285889, "learning_rate": 1.993049832880899e-05, "loss": 1.2237, "step": 272 }, { "epoch": 0.06673185040332437, "grad_norm": 7.024012088775635, "learning_rate": 1.992956339030355e-05, "loss": 1.2936, "step": 273 }, { "epoch": 0.06697628941579076, "grad_norm": 5.0647430419921875, "learning_rate": 1.992862222755498e-05, "loss": 1.3156, "step": 274 }, { "epoch": 0.06722072842825715, "grad_norm": 6.234722137451172, "learning_rate": 1.992767484115324e-05, "loss": 1.2481, "step": 275 }, { "epoch": 0.06746516744072353, "grad_norm": 5.06250524520874, "learning_rate": 1.992672123169219e-05, "loss": 1.177, "step": 276 }, { "epoch": 0.06770960645318994, "grad_norm": 6.165891647338867, "learning_rate": 1.992576139976959e-05, "loss": 1.2062, "step": 277 }, { "epoch": 0.06795404546565632, "grad_norm": 5.269135475158691, "learning_rate": 1.9924795345987103e-05, "loss": 1.2643, "step": 278 }, { "epoch": 0.06819848447812271, "grad_norm": 10.178529739379883, "learning_rate": 1.9923823070950282e-05, "loss": 1.3677, "step": 279 }, { "epoch": 0.0684429234905891, "grad_norm": 6.4800310134887695, "learning_rate": 1.99228445752686e-05, "loss": 1.2403, "step": 280 }, { "epoch": 0.06868736250305549, "grad_norm": 5.289521217346191, "learning_rate": 1.992185985955541e-05, "loss": 1.2473, "step": 281 }, { "epoch": 0.06893180151552188, "grad_norm": 4.943880081176758, "learning_rate": 1.9920868924427964e-05, "loss": 1.2794, "step": 282 }, { "epoch": 0.06917624052798826, "grad_norm": 6.226718425750732, "learning_rate": 1.991987177050743e-05, "loss": 1.3859, "step": 283 }, { "epoch": 0.06942067954045465, "grad_norm": 4.328200817108154, "learning_rate": 1.991886839841886e-05, "loss": 1.2455, "step": 284 }, { "epoch": 0.06966511855292104, "grad_norm": 4.170522689819336, "learning_rate": 1.9917858808791205e-05, "loss": 1.2064, "step": 285 }, { "epoch": 0.06990955756538744, "grad_norm": 4.448747158050537, "learning_rate": 1.9916843002257322e-05, "loss": 1.3702, "step": 286 }, { "epoch": 0.07015399657785383, "grad_norm": 3.4251625537872314, "learning_rate": 1.991582097945395e-05, "loss": 1.2967, "step": 287 }, { "epoch": 0.07039843559032022, "grad_norm": 4.255934715270996, "learning_rate": 1.9914792741021742e-05, "loss": 1.2506, "step": 288 }, { "epoch": 0.0706428746027866, "grad_norm": 3.717904567718506, "learning_rate": 1.9913758287605227e-05, "loss": 1.2631, "step": 289 }, { "epoch": 0.070887313615253, "grad_norm": 3.819291353225708, "learning_rate": 1.9912717619852856e-05, "loss": 1.2166, "step": 290 }, { "epoch": 0.07113175262771938, "grad_norm": 3.6381824016571045, "learning_rate": 1.991167073841695e-05, "loss": 1.1999, "step": 291 }, { "epoch": 0.07137619164018577, "grad_norm": 5.513364791870117, "learning_rate": 1.9910617643953737e-05, "loss": 1.2485, "step": 292 }, { "epoch": 0.07162063065265216, "grad_norm": 7.550871849060059, "learning_rate": 1.990955833712334e-05, "loss": 1.1929, "step": 293 }, { "epoch": 0.07186506966511855, "grad_norm": 5.627939224243164, "learning_rate": 1.9908492818589776e-05, "loss": 1.3323, "step": 294 }, { "epoch": 0.07210950867758495, "grad_norm": 3.952164649963379, "learning_rate": 1.9907421089020956e-05, "loss": 1.2111, "step": 295 }, { "epoch": 0.07235394769005134, "grad_norm": 5.329583168029785, "learning_rate": 1.9906343149088675e-05, "loss": 1.2314, "step": 296 }, { "epoch": 0.07259838670251773, "grad_norm": 8.233551025390625, "learning_rate": 1.9905258999468634e-05, "loss": 1.2911, "step": 297 }, { "epoch": 0.07284282571498411, "grad_norm": 5.11754035949707, "learning_rate": 1.990416864084042e-05, "loss": 1.189, "step": 298 }, { "epoch": 0.0730872647274505, "grad_norm": 6.454300880432129, "learning_rate": 1.9903072073887507e-05, "loss": 1.1986, "step": 299 }, { "epoch": 0.07333170373991689, "grad_norm": 4.868954181671143, "learning_rate": 1.9901969299297278e-05, "loss": 1.2302, "step": 300 }, { "epoch": 0.07357614275238328, "grad_norm": 4.498685359954834, "learning_rate": 1.990086031776099e-05, "loss": 1.1852, "step": 301 }, { "epoch": 0.07382058176484967, "grad_norm": 7.261220932006836, "learning_rate": 1.9899745129973792e-05, "loss": 1.261, "step": 302 }, { "epoch": 0.07406502077731605, "grad_norm": 3.6971499919891357, "learning_rate": 1.9898623736634738e-05, "loss": 1.1989, "step": 303 }, { "epoch": 0.07430945978978246, "grad_norm": 10.59734058380127, "learning_rate": 1.989749613844675e-05, "loss": 1.2304, "step": 304 }, { "epoch": 0.07455389880224884, "grad_norm": 4.786809921264648, "learning_rate": 1.989636233611666e-05, "loss": 1.2734, "step": 305 }, { "epoch": 0.07479833781471523, "grad_norm": 3.672269821166992, "learning_rate": 1.9895222330355174e-05, "loss": 1.2867, "step": 306 }, { "epoch": 0.07504277682718162, "grad_norm": 4.053799629211426, "learning_rate": 1.98940761218769e-05, "loss": 2.2605, "step": 307 }, { "epoch": 0.07528721583964801, "grad_norm": 5.109869480133057, "learning_rate": 1.989292371140032e-05, "loss": 1.2306, "step": 308 }, { "epoch": 0.0755316548521144, "grad_norm": 3.5280559062957764, "learning_rate": 1.989176509964781e-05, "loss": 1.2821, "step": 309 }, { "epoch": 0.07577609386458078, "grad_norm": 5.775945663452148, "learning_rate": 1.989060028734564e-05, "loss": 1.2577, "step": 310 }, { "epoch": 0.07602053287704717, "grad_norm": 4.2715864181518555, "learning_rate": 1.9889429275223958e-05, "loss": 1.2305, "step": 311 }, { "epoch": 0.07626497188951356, "grad_norm": 3.606166124343872, "learning_rate": 1.9888252064016798e-05, "loss": 1.2248, "step": 312 }, { "epoch": 0.07650941090197996, "grad_norm": 5.557638168334961, "learning_rate": 1.9887068654462088e-05, "loss": 1.2033, "step": 313 }, { "epoch": 0.07675384991444635, "grad_norm": 3.988684892654419, "learning_rate": 1.9885879047301628e-05, "loss": 2.2842, "step": 314 }, { "epoch": 0.07699828892691274, "grad_norm": 4.706129550933838, "learning_rate": 1.9884683243281117e-05, "loss": 1.257, "step": 315 }, { "epoch": 0.07724272793937913, "grad_norm": 8.253270149230957, "learning_rate": 1.988348124315013e-05, "loss": 1.3223, "step": 316 }, { "epoch": 0.07748716695184552, "grad_norm": 4.516219615936279, "learning_rate": 1.9882273047662128e-05, "loss": 1.2565, "step": 317 }, { "epoch": 0.0777316059643119, "grad_norm": 3.8688483238220215, "learning_rate": 1.9881058657574456e-05, "loss": 1.2674, "step": 318 }, { "epoch": 0.07797604497677829, "grad_norm": 4.626121520996094, "learning_rate": 1.9879838073648342e-05, "loss": 1.2049, "step": 319 }, { "epoch": 0.07822048398924468, "grad_norm": 5.420572757720947, "learning_rate": 1.98786112966489e-05, "loss": 1.2808, "step": 320 }, { "epoch": 0.07846492300171107, "grad_norm": 5.866894721984863, "learning_rate": 1.9877378327345115e-05, "loss": 1.3138, "step": 321 }, { "epoch": 0.07870936201417746, "grad_norm": 5.478062629699707, "learning_rate": 1.9876139166509863e-05, "loss": 1.2282, "step": 322 }, { "epoch": 0.07895380102664386, "grad_norm": 4.921603202819824, "learning_rate": 1.9874893814919908e-05, "loss": 1.2234, "step": 323 }, { "epoch": 0.07919824003911025, "grad_norm": 5.043973445892334, "learning_rate": 1.9873642273355873e-05, "loss": 1.1653, "step": 324 }, { "epoch": 0.07944267905157663, "grad_norm": 5.5380144119262695, "learning_rate": 1.9872384542602278e-05, "loss": 1.2197, "step": 325 }, { "epoch": 0.07968711806404302, "grad_norm": 4.780150413513184, "learning_rate": 1.987112062344752e-05, "loss": 1.1963, "step": 326 }, { "epoch": 0.07993155707650941, "grad_norm": 5.572015285491943, "learning_rate": 1.9869850516683875e-05, "loss": 1.15, "step": 327 }, { "epoch": 0.0801759960889758, "grad_norm": 6.987370491027832, "learning_rate": 1.9868574223107493e-05, "loss": 1.1748, "step": 328 }, { "epoch": 0.08042043510144219, "grad_norm": 6.550912380218506, "learning_rate": 1.9867291743518407e-05, "loss": 1.2382, "step": 329 }, { "epoch": 0.08066487411390857, "grad_norm": 6.358066082000732, "learning_rate": 1.9866003078720526e-05, "loss": 1.2318, "step": 330 }, { "epoch": 0.08090931312637496, "grad_norm": 8.014790534973145, "learning_rate": 1.9864708229521637e-05, "loss": 1.2417, "step": 331 }, { "epoch": 0.08115375213884136, "grad_norm": 6.014158248901367, "learning_rate": 1.9863407196733396e-05, "loss": 1.0414, "step": 332 }, { "epoch": 0.08139819115130775, "grad_norm": 6.411825180053711, "learning_rate": 1.986209998117135e-05, "loss": 1.2771, "step": 333 }, { "epoch": 0.08164263016377414, "grad_norm": 6.666935920715332, "learning_rate": 1.9860786583654913e-05, "loss": 1.3119, "step": 334 }, { "epoch": 0.08188706917624053, "grad_norm": 5.211427688598633, "learning_rate": 1.985946700500737e-05, "loss": 1.2362, "step": 335 }, { "epoch": 0.08213150818870692, "grad_norm": 5.023791313171387, "learning_rate": 1.9858141246055883e-05, "loss": 1.2924, "step": 336 }, { "epoch": 0.0823759472011733, "grad_norm": 4.761742115020752, "learning_rate": 1.9856809307631498e-05, "loss": 2.2858, "step": 337 }, { "epoch": 0.0826203862136397, "grad_norm": 5.831111907958984, "learning_rate": 1.9855471190569125e-05, "loss": 1.1394, "step": 338 }, { "epoch": 0.08286482522610608, "grad_norm": 5.819721698760986, "learning_rate": 1.985412689570754e-05, "loss": 1.2416, "step": 339 }, { "epoch": 0.08310926423857247, "grad_norm": 5.4629645347595215, "learning_rate": 1.9852776423889414e-05, "loss": 1.3436, "step": 340 }, { "epoch": 0.08335370325103887, "grad_norm": 5.885761260986328, "learning_rate": 1.9851419775961265e-05, "loss": 1.1212, "step": 341 }, { "epoch": 0.08359814226350526, "grad_norm": 3.1790401935577393, "learning_rate": 1.98500569527735e-05, "loss": 2.1646, "step": 342 }, { "epoch": 0.08384258127597165, "grad_norm": 6.93152379989624, "learning_rate": 1.9848687955180386e-05, "loss": 1.1123, "step": 343 }, { "epoch": 0.08408702028843804, "grad_norm": 6.305123805999756, "learning_rate": 1.9847312784040068e-05, "loss": 1.2377, "step": 344 }, { "epoch": 0.08433145930090442, "grad_norm": 11.483536720275879, "learning_rate": 1.9845931440214557e-05, "loss": 1.1454, "step": 345 }, { "epoch": 0.08457589831337081, "grad_norm": 4.77037239074707, "learning_rate": 1.984454392456973e-05, "loss": 1.3617, "step": 346 }, { "epoch": 0.0848203373258372, "grad_norm": 4.1926984786987305, "learning_rate": 1.9843150237975343e-05, "loss": 1.2383, "step": 347 }, { "epoch": 0.08506477633830359, "grad_norm": 4.398230075836182, "learning_rate": 1.984175038130501e-05, "loss": 1.2658, "step": 348 }, { "epoch": 0.08530921535076998, "grad_norm": 6.484598159790039, "learning_rate": 1.984034435543622e-05, "loss": 1.1849, "step": 349 }, { "epoch": 0.08555365436323638, "grad_norm": 4.679854393005371, "learning_rate": 1.9838932161250318e-05, "loss": 1.255, "step": 350 }, { "epoch": 0.08579809337570277, "grad_norm": 6.1571807861328125, "learning_rate": 1.9837513799632536e-05, "loss": 1.2073, "step": 351 }, { "epoch": 0.08604253238816915, "grad_norm": 5.911917686462402, "learning_rate": 1.9836089271471947e-05, "loss": 1.2561, "step": 352 }, { "epoch": 0.08628697140063554, "grad_norm": 4.896956920623779, "learning_rate": 1.9834658577661505e-05, "loss": 1.2099, "step": 353 }, { "epoch": 0.08653141041310193, "grad_norm": 4.391876220703125, "learning_rate": 1.983322171909803e-05, "loss": 1.304, "step": 354 }, { "epoch": 0.08677584942556832, "grad_norm": 4.319279670715332, "learning_rate": 1.9831778696682195e-05, "loss": 1.0834, "step": 355 }, { "epoch": 0.0870202884380347, "grad_norm": 4.364986896514893, "learning_rate": 1.983032951131855e-05, "loss": 1.1473, "step": 356 }, { "epoch": 0.0872647274505011, "grad_norm": 4.4287028312683105, "learning_rate": 1.9828874163915496e-05, "loss": 1.211, "step": 357 }, { "epoch": 0.08750916646296748, "grad_norm": 5.178832530975342, "learning_rate": 1.9827412655385308e-05, "loss": 1.2318, "step": 358 }, { "epoch": 0.08775360547543389, "grad_norm": 3.4217987060546875, "learning_rate": 1.9825944986644116e-05, "loss": 1.0763, "step": 359 }, { "epoch": 0.08799804448790027, "grad_norm": 6.738720417022705, "learning_rate": 1.9824471158611907e-05, "loss": 1.2174, "step": 360 }, { "epoch": 0.08824248350036666, "grad_norm": 4.281075954437256, "learning_rate": 1.982299117221254e-05, "loss": 1.1825, "step": 361 }, { "epoch": 0.08848692251283305, "grad_norm": 3.974607229232788, "learning_rate": 1.9821505028373727e-05, "loss": 1.1445, "step": 362 }, { "epoch": 0.08873136152529944, "grad_norm": 4.769040584564209, "learning_rate": 1.9820012728027044e-05, "loss": 1.128, "step": 363 }, { "epoch": 0.08897580053776583, "grad_norm": 4.660463333129883, "learning_rate": 1.9818514272107923e-05, "loss": 1.1685, "step": 364 }, { "epoch": 0.08922023955023221, "grad_norm": 4.7089338302612305, "learning_rate": 1.9817009661555654e-05, "loss": 1.2943, "step": 365 }, { "epoch": 0.0894646785626986, "grad_norm": 3.7988078594207764, "learning_rate": 1.9815498897313387e-05, "loss": 1.0695, "step": 366 }, { "epoch": 0.08970911757516499, "grad_norm": 4.49351692199707, "learning_rate": 1.981398198032813e-05, "loss": 1.2104, "step": 367 }, { "epoch": 0.08995355658763139, "grad_norm": 4.02097749710083, "learning_rate": 1.9812458911550748e-05, "loss": 1.2177, "step": 368 }, { "epoch": 0.09019799560009778, "grad_norm": 4.062883377075195, "learning_rate": 1.9810929691935956e-05, "loss": 1.1625, "step": 369 }, { "epoch": 0.09044243461256417, "grad_norm": 4.29163932800293, "learning_rate": 1.9809394322442333e-05, "loss": 1.1793, "step": 370 }, { "epoch": 0.09068687362503056, "grad_norm": 9.156668663024902, "learning_rate": 1.9807852804032306e-05, "loss": 1.1871, "step": 371 }, { "epoch": 0.09093131263749694, "grad_norm": 5.449754238128662, "learning_rate": 1.9806305137672165e-05, "loss": 1.1717, "step": 372 }, { "epoch": 0.09117575164996333, "grad_norm": 6.121524810791016, "learning_rate": 1.9804751324332043e-05, "loss": 1.0906, "step": 373 }, { "epoch": 0.09142019066242972, "grad_norm": 6.13816499710083, "learning_rate": 1.9803191364985937e-05, "loss": 1.1783, "step": 374 }, { "epoch": 0.09166462967489611, "grad_norm": 4.645517826080322, "learning_rate": 1.9801625260611684e-05, "loss": 1.1945, "step": 375 }, { "epoch": 0.0919090686873625, "grad_norm": 5.3838396072387695, "learning_rate": 1.9800053012190988e-05, "loss": 1.2118, "step": 376 }, { "epoch": 0.0921535076998289, "grad_norm": 9.308674812316895, "learning_rate": 1.9798474620709395e-05, "loss": 1.1243, "step": 377 }, { "epoch": 0.09239794671229529, "grad_norm": 4.490804672241211, "learning_rate": 1.9796890087156295e-05, "loss": 1.3031, "step": 378 }, { "epoch": 0.09264238572476167, "grad_norm": 4.418246746063232, "learning_rate": 1.9795299412524948e-05, "loss": 1.2028, "step": 379 }, { "epoch": 0.09288682473722806, "grad_norm": 4.358879089355469, "learning_rate": 1.9793702597812444e-05, "loss": 1.0912, "step": 380 }, { "epoch": 0.09313126374969445, "grad_norm": 4.636160850524902, "learning_rate": 1.979209964401973e-05, "loss": 1.1982, "step": 381 }, { "epoch": 0.09337570276216084, "grad_norm": 4.515685558319092, "learning_rate": 1.9790490552151605e-05, "loss": 1.1392, "step": 382 }, { "epoch": 0.09362014177462723, "grad_norm": 4.175724029541016, "learning_rate": 1.978887532321671e-05, "loss": 1.2269, "step": 383 }, { "epoch": 0.09386458078709362, "grad_norm": 3.605635404586792, "learning_rate": 1.9787253958227534e-05, "loss": 1.2386, "step": 384 }, { "epoch": 0.09410901979956, "grad_norm": 4.983905792236328, "learning_rate": 1.9785626458200413e-05, "loss": 1.2003, "step": 385 }, { "epoch": 0.0943534588120264, "grad_norm": 3.675171375274658, "learning_rate": 1.9783992824155523e-05, "loss": 1.2266, "step": 386 }, { "epoch": 0.0945978978244928, "grad_norm": 4.82205867767334, "learning_rate": 1.9782353057116902e-05, "loss": 1.1161, "step": 387 }, { "epoch": 0.09484233683695918, "grad_norm": 4.621769428253174, "learning_rate": 1.978070715811241e-05, "loss": 1.1658, "step": 388 }, { "epoch": 0.09508677584942557, "grad_norm": 4.2349982261657715, "learning_rate": 1.9779055128173766e-05, "loss": 1.269, "step": 389 }, { "epoch": 0.09533121486189196, "grad_norm": 3.450382947921753, "learning_rate": 1.9777396968336526e-05, "loss": 1.169, "step": 390 }, { "epoch": 0.09557565387435835, "grad_norm": 3.7534186840057373, "learning_rate": 1.9775732679640093e-05, "loss": 1.1282, "step": 391 }, { "epoch": 0.09582009288682473, "grad_norm": 4.103036880493164, "learning_rate": 1.9774062263127706e-05, "loss": 1.2434, "step": 392 }, { "epoch": 0.09606453189929112, "grad_norm": 4.361753463745117, "learning_rate": 1.9772385719846454e-05, "loss": 1.15, "step": 393 }, { "epoch": 0.09630897091175751, "grad_norm": 4.32541561126709, "learning_rate": 1.9770703050847254e-05, "loss": 1.2142, "step": 394 }, { "epoch": 0.09655340992422391, "grad_norm": 3.983166456222534, "learning_rate": 1.976901425718487e-05, "loss": 1.1632, "step": 395 }, { "epoch": 0.0967978489366903, "grad_norm": 3.596562385559082, "learning_rate": 1.9767319339917908e-05, "loss": 1.1292, "step": 396 }, { "epoch": 0.09704228794915669, "grad_norm": 4.378901481628418, "learning_rate": 1.976561830010881e-05, "loss": 1.2467, "step": 397 }, { "epoch": 0.09728672696162308, "grad_norm": 4.227049827575684, "learning_rate": 1.976391113882385e-05, "loss": 1.166, "step": 398 }, { "epoch": 0.09753116597408946, "grad_norm": 5.524839401245117, "learning_rate": 1.9762197857133148e-05, "loss": 1.2571, "step": 399 }, { "epoch": 0.09777560498655585, "grad_norm": 6.738865852355957, "learning_rate": 1.9760478456110655e-05, "loss": 1.23, "step": 400 }, { "epoch": 0.09802004399902224, "grad_norm": 5.322516918182373, "learning_rate": 1.975875293683416e-05, "loss": 1.2279, "step": 401 }, { "epoch": 0.09826448301148863, "grad_norm": 4.052256107330322, "learning_rate": 1.9757021300385288e-05, "loss": 1.2415, "step": 402 }, { "epoch": 0.09850892202395502, "grad_norm": 3.557264566421509, "learning_rate": 1.9755283547849496e-05, "loss": 1.2575, "step": 403 }, { "epoch": 0.09875336103642142, "grad_norm": 5.424764156341553, "learning_rate": 1.9753539680316077e-05, "loss": 1.2741, "step": 404 }, { "epoch": 0.09899780004888781, "grad_norm": 4.176656246185303, "learning_rate": 1.9751789698878153e-05, "loss": 1.2557, "step": 405 }, { "epoch": 0.0992422390613542, "grad_norm": 3.171238660812378, "learning_rate": 1.9750033604632683e-05, "loss": 1.0859, "step": 406 }, { "epoch": 0.09948667807382058, "grad_norm": 3.339172840118408, "learning_rate": 1.974827139868046e-05, "loss": 1.2761, "step": 407 }, { "epoch": 0.09973111708628697, "grad_norm": 3.803142786026001, "learning_rate": 1.9746503082126102e-05, "loss": 1.1748, "step": 408 }, { "epoch": 0.09997555609875336, "grad_norm": 2.872415781021118, "learning_rate": 1.974472865607806e-05, "loss": 2.1624, "step": 409 }, { "epoch": 0.10021999511121975, "grad_norm": 4.578094959259033, "learning_rate": 1.9742948121648614e-05, "loss": 1.1665, "step": 410 }, { "epoch": 0.10046443412368614, "grad_norm": 4.066554546356201, "learning_rate": 1.9741161479953872e-05, "loss": 1.2893, "step": 411 }, { "epoch": 0.10070887313615252, "grad_norm": 3.227421760559082, "learning_rate": 1.9739368732113773e-05, "loss": 1.2414, "step": 412 }, { "epoch": 0.10095331214861893, "grad_norm": 4.6151580810546875, "learning_rate": 1.9737569879252082e-05, "loss": 1.1905, "step": 413 }, { "epoch": 0.10119775116108531, "grad_norm": 3.6246869564056396, "learning_rate": 1.9735764922496393e-05, "loss": 1.2155, "step": 414 }, { "epoch": 0.1014421901735517, "grad_norm": 3.8012173175811768, "learning_rate": 1.9733953862978125e-05, "loss": 1.1976, "step": 415 }, { "epoch": 0.10168662918601809, "grad_norm": 6.171485424041748, "learning_rate": 1.9732136701832515e-05, "loss": 1.2181, "step": 416 }, { "epoch": 0.10193106819848448, "grad_norm": 3.9184229373931885, "learning_rate": 1.9730313440198637e-05, "loss": 1.156, "step": 417 }, { "epoch": 0.10217550721095087, "grad_norm": 4.018677711486816, "learning_rate": 1.9728484079219385e-05, "loss": 1.0827, "step": 418 }, { "epoch": 0.10241994622341725, "grad_norm": 3.981010675430298, "learning_rate": 1.972664862004147e-05, "loss": 1.315, "step": 419 }, { "epoch": 0.10266438523588364, "grad_norm": 3.377622604370117, "learning_rate": 1.9724807063815432e-05, "loss": 1.1715, "step": 420 }, { "epoch": 0.10290882424835003, "grad_norm": 2.6135737895965576, "learning_rate": 1.9722959411695636e-05, "loss": 2.1113, "step": 421 }, { "epoch": 0.10315326326081643, "grad_norm": 4.2256550788879395, "learning_rate": 1.9721105664840253e-05, "loss": 1.1828, "step": 422 }, { "epoch": 0.10339770227328282, "grad_norm": 4.993865966796875, "learning_rate": 1.9719245824411298e-05, "loss": 1.1503, "step": 423 }, { "epoch": 0.10364214128574921, "grad_norm": 3.423783540725708, "learning_rate": 1.9717379891574583e-05, "loss": 1.0649, "step": 424 }, { "epoch": 0.1038865802982156, "grad_norm": 3.6675407886505127, "learning_rate": 1.9715507867499754e-05, "loss": 1.0481, "step": 425 }, { "epoch": 0.10413101931068199, "grad_norm": 3.9582297801971436, "learning_rate": 1.9713629753360265e-05, "loss": 1.1743, "step": 426 }, { "epoch": 0.10437545832314837, "grad_norm": 4.515329837799072, "learning_rate": 1.9711745550333392e-05, "loss": 1.1082, "step": 427 }, { "epoch": 0.10461989733561476, "grad_norm": 4.547237873077393, "learning_rate": 1.9709855259600236e-05, "loss": 1.2417, "step": 428 }, { "epoch": 0.10486433634808115, "grad_norm": 4.494861125946045, "learning_rate": 1.9707958882345693e-05, "loss": 1.1873, "step": 429 }, { "epoch": 0.10510877536054754, "grad_norm": 7.8653059005737305, "learning_rate": 1.97060564197585e-05, "loss": 1.1605, "step": 430 }, { "epoch": 0.10535321437301394, "grad_norm": 4.853890895843506, "learning_rate": 1.970414787303119e-05, "loss": 1.1746, "step": 431 }, { "epoch": 0.10559765338548033, "grad_norm": 3.662851095199585, "learning_rate": 1.9702233243360114e-05, "loss": 1.154, "step": 432 }, { "epoch": 0.10584209239794672, "grad_norm": 4.81683349609375, "learning_rate": 1.9700312531945444e-05, "loss": 1.3095, "step": 433 }, { "epoch": 0.1060865314104131, "grad_norm": 4.331052780151367, "learning_rate": 1.9698385739991147e-05, "loss": 1.1571, "step": 434 }, { "epoch": 0.10633097042287949, "grad_norm": 4.024257659912109, "learning_rate": 1.9696452868705026e-05, "loss": 1.1725, "step": 435 }, { "epoch": 0.10657540943534588, "grad_norm": 6.647060871124268, "learning_rate": 1.969451391929867e-05, "loss": 1.1375, "step": 436 }, { "epoch": 0.10681984844781227, "grad_norm": 4.554993152618408, "learning_rate": 1.9692568892987494e-05, "loss": 1.1222, "step": 437 }, { "epoch": 0.10706428746027866, "grad_norm": 3.893545627593994, "learning_rate": 1.9690617790990717e-05, "loss": 1.2332, "step": 438 }, { "epoch": 0.10730872647274504, "grad_norm": 2.6785714626312256, "learning_rate": 1.968866061453137e-05, "loss": 2.1153, "step": 439 }, { "epoch": 0.10755316548521145, "grad_norm": 5.624014854431152, "learning_rate": 1.9686697364836278e-05, "loss": 1.2113, "step": 440 }, { "epoch": 0.10779760449767783, "grad_norm": 4.759518146514893, "learning_rate": 1.9684728043136093e-05, "loss": 1.1685, "step": 441 }, { "epoch": 0.10804204351014422, "grad_norm": 4.094371318817139, "learning_rate": 1.9682752650665268e-05, "loss": 1.2696, "step": 442 }, { "epoch": 0.10828648252261061, "grad_norm": 4.044309139251709, "learning_rate": 1.9680771188662044e-05, "loss": 1.1713, "step": 443 }, { "epoch": 0.108530921535077, "grad_norm": 3.826819658279419, "learning_rate": 1.967878365836849e-05, "loss": 1.1808, "step": 444 }, { "epoch": 0.10877536054754339, "grad_norm": 4.536250591278076, "learning_rate": 1.9676790061030466e-05, "loss": 1.3316, "step": 445 }, { "epoch": 0.10901979956000978, "grad_norm": 3.299072742462158, "learning_rate": 1.9674790397897633e-05, "loss": 1.1167, "step": 446 }, { "epoch": 0.10926423857247616, "grad_norm": 3.9642040729522705, "learning_rate": 1.9672784670223465e-05, "loss": 1.2412, "step": 447 }, { "epoch": 0.10950867758494255, "grad_norm": 3.646501302719116, "learning_rate": 1.967077287926523e-05, "loss": 1.1585, "step": 448 }, { "epoch": 0.10975311659740894, "grad_norm": 4.142199993133545, "learning_rate": 1.9668755026283995e-05, "loss": 1.2492, "step": 449 }, { "epoch": 0.10999755560987534, "grad_norm": 4.070352554321289, "learning_rate": 1.9666731112544638e-05, "loss": 1.2055, "step": 450 }, { "epoch": 0.11024199462234173, "grad_norm": 3.4566304683685303, "learning_rate": 1.966470113931582e-05, "loss": 1.1793, "step": 451 }, { "epoch": 0.11048643363480812, "grad_norm": 3.860222578048706, "learning_rate": 1.9662665107870007e-05, "loss": 1.1789, "step": 452 }, { "epoch": 0.1107308726472745, "grad_norm": 3.7197413444519043, "learning_rate": 1.966062301948347e-05, "loss": 1.1849, "step": 453 }, { "epoch": 0.1109753116597409, "grad_norm": 4.1757965087890625, "learning_rate": 1.965857487543627e-05, "loss": 1.2825, "step": 454 }, { "epoch": 0.11121975067220728, "grad_norm": 4.345134258270264, "learning_rate": 1.9656520677012264e-05, "loss": 1.1732, "step": 455 }, { "epoch": 0.11146418968467367, "grad_norm": 4.474052429199219, "learning_rate": 1.965446042549911e-05, "loss": 1.1597, "step": 456 }, { "epoch": 0.11170862869714006, "grad_norm": 3.8579421043395996, "learning_rate": 1.9652394122188242e-05, "loss": 1.2017, "step": 457 }, { "epoch": 0.11195306770960645, "grad_norm": 5.020146369934082, "learning_rate": 1.9650321768374908e-05, "loss": 1.3404, "step": 458 }, { "epoch": 0.11219750672207285, "grad_norm": 4.499261379241943, "learning_rate": 1.9648243365358145e-05, "loss": 1.298, "step": 459 }, { "epoch": 0.11244194573453924, "grad_norm": 4.317763328552246, "learning_rate": 1.9646158914440775e-05, "loss": 1.2115, "step": 460 }, { "epoch": 0.11268638474700562, "grad_norm": 3.9400923252105713, "learning_rate": 1.9644068416929417e-05, "loss": 1.0254, "step": 461 }, { "epoch": 0.11293082375947201, "grad_norm": 4.948456764221191, "learning_rate": 1.964197187413447e-05, "loss": 1.3035, "step": 462 }, { "epoch": 0.1131752627719384, "grad_norm": 6.067734718322754, "learning_rate": 1.963986928737014e-05, "loss": 1.1723, "step": 463 }, { "epoch": 0.11341970178440479, "grad_norm": 3.6457936763763428, "learning_rate": 1.96377606579544e-05, "loss": 1.2246, "step": 464 }, { "epoch": 0.11366414079687118, "grad_norm": 4.135344505310059, "learning_rate": 1.963564598720903e-05, "loss": 1.1155, "step": 465 }, { "epoch": 0.11390857980933757, "grad_norm": 3.4885904788970947, "learning_rate": 1.963352527645959e-05, "loss": 1.198, "step": 466 }, { "epoch": 0.11415301882180395, "grad_norm": 5.2463483810424805, "learning_rate": 1.9631398527035424e-05, "loss": 1.1427, "step": 467 }, { "epoch": 0.11439745783427036, "grad_norm": 4.414389610290527, "learning_rate": 1.962926574026966e-05, "loss": 1.1839, "step": 468 }, { "epoch": 0.11464189684673674, "grad_norm": 3.498870849609375, "learning_rate": 1.962712691749922e-05, "loss": 1.1213, "step": 469 }, { "epoch": 0.11488633585920313, "grad_norm": 3.4490199089050293, "learning_rate": 1.962498206006479e-05, "loss": 2.1561, "step": 470 }, { "epoch": 0.11513077487166952, "grad_norm": 2.884737730026245, "learning_rate": 1.9622831169310864e-05, "loss": 1.133, "step": 471 }, { "epoch": 0.11537521388413591, "grad_norm": 4.203254699707031, "learning_rate": 1.96206742465857e-05, "loss": 1.2361, "step": 472 }, { "epoch": 0.1156196528966023, "grad_norm": 3.780165195465088, "learning_rate": 1.961851129324134e-05, "loss": 1.1643, "step": 473 }, { "epoch": 0.11586409190906868, "grad_norm": 6.365428447723389, "learning_rate": 1.9616342310633615e-05, "loss": 1.1947, "step": 474 }, { "epoch": 0.11610853092153507, "grad_norm": 4.483095169067383, "learning_rate": 1.9614167300122126e-05, "loss": 1.0783, "step": 475 }, { "epoch": 0.11635296993400146, "grad_norm": 4.36859130859375, "learning_rate": 1.9611986263070256e-05, "loss": 1.0763, "step": 476 }, { "epoch": 0.11659740894646786, "grad_norm": 6.242516994476318, "learning_rate": 1.9609799200845163e-05, "loss": 1.284, "step": 477 }, { "epoch": 0.11684184795893425, "grad_norm": 4.574044227600098, "learning_rate": 1.9607606114817787e-05, "loss": 1.0097, "step": 478 }, { "epoch": 0.11708628697140064, "grad_norm": 4.711122989654541, "learning_rate": 1.9605407006362843e-05, "loss": 1.221, "step": 479 }, { "epoch": 0.11733072598386703, "grad_norm": 4.41146993637085, "learning_rate": 1.9603201876858812e-05, "loss": 1.1135, "step": 480 }, { "epoch": 0.11757516499633341, "grad_norm": 3.8882482051849365, "learning_rate": 1.9600990727687964e-05, "loss": 1.1542, "step": 481 }, { "epoch": 0.1178196040087998, "grad_norm": 3.4748640060424805, "learning_rate": 1.959877356023633e-05, "loss": 1.0965, "step": 482 }, { "epoch": 0.11806404302126619, "grad_norm": 4.37790060043335, "learning_rate": 1.9596550375893718e-05, "loss": 1.1817, "step": 483 }, { "epoch": 0.11830848203373258, "grad_norm": 3.8884594440460205, "learning_rate": 1.959432117605371e-05, "loss": 1.2542, "step": 484 }, { "epoch": 0.11855292104619897, "grad_norm": 3.711740732192993, "learning_rate": 1.959208596211366e-05, "loss": 1.0646, "step": 485 }, { "epoch": 0.11879736005866537, "grad_norm": 3.075740337371826, "learning_rate": 1.958984473547468e-05, "loss": 2.1749, "step": 486 }, { "epoch": 0.11904179907113176, "grad_norm": 4.204500198364258, "learning_rate": 1.958759749754167e-05, "loss": 1.181, "step": 487 }, { "epoch": 0.11928623808359815, "grad_norm": 4.3053741455078125, "learning_rate": 1.958534424972328e-05, "loss": 1.1511, "step": 488 }, { "epoch": 0.11953067709606453, "grad_norm": 7.22095251083374, "learning_rate": 1.9583084993431935e-05, "loss": 1.2187, "step": 489 }, { "epoch": 0.11977511610853092, "grad_norm": 4.425522327423096, "learning_rate": 1.958081973008383e-05, "loss": 1.1437, "step": 490 }, { "epoch": 0.12001955512099731, "grad_norm": 4.114257335662842, "learning_rate": 1.9578548461098912e-05, "loss": 1.1693, "step": 491 }, { "epoch": 0.1202639941334637, "grad_norm": 4.519433498382568, "learning_rate": 1.957627118790092e-05, "loss": 1.2113, "step": 492 }, { "epoch": 0.12050843314593009, "grad_norm": 4.922794342041016, "learning_rate": 1.9573987911917325e-05, "loss": 1.0789, "step": 493 }, { "epoch": 0.12075287215839647, "grad_norm": 4.798068046569824, "learning_rate": 1.9571698634579377e-05, "loss": 1.2138, "step": 494 }, { "epoch": 0.12099731117086288, "grad_norm": 4.2914533615112305, "learning_rate": 1.956940335732209e-05, "loss": 1.1323, "step": 495 }, { "epoch": 0.12124175018332926, "grad_norm": 5.3744401931762695, "learning_rate": 1.956710208158423e-05, "loss": 1.2722, "step": 496 }, { "epoch": 0.12148618919579565, "grad_norm": 3.724177122116089, "learning_rate": 1.956479480880833e-05, "loss": 1.0054, "step": 497 }, { "epoch": 0.12173062820826204, "grad_norm": 22.230966567993164, "learning_rate": 1.9562481540440677e-05, "loss": 1.0581, "step": 498 }, { "epoch": 0.12197506722072843, "grad_norm": 5.926990509033203, "learning_rate": 1.9560162277931326e-05, "loss": 1.128, "step": 499 }, { "epoch": 0.12221950623319482, "grad_norm": 7.3720784187316895, "learning_rate": 1.9557837022734075e-05, "loss": 1.1306, "step": 500 }, { "epoch": 0.1224639452456612, "grad_norm": 14.880962371826172, "learning_rate": 1.9555505776306492e-05, "loss": 1.2239, "step": 501 }, { "epoch": 0.12270838425812759, "grad_norm": 4.3613200187683105, "learning_rate": 1.9553168540109888e-05, "loss": 1.1388, "step": 502 }, { "epoch": 0.12295282327059398, "grad_norm": 5.127346992492676, "learning_rate": 1.955082531560934e-05, "loss": 1.0958, "step": 503 }, { "epoch": 0.12319726228306038, "grad_norm": 4.747497081756592, "learning_rate": 1.954847610427367e-05, "loss": 1.0396, "step": 504 }, { "epoch": 0.12344170129552677, "grad_norm": 4.675597667694092, "learning_rate": 1.954612090757546e-05, "loss": 1.1731, "step": 505 }, { "epoch": 0.12368614030799316, "grad_norm": 34.22259521484375, "learning_rate": 1.9543759726991037e-05, "loss": 1.2323, "step": 506 }, { "epoch": 0.12393057932045955, "grad_norm": 6.397866725921631, "learning_rate": 1.954139256400049e-05, "loss": 0.9992, "step": 507 }, { "epoch": 0.12417501833292593, "grad_norm": 17.597700119018555, "learning_rate": 1.9539019420087643e-05, "loss": 1.1705, "step": 508 }, { "epoch": 0.12441945734539232, "grad_norm": 8.879899978637695, "learning_rate": 1.9536640296740078e-05, "loss": 1.249, "step": 509 }, { "epoch": 0.12466389635785871, "grad_norm": 14.198099136352539, "learning_rate": 1.9534255195449123e-05, "loss": 1.2276, "step": 510 }, { "epoch": 0.1249083353703251, "grad_norm": 6.117961883544922, "learning_rate": 1.9531864117709855e-05, "loss": 1.2236, "step": 511 }, { "epoch": 0.1251527743827915, "grad_norm": 7.427270412445068, "learning_rate": 1.9529467065021098e-05, "loss": 1.1852, "step": 512 }, { "epoch": 0.1253972133952579, "grad_norm": 8.58948802947998, "learning_rate": 1.952706403888542e-05, "loss": 1.1082, "step": 513 }, { "epoch": 0.12564165240772426, "grad_norm": 4.219729423522949, "learning_rate": 1.9524655040809126e-05, "loss": 1.1912, "step": 514 }, { "epoch": 0.12588609142019067, "grad_norm": 4.362913608551025, "learning_rate": 1.9522240072302275e-05, "loss": 1.1358, "step": 515 }, { "epoch": 0.12613053043265704, "grad_norm": 8.792346954345703, "learning_rate": 1.9519819134878668e-05, "loss": 1.1447, "step": 516 }, { "epoch": 0.12637496944512344, "grad_norm": 7.836843490600586, "learning_rate": 1.951739223005584e-05, "loss": 1.2419, "step": 517 }, { "epoch": 0.12661940845758984, "grad_norm": 4.472485542297363, "learning_rate": 1.9514959359355076e-05, "loss": 1.2295, "step": 518 }, { "epoch": 0.12686384747005622, "grad_norm": 5.330399513244629, "learning_rate": 1.9512520524301388e-05, "loss": 1.2025, "step": 519 }, { "epoch": 0.12710828648252262, "grad_norm": 5.496823310852051, "learning_rate": 1.951007572642354e-05, "loss": 2.1139, "step": 520 }, { "epoch": 0.127352725494989, "grad_norm": 5.915637016296387, "learning_rate": 1.950762496725403e-05, "loss": 1.2428, "step": 521 }, { "epoch": 0.1275971645074554, "grad_norm": 11.7316255569458, "learning_rate": 1.9505168248329084e-05, "loss": 1.1854, "step": 522 }, { "epoch": 0.12784160351992177, "grad_norm": 8.459104537963867, "learning_rate": 1.9502705571188675e-05, "loss": 1.2272, "step": 523 }, { "epoch": 0.12808604253238817, "grad_norm": 4.567411422729492, "learning_rate": 1.9500236937376504e-05, "loss": 1.0839, "step": 524 }, { "epoch": 0.12833048154485455, "grad_norm": 4.574172019958496, "learning_rate": 1.949776234844001e-05, "loss": 1.0274, "step": 525 }, { "epoch": 0.12857492055732095, "grad_norm": 7.426229476928711, "learning_rate": 1.949528180593037e-05, "loss": 1.1825, "step": 526 }, { "epoch": 0.12881935956978735, "grad_norm": 4.16442346572876, "learning_rate": 1.949279531140247e-05, "loss": 1.1525, "step": 527 }, { "epoch": 0.12906379858225372, "grad_norm": 4.447491645812988, "learning_rate": 1.949030286641496e-05, "loss": 1.2678, "step": 528 }, { "epoch": 0.12930823759472013, "grad_norm": 6.830371379852295, "learning_rate": 1.948780447253019e-05, "loss": 1.2301, "step": 529 }, { "epoch": 0.1295526766071865, "grad_norm": 3.7037405967712402, "learning_rate": 1.9485300131314262e-05, "loss": 2.1888, "step": 530 }, { "epoch": 0.1297971156196529, "grad_norm": 6.7080912590026855, "learning_rate": 1.948278984433699e-05, "loss": 1.3056, "step": 531 }, { "epoch": 0.13004155463211928, "grad_norm": 3.5571670532226562, "learning_rate": 1.948027361317192e-05, "loss": 2.1244, "step": 532 }, { "epoch": 0.13028599364458568, "grad_norm": 5.456795692443848, "learning_rate": 1.947775143939633e-05, "loss": 1.2952, "step": 533 }, { "epoch": 0.13053043265705205, "grad_norm": 6.878313064575195, "learning_rate": 1.947522332459122e-05, "loss": 1.18, "step": 534 }, { "epoch": 0.13077487166951846, "grad_norm": 4.375424385070801, "learning_rate": 1.9472689270341303e-05, "loss": 1.1559, "step": 535 }, { "epoch": 0.13101931068198486, "grad_norm": 6.418196201324463, "learning_rate": 1.9470149278235032e-05, "loss": 1.1485, "step": 536 }, { "epoch": 0.13126374969445123, "grad_norm": 4.910287380218506, "learning_rate": 1.946760334986458e-05, "loss": 1.0982, "step": 537 }, { "epoch": 0.13150818870691763, "grad_norm": 4.6932549476623535, "learning_rate": 1.946505148682582e-05, "loss": 1.1912, "step": 538 }, { "epoch": 0.131752627719384, "grad_norm": 5.659276962280273, "learning_rate": 1.9462493690718373e-05, "loss": 1.1048, "step": 539 }, { "epoch": 0.1319970667318504, "grad_norm": 7.565642833709717, "learning_rate": 1.9459929963145558e-05, "loss": 1.1765, "step": 540 }, { "epoch": 0.13224150574431678, "grad_norm": 7.787476062774658, "learning_rate": 1.945736030571443e-05, "loss": 1.2969, "step": 541 }, { "epoch": 0.13248594475678319, "grad_norm": 4.8547773361206055, "learning_rate": 1.9454784720035747e-05, "loss": 1.0794, "step": 542 }, { "epoch": 0.13273038376924956, "grad_norm": 7.971114635467529, "learning_rate": 1.945220320772399e-05, "loss": 1.0347, "step": 543 }, { "epoch": 0.13297482278171596, "grad_norm": 4.092767238616943, "learning_rate": 1.944961577039735e-05, "loss": 2.2551, "step": 544 }, { "epoch": 0.13321926179418236, "grad_norm": 6.591254234313965, "learning_rate": 1.9447022409677736e-05, "loss": 1.1544, "step": 545 }, { "epoch": 0.13346370080664874, "grad_norm": 5.323657035827637, "learning_rate": 1.944442312719077e-05, "loss": 1.1388, "step": 546 }, { "epoch": 0.13370813981911514, "grad_norm": 4.295138835906982, "learning_rate": 1.9441817924565787e-05, "loss": 1.243, "step": 547 }, { "epoch": 0.13395257883158151, "grad_norm": 3.4274163246154785, "learning_rate": 1.9439206803435828e-05, "loss": 2.145, "step": 548 }, { "epoch": 0.13419701784404792, "grad_norm": 6.668155670166016, "learning_rate": 1.9436589765437646e-05, "loss": 1.3672, "step": 549 }, { "epoch": 0.1344414568565143, "grad_norm": 4.891482830047607, "learning_rate": 1.9433966812211712e-05, "loss": 1.2499, "step": 550 }, { "epoch": 0.1346858958689807, "grad_norm": 3.2084758281707764, "learning_rate": 1.9431337945402186e-05, "loss": 2.0341, "step": 551 }, { "epoch": 0.13493033488144707, "grad_norm": 5.467001438140869, "learning_rate": 1.9428703166656952e-05, "loss": 1.059, "step": 552 }, { "epoch": 0.13517477389391347, "grad_norm": 9.086297035217285, "learning_rate": 1.942606247762759e-05, "loss": 1.0457, "step": 553 }, { "epoch": 0.13541921290637987, "grad_norm": 7.168754577636719, "learning_rate": 1.9423415879969395e-05, "loss": 1.2066, "step": 554 }, { "epoch": 0.13566365191884625, "grad_norm": 5.245144844055176, "learning_rate": 1.942076337534135e-05, "loss": 1.1832, "step": 555 }, { "epoch": 0.13590809093131265, "grad_norm": 5.616425037384033, "learning_rate": 1.9418104965406158e-05, "loss": 1.1588, "step": 556 }, { "epoch": 0.13615252994377902, "grad_norm": 4.895685195922852, "learning_rate": 1.941544065183021e-05, "loss": 1.1814, "step": 557 }, { "epoch": 0.13639696895624542, "grad_norm": 4.542455673217773, "learning_rate": 1.9412770436283604e-05, "loss": 1.2014, "step": 558 }, { "epoch": 0.1366414079687118, "grad_norm": 6.379021167755127, "learning_rate": 1.941009432044014e-05, "loss": 1.2843, "step": 559 }, { "epoch": 0.1368858469811782, "grad_norm": 9.439204216003418, "learning_rate": 1.940741230597731e-05, "loss": 1.1331, "step": 560 }, { "epoch": 0.13713028599364457, "grad_norm": 5.6895856857299805, "learning_rate": 1.9404724394576305e-05, "loss": 1.1366, "step": 561 }, { "epoch": 0.13737472500611098, "grad_norm": 3.5664186477661133, "learning_rate": 1.940203058792202e-05, "loss": 1.0722, "step": 562 }, { "epoch": 0.13761916401857738, "grad_norm": 11.332030296325684, "learning_rate": 1.9399330887703035e-05, "loss": 1.0417, "step": 563 }, { "epoch": 0.13786360303104375, "grad_norm": 3.848562717437744, "learning_rate": 1.9396625295611634e-05, "loss": 1.062, "step": 564 }, { "epoch": 0.13810804204351015, "grad_norm": 5.241700649261475, "learning_rate": 1.9393913813343784e-05, "loss": 1.0942, "step": 565 }, { "epoch": 0.13835248105597653, "grad_norm": 3.3595571517944336, "learning_rate": 1.939119644259915e-05, "loss": 1.12, "step": 566 }, { "epoch": 0.13859692006844293, "grad_norm": 6.70857048034668, "learning_rate": 1.938847318508109e-05, "loss": 1.0577, "step": 567 }, { "epoch": 0.1388413590809093, "grad_norm": 6.597379684448242, "learning_rate": 1.9385744042496646e-05, "loss": 1.1119, "step": 568 }, { "epoch": 0.1390857980933757, "grad_norm": 6.321605682373047, "learning_rate": 1.9383009016556555e-05, "loss": 1.0859, "step": 569 }, { "epoch": 0.13933023710584208, "grad_norm": 5.733395576477051, "learning_rate": 1.9380268108975238e-05, "loss": 1.1092, "step": 570 }, { "epoch": 0.13957467611830848, "grad_norm": 4.5978922843933105, "learning_rate": 1.9377521321470806e-05, "loss": 1.1017, "step": 571 }, { "epoch": 0.13981911513077488, "grad_norm": 6.471620082855225, "learning_rate": 1.9374768655765056e-05, "loss": 1.143, "step": 572 }, { "epoch": 0.14006355414324126, "grad_norm": 3.8329930305480957, "learning_rate": 1.937201011358346e-05, "loss": 2.1819, "step": 573 }, { "epoch": 0.14030799315570766, "grad_norm": 5.367978096008301, "learning_rate": 1.9369245696655187e-05, "loss": 1.0577, "step": 574 }, { "epoch": 0.14055243216817404, "grad_norm": 5.2052435874938965, "learning_rate": 1.9366475406713085e-05, "loss": 1.1255, "step": 575 }, { "epoch": 0.14079687118064044, "grad_norm": 9.861551284790039, "learning_rate": 1.9363699245493678e-05, "loss": 1.1107, "step": 576 }, { "epoch": 0.1410413101931068, "grad_norm": 6.768249034881592, "learning_rate": 1.9360917214737177e-05, "loss": 1.1755, "step": 577 }, { "epoch": 0.1412857492055732, "grad_norm": 2.9812920093536377, "learning_rate": 1.935812931618746e-05, "loss": 2.1717, "step": 578 }, { "epoch": 0.1415301882180396, "grad_norm": 4.798773288726807, "learning_rate": 1.9355335551592104e-05, "loss": 0.9572, "step": 579 }, { "epoch": 0.141774627230506, "grad_norm": 5.733400821685791, "learning_rate": 1.9352535922702345e-05, "loss": 1.0916, "step": 580 }, { "epoch": 0.1420190662429724, "grad_norm": 3.7753677368164062, "learning_rate": 1.93497304312731e-05, "loss": 2.0032, "step": 581 }, { "epoch": 0.14226350525543877, "grad_norm": 3.9927875995635986, "learning_rate": 1.9346919079062965e-05, "loss": 1.1038, "step": 582 }, { "epoch": 0.14250794426790517, "grad_norm": 3.727595329284668, "learning_rate": 1.9344101867834202e-05, "loss": 1.0979, "step": 583 }, { "epoch": 0.14275238328037154, "grad_norm": 5.125797271728516, "learning_rate": 1.9341278799352755e-05, "loss": 1.0498, "step": 584 }, { "epoch": 0.14299682229283794, "grad_norm": 6.8079304695129395, "learning_rate": 1.9338449875388234e-05, "loss": 1.2344, "step": 585 }, { "epoch": 0.14324126130530432, "grad_norm": 6.831772327423096, "learning_rate": 1.9335615097713923e-05, "loss": 1.1498, "step": 586 }, { "epoch": 0.14348570031777072, "grad_norm": 3.879077196121216, "learning_rate": 1.933277446810677e-05, "loss": 1.0912, "step": 587 }, { "epoch": 0.1437301393302371, "grad_norm": 5.787622451782227, "learning_rate": 1.932992798834739e-05, "loss": 1.2058, "step": 588 }, { "epoch": 0.1439745783427035, "grad_norm": 3.4940860271453857, "learning_rate": 1.9327075660220076e-05, "loss": 1.2182, "step": 589 }, { "epoch": 0.1442190173551699, "grad_norm": 4.091410160064697, "learning_rate": 1.9324217485512777e-05, "loss": 1.0812, "step": 590 }, { "epoch": 0.14446345636763627, "grad_norm": 6.517455577850342, "learning_rate": 1.932135346601711e-05, "loss": 1.083, "step": 591 }, { "epoch": 0.14470789538010267, "grad_norm": 7.640650272369385, "learning_rate": 1.9318483603528358e-05, "loss": 1.212, "step": 592 }, { "epoch": 0.14495233439256905, "grad_norm": 5.1665472984313965, "learning_rate": 1.931560789984546e-05, "loss": 1.1266, "step": 593 }, { "epoch": 0.14519677340503545, "grad_norm": 10.023889541625977, "learning_rate": 1.931272635677102e-05, "loss": 1.2247, "step": 594 }, { "epoch": 0.14544121241750183, "grad_norm": 4.338187217712402, "learning_rate": 1.9309838976111312e-05, "loss": 1.0899, "step": 595 }, { "epoch": 0.14568565142996823, "grad_norm": 7.103562831878662, "learning_rate": 1.9306945759676254e-05, "loss": 1.3005, "step": 596 }, { "epoch": 0.1459300904424346, "grad_norm": 6.248996734619141, "learning_rate": 1.9304046709279426e-05, "loss": 1.2193, "step": 597 }, { "epoch": 0.146174529454901, "grad_norm": 3.4053194522857666, "learning_rate": 1.9301141826738074e-05, "loss": 1.1569, "step": 598 }, { "epoch": 0.1464189684673674, "grad_norm": 3.0260212421417236, "learning_rate": 1.9298231113873086e-05, "loss": 2.0633, "step": 599 }, { "epoch": 0.14666340747983378, "grad_norm": 4.802820205688477, "learning_rate": 1.929531457250902e-05, "loss": 1.1888, "step": 600 }, { "epoch": 0.14690784649230018, "grad_norm": 4.366469383239746, "learning_rate": 1.9292392204474075e-05, "loss": 1.1144, "step": 601 }, { "epoch": 0.14715228550476656, "grad_norm": 6.060830593109131, "learning_rate": 1.9289464011600106e-05, "loss": 1.0817, "step": 602 }, { "epoch": 0.14739672451723296, "grad_norm": 10.182448387145996, "learning_rate": 1.9286529995722624e-05, "loss": 1.0787, "step": 603 }, { "epoch": 0.14764116352969933, "grad_norm": 3.8352510929107666, "learning_rate": 1.928359015868078e-05, "loss": 1.1872, "step": 604 }, { "epoch": 0.14788560254216573, "grad_norm": 3.9045841693878174, "learning_rate": 1.928064450231739e-05, "loss": 1.2423, "step": 605 }, { "epoch": 0.1481300415546321, "grad_norm": 4.883039951324463, "learning_rate": 1.9277693028478895e-05, "loss": 1.2389, "step": 606 }, { "epoch": 0.1483744805670985, "grad_norm": 4.2947773933410645, "learning_rate": 1.9274735739015405e-05, "loss": 1.1433, "step": 607 }, { "epoch": 0.1486189195795649, "grad_norm": 4.876698970794678, "learning_rate": 1.9271772635780664e-05, "loss": 1.1115, "step": 608 }, { "epoch": 0.1488633585920313, "grad_norm": 4.491443157196045, "learning_rate": 1.926880372063206e-05, "loss": 1.1474, "step": 609 }, { "epoch": 0.1491077976044977, "grad_norm": 3.7161436080932617, "learning_rate": 1.926582899543063e-05, "loss": 1.1142, "step": 610 }, { "epoch": 0.14935223661696406, "grad_norm": 4.214755058288574, "learning_rate": 1.9262848462041046e-05, "loss": 1.0922, "step": 611 }, { "epoch": 0.14959667562943046, "grad_norm": 9.900253295898438, "learning_rate": 1.9259862122331627e-05, "loss": 1.0447, "step": 612 }, { "epoch": 0.14984111464189684, "grad_norm": 5.963320732116699, "learning_rate": 1.9256869978174326e-05, "loss": 1.0681, "step": 613 }, { "epoch": 0.15008555365436324, "grad_norm": 4.601866722106934, "learning_rate": 1.9253872031444742e-05, "loss": 0.944, "step": 614 }, { "epoch": 0.15032999266682961, "grad_norm": 4.92673921585083, "learning_rate": 1.9250868284022102e-05, "loss": 1.0355, "step": 615 }, { "epoch": 0.15057443167929602, "grad_norm": 5.88585090637207, "learning_rate": 1.9247858737789276e-05, "loss": 2.0564, "step": 616 }, { "epoch": 0.15081887069176242, "grad_norm": 9.797651290893555, "learning_rate": 1.9244843394632765e-05, "loss": 1.2278, "step": 617 }, { "epoch": 0.1510633097042288, "grad_norm": 5.198269844055176, "learning_rate": 1.9241822256442707e-05, "loss": 1.172, "step": 618 }, { "epoch": 0.1513077487166952, "grad_norm": 3.963874340057373, "learning_rate": 1.9238795325112867e-05, "loss": 1.1404, "step": 619 }, { "epoch": 0.15155218772916157, "grad_norm": 5.644730091094971, "learning_rate": 1.9235762602540654e-05, "loss": 1.0872, "step": 620 }, { "epoch": 0.15179662674162797, "grad_norm": 4.740919589996338, "learning_rate": 1.923272409062709e-05, "loss": 1.1895, "step": 621 }, { "epoch": 0.15204106575409435, "grad_norm": 8.003108024597168, "learning_rate": 1.922967979127684e-05, "loss": 1.219, "step": 622 }, { "epoch": 0.15228550476656075, "grad_norm": 14.207559585571289, "learning_rate": 1.9226629706398186e-05, "loss": 1.1041, "step": 623 }, { "epoch": 0.15252994377902712, "grad_norm": 3.9512312412261963, "learning_rate": 1.9223573837903043e-05, "loss": 1.246, "step": 624 }, { "epoch": 0.15277438279149352, "grad_norm": 5.806921005249023, "learning_rate": 1.9220512187706956e-05, "loss": 1.0669, "step": 625 }, { "epoch": 0.15301882180395993, "grad_norm": 6.516802787780762, "learning_rate": 1.921744475772908e-05, "loss": 1.1033, "step": 626 }, { "epoch": 0.1532632608164263, "grad_norm": 6.30975341796875, "learning_rate": 1.921437154989221e-05, "loss": 1.1297, "step": 627 }, { "epoch": 0.1535076998288927, "grad_norm": 7.9424147605896, "learning_rate": 1.921129256612275e-05, "loss": 1.1077, "step": 628 }, { "epoch": 0.15375213884135908, "grad_norm": 5.780880928039551, "learning_rate": 1.9208207808350727e-05, "loss": 1.1417, "step": 629 }, { "epoch": 0.15399657785382548, "grad_norm": 8.841265678405762, "learning_rate": 1.920511727850979e-05, "loss": 1.0083, "step": 630 }, { "epoch": 0.15424101686629185, "grad_norm": 13.319135665893555, "learning_rate": 1.920202097853721e-05, "loss": 1.2188, "step": 631 }, { "epoch": 0.15448545587875825, "grad_norm": 5.12705659866333, "learning_rate": 1.9198918910373864e-05, "loss": 1.1483, "step": 632 }, { "epoch": 0.15472989489122463, "grad_norm": 7.4151835441589355, "learning_rate": 1.9195811075964253e-05, "loss": 1.2291, "step": 633 }, { "epoch": 0.15497433390369103, "grad_norm": 5.743813514709473, "learning_rate": 1.919269747725649e-05, "loss": 1.0871, "step": 634 }, { "epoch": 0.15521877291615743, "grad_norm": 6.581920146942139, "learning_rate": 1.918957811620231e-05, "loss": 1.1378, "step": 635 }, { "epoch": 0.1554632119286238, "grad_norm": 8.724961280822754, "learning_rate": 1.918645299475704e-05, "loss": 1.0898, "step": 636 }, { "epoch": 0.1557076509410902, "grad_norm": 7.256657600402832, "learning_rate": 1.9183322114879633e-05, "loss": 1.015, "step": 637 }, { "epoch": 0.15595208995355658, "grad_norm": 5.943508148193359, "learning_rate": 1.9180185478532655e-05, "loss": 1.2436, "step": 638 }, { "epoch": 0.15619652896602298, "grad_norm": 14.481973648071289, "learning_rate": 1.9177043087682266e-05, "loss": 1.0872, "step": 639 }, { "epoch": 0.15644096797848936, "grad_norm": 11.159737586975098, "learning_rate": 1.9173894944298245e-05, "loss": 1.2359, "step": 640 }, { "epoch": 0.15668540699095576, "grad_norm": 7.978509426116943, "learning_rate": 1.917074105035397e-05, "loss": 1.1235, "step": 641 }, { "epoch": 0.15692984600342214, "grad_norm": 12.450584411621094, "learning_rate": 1.916758140782643e-05, "loss": 1.1161, "step": 642 }, { "epoch": 0.15717428501588854, "grad_norm": 4.7206525802612305, "learning_rate": 1.916441601869621e-05, "loss": 1.1474, "step": 643 }, { "epoch": 0.1574187240283549, "grad_norm": 12.776074409484863, "learning_rate": 1.9161244884947503e-05, "loss": 1.1997, "step": 644 }, { "epoch": 0.1576631630408213, "grad_norm": 14.507882118225098, "learning_rate": 1.9158068008568102e-05, "loss": 1.084, "step": 645 }, { "epoch": 0.15790760205328772, "grad_norm": 18.697959899902344, "learning_rate": 1.9154885391549397e-05, "loss": 1.0914, "step": 646 }, { "epoch": 0.1581520410657541, "grad_norm": 5.158618450164795, "learning_rate": 1.9151697035886384e-05, "loss": 1.2494, "step": 647 }, { "epoch": 0.1583964800782205, "grad_norm": 6.806406497955322, "learning_rate": 1.914850294357764e-05, "loss": 1.1049, "step": 648 }, { "epoch": 0.15864091909068687, "grad_norm": 13.390410423278809, "learning_rate": 1.9145303116625357e-05, "loss": 1.1129, "step": 649 }, { "epoch": 0.15888535810315327, "grad_norm": 4.751810550689697, "learning_rate": 1.914209755703531e-05, "loss": 1.1234, "step": 650 }, { "epoch": 0.15912979711561964, "grad_norm": 6.872396469116211, "learning_rate": 1.9138886266816868e-05, "loss": 1.0624, "step": 651 }, { "epoch": 0.15937423612808604, "grad_norm": 9.056709289550781, "learning_rate": 1.9135669247983e-05, "loss": 1.1375, "step": 652 }, { "epoch": 0.15961867514055242, "grad_norm": 8.196880340576172, "learning_rate": 1.9132446502550257e-05, "loss": 1.1987, "step": 653 }, { "epoch": 0.15986311415301882, "grad_norm": 6.337855815887451, "learning_rate": 1.9129218032538787e-05, "loss": 1.1538, "step": 654 }, { "epoch": 0.16010755316548522, "grad_norm": 9.169740676879883, "learning_rate": 1.9125983839972317e-05, "loss": 1.1671, "step": 655 }, { "epoch": 0.1603519921779516, "grad_norm": 4.739553928375244, "learning_rate": 1.9122743926878173e-05, "loss": 1.0563, "step": 656 }, { "epoch": 0.160596431190418, "grad_norm": 10.157617568969727, "learning_rate": 1.9119498295287256e-05, "loss": 1.174, "step": 657 }, { "epoch": 0.16084087020288437, "grad_norm": 6.214486598968506, "learning_rate": 1.911624694723406e-05, "loss": 1.1447, "step": 658 }, { "epoch": 0.16108530921535077, "grad_norm": 4.878256320953369, "learning_rate": 1.9112989884756655e-05, "loss": 1.1352, "step": 659 }, { "epoch": 0.16132974822781715, "grad_norm": 5.432155609130859, "learning_rate": 1.9109727109896695e-05, "loss": 1.2467, "step": 660 }, { "epoch": 0.16157418724028355, "grad_norm": 5.435805320739746, "learning_rate": 1.9106458624699425e-05, "loss": 1.203, "step": 661 }, { "epoch": 0.16181862625274993, "grad_norm": 4.5998759269714355, "learning_rate": 1.910318443121365e-05, "loss": 1.1275, "step": 662 }, { "epoch": 0.16206306526521633, "grad_norm": 9.564048767089844, "learning_rate": 1.9099904531491776e-05, "loss": 1.1447, "step": 663 }, { "epoch": 0.16230750427768273, "grad_norm": 6.536534309387207, "learning_rate": 1.909661892758976e-05, "loss": 1.1537, "step": 664 }, { "epoch": 0.1625519432901491, "grad_norm": 4.478946685791016, "learning_rate": 1.9093327621567162e-05, "loss": 1.1499, "step": 665 }, { "epoch": 0.1627963823026155, "grad_norm": 8.710973739624023, "learning_rate": 1.9090030615487093e-05, "loss": 1.1229, "step": 666 }, { "epoch": 0.16304082131508188, "grad_norm": 3.5803980827331543, "learning_rate": 1.908672791141625e-05, "loss": 2.077, "step": 667 }, { "epoch": 0.16328526032754828, "grad_norm": 7.929402828216553, "learning_rate": 1.90834195114249e-05, "loss": 1.1464, "step": 668 }, { "epoch": 0.16352969934001466, "grad_norm": 5.926062107086182, "learning_rate": 1.9080105417586878e-05, "loss": 1.196, "step": 669 }, { "epoch": 0.16377413835248106, "grad_norm": 4.997158050537109, "learning_rate": 1.9076785631979592e-05, "loss": 1.192, "step": 670 }, { "epoch": 0.16401857736494743, "grad_norm": 4.977759838104248, "learning_rate": 1.907346015668401e-05, "loss": 1.0964, "step": 671 }, { "epoch": 0.16426301637741383, "grad_norm": 17.051420211791992, "learning_rate": 1.907012899378467e-05, "loss": 1.0237, "step": 672 }, { "epoch": 0.16450745538988024, "grad_norm": 5.6634297370910645, "learning_rate": 1.9066792145369685e-05, "loss": 1.1316, "step": 673 }, { "epoch": 0.1647518944023466, "grad_norm": 4.748595714569092, "learning_rate": 1.9063449613530712e-05, "loss": 1.0597, "step": 674 }, { "epoch": 0.164996333414813, "grad_norm": 7.550394058227539, "learning_rate": 1.9060101400362998e-05, "loss": 1.0291, "step": 675 }, { "epoch": 0.1652407724272794, "grad_norm": 7.80336856842041, "learning_rate": 1.905674750796532e-05, "loss": 1.1255, "step": 676 }, { "epoch": 0.1654852114397458, "grad_norm": 4.666580677032471, "learning_rate": 1.905338793844004e-05, "loss": 1.0285, "step": 677 }, { "epoch": 0.16572965045221216, "grad_norm": 4.3687896728515625, "learning_rate": 1.9050022693893062e-05, "loss": 1.1289, "step": 678 }, { "epoch": 0.16597408946467856, "grad_norm": 3.923128366470337, "learning_rate": 1.9046651776433863e-05, "loss": 1.1788, "step": 679 }, { "epoch": 0.16621852847714494, "grad_norm": 5.142563343048096, "learning_rate": 1.904327518817546e-05, "loss": 1.0525, "step": 680 }, { "epoch": 0.16646296748961134, "grad_norm": 4.094109535217285, "learning_rate": 1.9039892931234434e-05, "loss": 1.2607, "step": 681 }, { "epoch": 0.16670740650207774, "grad_norm": 4.689309120178223, "learning_rate": 1.903650500773092e-05, "loss": 0.9916, "step": 682 }, { "epoch": 0.16695184551454412, "grad_norm": 7.4928693771362305, "learning_rate": 1.90331114197886e-05, "loss": 1.1558, "step": 683 }, { "epoch": 0.16719628452701052, "grad_norm": 7.847235679626465, "learning_rate": 1.902971216953471e-05, "loss": 1.105, "step": 684 }, { "epoch": 0.1674407235394769, "grad_norm": 4.171519756317139, "learning_rate": 1.9026307259100037e-05, "loss": 1.0784, "step": 685 }, { "epoch": 0.1676851625519433, "grad_norm": 4.570870876312256, "learning_rate": 1.9022896690618908e-05, "loss": 1.118, "step": 686 }, { "epoch": 0.16792960156440967, "grad_norm": 4.4747748374938965, "learning_rate": 1.9019480466229203e-05, "loss": 1.0661, "step": 687 }, { "epoch": 0.16817404057687607, "grad_norm": 4.1525959968566895, "learning_rate": 1.9016058588072354e-05, "loss": 1.0394, "step": 688 }, { "epoch": 0.16841847958934245, "grad_norm": 4.377383232116699, "learning_rate": 1.9012631058293322e-05, "loss": 1.137, "step": 689 }, { "epoch": 0.16866291860180885, "grad_norm": 4.25969123840332, "learning_rate": 1.9009197879040617e-05, "loss": 1.1238, "step": 690 }, { "epoch": 0.16890735761427525, "grad_norm": 4.3925299644470215, "learning_rate": 1.9005759052466303e-05, "loss": 1.0514, "step": 691 }, { "epoch": 0.16915179662674162, "grad_norm": 6.636261940002441, "learning_rate": 1.900231458072596e-05, "loss": 1.0674, "step": 692 }, { "epoch": 0.16939623563920803, "grad_norm": 5.294342517852783, "learning_rate": 1.899886446597872e-05, "loss": 1.1533, "step": 693 }, { "epoch": 0.1696406746516744, "grad_norm": 4.914679527282715, "learning_rate": 1.899540871038726e-05, "loss": 1.0943, "step": 694 }, { "epoch": 0.1698851136641408, "grad_norm": 4.830330848693848, "learning_rate": 1.899194731611778e-05, "loss": 1.0313, "step": 695 }, { "epoch": 0.17012955267660718, "grad_norm": 4.74468994140625, "learning_rate": 1.8988480285340018e-05, "loss": 1.1535, "step": 696 }, { "epoch": 0.17037399168907358, "grad_norm": 6.699695110321045, "learning_rate": 1.8985007620227246e-05, "loss": 1.0901, "step": 697 }, { "epoch": 0.17061843070153995, "grad_norm": 7.729903697967529, "learning_rate": 1.898152932295627e-05, "loss": 1.0657, "step": 698 }, { "epoch": 0.17086286971400635, "grad_norm": 4.743852138519287, "learning_rate": 1.897804539570742e-05, "loss": 1.0182, "step": 699 }, { "epoch": 0.17110730872647276, "grad_norm": 4.740683555603027, "learning_rate": 1.897455584066456e-05, "loss": 1.0612, "step": 700 }, { "epoch": 0.17135174773893913, "grad_norm": 3.3871266841888428, "learning_rate": 1.897106066001509e-05, "loss": 1.0377, "step": 701 }, { "epoch": 0.17159618675140553, "grad_norm": 5.6066083908081055, "learning_rate": 1.8967559855949913e-05, "loss": 1.2205, "step": 702 }, { "epoch": 0.1718406257638719, "grad_norm": 7.7547101974487305, "learning_rate": 1.8964053430663485e-05, "loss": 1.1098, "step": 703 }, { "epoch": 0.1720850647763383, "grad_norm": 3.957638740539551, "learning_rate": 1.8960541386353765e-05, "loss": 1.0606, "step": 704 }, { "epoch": 0.17232950378880468, "grad_norm": 5.753738880157471, "learning_rate": 1.8957023725222238e-05, "loss": 1.081, "step": 705 }, { "epoch": 0.17257394280127109, "grad_norm": 5.259015083312988, "learning_rate": 1.8953500449473924e-05, "loss": 1.0121, "step": 706 }, { "epoch": 0.17281838181373746, "grad_norm": 4.496668338775635, "learning_rate": 1.894997156131734e-05, "loss": 1.0718, "step": 707 }, { "epoch": 0.17306282082620386, "grad_norm": 14.881460189819336, "learning_rate": 1.8946437062964543e-05, "loss": 1.1767, "step": 708 }, { "epoch": 0.17330725983867026, "grad_norm": 5.021279811859131, "learning_rate": 1.894289695663109e-05, "loss": 1.1781, "step": 709 }, { "epoch": 0.17355169885113664, "grad_norm": 5.705527305603027, "learning_rate": 1.893935124453606e-05, "loss": 1.0158, "step": 710 }, { "epoch": 0.17379613786360304, "grad_norm": 4.499573230743408, "learning_rate": 1.8935799928902046e-05, "loss": 1.168, "step": 711 }, { "epoch": 0.1740405768760694, "grad_norm": 4.802264213562012, "learning_rate": 1.8932243011955154e-05, "loss": 1.1318, "step": 712 }, { "epoch": 0.17428501588853582, "grad_norm": 12.059589385986328, "learning_rate": 1.8928680495925e-05, "loss": 0.9681, "step": 713 }, { "epoch": 0.1745294549010022, "grad_norm": 5.503899574279785, "learning_rate": 1.8925112383044713e-05, "loss": 1.1116, "step": 714 }, { "epoch": 0.1747738939134686, "grad_norm": 5.018351078033447, "learning_rate": 1.8921538675550923e-05, "loss": 1.098, "step": 715 }, { "epoch": 0.17501833292593497, "grad_norm": 6.688838958740234, "learning_rate": 1.891795937568377e-05, "loss": 1.1197, "step": 716 }, { "epoch": 0.17526277193840137, "grad_norm": 13.338314056396484, "learning_rate": 1.8914374485686905e-05, "loss": 1.0497, "step": 717 }, { "epoch": 0.17550721095086777, "grad_norm": 11.007012367248535, "learning_rate": 1.8910784007807477e-05, "loss": 1.1531, "step": 718 }, { "epoch": 0.17575164996333414, "grad_norm": 9.250398635864258, "learning_rate": 1.890718794429614e-05, "loss": 1.047, "step": 719 }, { "epoch": 0.17599608897580055, "grad_norm": 12.425487518310547, "learning_rate": 1.8903586297407052e-05, "loss": 1.0937, "step": 720 }, { "epoch": 0.17624052798826692, "grad_norm": 5.152442932128906, "learning_rate": 1.8899979069397858e-05, "loss": 1.1518, "step": 721 }, { "epoch": 0.17648496700073332, "grad_norm": 6.493432998657227, "learning_rate": 1.8896366262529718e-05, "loss": 1.0677, "step": 722 }, { "epoch": 0.1767294060131997, "grad_norm": 4.106019020080566, "learning_rate": 1.8892747879067284e-05, "loss": 1.1456, "step": 723 }, { "epoch": 0.1769738450256661, "grad_norm": 8.680675506591797, "learning_rate": 1.8889123921278703e-05, "loss": 1.1829, "step": 724 }, { "epoch": 0.17721828403813247, "grad_norm": 3.9373128414154053, "learning_rate": 1.8885494391435607e-05, "loss": 1.069, "step": 725 }, { "epoch": 0.17746272305059888, "grad_norm": 4.0960822105407715, "learning_rate": 1.8881859291813137e-05, "loss": 1.258, "step": 726 }, { "epoch": 0.17770716206306528, "grad_norm": 7.045689105987549, "learning_rate": 1.8878218624689913e-05, "loss": 1.172, "step": 727 }, { "epoch": 0.17795160107553165, "grad_norm": 4.111807823181152, "learning_rate": 1.887457239234805e-05, "loss": 1.1729, "step": 728 }, { "epoch": 0.17819604008799805, "grad_norm": 4.325777053833008, "learning_rate": 1.887092059707315e-05, "loss": 1.1321, "step": 729 }, { "epoch": 0.17844047910046443, "grad_norm": 4.5247578620910645, "learning_rate": 1.8867263241154302e-05, "loss": 1.0469, "step": 730 }, { "epoch": 0.17868491811293083, "grad_norm": 3.589571237564087, "learning_rate": 1.8863600326884085e-05, "loss": 1.198, "step": 731 }, { "epoch": 0.1789293571253972, "grad_norm": 3.886864185333252, "learning_rate": 1.8859931856558555e-05, "loss": 1.0563, "step": 732 }, { "epoch": 0.1791737961378636, "grad_norm": 4.039812088012695, "learning_rate": 1.8856257832477255e-05, "loss": 1.0757, "step": 733 }, { "epoch": 0.17941823515032998, "grad_norm": 4.089430809020996, "learning_rate": 1.8852578256943212e-05, "loss": 1.1413, "step": 734 }, { "epoch": 0.17966267416279638, "grad_norm": 4.228828430175781, "learning_rate": 1.8848893132262927e-05, "loss": 1.2184, "step": 735 }, { "epoch": 0.17990711317526278, "grad_norm": 6.709991931915283, "learning_rate": 1.884520246074638e-05, "loss": 1.2658, "step": 736 }, { "epoch": 0.18015155218772916, "grad_norm": 5.3593430519104, "learning_rate": 1.884150624470703e-05, "loss": 1.1778, "step": 737 }, { "epoch": 0.18039599120019556, "grad_norm": 4.252145767211914, "learning_rate": 1.8837804486461818e-05, "loss": 1.0652, "step": 738 }, { "epoch": 0.18064043021266193, "grad_norm": 4.08242130279541, "learning_rate": 1.8834097188331143e-05, "loss": 1.1207, "step": 739 }, { "epoch": 0.18088486922512834, "grad_norm": 6.409980297088623, "learning_rate": 1.8830384352638896e-05, "loss": 1.1498, "step": 740 }, { "epoch": 0.1811293082375947, "grad_norm": 5.007962226867676, "learning_rate": 1.882666598171242e-05, "loss": 1.1071, "step": 741 }, { "epoch": 0.1813737472500611, "grad_norm": 3.8679728507995605, "learning_rate": 1.8822942077882542e-05, "loss": 1.0612, "step": 742 }, { "epoch": 0.1816181862625275, "grad_norm": 4.374696731567383, "learning_rate": 1.881921264348355e-05, "loss": 1.1656, "step": 743 }, { "epoch": 0.1818626252749939, "grad_norm": 3.6539933681488037, "learning_rate": 1.8815477680853208e-05, "loss": 1.0958, "step": 744 }, { "epoch": 0.1821070642874603, "grad_norm": 3.358370542526245, "learning_rate": 1.8811737192332725e-05, "loss": 1.0235, "step": 745 }, { "epoch": 0.18235150329992666, "grad_norm": 4.081916332244873, "learning_rate": 1.8807991180266798e-05, "loss": 1.0765, "step": 746 }, { "epoch": 0.18259594231239307, "grad_norm": 4.941745758056641, "learning_rate": 1.8804239647003574e-05, "loss": 1.1526, "step": 747 }, { "epoch": 0.18284038132485944, "grad_norm": 4.099913597106934, "learning_rate": 1.880048259489466e-05, "loss": 1.2486, "step": 748 }, { "epoch": 0.18308482033732584, "grad_norm": 3.721834897994995, "learning_rate": 1.8796720026295122e-05, "loss": 1.1316, "step": 749 }, { "epoch": 0.18332925934979222, "grad_norm": 3.963992118835449, "learning_rate": 1.8792951943563492e-05, "loss": 0.9799, "step": 750 }, { "epoch": 0.18357369836225862, "grad_norm": 3.802098274230957, "learning_rate": 1.8789178349061755e-05, "loss": 1.1526, "step": 751 }, { "epoch": 0.183818137374725, "grad_norm": 3.5054547786712646, "learning_rate": 1.8785399245155345e-05, "loss": 1.0229, "step": 752 }, { "epoch": 0.1840625763871914, "grad_norm": 3.7871315479278564, "learning_rate": 1.8781614634213153e-05, "loss": 1.0775, "step": 753 }, { "epoch": 0.1843070153996578, "grad_norm": 3.387770891189575, "learning_rate": 1.8777824518607527e-05, "loss": 1.0812, "step": 754 }, { "epoch": 0.18455145441212417, "grad_norm": 3.360201120376587, "learning_rate": 1.8774028900714257e-05, "loss": 1.0478, "step": 755 }, { "epoch": 0.18479589342459057, "grad_norm": 5.499178409576416, "learning_rate": 1.877022778291259e-05, "loss": 1.1209, "step": 756 }, { "epoch": 0.18504033243705695, "grad_norm": 3.671175718307495, "learning_rate": 1.8766421167585216e-05, "loss": 1.1505, "step": 757 }, { "epoch": 0.18528477144952335, "grad_norm": 4.1667256355285645, "learning_rate": 1.8762609057118268e-05, "loss": 1.1308, "step": 758 }, { "epoch": 0.18552921046198972, "grad_norm": 4.283543109893799, "learning_rate": 1.8758791453901337e-05, "loss": 1.1366, "step": 759 }, { "epoch": 0.18577364947445613, "grad_norm": 5.065645217895508, "learning_rate": 1.8754968360327438e-05, "loss": 1.2085, "step": 760 }, { "epoch": 0.1860180884869225, "grad_norm": 5.341877460479736, "learning_rate": 1.8751139778793043e-05, "loss": 1.1089, "step": 761 }, { "epoch": 0.1862625274993889, "grad_norm": 4.320544242858887, "learning_rate": 1.8747305711698058e-05, "loss": 1.1308, "step": 762 }, { "epoch": 0.1865069665118553, "grad_norm": 3.871718168258667, "learning_rate": 1.8743466161445823e-05, "loss": 1.0672, "step": 763 }, { "epoch": 0.18675140552432168, "grad_norm": 4.854724884033203, "learning_rate": 1.8739621130443125e-05, "loss": 1.0795, "step": 764 }, { "epoch": 0.18699584453678808, "grad_norm": 4.116591930389404, "learning_rate": 1.8735770621100182e-05, "loss": 1.1431, "step": 765 }, { "epoch": 0.18724028354925445, "grad_norm": 11.736309051513672, "learning_rate": 1.8731914635830643e-05, "loss": 1.0883, "step": 766 }, { "epoch": 0.18748472256172086, "grad_norm": 3.832092046737671, "learning_rate": 1.872805317705159e-05, "loss": 1.1814, "step": 767 }, { "epoch": 0.18772916157418723, "grad_norm": 4.503589153289795, "learning_rate": 1.872418624718354e-05, "loss": 1.1775, "step": 768 }, { "epoch": 0.18797360058665363, "grad_norm": 3.3327720165252686, "learning_rate": 1.8720313848650448e-05, "loss": 1.158, "step": 769 }, { "epoch": 0.18821803959912, "grad_norm": 6.3881754875183105, "learning_rate": 1.871643598387967e-05, "loss": 1.1307, "step": 770 }, { "epoch": 0.1884624786115864, "grad_norm": 4.862054824829102, "learning_rate": 1.871255265530201e-05, "loss": 1.1961, "step": 771 }, { "epoch": 0.1887069176240528, "grad_norm": 3.6286745071411133, "learning_rate": 1.8708663865351696e-05, "loss": 0.9553, "step": 772 }, { "epoch": 0.18895135663651919, "grad_norm": 3.490654706954956, "learning_rate": 1.8704769616466377e-05, "loss": 1.1614, "step": 773 }, { "epoch": 0.1891957956489856, "grad_norm": 3.3287625312805176, "learning_rate": 1.8700869911087115e-05, "loss": 2.1641, "step": 774 }, { "epoch": 0.18944023466145196, "grad_norm": 4.803013324737549, "learning_rate": 1.8696964751658405e-05, "loss": 1.0372, "step": 775 }, { "epoch": 0.18968467367391836, "grad_norm": 3.6098735332489014, "learning_rate": 1.8693054140628154e-05, "loss": 1.0666, "step": 776 }, { "epoch": 0.18992911268638474, "grad_norm": 4.2922749519348145, "learning_rate": 1.8689138080447687e-05, "loss": 1.1183, "step": 777 }, { "epoch": 0.19017355169885114, "grad_norm": 4.80598258972168, "learning_rate": 1.8685216573571744e-05, "loss": 1.1296, "step": 778 }, { "epoch": 0.19041799071131751, "grad_norm": 3.6623916625976562, "learning_rate": 1.8681289622458485e-05, "loss": 1.0789, "step": 779 }, { "epoch": 0.19066242972378392, "grad_norm": 4.169750690460205, "learning_rate": 1.8677357229569474e-05, "loss": 1.1334, "step": 780 }, { "epoch": 0.19090686873625032, "grad_norm": 5.41149377822876, "learning_rate": 1.8673419397369693e-05, "loss": 1.1136, "step": 781 }, { "epoch": 0.1911513077487167, "grad_norm": 3.4953672885894775, "learning_rate": 1.8669476128327534e-05, "loss": 2.1952, "step": 782 }, { "epoch": 0.1913957467611831, "grad_norm": 4.066995620727539, "learning_rate": 1.8665527424914786e-05, "loss": 1.1636, "step": 783 }, { "epoch": 0.19164018577364947, "grad_norm": 4.137174129486084, "learning_rate": 1.8661573289606657e-05, "loss": 1.0459, "step": 784 }, { "epoch": 0.19188462478611587, "grad_norm": 4.325610637664795, "learning_rate": 1.8657613724881755e-05, "loss": 0.9766, "step": 785 }, { "epoch": 0.19212906379858224, "grad_norm": 3.6171977519989014, "learning_rate": 1.8653648733222096e-05, "loss": 1.042, "step": 786 }, { "epoch": 0.19237350281104865, "grad_norm": 3.619912624359131, "learning_rate": 1.8649678317113084e-05, "loss": 1.1192, "step": 787 }, { "epoch": 0.19261794182351502, "grad_norm": 4.169992923736572, "learning_rate": 1.864570247904354e-05, "loss": 1.1468, "step": 788 }, { "epoch": 0.19286238083598142, "grad_norm": 4.603641510009766, "learning_rate": 1.8641721221505674e-05, "loss": 1.1097, "step": 789 }, { "epoch": 0.19310681984844782, "grad_norm": 4.259536266326904, "learning_rate": 1.86377345469951e-05, "loss": 1.1036, "step": 790 }, { "epoch": 0.1933512588609142, "grad_norm": 3.3828346729278564, "learning_rate": 1.863374245801082e-05, "loss": 1.0612, "step": 791 }, { "epoch": 0.1935956978733806, "grad_norm": 4.6406941413879395, "learning_rate": 1.862974495705523e-05, "loss": 0.9907, "step": 792 }, { "epoch": 0.19384013688584698, "grad_norm": 3.43007755279541, "learning_rate": 1.862574204663413e-05, "loss": 1.0577, "step": 793 }, { "epoch": 0.19408457589831338, "grad_norm": 8.210151672363281, "learning_rate": 1.8621733729256694e-05, "loss": 1.052, "step": 794 }, { "epoch": 0.19432901491077975, "grad_norm": 3.7519443035125732, "learning_rate": 1.8617720007435497e-05, "loss": 1.1144, "step": 795 }, { "epoch": 0.19457345392324615, "grad_norm": 5.4279069900512695, "learning_rate": 1.8613700883686503e-05, "loss": 1.1851, "step": 796 }, { "epoch": 0.19481789293571253, "grad_norm": 4.4938130378723145, "learning_rate": 1.860967636052905e-05, "loss": 1.1729, "step": 797 }, { "epoch": 0.19506233194817893, "grad_norm": 3.714663028717041, "learning_rate": 1.8605646440485877e-05, "loss": 1.0932, "step": 798 }, { "epoch": 0.19530677096064533, "grad_norm": 3.895505666732788, "learning_rate": 1.860161112608309e-05, "loss": 1.1341, "step": 799 }, { "epoch": 0.1955512099731117, "grad_norm": 5.3650641441345215, "learning_rate": 1.8597570419850184e-05, "loss": 1.0481, "step": 800 }, { "epoch": 0.1957956489855781, "grad_norm": 4.210098743438721, "learning_rate": 1.8593524324320035e-05, "loss": 1.1164, "step": 801 }, { "epoch": 0.19604008799804448, "grad_norm": 4.8966851234436035, "learning_rate": 1.85894728420289e-05, "loss": 1.1498, "step": 802 }, { "epoch": 0.19628452701051088, "grad_norm": 4.228818416595459, "learning_rate": 1.8585415975516405e-05, "loss": 1.0885, "step": 803 }, { "epoch": 0.19652896602297726, "grad_norm": 3.544618606567383, "learning_rate": 1.858135372732555e-05, "loss": 1.1276, "step": 804 }, { "epoch": 0.19677340503544366, "grad_norm": 3.545396327972412, "learning_rate": 1.8577286100002723e-05, "loss": 1.0031, "step": 805 }, { "epoch": 0.19701784404791003, "grad_norm": 3.896831512451172, "learning_rate": 1.8573213096097666e-05, "loss": 0.9596, "step": 806 }, { "epoch": 0.19726228306037644, "grad_norm": 5.555988788604736, "learning_rate": 1.8569134718163496e-05, "loss": 1.1155, "step": 807 }, { "epoch": 0.19750672207284284, "grad_norm": 4.680154800415039, "learning_rate": 1.8565050968756714e-05, "loss": 1.1159, "step": 808 }, { "epoch": 0.1977511610853092, "grad_norm": 3.59993314743042, "learning_rate": 1.8560961850437163e-05, "loss": 2.0246, "step": 809 }, { "epoch": 0.19799560009777561, "grad_norm": 6.578675746917725, "learning_rate": 1.8556867365768073e-05, "loss": 1.057, "step": 810 }, { "epoch": 0.198240039110242, "grad_norm": 4.831800937652588, "learning_rate": 1.855276751731602e-05, "loss": 1.0949, "step": 811 }, { "epoch": 0.1984844781227084, "grad_norm": 4.584733963012695, "learning_rate": 1.854866230765096e-05, "loss": 1.1775, "step": 812 }, { "epoch": 0.19872891713517477, "grad_norm": 3.855050563812256, "learning_rate": 1.85445517393462e-05, "loss": 1.1027, "step": 813 }, { "epoch": 0.19897335614764117, "grad_norm": 4.348968029022217, "learning_rate": 1.85404358149784e-05, "loss": 1.091, "step": 814 }, { "epoch": 0.19921779516010754, "grad_norm": 5.458604335784912, "learning_rate": 1.8536314537127586e-05, "loss": 1.0569, "step": 815 }, { "epoch": 0.19946223417257394, "grad_norm": 4.697046279907227, "learning_rate": 1.8532187908377146e-05, "loss": 1.1723, "step": 816 }, { "epoch": 0.19970667318504035, "grad_norm": 3.4961538314819336, "learning_rate": 1.8528055931313803e-05, "loss": 2.107, "step": 817 }, { "epoch": 0.19995111219750672, "grad_norm": 13.489090919494629, "learning_rate": 1.8523918608527653e-05, "loss": 1.0733, "step": 818 }, { "epoch": 0.20019555120997312, "grad_norm": 3.1647040843963623, "learning_rate": 1.8519775942612128e-05, "loss": 1.061, "step": 819 }, { "epoch": 0.2004399902224395, "grad_norm": 5.267470359802246, "learning_rate": 1.8515627936164015e-05, "loss": 1.07, "step": 820 }, { "epoch": 0.2006844292349059, "grad_norm": 4.1006951332092285, "learning_rate": 1.8511474591783454e-05, "loss": 1.1726, "step": 821 }, { "epoch": 0.20092886824737227, "grad_norm": 3.6335432529449463, "learning_rate": 1.850731591207392e-05, "loss": 1.0347, "step": 822 }, { "epoch": 0.20117330725983867, "grad_norm": 4.431300163269043, "learning_rate": 1.8503151899642237e-05, "loss": 2.1155, "step": 823 }, { "epoch": 0.20141774627230505, "grad_norm": 6.050295829772949, "learning_rate": 1.8498982557098578e-05, "loss": 1.1272, "step": 824 }, { "epoch": 0.20166218528477145, "grad_norm": 4.173649311065674, "learning_rate": 1.849480788705645e-05, "loss": 1.1039, "step": 825 }, { "epoch": 0.20190662429723785, "grad_norm": 3.55676531791687, "learning_rate": 1.8490627892132697e-05, "loss": 1.1069, "step": 826 }, { "epoch": 0.20215106330970423, "grad_norm": 3.58681583404541, "learning_rate": 1.848644257494751e-05, "loss": 0.963, "step": 827 }, { "epoch": 0.20239550232217063, "grad_norm": 3.541001558303833, "learning_rate": 1.8482251938124412e-05, "loss": 1.0795, "step": 828 }, { "epoch": 0.202639941334637, "grad_norm": 4.774433135986328, "learning_rate": 1.8478055984290258e-05, "loss": 1.0283, "step": 829 }, { "epoch": 0.2028843803471034, "grad_norm": 3.2330007553100586, "learning_rate": 1.8473854716075233e-05, "loss": 2.0566, "step": 830 }, { "epoch": 0.20312881935956978, "grad_norm": 10.420342445373535, "learning_rate": 1.8469648136112867e-05, "loss": 1.0285, "step": 831 }, { "epoch": 0.20337325837203618, "grad_norm": 9.658334732055664, "learning_rate": 1.8465436247040005e-05, "loss": 1.0787, "step": 832 }, { "epoch": 0.20361769738450256, "grad_norm": 6.110720634460449, "learning_rate": 1.846121905149683e-05, "loss": 1.0487, "step": 833 }, { "epoch": 0.20386213639696896, "grad_norm": 3.708970785140991, "learning_rate": 1.845699655212684e-05, "loss": 1.1554, "step": 834 }, { "epoch": 0.20410657540943536, "grad_norm": 5.930639266967773, "learning_rate": 1.845276875157687e-05, "loss": 1.0547, "step": 835 }, { "epoch": 0.20435101442190173, "grad_norm": 3.6849205493927, "learning_rate": 1.8448535652497073e-05, "loss": 1.0065, "step": 836 }, { "epoch": 0.20459545343436814, "grad_norm": 3.9739530086517334, "learning_rate": 1.844429725754092e-05, "loss": 1.0815, "step": 837 }, { "epoch": 0.2048398924468345, "grad_norm": 5.715768337249756, "learning_rate": 1.8440053569365204e-05, "loss": 1.0391, "step": 838 }, { "epoch": 0.2050843314593009, "grad_norm": 5.876922130584717, "learning_rate": 1.8435804590630045e-05, "loss": 1.0286, "step": 839 }, { "epoch": 0.20532877047176729, "grad_norm": 4.035728931427002, "learning_rate": 1.8431550323998862e-05, "loss": 1.1854, "step": 840 }, { "epoch": 0.2055732094842337, "grad_norm": 3.5158615112304688, "learning_rate": 1.8427290772138397e-05, "loss": 1.0352, "step": 841 }, { "epoch": 0.20581764849670006, "grad_norm": 5.412172794342041, "learning_rate": 1.8423025937718712e-05, "loss": 1.0675, "step": 842 }, { "epoch": 0.20606208750916646, "grad_norm": 4.868218421936035, "learning_rate": 1.841875582341317e-05, "loss": 1.1832, "step": 843 }, { "epoch": 0.20630652652163287, "grad_norm": 3.841078281402588, "learning_rate": 1.841448043189845e-05, "loss": 1.1062, "step": 844 }, { "epoch": 0.20655096553409924, "grad_norm": 4.0944743156433105, "learning_rate": 1.8410199765854532e-05, "loss": 1.0201, "step": 845 }, { "epoch": 0.20679540454656564, "grad_norm": 10.717774391174316, "learning_rate": 1.8405913827964713e-05, "loss": 1.1188, "step": 846 }, { "epoch": 0.20703984355903202, "grad_norm": 4.635489463806152, "learning_rate": 1.840162262091558e-05, "loss": 1.0325, "step": 847 }, { "epoch": 0.20728428257149842, "grad_norm": 6.215915679931641, "learning_rate": 1.8397326147397043e-05, "loss": 1.0543, "step": 848 }, { "epoch": 0.2075287215839648, "grad_norm": 4.547093391418457, "learning_rate": 1.839302441010229e-05, "loss": 1.1416, "step": 849 }, { "epoch": 0.2077731605964312, "grad_norm": 3.299689531326294, "learning_rate": 1.838871741172783e-05, "loss": 1.1036, "step": 850 }, { "epoch": 0.20801759960889757, "grad_norm": 3.8489363193511963, "learning_rate": 1.838440515497345e-05, "loss": 1.1148, "step": 851 }, { "epoch": 0.20826203862136397, "grad_norm": 3.808192014694214, "learning_rate": 1.838008764254225e-05, "loss": 1.1627, "step": 852 }, { "epoch": 0.20850647763383037, "grad_norm": 4.016510009765625, "learning_rate": 1.8375764877140617e-05, "loss": 1.0187, "step": 853 }, { "epoch": 0.20875091664629675, "grad_norm": 14.063336372375488, "learning_rate": 1.837143686147823e-05, "loss": 1.083, "step": 854 }, { "epoch": 0.20899535565876315, "grad_norm": 5.032708168029785, "learning_rate": 1.836710359826806e-05, "loss": 1.1282, "step": 855 }, { "epoch": 0.20923979467122952, "grad_norm": 5.023953914642334, "learning_rate": 1.836276509022637e-05, "loss": 1.0746, "step": 856 }, { "epoch": 0.20948423368369593, "grad_norm": 4.224484920501709, "learning_rate": 1.8358421340072704e-05, "loss": 1.0772, "step": 857 }, { "epoch": 0.2097286726961623, "grad_norm": 4.803196430206299, "learning_rate": 1.83540723505299e-05, "loss": 2.1353, "step": 858 }, { "epoch": 0.2099731117086287, "grad_norm": 4.721566200256348, "learning_rate": 1.8349718124324075e-05, "loss": 0.9563, "step": 859 }, { "epoch": 0.21021755072109508, "grad_norm": 3.717606782913208, "learning_rate": 1.834535866418463e-05, "loss": 1.1166, "step": 860 }, { "epoch": 0.21046198973356148, "grad_norm": 4.486362934112549, "learning_rate": 1.8340993972844252e-05, "loss": 1.0245, "step": 861 }, { "epoch": 0.21070642874602788, "grad_norm": 3.6472747325897217, "learning_rate": 1.8336624053038898e-05, "loss": 0.9977, "step": 862 }, { "epoch": 0.21095086775849425, "grad_norm": 6.7217230796813965, "learning_rate": 1.8332248907507808e-05, "loss": 1.1201, "step": 863 }, { "epoch": 0.21119530677096066, "grad_norm": 5.946406841278076, "learning_rate": 1.8327868538993492e-05, "loss": 1.1357, "step": 864 }, { "epoch": 0.21143974578342703, "grad_norm": 3.6695261001586914, "learning_rate": 1.8323482950241743e-05, "loss": 1.0761, "step": 865 }, { "epoch": 0.21168418479589343, "grad_norm": 3.1539971828460693, "learning_rate": 1.831909214400162e-05, "loss": 1.0073, "step": 866 }, { "epoch": 0.2119286238083598, "grad_norm": 4.463747501373291, "learning_rate": 1.8314696123025456e-05, "loss": 1.1237, "step": 867 }, { "epoch": 0.2121730628208262, "grad_norm": 4.581067085266113, "learning_rate": 1.8310294890068845e-05, "loss": 1.0942, "step": 868 }, { "epoch": 0.21241750183329258, "grad_norm": 3.73099684715271, "learning_rate": 1.830588844789066e-05, "loss": 1.0187, "step": 869 }, { "epoch": 0.21266194084575898, "grad_norm": 3.665910243988037, "learning_rate": 1.830147679925303e-05, "loss": 1.0703, "step": 870 }, { "epoch": 0.2129063798582254, "grad_norm": 4.117025852203369, "learning_rate": 1.8297059946921357e-05, "loss": 0.9757, "step": 871 }, { "epoch": 0.21315081887069176, "grad_norm": 3.8741068840026855, "learning_rate": 1.829263789366429e-05, "loss": 1.1434, "step": 872 }, { "epoch": 0.21339525788315816, "grad_norm": 3.788287401199341, "learning_rate": 1.828821064225375e-05, "loss": 0.9807, "step": 873 }, { "epoch": 0.21363969689562454, "grad_norm": 3.6013906002044678, "learning_rate": 1.8283778195464913e-05, "loss": 1.0818, "step": 874 }, { "epoch": 0.21388413590809094, "grad_norm": 4.018670558929443, "learning_rate": 1.8279340556076218e-05, "loss": 1.1252, "step": 875 }, { "epoch": 0.2141285749205573, "grad_norm": 4.4719557762146, "learning_rate": 1.8274897726869342e-05, "loss": 1.127, "step": 876 }, { "epoch": 0.21437301393302371, "grad_norm": 6.977454662322998, "learning_rate": 1.8270449710629236e-05, "loss": 1.092, "step": 877 }, { "epoch": 0.2146174529454901, "grad_norm": 4.391088485717773, "learning_rate": 1.8265996510144085e-05, "loss": 1.1899, "step": 878 }, { "epoch": 0.2148618919579565, "grad_norm": 4.422701358795166, "learning_rate": 1.8261538128205337e-05, "loss": 1.0594, "step": 879 }, { "epoch": 0.2151063309704229, "grad_norm": 6.5446953773498535, "learning_rate": 1.825707456760768e-05, "loss": 1.0841, "step": 880 }, { "epoch": 0.21535076998288927, "grad_norm": 4.518584251403809, "learning_rate": 1.8252605831149052e-05, "loss": 1.0258, "step": 881 }, { "epoch": 0.21559520899535567, "grad_norm": 3.3467774391174316, "learning_rate": 1.824813192163063e-05, "loss": 2.1104, "step": 882 }, { "epoch": 0.21583964800782204, "grad_norm": 4.550435543060303, "learning_rate": 1.8243652841856842e-05, "loss": 1.093, "step": 883 }, { "epoch": 0.21608408702028845, "grad_norm": 5.360644817352295, "learning_rate": 1.823916859463535e-05, "loss": 1.1999, "step": 884 }, { "epoch": 0.21632852603275482, "grad_norm": 4.890727519989014, "learning_rate": 1.8234679182777055e-05, "loss": 0.9747, "step": 885 }, { "epoch": 0.21657296504522122, "grad_norm": 4.1767659187316895, "learning_rate": 1.82301846090961e-05, "loss": 1.0555, "step": 886 }, { "epoch": 0.2168174040576876, "grad_norm": 3.9458184242248535, "learning_rate": 1.8225684876409863e-05, "loss": 1.1683, "step": 887 }, { "epoch": 0.217061843070154, "grad_norm": 3.597496747970581, "learning_rate": 1.8221179987538952e-05, "loss": 1.1083, "step": 888 }, { "epoch": 0.2173062820826204, "grad_norm": 4.229506015777588, "learning_rate": 1.8216669945307214e-05, "loss": 1.1957, "step": 889 }, { "epoch": 0.21755072109508677, "grad_norm": 3.556558132171631, "learning_rate": 1.8212154752541718e-05, "loss": 1.1348, "step": 890 }, { "epoch": 0.21779516010755318, "grad_norm": 3.491957664489746, "learning_rate": 1.8207634412072765e-05, "loss": 1.1947, "step": 891 }, { "epoch": 0.21803959912001955, "grad_norm": 4.08765983581543, "learning_rate": 1.8203108926733886e-05, "loss": 1.0128, "step": 892 }, { "epoch": 0.21828403813248595, "grad_norm": 4.435181617736816, "learning_rate": 1.8198578299361832e-05, "loss": 1.0455, "step": 893 }, { "epoch": 0.21852847714495233, "grad_norm": 4.2444634437561035, "learning_rate": 1.819404253279658e-05, "loss": 1.0579, "step": 894 }, { "epoch": 0.21877291615741873, "grad_norm": 3.8789174556732178, "learning_rate": 1.818950162988133e-05, "loss": 1.0078, "step": 895 }, { "epoch": 0.2190173551698851, "grad_norm": 4.655381679534912, "learning_rate": 1.8184955593462494e-05, "loss": 1.1297, "step": 896 }, { "epoch": 0.2192617941823515, "grad_norm": 16.207721710205078, "learning_rate": 1.818040442638971e-05, "loss": 1.1169, "step": 897 }, { "epoch": 0.21950623319481788, "grad_norm": 4.304232120513916, "learning_rate": 1.817584813151584e-05, "loss": 1.1648, "step": 898 }, { "epoch": 0.21975067220728428, "grad_norm": 2.8558313846588135, "learning_rate": 1.8171286711696935e-05, "loss": 2.0964, "step": 899 }, { "epoch": 0.21999511121975068, "grad_norm": 6.2625322341918945, "learning_rate": 1.8166720169792276e-05, "loss": 1.1234, "step": 900 }, { "epoch": 0.22023955023221706, "grad_norm": 3.8653852939605713, "learning_rate": 1.816214850866436e-05, "loss": 0.9417, "step": 901 }, { "epoch": 0.22048398924468346, "grad_norm": 4.717000961303711, "learning_rate": 1.815757173117888e-05, "loss": 1.0726, "step": 902 }, { "epoch": 0.22072842825714983, "grad_norm": 4.075290203094482, "learning_rate": 1.8152989840204734e-05, "loss": 1.1515, "step": 903 }, { "epoch": 0.22097286726961624, "grad_norm": 3.7050132751464844, "learning_rate": 1.814840283861405e-05, "loss": 1.0404, "step": 904 }, { "epoch": 0.2212173062820826, "grad_norm": 4.526838302612305, "learning_rate": 1.8143810729282125e-05, "loss": 1.0268, "step": 905 }, { "epoch": 0.221461745294549, "grad_norm": 4.461960315704346, "learning_rate": 1.8139213515087484e-05, "loss": 1.0661, "step": 906 }, { "epoch": 0.22170618430701539, "grad_norm": 2.5100247859954834, "learning_rate": 1.813461119891184e-05, "loss": 2.0317, "step": 907 }, { "epoch": 0.2219506233194818, "grad_norm": 4.275705337524414, "learning_rate": 1.8130003783640106e-05, "loss": 1.1287, "step": 908 }, { "epoch": 0.2221950623319482, "grad_norm": 26.058462142944336, "learning_rate": 1.8125391272160397e-05, "loss": 1.1154, "step": 909 }, { "epoch": 0.22243950134441456, "grad_norm": 5.324725151062012, "learning_rate": 1.8120773667364012e-05, "loss": 1.0256, "step": 910 }, { "epoch": 0.22268394035688097, "grad_norm": 4.943587303161621, "learning_rate": 1.811615097214545e-05, "loss": 1.1156, "step": 911 }, { "epoch": 0.22292837936934734, "grad_norm": 3.439422845840454, "learning_rate": 1.81115231894024e-05, "loss": 1.0305, "step": 912 }, { "epoch": 0.22317281838181374, "grad_norm": 4.746899604797363, "learning_rate": 1.8106890322035737e-05, "loss": 1.0388, "step": 913 }, { "epoch": 0.22341725739428012, "grad_norm": 3.8092880249023438, "learning_rate": 1.810225237294953e-05, "loss": 0.9694, "step": 914 }, { "epoch": 0.22366169640674652, "grad_norm": 4.545948028564453, "learning_rate": 1.8097609345051027e-05, "loss": 1.0253, "step": 915 }, { "epoch": 0.2239061354192129, "grad_norm": 3.7231202125549316, "learning_rate": 1.8092961241250658e-05, "loss": 1.0292, "step": 916 }, { "epoch": 0.2241505744316793, "grad_norm": 4.942643642425537, "learning_rate": 1.8088308064462043e-05, "loss": 1.1375, "step": 917 }, { "epoch": 0.2243950134441457, "grad_norm": 5.857900619506836, "learning_rate": 1.8083649817601976e-05, "loss": 1.0794, "step": 918 }, { "epoch": 0.22463945245661207, "grad_norm": 4.76950740814209, "learning_rate": 1.807898650359043e-05, "loss": 1.1841, "step": 919 }, { "epoch": 0.22488389146907847, "grad_norm": 3.846597194671631, "learning_rate": 1.8074318125350553e-05, "loss": 1.1022, "step": 920 }, { "epoch": 0.22512833048154485, "grad_norm": 3.865078926086426, "learning_rate": 1.8069644685808673e-05, "loss": 0.996, "step": 921 }, { "epoch": 0.22537276949401125, "grad_norm": 5.097954273223877, "learning_rate": 1.8064966187894284e-05, "loss": 1.1416, "step": 922 }, { "epoch": 0.22561720850647762, "grad_norm": 5.553801536560059, "learning_rate": 1.8060282634540053e-05, "loss": 0.9644, "step": 923 }, { "epoch": 0.22586164751894403, "grad_norm": 5.003941059112549, "learning_rate": 1.8055594028681817e-05, "loss": 1.0668, "step": 924 }, { "epoch": 0.2261060865314104, "grad_norm": 4.66826057434082, "learning_rate": 1.805090037325858e-05, "loss": 1.1768, "step": 925 }, { "epoch": 0.2263505255438768, "grad_norm": 6.5088396072387695, "learning_rate": 1.8046201671212515e-05, "loss": 1.033, "step": 926 }, { "epoch": 0.2265949645563432, "grad_norm": 4.333165645599365, "learning_rate": 1.804149792548895e-05, "loss": 1.0737, "step": 927 }, { "epoch": 0.22683940356880958, "grad_norm": 5.561150074005127, "learning_rate": 1.8036789139036375e-05, "loss": 1.0841, "step": 928 }, { "epoch": 0.22708384258127598, "grad_norm": 4.797734260559082, "learning_rate": 1.803207531480645e-05, "loss": 1.1351, "step": 929 }, { "epoch": 0.22732828159374235, "grad_norm": 5.050440311431885, "learning_rate": 1.8027356455753984e-05, "loss": 1.0342, "step": 930 }, { "epoch": 0.22757272060620876, "grad_norm": 5.201828479766846, "learning_rate": 1.8022632564836948e-05, "loss": 1.0864, "step": 931 }, { "epoch": 0.22781715961867513, "grad_norm": 4.372162342071533, "learning_rate": 1.801790364501646e-05, "loss": 0.9967, "step": 932 }, { "epoch": 0.22806159863114153, "grad_norm": 3.2812650203704834, "learning_rate": 1.801316969925679e-05, "loss": 2.1303, "step": 933 }, { "epoch": 0.2283060376436079, "grad_norm": 4.838235378265381, "learning_rate": 1.8008430730525372e-05, "loss": 1.084, "step": 934 }, { "epoch": 0.2285504766560743, "grad_norm": 6.401941776275635, "learning_rate": 1.8003686741792773e-05, "loss": 0.9873, "step": 935 }, { "epoch": 0.2287949156685407, "grad_norm": 5.116730690002441, "learning_rate": 1.7998937736032714e-05, "loss": 1.028, "step": 936 }, { "epoch": 0.22903935468100708, "grad_norm": 3.7749273777008057, "learning_rate": 1.799418371622206e-05, "loss": 1.1115, "step": 937 }, { "epoch": 0.2292837936934735, "grad_norm": 6.529001235961914, "learning_rate": 1.798942468534082e-05, "loss": 1.0567, "step": 938 }, { "epoch": 0.22952823270593986, "grad_norm": 5.243895530700684, "learning_rate": 1.7984660646372138e-05, "loss": 0.9885, "step": 939 }, { "epoch": 0.22977267171840626, "grad_norm": 6.02579927444458, "learning_rate": 1.7979891602302305e-05, "loss": 1.0929, "step": 940 }, { "epoch": 0.23001711073087264, "grad_norm": 10.071123123168945, "learning_rate": 1.797511755612075e-05, "loss": 0.9637, "step": 941 }, { "epoch": 0.23026154974333904, "grad_norm": 5.939101696014404, "learning_rate": 1.797033851082003e-05, "loss": 1.1494, "step": 942 }, { "epoch": 0.2305059887558054, "grad_norm": 4.889463424682617, "learning_rate": 1.7965554469395837e-05, "loss": 1.1101, "step": 943 }, { "epoch": 0.23075042776827182, "grad_norm": 5.173099517822266, "learning_rate": 1.7960765434847007e-05, "loss": 1.0926, "step": 944 }, { "epoch": 0.23099486678073822, "grad_norm": 9.772001266479492, "learning_rate": 1.7955971410175484e-05, "loss": 1.0194, "step": 945 }, { "epoch": 0.2312393057932046, "grad_norm": 4.739356517791748, "learning_rate": 1.7951172398386363e-05, "loss": 1.0346, "step": 946 }, { "epoch": 0.231483744805671, "grad_norm": 11.065056800842285, "learning_rate": 1.7946368402487845e-05, "loss": 1.091, "step": 947 }, { "epoch": 0.23172818381813737, "grad_norm": 4.65624475479126, "learning_rate": 1.7941559425491276e-05, "loss": 1.1244, "step": 948 }, { "epoch": 0.23197262283060377, "grad_norm": 6.162562370300293, "learning_rate": 1.79367454704111e-05, "loss": 1.0596, "step": 949 }, { "epoch": 0.23221706184307014, "grad_norm": 4.951538562774658, "learning_rate": 1.7931926540264907e-05, "loss": 1.0474, "step": 950 }, { "epoch": 0.23246150085553655, "grad_norm": 4.932707786560059, "learning_rate": 1.7927102638073384e-05, "loss": 1.0991, "step": 951 }, { "epoch": 0.23270593986800292, "grad_norm": 5.941732883453369, "learning_rate": 1.7922273766860346e-05, "loss": 1.054, "step": 952 }, { "epoch": 0.23295037888046932, "grad_norm": 5.31721830368042, "learning_rate": 1.7917439929652723e-05, "loss": 1.1043, "step": 953 }, { "epoch": 0.23319481789293572, "grad_norm": 6.00404691696167, "learning_rate": 1.7912601129480554e-05, "loss": 1.0955, "step": 954 }, { "epoch": 0.2334392569054021, "grad_norm": 5.6774187088012695, "learning_rate": 1.7907757369376984e-05, "loss": 1.0019, "step": 955 }, { "epoch": 0.2336836959178685, "grad_norm": 11.581239700317383, "learning_rate": 1.7902908652378287e-05, "loss": 1.0713, "step": 956 }, { "epoch": 0.23392813493033487, "grad_norm": 8.59800910949707, "learning_rate": 1.789805498152382e-05, "loss": 1.0113, "step": 957 }, { "epoch": 0.23417257394280128, "grad_norm": 6.038421154022217, "learning_rate": 1.7893196359856058e-05, "loss": 1.0019, "step": 958 }, { "epoch": 0.23441701295526765, "grad_norm": 15.316848754882812, "learning_rate": 1.7888332790420582e-05, "loss": 1.1193, "step": 959 }, { "epoch": 0.23466145196773405, "grad_norm": 5.798333168029785, "learning_rate": 1.7883464276266064e-05, "loss": 0.8781, "step": 960 }, { "epoch": 0.23490589098020043, "grad_norm": 9.919295310974121, "learning_rate": 1.7878590820444283e-05, "loss": 1.08, "step": 961 }, { "epoch": 0.23515032999266683, "grad_norm": 4.5314812660217285, "learning_rate": 1.7873712426010112e-05, "loss": 0.9922, "step": 962 }, { "epoch": 0.23539476900513323, "grad_norm": 5.200216770172119, "learning_rate": 1.786882909602153e-05, "loss": 1.059, "step": 963 }, { "epoch": 0.2356392080175996, "grad_norm": 12.14441967010498, "learning_rate": 1.786394083353959e-05, "loss": 0.9602, "step": 964 }, { "epoch": 0.235883647030066, "grad_norm": 6.1346964836120605, "learning_rate": 1.7859047641628452e-05, "loss": 1.0439, "step": 965 }, { "epoch": 0.23612808604253238, "grad_norm": 11.949640274047852, "learning_rate": 1.7854149523355366e-05, "loss": 1.1008, "step": 966 }, { "epoch": 0.23637252505499878, "grad_norm": 5.413366794586182, "learning_rate": 1.7849246481790664e-05, "loss": 1.0515, "step": 967 }, { "epoch": 0.23661696406746516, "grad_norm": 4.006432056427002, "learning_rate": 1.7844338520007765e-05, "loss": 1.0281, "step": 968 }, { "epoch": 0.23686140307993156, "grad_norm": 4.715639114379883, "learning_rate": 1.783942564108318e-05, "loss": 1.0322, "step": 969 }, { "epoch": 0.23710584209239793, "grad_norm": 6.650883674621582, "learning_rate": 1.7834507848096488e-05, "loss": 1.1547, "step": 970 }, { "epoch": 0.23735028110486434, "grad_norm": 3.7897934913635254, "learning_rate": 1.7829585144130356e-05, "loss": 1.0635, "step": 971 }, { "epoch": 0.23759472011733074, "grad_norm": 7.736684799194336, "learning_rate": 1.7824657532270535e-05, "loss": 1.0343, "step": 972 }, { "epoch": 0.2378391591297971, "grad_norm": 4.9490275382995605, "learning_rate": 1.7819725015605843e-05, "loss": 1.0995, "step": 973 }, { "epoch": 0.23808359814226351, "grad_norm": 4.268904209136963, "learning_rate": 1.781478759722818e-05, "loss": 1.0989, "step": 974 }, { "epoch": 0.2383280371547299, "grad_norm": 4.651836395263672, "learning_rate": 1.7809845280232512e-05, "loss": 1.0715, "step": 975 }, { "epoch": 0.2385724761671963, "grad_norm": 11.762102127075195, "learning_rate": 1.7804898067716883e-05, "loss": 1.0578, "step": 976 }, { "epoch": 0.23881691517966266, "grad_norm": 5.454673767089844, "learning_rate": 1.7799945962782397e-05, "loss": 1.0425, "step": 977 }, { "epoch": 0.23906135419212907, "grad_norm": 5.125723838806152, "learning_rate": 1.7794988968533232e-05, "loss": 1.0211, "step": 978 }, { "epoch": 0.23930579320459544, "grad_norm": 5.660276412963867, "learning_rate": 1.779002708807662e-05, "loss": 1.0719, "step": 979 }, { "epoch": 0.23955023221706184, "grad_norm": 4.565948963165283, "learning_rate": 1.778506032452288e-05, "loss": 1.0852, "step": 980 }, { "epoch": 0.23979467122952824, "grad_norm": 5.478196144104004, "learning_rate": 1.7780088680985365e-05, "loss": 1.0543, "step": 981 }, { "epoch": 0.24003911024199462, "grad_norm": 4.445716381072998, "learning_rate": 1.7775112160580497e-05, "loss": 1.1229, "step": 982 }, { "epoch": 0.24028354925446102, "grad_norm": 5.014064311981201, "learning_rate": 1.777013076642776e-05, "loss": 1.0659, "step": 983 }, { "epoch": 0.2405279882669274, "grad_norm": 6.775639057159424, "learning_rate": 1.7765144501649687e-05, "loss": 1.0905, "step": 984 }, { "epoch": 0.2407724272793938, "grad_norm": 7.390527725219727, "learning_rate": 1.776015336937187e-05, "loss": 1.008, "step": 985 }, { "epoch": 0.24101686629186017, "grad_norm": 5.768934726715088, "learning_rate": 1.7755157372722938e-05, "loss": 0.9349, "step": 986 }, { "epoch": 0.24126130530432657, "grad_norm": 3.9665780067443848, "learning_rate": 1.775015651483459e-05, "loss": 0.9906, "step": 987 }, { "epoch": 0.24150574431679295, "grad_norm": 5.45059061050415, "learning_rate": 1.7745150798841554e-05, "loss": 1.0664, "step": 988 }, { "epoch": 0.24175018332925935, "grad_norm": 8.767844200134277, "learning_rate": 1.774014022788162e-05, "loss": 1.1122, "step": 989 }, { "epoch": 0.24199462234172575, "grad_norm": 5.439845085144043, "learning_rate": 1.7735124805095602e-05, "loss": 0.9101, "step": 990 }, { "epoch": 0.24223906135419213, "grad_norm": 4.81063175201416, "learning_rate": 1.773010453362737e-05, "loss": 1.0126, "step": 991 }, { "epoch": 0.24248350036665853, "grad_norm": 6.504293441772461, "learning_rate": 1.7725079416623828e-05, "loss": 0.9735, "step": 992 }, { "epoch": 0.2427279393791249, "grad_norm": 9.619270324707031, "learning_rate": 1.772004945723492e-05, "loss": 1.1093, "step": 993 }, { "epoch": 0.2429723783915913, "grad_norm": 4.560388565063477, "learning_rate": 1.7715014658613626e-05, "loss": 1.0558, "step": 994 }, { "epoch": 0.24321681740405768, "grad_norm": 7.408061981201172, "learning_rate": 1.7709975023915948e-05, "loss": 1.1495, "step": 995 }, { "epoch": 0.24346125641652408, "grad_norm": 4.253489017486572, "learning_rate": 1.7704930556300936e-05, "loss": 1.0827, "step": 996 }, { "epoch": 0.24370569542899045, "grad_norm": 4.0216546058654785, "learning_rate": 1.7699881258930658e-05, "loss": 1.0846, "step": 997 }, { "epoch": 0.24395013444145686, "grad_norm": 5.678346633911133, "learning_rate": 1.7694827134970217e-05, "loss": 1.1458, "step": 998 }, { "epoch": 0.24419457345392326, "grad_norm": 4.36787748336792, "learning_rate": 1.7689768187587732e-05, "loss": 1.0336, "step": 999 }, { "epoch": 0.24443901246638963, "grad_norm": 3.6349778175354004, "learning_rate": 1.7684704419954356e-05, "loss": 1.07, "step": 1000 }, { "epoch": 0.24468345147885603, "grad_norm": 4.574316501617432, "learning_rate": 1.7679635835244256e-05, "loss": 1.0281, "step": 1001 }, { "epoch": 0.2449278904913224, "grad_norm": 4.44944953918457, "learning_rate": 1.7674562436634627e-05, "loss": 1.1225, "step": 1002 }, { "epoch": 0.2451723295037888, "grad_norm": 3.7810378074645996, "learning_rate": 1.766948422730567e-05, "loss": 2.1173, "step": 1003 }, { "epoch": 0.24541676851625518, "grad_norm": 6.347598552703857, "learning_rate": 1.766440121044061e-05, "loss": 1.0719, "step": 1004 }, { "epoch": 0.2456612075287216, "grad_norm": 4.953799247741699, "learning_rate": 1.765931338922568e-05, "loss": 0.9977, "step": 1005 }, { "epoch": 0.24590564654118796, "grad_norm": 6.32512092590332, "learning_rate": 1.7654220766850134e-05, "loss": 1.044, "step": 1006 }, { "epoch": 0.24615008555365436, "grad_norm": 2.5741055011749268, "learning_rate": 1.7649123346506223e-05, "loss": 2.0905, "step": 1007 }, { "epoch": 0.24639452456612077, "grad_norm": 2.821791410446167, "learning_rate": 1.7644021131389216e-05, "loss": 2.1265, "step": 1008 }, { "epoch": 0.24663896357858714, "grad_norm": 3.9031708240509033, "learning_rate": 1.763891412469738e-05, "loss": 1.0813, "step": 1009 }, { "epoch": 0.24688340259105354, "grad_norm": 4.954384803771973, "learning_rate": 1.763380232963199e-05, "loss": 1.0133, "step": 1010 }, { "epoch": 0.24712784160351992, "grad_norm": 3.2676358222961426, "learning_rate": 1.762868574939732e-05, "loss": 1.0316, "step": 1011 }, { "epoch": 0.24737228061598632, "grad_norm": 7.159889221191406, "learning_rate": 1.7623564387200648e-05, "loss": 1.119, "step": 1012 }, { "epoch": 0.2476167196284527, "grad_norm": 3.8670740127563477, "learning_rate": 1.761843824625225e-05, "loss": 1.1125, "step": 1013 }, { "epoch": 0.2478611586409191, "grad_norm": 4.3193817138671875, "learning_rate": 1.7613307329765382e-05, "loss": 1.0935, "step": 1014 }, { "epoch": 0.24810559765338547, "grad_norm": 4.301325798034668, "learning_rate": 1.7608171640956312e-05, "loss": 0.9862, "step": 1015 }, { "epoch": 0.24835003666585187, "grad_norm": 3.985088348388672, "learning_rate": 1.7603031183044296e-05, "loss": 1.046, "step": 1016 }, { "epoch": 0.24859447567831827, "grad_norm": 6.649075031280518, "learning_rate": 1.7597885959251574e-05, "loss": 1.104, "step": 1017 }, { "epoch": 0.24883891469078465, "grad_norm": 3.7859950065612793, "learning_rate": 1.759273597280337e-05, "loss": 0.8895, "step": 1018 }, { "epoch": 0.24908335370325105, "grad_norm": 4.199967384338379, "learning_rate": 1.758758122692791e-05, "loss": 0.9918, "step": 1019 }, { "epoch": 0.24932779271571742, "grad_norm": 4.479875087738037, "learning_rate": 1.7582421724856383e-05, "loss": 1.0098, "step": 1020 }, { "epoch": 0.24957223172818382, "grad_norm": 3.7866857051849365, "learning_rate": 1.7577257469822976e-05, "loss": 1.0731, "step": 1021 }, { "epoch": 0.2498166707406502, "grad_norm": 3.81478214263916, "learning_rate": 1.7572088465064847e-05, "loss": 0.9534, "step": 1022 }, { "epoch": 0.2500611097531166, "grad_norm": 4.332945346832275, "learning_rate": 1.756691471382213e-05, "loss": 1.047, "step": 1023 }, { "epoch": 0.250305548765583, "grad_norm": 4.55991268157959, "learning_rate": 1.7561736219337935e-05, "loss": 1.0302, "step": 1024 }, { "epoch": 0.2505499877780494, "grad_norm": 3.5179877281188965, "learning_rate": 1.7556552984858353e-05, "loss": 1.0994, "step": 1025 }, { "epoch": 0.2507944267905158, "grad_norm": 5.125848770141602, "learning_rate": 1.7551365013632437e-05, "loss": 1.0657, "step": 1026 }, { "epoch": 0.2510388658029822, "grad_norm": 4.182981014251709, "learning_rate": 1.7546172308912213e-05, "loss": 1.0652, "step": 1027 }, { "epoch": 0.2512833048154485, "grad_norm": 5.154917240142822, "learning_rate": 1.7540974873952673e-05, "loss": 1.11, "step": 1028 }, { "epoch": 0.25152774382791493, "grad_norm": 4.645609378814697, "learning_rate": 1.7535772712011775e-05, "loss": 1.0179, "step": 1029 }, { "epoch": 0.25177218284038133, "grad_norm": 4.523293495178223, "learning_rate": 1.7530565826350444e-05, "loss": 1.0007, "step": 1030 }, { "epoch": 0.25201662185284773, "grad_norm": 4.768362998962402, "learning_rate": 1.7525354220232558e-05, "loss": 1.0095, "step": 1031 }, { "epoch": 0.2522610608653141, "grad_norm": 4.03495454788208, "learning_rate": 1.752013789692496e-05, "loss": 1.0007, "step": 1032 }, { "epoch": 0.2525054998777805, "grad_norm": 4.036981105804443, "learning_rate": 1.7514916859697443e-05, "loss": 1.0643, "step": 1033 }, { "epoch": 0.2527499388902469, "grad_norm": 4.7859978675842285, "learning_rate": 1.7509691111822764e-05, "loss": 1.0359, "step": 1034 }, { "epoch": 0.2529943779027133, "grad_norm": 4.358851909637451, "learning_rate": 1.750446065657663e-05, "loss": 0.9784, "step": 1035 }, { "epoch": 0.2532388169151797, "grad_norm": 4.325262546539307, "learning_rate": 1.7499225497237686e-05, "loss": 1.0047, "step": 1036 }, { "epoch": 0.25348325592764603, "grad_norm": 5.185552597045898, "learning_rate": 1.7493985637087555e-05, "loss": 1.14, "step": 1037 }, { "epoch": 0.25372769494011244, "grad_norm": 4.701370716094971, "learning_rate": 1.7488741079410775e-05, "loss": 2.2303, "step": 1038 }, { "epoch": 0.25397213395257884, "grad_norm": 6.095943927764893, "learning_rate": 1.7483491827494846e-05, "loss": 1.1368, "step": 1039 }, { "epoch": 0.25421657296504524, "grad_norm": 4.193117141723633, "learning_rate": 1.7478237884630208e-05, "loss": 1.2079, "step": 1040 }, { "epoch": 0.2544610119775116, "grad_norm": 4.067698001861572, "learning_rate": 1.747297925411024e-05, "loss": 1.045, "step": 1041 }, { "epoch": 0.254705450989978, "grad_norm": 4.788492202758789, "learning_rate": 1.746771593923126e-05, "loss": 0.9934, "step": 1042 }, { "epoch": 0.2549498900024444, "grad_norm": 8.858131408691406, "learning_rate": 1.7462447943292522e-05, "loss": 1.0462, "step": 1043 }, { "epoch": 0.2551943290149108, "grad_norm": 3.593235492706299, "learning_rate": 1.745717526959621e-05, "loss": 1.0151, "step": 1044 }, { "epoch": 0.2554387680273772, "grad_norm": 4.977777004241943, "learning_rate": 1.7451897921447448e-05, "loss": 1.0593, "step": 1045 }, { "epoch": 0.25568320703984354, "grad_norm": 4.215940952301025, "learning_rate": 1.7446615902154288e-05, "loss": 1.0583, "step": 1046 }, { "epoch": 0.25592764605230994, "grad_norm": 3.8196914196014404, "learning_rate": 1.7441329215027707e-05, "loss": 1.0829, "step": 1047 }, { "epoch": 0.25617208506477634, "grad_norm": 4.272761821746826, "learning_rate": 1.743603786338161e-05, "loss": 1.108, "step": 1048 }, { "epoch": 0.25641652407724275, "grad_norm": 3.1501529216766357, "learning_rate": 1.7430741850532824e-05, "loss": 0.9738, "step": 1049 }, { "epoch": 0.2566609630897091, "grad_norm": 3.941812753677368, "learning_rate": 1.74254411798011e-05, "loss": 1.1132, "step": 1050 }, { "epoch": 0.2569054021021755, "grad_norm": 4.7528276443481445, "learning_rate": 1.742013585450911e-05, "loss": 1.144, "step": 1051 }, { "epoch": 0.2571498411146419, "grad_norm": 7.925037384033203, "learning_rate": 1.741482587798244e-05, "loss": 2.2551, "step": 1052 }, { "epoch": 0.2573942801271083, "grad_norm": 5.1193318367004395, "learning_rate": 1.7409511253549592e-05, "loss": 1.0979, "step": 1053 }, { "epoch": 0.2576387191395747, "grad_norm": 5.097321033477783, "learning_rate": 1.7404191984541987e-05, "loss": 1.0282, "step": 1054 }, { "epoch": 0.25788315815204105, "grad_norm": 4.32500696182251, "learning_rate": 1.7398868074293947e-05, "loss": 1.1019, "step": 1055 }, { "epoch": 0.25812759716450745, "grad_norm": 6.229393005371094, "learning_rate": 1.7393539526142717e-05, "loss": 1.034, "step": 1056 }, { "epoch": 0.25837203617697385, "grad_norm": 4.147944927215576, "learning_rate": 1.7388206343428438e-05, "loss": 1.0432, "step": 1057 }, { "epoch": 0.25861647518944025, "grad_norm": 3.581871271133423, "learning_rate": 1.7382868529494158e-05, "loss": 1.0871, "step": 1058 }, { "epoch": 0.2588609142019066, "grad_norm": 5.094482421875, "learning_rate": 1.7377526087685832e-05, "loss": 1.0374, "step": 1059 }, { "epoch": 0.259105353214373, "grad_norm": 3.791944980621338, "learning_rate": 1.7372179021352317e-05, "loss": 1.1247, "step": 1060 }, { "epoch": 0.2593497922268394, "grad_norm": 4.173458099365234, "learning_rate": 1.736682733384536e-05, "loss": 0.9955, "step": 1061 }, { "epoch": 0.2595942312393058, "grad_norm": 3.895758628845215, "learning_rate": 1.7361471028519613e-05, "loss": 1.1296, "step": 1062 }, { "epoch": 0.2598386702517722, "grad_norm": 3.5435473918914795, "learning_rate": 1.735611010873262e-05, "loss": 1.0162, "step": 1063 }, { "epoch": 0.26008310926423855, "grad_norm": 6.77610969543457, "learning_rate": 1.7350744577844816e-05, "loss": 0.95, "step": 1064 }, { "epoch": 0.26032754827670496, "grad_norm": 7.926026344299316, "learning_rate": 1.734537443921953e-05, "loss": 1.0358, "step": 1065 }, { "epoch": 0.26057198728917136, "grad_norm": 3.391551971435547, "learning_rate": 1.733999969622298e-05, "loss": 0.9246, "step": 1066 }, { "epoch": 0.26081642630163776, "grad_norm": 4.00184440612793, "learning_rate": 1.733462035222426e-05, "loss": 1.0923, "step": 1067 }, { "epoch": 0.2610608653141041, "grad_norm": 6.604822158813477, "learning_rate": 1.7329236410595364e-05, "loss": 0.9171, "step": 1068 }, { "epoch": 0.2613053043265705, "grad_norm": 4.701949596405029, "learning_rate": 1.7323847874711152e-05, "loss": 1.0092, "step": 1069 }, { "epoch": 0.2615497433390369, "grad_norm": 4.597169399261475, "learning_rate": 1.7318454747949375e-05, "loss": 1.0115, "step": 1070 }, { "epoch": 0.2617941823515033, "grad_norm": 5.872190952301025, "learning_rate": 1.7313057033690662e-05, "loss": 1.0412, "step": 1071 }, { "epoch": 0.2620386213639697, "grad_norm": 5.632815837860107, "learning_rate": 1.7307654735318505e-05, "loss": 1.0499, "step": 1072 }, { "epoch": 0.26228306037643606, "grad_norm": 5.916876792907715, "learning_rate": 1.730224785621929e-05, "loss": 1.0165, "step": 1073 }, { "epoch": 0.26252749938890246, "grad_norm": 3.9583590030670166, "learning_rate": 1.7296836399782253e-05, "loss": 0.983, "step": 1074 }, { "epoch": 0.26277193840136887, "grad_norm": 3.7609782218933105, "learning_rate": 1.729142036939951e-05, "loss": 1.0606, "step": 1075 }, { "epoch": 0.26301637741383527, "grad_norm": 8.059402465820312, "learning_rate": 1.7285999768466048e-05, "loss": 1.0308, "step": 1076 }, { "epoch": 0.2632608164263016, "grad_norm": 4.188074111938477, "learning_rate": 1.728057460037971e-05, "loss": 1.0676, "step": 1077 }, { "epoch": 0.263505255438768, "grad_norm": 3.7996623516082764, "learning_rate": 1.7275144868541208e-05, "loss": 1.0604, "step": 1078 }, { "epoch": 0.2637496944512344, "grad_norm": 5.376227378845215, "learning_rate": 1.7269710576354112e-05, "loss": 1.1378, "step": 1079 }, { "epoch": 0.2639941334637008, "grad_norm": 4.061453819274902, "learning_rate": 1.7264271727224847e-05, "loss": 1.0861, "step": 1080 }, { "epoch": 0.2642385724761672, "grad_norm": 3.80562686920166, "learning_rate": 1.7258828324562705e-05, "loss": 1.0534, "step": 1081 }, { "epoch": 0.26448301148863357, "grad_norm": 4.415918350219727, "learning_rate": 1.725338037177983e-05, "loss": 0.9853, "step": 1082 }, { "epoch": 0.26472745050109997, "grad_norm": 4.435208797454834, "learning_rate": 1.72479278722912e-05, "loss": 1.1745, "step": 1083 }, { "epoch": 0.26497188951356637, "grad_norm": 3.257791042327881, "learning_rate": 1.7242470829514674e-05, "loss": 2.1818, "step": 1084 }, { "epoch": 0.2652163285260328, "grad_norm": 4.946976184844971, "learning_rate": 1.723700924687093e-05, "loss": 1.2528, "step": 1085 }, { "epoch": 0.2654607675384991, "grad_norm": 4.8052802085876465, "learning_rate": 1.7231543127783506e-05, "loss": 1.034, "step": 1086 }, { "epoch": 0.2657052065509655, "grad_norm": 3.2481181621551514, "learning_rate": 1.7226072475678783e-05, "loss": 1.1201, "step": 1087 }, { "epoch": 0.2659496455634319, "grad_norm": 3.8651785850524902, "learning_rate": 1.7220597293985987e-05, "loss": 0.9795, "step": 1088 }, { "epoch": 0.2661940845758983, "grad_norm": 4.134873867034912, "learning_rate": 1.7215117586137174e-05, "loss": 2.0171, "step": 1089 }, { "epoch": 0.26643852358836473, "grad_norm": 3.266214609146118, "learning_rate": 1.7209633355567242e-05, "loss": 0.9662, "step": 1090 }, { "epoch": 0.2666829626008311, "grad_norm": 4.433269023895264, "learning_rate": 1.7204144605713922e-05, "loss": 1.0065, "step": 1091 }, { "epoch": 0.2669274016132975, "grad_norm": 3.932711601257324, "learning_rate": 1.7198651340017783e-05, "loss": 0.9736, "step": 1092 }, { "epoch": 0.2671718406257639, "grad_norm": 4.1653733253479, "learning_rate": 1.7193153561922214e-05, "loss": 0.9304, "step": 1093 }, { "epoch": 0.2674162796382303, "grad_norm": 6.558185577392578, "learning_rate": 1.718765127487345e-05, "loss": 1.0107, "step": 1094 }, { "epoch": 0.2676607186506966, "grad_norm": 4.074313640594482, "learning_rate": 1.718214448232054e-05, "loss": 1.1244, "step": 1095 }, { "epoch": 0.26790515766316303, "grad_norm": 4.0094733238220215, "learning_rate": 1.7176633187715353e-05, "loss": 1.0114, "step": 1096 }, { "epoch": 0.26814959667562943, "grad_norm": 4.766077041625977, "learning_rate": 1.7171117394512594e-05, "loss": 1.0161, "step": 1097 }, { "epoch": 0.26839403568809583, "grad_norm": 3.321650743484497, "learning_rate": 1.7165597106169775e-05, "loss": 0.9981, "step": 1098 }, { "epoch": 0.26863847470056224, "grad_norm": 3.749901056289673, "learning_rate": 1.716007232614723e-05, "loss": 1.0639, "step": 1099 }, { "epoch": 0.2688829137130286, "grad_norm": 4.82611608505249, "learning_rate": 1.7154543057908116e-05, "loss": 1.0381, "step": 1100 }, { "epoch": 0.269127352725495, "grad_norm": 7.40592622756958, "learning_rate": 1.7149009304918392e-05, "loss": 0.9762, "step": 1101 }, { "epoch": 0.2693717917379614, "grad_norm": 3.8617756366729736, "learning_rate": 1.7143471070646834e-05, "loss": 1.0359, "step": 1102 }, { "epoch": 0.2696162307504278, "grad_norm": 3.7181596755981445, "learning_rate": 1.713792835856503e-05, "loss": 1.0627, "step": 1103 }, { "epoch": 0.26986066976289413, "grad_norm": 5.8879714012146, "learning_rate": 1.7132381172147368e-05, "loss": 0.9501, "step": 1104 }, { "epoch": 0.27010510877536054, "grad_norm": 3.9147109985351562, "learning_rate": 1.7126829514871043e-05, "loss": 1.0481, "step": 1105 }, { "epoch": 0.27034954778782694, "grad_norm": 3.318326234817505, "learning_rate": 1.7121273390216052e-05, "loss": 1.0857, "step": 1106 }, { "epoch": 0.27059398680029334, "grad_norm": 3.6183900833129883, "learning_rate": 1.71157128016652e-05, "loss": 1.0572, "step": 1107 }, { "epoch": 0.27083842581275974, "grad_norm": 15.032169342041016, "learning_rate": 1.711014775270408e-05, "loss": 1.0478, "step": 1108 }, { "epoch": 0.2710828648252261, "grad_norm": 3.319171190261841, "learning_rate": 1.7104578246821083e-05, "loss": 0.9196, "step": 1109 }, { "epoch": 0.2713273038376925, "grad_norm": 3.304271697998047, "learning_rate": 1.70990042875074e-05, "loss": 2.085, "step": 1110 }, { "epoch": 0.2715717428501589, "grad_norm": 4.110102653503418, "learning_rate": 1.7093425878257007e-05, "loss": 0.9727, "step": 1111 }, { "epoch": 0.2718161818626253, "grad_norm": 6.215137958526611, "learning_rate": 1.708784302256667e-05, "loss": 1.1239, "step": 1112 }, { "epoch": 0.27206062087509164, "grad_norm": 5.920506477355957, "learning_rate": 1.7082255723935944e-05, "loss": 1.1007, "step": 1113 }, { "epoch": 0.27230505988755804, "grad_norm": 6.019047260284424, "learning_rate": 1.7076663985867174e-05, "loss": 1.1096, "step": 1114 }, { "epoch": 0.27254949890002445, "grad_norm": 3.635051727294922, "learning_rate": 1.7071067811865477e-05, "loss": 1.1436, "step": 1115 }, { "epoch": 0.27279393791249085, "grad_norm": 4.673436164855957, "learning_rate": 1.706546720543876e-05, "loss": 0.9838, "step": 1116 }, { "epoch": 0.27303837692495725, "grad_norm": 2.6015164852142334, "learning_rate": 1.7059862170097696e-05, "loss": 2.0027, "step": 1117 }, { "epoch": 0.2732828159374236, "grad_norm": 4.968930244445801, "learning_rate": 1.7054252709355754e-05, "loss": 0.9975, "step": 1118 }, { "epoch": 0.27352725494989, "grad_norm": 2.573199510574341, "learning_rate": 1.7048638826729163e-05, "loss": 2.0794, "step": 1119 }, { "epoch": 0.2737716939623564, "grad_norm": 4.521797180175781, "learning_rate": 1.704302052573692e-05, "loss": 0.9813, "step": 1120 }, { "epoch": 0.2740161329748228, "grad_norm": 3.8551506996154785, "learning_rate": 1.7037397809900807e-05, "loss": 1.1006, "step": 1121 }, { "epoch": 0.27426057198728915, "grad_norm": 3.2936413288116455, "learning_rate": 1.703177068274536e-05, "loss": 1.0544, "step": 1122 }, { "epoch": 0.27450501099975555, "grad_norm": 3.572744846343994, "learning_rate": 1.7026139147797887e-05, "loss": 1.0277, "step": 1123 }, { "epoch": 0.27474945001222195, "grad_norm": 3.7930381298065186, "learning_rate": 1.7020503208588456e-05, "loss": 1.0485, "step": 1124 }, { "epoch": 0.27499388902468835, "grad_norm": 3.3152592182159424, "learning_rate": 1.7014862868649894e-05, "loss": 0.9606, "step": 1125 }, { "epoch": 0.27523832803715476, "grad_norm": 4.435396671295166, "learning_rate": 1.7009218131517795e-05, "loss": 1.0386, "step": 1126 }, { "epoch": 0.2754827670496211, "grad_norm": 3.8358654975891113, "learning_rate": 1.7003569000730498e-05, "loss": 1.0104, "step": 1127 }, { "epoch": 0.2757272060620875, "grad_norm": 4.017136096954346, "learning_rate": 1.6997915479829104e-05, "loss": 0.9864, "step": 1128 }, { "epoch": 0.2759716450745539, "grad_norm": 4.429943084716797, "learning_rate": 1.6992257572357465e-05, "loss": 1.0166, "step": 1129 }, { "epoch": 0.2762160840870203, "grad_norm": 4.847922325134277, "learning_rate": 1.6986595281862177e-05, "loss": 1.0326, "step": 1130 }, { "epoch": 0.27646052309948665, "grad_norm": 2.732870101928711, "learning_rate": 1.698092861189259e-05, "loss": 2.1184, "step": 1131 }, { "epoch": 0.27670496211195306, "grad_norm": 3.975236654281616, "learning_rate": 1.69752575660008e-05, "loss": 1.0131, "step": 1132 }, { "epoch": 0.27694940112441946, "grad_norm": 4.296026229858398, "learning_rate": 1.696958214774164e-05, "loss": 0.9962, "step": 1133 }, { "epoch": 0.27719384013688586, "grad_norm": 3.756542921066284, "learning_rate": 1.6963902360672684e-05, "loss": 1.0905, "step": 1134 }, { "epoch": 0.27743827914935226, "grad_norm": 3.982203960418701, "learning_rate": 1.695821820835425e-05, "loss": 1.1649, "step": 1135 }, { "epoch": 0.2776827181618186, "grad_norm": 4.048048973083496, "learning_rate": 1.6952529694349397e-05, "loss": 1.0912, "step": 1136 }, { "epoch": 0.277927157174285, "grad_norm": 3.5503664016723633, "learning_rate": 1.6946836822223895e-05, "loss": 0.9794, "step": 1137 }, { "epoch": 0.2781715961867514, "grad_norm": 2.99806547164917, "learning_rate": 1.6941139595546273e-05, "loss": 0.9834, "step": 1138 }, { "epoch": 0.2784160351992178, "grad_norm": 3.380628824234009, "learning_rate": 1.693543801788777e-05, "loss": 1.0707, "step": 1139 }, { "epoch": 0.27866047421168416, "grad_norm": 4.29387903213501, "learning_rate": 1.6929732092822372e-05, "loss": 1.0751, "step": 1140 }, { "epoch": 0.27890491322415056, "grad_norm": 3.4037857055664062, "learning_rate": 1.6924021823926766e-05, "loss": 1.031, "step": 1141 }, { "epoch": 0.27914935223661697, "grad_norm": 4.047512054443359, "learning_rate": 1.6918307214780382e-05, "loss": 1.1181, "step": 1142 }, { "epoch": 0.27939379124908337, "grad_norm": 3.380204677581787, "learning_rate": 1.6912588268965363e-05, "loss": 1.1261, "step": 1143 }, { "epoch": 0.27963823026154977, "grad_norm": 3.7261037826538086, "learning_rate": 1.6906864990066563e-05, "loss": 1.0642, "step": 1144 }, { "epoch": 0.2798826692740161, "grad_norm": 3.2923035621643066, "learning_rate": 1.690113738167157e-05, "loss": 1.1016, "step": 1145 }, { "epoch": 0.2801271082864825, "grad_norm": 3.73152756690979, "learning_rate": 1.689540544737067e-05, "loss": 0.9956, "step": 1146 }, { "epoch": 0.2803715472989489, "grad_norm": 3.534151077270508, "learning_rate": 1.688966919075687e-05, "loss": 1.0825, "step": 1147 }, { "epoch": 0.2806159863114153, "grad_norm": 3.604247808456421, "learning_rate": 1.688392861542587e-05, "loss": 0.9772, "step": 1148 }, { "epoch": 0.28086042532388167, "grad_norm": 2.977870464324951, "learning_rate": 1.6878183724976107e-05, "loss": 0.9825, "step": 1149 }, { "epoch": 0.28110486433634807, "grad_norm": 4.251736164093018, "learning_rate": 1.6872434523008697e-05, "loss": 0.9403, "step": 1150 }, { "epoch": 0.2813493033488145, "grad_norm": 5.256099700927734, "learning_rate": 1.6866681013127466e-05, "loss": 1.0369, "step": 1151 }, { "epoch": 0.2815937423612809, "grad_norm": 4.620885848999023, "learning_rate": 1.6860923198938945e-05, "loss": 1.1105, "step": 1152 }, { "epoch": 0.2818381813737473, "grad_norm": 5.067963600158691, "learning_rate": 1.6855161084052358e-05, "loss": 1.02, "step": 1153 }, { "epoch": 0.2820826203862136, "grad_norm": 3.8163278102874756, "learning_rate": 1.6849394672079625e-05, "loss": 0.9542, "step": 1154 }, { "epoch": 0.28232705939868, "grad_norm": 4.469657897949219, "learning_rate": 1.6843623966635364e-05, "loss": 0.9513, "step": 1155 }, { "epoch": 0.2825714984111464, "grad_norm": 3.72050404548645, "learning_rate": 1.683784897133688e-05, "loss": 1.0148, "step": 1156 }, { "epoch": 0.28281593742361283, "grad_norm": 5.091887474060059, "learning_rate": 1.683206968980417e-05, "loss": 0.9674, "step": 1157 }, { "epoch": 0.2830603764360792, "grad_norm": 3.263423442840576, "learning_rate": 1.6826286125659913e-05, "loss": 1.0073, "step": 1158 }, { "epoch": 0.2833048154485456, "grad_norm": 4.400149345397949, "learning_rate": 1.682049828252948e-05, "loss": 1.0492, "step": 1159 }, { "epoch": 0.283549254461012, "grad_norm": 3.1265735626220703, "learning_rate": 1.6814706164040913e-05, "loss": 1.0499, "step": 1160 }, { "epoch": 0.2837936934734784, "grad_norm": 4.412779331207275, "learning_rate": 1.6808909773824952e-05, "loss": 0.9605, "step": 1161 }, { "epoch": 0.2840381324859448, "grad_norm": 4.249974250793457, "learning_rate": 1.6803109115515e-05, "loss": 1.0151, "step": 1162 }, { "epoch": 0.28428257149841113, "grad_norm": 3.751030445098877, "learning_rate": 1.679730419274713e-05, "loss": 1.029, "step": 1163 }, { "epoch": 0.28452701051087753, "grad_norm": 2.881483316421509, "learning_rate": 1.679149500916011e-05, "loss": 2.0311, "step": 1164 }, { "epoch": 0.28477144952334393, "grad_norm": 6.882498264312744, "learning_rate": 1.6785681568395353e-05, "loss": 1.0385, "step": 1165 }, { "epoch": 0.28501588853581034, "grad_norm": 3.647284507751465, "learning_rate": 1.677986387409696e-05, "loss": 1.1402, "step": 1166 }, { "epoch": 0.2852603275482767, "grad_norm": 4.305856227874756, "learning_rate": 1.67740419299117e-05, "loss": 1.0261, "step": 1167 }, { "epoch": 0.2855047665607431, "grad_norm": 3.294780731201172, "learning_rate": 1.6768215739488986e-05, "loss": 0.9288, "step": 1168 }, { "epoch": 0.2857492055732095, "grad_norm": 3.736102342605591, "learning_rate": 1.676238530648091e-05, "loss": 0.8749, "step": 1169 }, { "epoch": 0.2859936445856759, "grad_norm": 4.358032703399658, "learning_rate": 1.6756550634542218e-05, "loss": 1.0092, "step": 1170 }, { "epoch": 0.2862380835981423, "grad_norm": 3.4615347385406494, "learning_rate": 1.675071172733031e-05, "loss": 1.1267, "step": 1171 }, { "epoch": 0.28648252261060864, "grad_norm": 3.5754973888397217, "learning_rate": 1.674486858850525e-05, "loss": 2.041, "step": 1172 }, { "epoch": 0.28672696162307504, "grad_norm": 3.58276629447937, "learning_rate": 1.6739021221729743e-05, "loss": 0.9799, "step": 1173 }, { "epoch": 0.28697140063554144, "grad_norm": 2.8018088340759277, "learning_rate": 1.673316963066915e-05, "loss": 0.9621, "step": 1174 }, { "epoch": 0.28721583964800784, "grad_norm": 3.283891201019287, "learning_rate": 1.6727313818991485e-05, "loss": 1.0314, "step": 1175 }, { "epoch": 0.2874602786604742, "grad_norm": 3.2494401931762695, "learning_rate": 1.6721453790367396e-05, "loss": 1.0646, "step": 1176 }, { "epoch": 0.2877047176729406, "grad_norm": 3.952605724334717, "learning_rate": 1.6715589548470187e-05, "loss": 1.0435, "step": 1177 }, { "epoch": 0.287949156685407, "grad_norm": 5.484105110168457, "learning_rate": 1.670972109697579e-05, "loss": 0.9748, "step": 1178 }, { "epoch": 0.2881935956978734, "grad_norm": 3.5420234203338623, "learning_rate": 1.6703848439562787e-05, "loss": 0.9723, "step": 1179 }, { "epoch": 0.2884380347103398, "grad_norm": 3.4928793907165527, "learning_rate": 1.6697971579912392e-05, "loss": 1.0251, "step": 1180 }, { "epoch": 0.28868247372280614, "grad_norm": 3.41666579246521, "learning_rate": 1.669209052170845e-05, "loss": 1.0639, "step": 1181 }, { "epoch": 0.28892691273527255, "grad_norm": 4.020567417144775, "learning_rate": 1.6686205268637444e-05, "loss": 1.0303, "step": 1182 }, { "epoch": 0.28917135174773895, "grad_norm": 3.600407361984253, "learning_rate": 1.668031582438848e-05, "loss": 1.0419, "step": 1183 }, { "epoch": 0.28941579076020535, "grad_norm": 3.676823377609253, "learning_rate": 1.6674422192653305e-05, "loss": 1.0094, "step": 1184 }, { "epoch": 0.2896602297726717, "grad_norm": 3.4822778701782227, "learning_rate": 1.666852437712626e-05, "loss": 1.0277, "step": 1185 }, { "epoch": 0.2899046687851381, "grad_norm": 3.104646921157837, "learning_rate": 1.6662622381504353e-05, "loss": 0.9741, "step": 1186 }, { "epoch": 0.2901491077976045, "grad_norm": 3.453314781188965, "learning_rate": 1.6656716209487175e-05, "loss": 0.9603, "step": 1187 }, { "epoch": 0.2903935468100709, "grad_norm": 4.3645524978637695, "learning_rate": 1.6650805864776946e-05, "loss": 1.0422, "step": 1188 }, { "epoch": 0.2906379858225373, "grad_norm": 3.804931402206421, "learning_rate": 1.664489135107851e-05, "loss": 0.9541, "step": 1189 }, { "epoch": 0.29088242483500365, "grad_norm": 3.7183544635772705, "learning_rate": 1.6638972672099317e-05, "loss": 0.9893, "step": 1190 }, { "epoch": 0.29112686384747005, "grad_norm": 3.7971720695495605, "learning_rate": 1.6633049831549424e-05, "loss": 1.1019, "step": 1191 }, { "epoch": 0.29137130285993645, "grad_norm": 3.7743401527404785, "learning_rate": 1.6627122833141514e-05, "loss": 0.9558, "step": 1192 }, { "epoch": 0.29161574187240286, "grad_norm": 4.806497097015381, "learning_rate": 1.6621191680590853e-05, "loss": 0.9999, "step": 1193 }, { "epoch": 0.2918601808848692, "grad_norm": 3.7626373767852783, "learning_rate": 1.6615256377615324e-05, "loss": 1.0013, "step": 1194 }, { "epoch": 0.2921046198973356, "grad_norm": 44.19645690917969, "learning_rate": 1.660931692793541e-05, "loss": 1.0507, "step": 1195 }, { "epoch": 0.292349058909802, "grad_norm": 3.0135140419006348, "learning_rate": 1.6603373335274194e-05, "loss": 2.162, "step": 1196 }, { "epoch": 0.2925934979222684, "grad_norm": 3.414067029953003, "learning_rate": 1.6597425603357357e-05, "loss": 1.0797, "step": 1197 }, { "epoch": 0.2928379369347348, "grad_norm": 4.47196626663208, "learning_rate": 1.659147373591317e-05, "loss": 0.865, "step": 1198 }, { "epoch": 0.29308237594720116, "grad_norm": 4.882946014404297, "learning_rate": 1.6585517736672498e-05, "loss": 0.9721, "step": 1199 }, { "epoch": 0.29332681495966756, "grad_norm": 4.430570125579834, "learning_rate": 1.6579557609368803e-05, "loss": 1.0138, "step": 1200 }, { "epoch": 0.29357125397213396, "grad_norm": 3.383298397064209, "learning_rate": 1.657359335773812e-05, "loss": 0.9532, "step": 1201 }, { "epoch": 0.29381569298460036, "grad_norm": 4.805011749267578, "learning_rate": 1.656762498551908e-05, "loss": 1.0646, "step": 1202 }, { "epoch": 0.2940601319970667, "grad_norm": 3.1969213485717773, "learning_rate": 1.65616524964529e-05, "loss": 1.0311, "step": 1203 }, { "epoch": 0.2943045710095331, "grad_norm": 3.598453998565674, "learning_rate": 1.655567589428337e-05, "loss": 0.9863, "step": 1204 }, { "epoch": 0.2945490100219995, "grad_norm": 2.7105185985565186, "learning_rate": 1.6549695182756857e-05, "loss": 2.0363, "step": 1205 }, { "epoch": 0.2947934490344659, "grad_norm": 4.3872599601745605, "learning_rate": 1.654371036562231e-05, "loss": 0.9133, "step": 1206 }, { "epoch": 0.2950378880469323, "grad_norm": 4.585229873657227, "learning_rate": 1.653772144663125e-05, "loss": 1.1049, "step": 1207 }, { "epoch": 0.29528232705939866, "grad_norm": 4.361045837402344, "learning_rate": 1.6531728429537766e-05, "loss": 1.1456, "step": 1208 }, { "epoch": 0.29552676607186507, "grad_norm": 4.613654136657715, "learning_rate": 1.6525731318098526e-05, "loss": 1.0033, "step": 1209 }, { "epoch": 0.29577120508433147, "grad_norm": 3.5829994678497314, "learning_rate": 1.651973011607275e-05, "loss": 0.9729, "step": 1210 }, { "epoch": 0.29601564409679787, "grad_norm": 3.3050808906555176, "learning_rate": 1.6513724827222225e-05, "loss": 1.0571, "step": 1211 }, { "epoch": 0.2962600831092642, "grad_norm": 3.048600673675537, "learning_rate": 1.6507715455311308e-05, "loss": 0.9471, "step": 1212 }, { "epoch": 0.2965045221217306, "grad_norm": 3.568232297897339, "learning_rate": 1.6501702004106914e-05, "loss": 0.9478, "step": 1213 }, { "epoch": 0.296748961134197, "grad_norm": 3.4384896755218506, "learning_rate": 1.6495684477378507e-05, "loss": 1.0493, "step": 1214 }, { "epoch": 0.2969934001466634, "grad_norm": 3.981926918029785, "learning_rate": 1.6489662878898115e-05, "loss": 0.9608, "step": 1215 }, { "epoch": 0.2972378391591298, "grad_norm": 3.6841025352478027, "learning_rate": 1.6483637212440312e-05, "loss": 0.9997, "step": 1216 }, { "epoch": 0.29748227817159617, "grad_norm": 3.6886446475982666, "learning_rate": 1.6477607481782222e-05, "loss": 0.9751, "step": 1217 }, { "epoch": 0.2977267171840626, "grad_norm": 3.5062296390533447, "learning_rate": 1.6471573690703523e-05, "loss": 0.9781, "step": 1218 }, { "epoch": 0.297971156196529, "grad_norm": 4.244730472564697, "learning_rate": 1.6465535842986433e-05, "loss": 1.0561, "step": 1219 }, { "epoch": 0.2982155952089954, "grad_norm": 3.478639602661133, "learning_rate": 1.645949394241571e-05, "loss": 1.0163, "step": 1220 }, { "epoch": 0.2984600342214617, "grad_norm": 3.335463523864746, "learning_rate": 1.645344799277866e-05, "loss": 0.9661, "step": 1221 }, { "epoch": 0.2987044732339281, "grad_norm": 3.1549289226531982, "learning_rate": 1.6447397997865124e-05, "loss": 0.902, "step": 1222 }, { "epoch": 0.2989489122463945, "grad_norm": 3.6085996627807617, "learning_rate": 1.6441343961467473e-05, "loss": 1.0333, "step": 1223 }, { "epoch": 0.29919335125886093, "grad_norm": 4.483367443084717, "learning_rate": 1.643528588738062e-05, "loss": 0.9929, "step": 1224 }, { "epoch": 0.29943779027132733, "grad_norm": 3.825385808944702, "learning_rate": 1.642922377940201e-05, "loss": 1.0265, "step": 1225 }, { "epoch": 0.2996822292837937, "grad_norm": 3.291959285736084, "learning_rate": 1.6423157641331604e-05, "loss": 0.9699, "step": 1226 }, { "epoch": 0.2999266682962601, "grad_norm": 4.078056335449219, "learning_rate": 1.64170874769719e-05, "loss": 0.9933, "step": 1227 }, { "epoch": 0.3001711073087265, "grad_norm": 3.92166805267334, "learning_rate": 1.641101329012792e-05, "loss": 0.9115, "step": 1228 }, { "epoch": 0.3004155463211929, "grad_norm": 4.900821208953857, "learning_rate": 1.6404935084607202e-05, "loss": 0.9236, "step": 1229 }, { "epoch": 0.30065998533365923, "grad_norm": 2.878377914428711, "learning_rate": 1.6398852864219807e-05, "loss": 2.0352, "step": 1230 }, { "epoch": 0.30090442434612563, "grad_norm": 3.8259389400482178, "learning_rate": 1.639276663277831e-05, "loss": 1.0943, "step": 1231 }, { "epoch": 0.30114886335859203, "grad_norm": 4.231867790222168, "learning_rate": 1.63866763940978e-05, "loss": 0.9894, "step": 1232 }, { "epoch": 0.30139330237105844, "grad_norm": 5.750034809112549, "learning_rate": 1.6380582151995886e-05, "loss": 1.0019, "step": 1233 }, { "epoch": 0.30163774138352484, "grad_norm": 3.387482166290283, "learning_rate": 1.6374483910292675e-05, "loss": 0.9352, "step": 1234 }, { "epoch": 0.3018821803959912, "grad_norm": 4.67958402633667, "learning_rate": 1.6368381672810788e-05, "loss": 1.0685, "step": 1235 }, { "epoch": 0.3021266194084576, "grad_norm": 4.168334007263184, "learning_rate": 1.6362275443375343e-05, "loss": 0.8781, "step": 1236 }, { "epoch": 0.302371058420924, "grad_norm": 3.400838851928711, "learning_rate": 1.6356165225813975e-05, "loss": 0.9975, "step": 1237 }, { "epoch": 0.3026154974333904, "grad_norm": 3.742551326751709, "learning_rate": 1.6350051023956806e-05, "loss": 0.9164, "step": 1238 }, { "epoch": 0.30285993644585674, "grad_norm": 4.234443187713623, "learning_rate": 1.6343932841636455e-05, "loss": 1.1713, "step": 1239 }, { "epoch": 0.30310437545832314, "grad_norm": 3.2706210613250732, "learning_rate": 1.633781068268805e-05, "loss": 1.0239, "step": 1240 }, { "epoch": 0.30334881447078954, "grad_norm": 3.5965492725372314, "learning_rate": 1.6331684550949197e-05, "loss": 0.9647, "step": 1241 }, { "epoch": 0.30359325348325594, "grad_norm": 3.227299213409424, "learning_rate": 1.6325554450259997e-05, "loss": 1.0027, "step": 1242 }, { "epoch": 0.30383769249572234, "grad_norm": 4.168364524841309, "learning_rate": 1.631942038446304e-05, "loss": 0.9438, "step": 1243 }, { "epoch": 0.3040821315081887, "grad_norm": 3.197082042694092, "learning_rate": 1.6313282357403407e-05, "loss": 1.0672, "step": 1244 }, { "epoch": 0.3043265705206551, "grad_norm": 3.9803450107574463, "learning_rate": 1.630714037292865e-05, "loss": 1.0494, "step": 1245 }, { "epoch": 0.3045710095331215, "grad_norm": 3.74851393699646, "learning_rate": 1.6300994434888805e-05, "loss": 1.072, "step": 1246 }, { "epoch": 0.3048154485455879, "grad_norm": 3.272641658782959, "learning_rate": 1.6294844547136394e-05, "loss": 1.0109, "step": 1247 }, { "epoch": 0.30505988755805424, "grad_norm": 3.4987287521362305, "learning_rate": 1.628869071352641e-05, "loss": 0.9238, "step": 1248 }, { "epoch": 0.30530432657052065, "grad_norm": 3.6942923069000244, "learning_rate": 1.628253293791632e-05, "loss": 0.9616, "step": 1249 }, { "epoch": 0.30554876558298705, "grad_norm": 4.471048355102539, "learning_rate": 1.627637122416606e-05, "loss": 0.9523, "step": 1250 }, { "epoch": 0.30579320459545345, "grad_norm": 3.160109758377075, "learning_rate": 1.627020557613803e-05, "loss": 0.9743, "step": 1251 }, { "epoch": 0.30603764360791985, "grad_norm": 3.819319248199463, "learning_rate": 1.6264035997697118e-05, "loss": 1.0451, "step": 1252 }, { "epoch": 0.3062820826203862, "grad_norm": 4.501501083374023, "learning_rate": 1.625786249271064e-05, "loss": 1.0177, "step": 1253 }, { "epoch": 0.3065265216328526, "grad_norm": 3.6937527656555176, "learning_rate": 1.6251685065048408e-05, "loss": 0.9424, "step": 1254 }, { "epoch": 0.306770960645319, "grad_norm": 3.759077548980713, "learning_rate": 1.624550371858267e-05, "loss": 1.1124, "step": 1255 }, { "epoch": 0.3070153996577854, "grad_norm": 2.905395984649658, "learning_rate": 1.6239318457188146e-05, "loss": 2.2317, "step": 1256 }, { "epoch": 0.30725983867025175, "grad_norm": 3.678001880645752, "learning_rate": 1.6233129284741992e-05, "loss": 1.1202, "step": 1257 }, { "epoch": 0.30750427768271815, "grad_norm": 4.4421186447143555, "learning_rate": 1.6226936205123835e-05, "loss": 1.0008, "step": 1258 }, { "epoch": 0.30774871669518455, "grad_norm": 3.2402305603027344, "learning_rate": 1.6220739222215738e-05, "loss": 0.9972, "step": 1259 }, { "epoch": 0.30799315570765096, "grad_norm": 3.374966859817505, "learning_rate": 1.621453833990222e-05, "loss": 0.9716, "step": 1260 }, { "epoch": 0.30823759472011736, "grad_norm": 3.3255765438079834, "learning_rate": 1.6208333562070232e-05, "loss": 1.0439, "step": 1261 }, { "epoch": 0.3084820337325837, "grad_norm": 3.0071613788604736, "learning_rate": 1.6202124892609175e-05, "loss": 0.8857, "step": 1262 }, { "epoch": 0.3087264727450501, "grad_norm": 3.1655237674713135, "learning_rate": 1.6195912335410896e-05, "loss": 0.9687, "step": 1263 }, { "epoch": 0.3089709117575165, "grad_norm": 3.5941660404205322, "learning_rate": 1.618969589436967e-05, "loss": 1.0306, "step": 1264 }, { "epoch": 0.3092153507699829, "grad_norm": 3.753988742828369, "learning_rate": 1.6183475573382196e-05, "loss": 0.9762, "step": 1265 }, { "epoch": 0.30945978978244926, "grad_norm": 3.202784538269043, "learning_rate": 1.6177251376347632e-05, "loss": 1.0616, "step": 1266 }, { "epoch": 0.30970422879491566, "grad_norm": 2.9768574237823486, "learning_rate": 1.6171023307167545e-05, "loss": 1.047, "step": 1267 }, { "epoch": 0.30994866780738206, "grad_norm": 3.3893463611602783, "learning_rate": 1.6164791369745935e-05, "loss": 0.9366, "step": 1268 }, { "epoch": 0.31019310681984846, "grad_norm": 3.303011178970337, "learning_rate": 1.615855556798923e-05, "loss": 0.9906, "step": 1269 }, { "epoch": 0.31043754583231487, "grad_norm": 2.891843557357788, "learning_rate": 1.615231590580627e-05, "loss": 0.8647, "step": 1270 }, { "epoch": 0.3106819848447812, "grad_norm": 3.257838487625122, "learning_rate": 1.614607238710833e-05, "loss": 0.8725, "step": 1271 }, { "epoch": 0.3109264238572476, "grad_norm": 4.155395984649658, "learning_rate": 1.613982501580909e-05, "loss": 0.9151, "step": 1272 }, { "epoch": 0.311170862869714, "grad_norm": 3.5042197704315186, "learning_rate": 1.613357379582465e-05, "loss": 0.9406, "step": 1273 }, { "epoch": 0.3114153018821804, "grad_norm": 4.06215763092041, "learning_rate": 1.612731873107352e-05, "loss": 1.0614, "step": 1274 }, { "epoch": 0.31165974089464676, "grad_norm": 3.6693708896636963, "learning_rate": 1.612105982547663e-05, "loss": 0.9339, "step": 1275 }, { "epoch": 0.31190417990711317, "grad_norm": 3.698263645172119, "learning_rate": 1.6114797082957304e-05, "loss": 0.9648, "step": 1276 }, { "epoch": 0.31214861891957957, "grad_norm": 3.2180988788604736, "learning_rate": 1.6108530507441273e-05, "loss": 2.0458, "step": 1277 }, { "epoch": 0.31239305793204597, "grad_norm": 4.171623229980469, "learning_rate": 1.610226010285668e-05, "loss": 1.0276, "step": 1278 }, { "epoch": 0.3126374969445123, "grad_norm": 3.3476104736328125, "learning_rate": 1.6095985873134063e-05, "loss": 1.054, "step": 1279 }, { "epoch": 0.3128819359569787, "grad_norm": 3.6903676986694336, "learning_rate": 1.6089707822206353e-05, "loss": 0.8384, "step": 1280 }, { "epoch": 0.3131263749694451, "grad_norm": 3.4239535331726074, "learning_rate": 1.6083425954008883e-05, "loss": 1.0724, "step": 1281 }, { "epoch": 0.3133708139819115, "grad_norm": 3.497863292694092, "learning_rate": 1.6077140272479378e-05, "loss": 1.0717, "step": 1282 }, { "epoch": 0.3136152529943779, "grad_norm": 3.6051435470581055, "learning_rate": 1.607085078155795e-05, "loss": 0.9298, "step": 1283 }, { "epoch": 0.31385969200684427, "grad_norm": 3.1399309635162354, "learning_rate": 1.6064557485187104e-05, "loss": 0.9527, "step": 1284 }, { "epoch": 0.3141041310193107, "grad_norm": 3.3739049434661865, "learning_rate": 1.605826038731172e-05, "loss": 0.9738, "step": 1285 }, { "epoch": 0.3143485700317771, "grad_norm": 3.3782050609588623, "learning_rate": 1.6051959491879072e-05, "loss": 0.9883, "step": 1286 }, { "epoch": 0.3145930090442435, "grad_norm": 3.374809503555298, "learning_rate": 1.6045654802838812e-05, "loss": 0.9174, "step": 1287 }, { "epoch": 0.3148374480567098, "grad_norm": 3.3350439071655273, "learning_rate": 1.603934632414297e-05, "loss": 0.9236, "step": 1288 }, { "epoch": 0.3150818870691762, "grad_norm": 3.8035969734191895, "learning_rate": 1.6033034059745943e-05, "loss": 0.8416, "step": 1289 }, { "epoch": 0.3153263260816426, "grad_norm": 3.8607382774353027, "learning_rate": 1.6026718013604514e-05, "loss": 1.0511, "step": 1290 }, { "epoch": 0.31557076509410903, "grad_norm": 3.493785858154297, "learning_rate": 1.602039818967783e-05, "loss": 1.0373, "step": 1291 }, { "epoch": 0.31581520410657543, "grad_norm": 3.8863096237182617, "learning_rate": 1.6014074591927405e-05, "loss": 0.9358, "step": 1292 }, { "epoch": 0.3160596431190418, "grad_norm": 3.0080204010009766, "learning_rate": 1.6007747224317124e-05, "loss": 0.9329, "step": 1293 }, { "epoch": 0.3163040821315082, "grad_norm": 4.005444526672363, "learning_rate": 1.6001416090813224e-05, "loss": 1.0229, "step": 1294 }, { "epoch": 0.3165485211439746, "grad_norm": 3.2476863861083984, "learning_rate": 1.5995081195384314e-05, "loss": 0.9798, "step": 1295 }, { "epoch": 0.316792960156441, "grad_norm": 3.350329637527466, "learning_rate": 1.598874254200136e-05, "loss": 1.0878, "step": 1296 }, { "epoch": 0.31703739916890733, "grad_norm": 3.887662649154663, "learning_rate": 1.5982400134637674e-05, "loss": 1.1045, "step": 1297 }, { "epoch": 0.31728183818137373, "grad_norm": 3.1978490352630615, "learning_rate": 1.5976053977268934e-05, "loss": 0.9608, "step": 1298 }, { "epoch": 0.31752627719384013, "grad_norm": 3.234931707382202, "learning_rate": 1.5969704073873157e-05, "loss": 1.1088, "step": 1299 }, { "epoch": 0.31777071620630654, "grad_norm": 3.477632761001587, "learning_rate": 1.5963350428430722e-05, "loss": 0.9836, "step": 1300 }, { "epoch": 0.31801515521877294, "grad_norm": 3.838231325149536, "learning_rate": 1.5956993044924334e-05, "loss": 1.0023, "step": 1301 }, { "epoch": 0.3182595942312393, "grad_norm": 3.980022430419922, "learning_rate": 1.595063192733906e-05, "loss": 1.0221, "step": 1302 }, { "epoch": 0.3185040332437057, "grad_norm": 3.580064296722412, "learning_rate": 1.5944267079662302e-05, "loss": 1.0318, "step": 1303 }, { "epoch": 0.3187484722561721, "grad_norm": 3.1863508224487305, "learning_rate": 1.5937898505883794e-05, "loss": 1.0015, "step": 1304 }, { "epoch": 0.3189929112686385, "grad_norm": 3.451911211013794, "learning_rate": 1.5931526209995608e-05, "loss": 0.9728, "step": 1305 }, { "epoch": 0.31923735028110484, "grad_norm": 2.908261299133301, "learning_rate": 1.5925150195992156e-05, "loss": 2.0991, "step": 1306 }, { "epoch": 0.31948178929357124, "grad_norm": 3.784224033355713, "learning_rate": 1.5918770467870174e-05, "loss": 1.0714, "step": 1307 }, { "epoch": 0.31972622830603764, "grad_norm": 3.7832584381103516, "learning_rate": 1.5912387029628726e-05, "loss": 1.0196, "step": 1308 }, { "epoch": 0.31997066731850404, "grad_norm": 3.057114601135254, "learning_rate": 1.5905999885269208e-05, "loss": 1.0451, "step": 1309 }, { "epoch": 0.32021510633097044, "grad_norm": 3.717481851577759, "learning_rate": 1.5899609038795333e-05, "loss": 0.8955, "step": 1310 }, { "epoch": 0.3204595453434368, "grad_norm": 3.5934057235717773, "learning_rate": 1.589321449421313e-05, "loss": 0.9301, "step": 1311 }, { "epoch": 0.3207039843559032, "grad_norm": 3.7299716472625732, "learning_rate": 1.5886816255530964e-05, "loss": 0.9594, "step": 1312 }, { "epoch": 0.3209484233683696, "grad_norm": 3.0106871128082275, "learning_rate": 1.5880414326759493e-05, "loss": 1.0206, "step": 1313 }, { "epoch": 0.321192862380836, "grad_norm": 3.8183209896087646, "learning_rate": 1.5874008711911706e-05, "loss": 0.9222, "step": 1314 }, { "epoch": 0.32143730139330234, "grad_norm": 3.292402744293213, "learning_rate": 1.5867599415002896e-05, "loss": 0.9671, "step": 1315 }, { "epoch": 0.32168174040576875, "grad_norm": 3.2833142280578613, "learning_rate": 1.5861186440050663e-05, "loss": 1.0283, "step": 1316 }, { "epoch": 0.32192617941823515, "grad_norm": 3.5037293434143066, "learning_rate": 1.5854769791074906e-05, "loss": 1.0508, "step": 1317 }, { "epoch": 0.32217061843070155, "grad_norm": 3.340787172317505, "learning_rate": 1.5848349472097848e-05, "loss": 0.9573, "step": 1318 }, { "epoch": 0.32241505744316795, "grad_norm": 8.840167999267578, "learning_rate": 1.5841925487143986e-05, "loss": 0.9666, "step": 1319 }, { "epoch": 0.3226594964556343, "grad_norm": 3.3367087841033936, "learning_rate": 1.583549784024014e-05, "loss": 0.9763, "step": 1320 }, { "epoch": 0.3229039354681007, "grad_norm": 3.5148043632507324, "learning_rate": 1.5829066535415402e-05, "loss": 0.9383, "step": 1321 }, { "epoch": 0.3231483744805671, "grad_norm": 3.5361037254333496, "learning_rate": 1.5822631576701174e-05, "loss": 0.977, "step": 1322 }, { "epoch": 0.3233928134930335, "grad_norm": 3.2863235473632812, "learning_rate": 1.5816192968131138e-05, "loss": 0.9872, "step": 1323 }, { "epoch": 0.32363725250549985, "grad_norm": 3.653019428253174, "learning_rate": 1.5809750713741277e-05, "loss": 0.8894, "step": 1324 }, { "epoch": 0.32388169151796625, "grad_norm": 2.8651626110076904, "learning_rate": 1.5803304817569847e-05, "loss": 0.9137, "step": 1325 }, { "epoch": 0.32412613053043265, "grad_norm": 2.9790525436401367, "learning_rate": 1.579685528365739e-05, "loss": 0.9451, "step": 1326 }, { "epoch": 0.32437056954289906, "grad_norm": 3.41646146774292, "learning_rate": 1.5790402116046725e-05, "loss": 1.0006, "step": 1327 }, { "epoch": 0.32461500855536546, "grad_norm": 3.762275218963623, "learning_rate": 1.578394531878296e-05, "loss": 0.9529, "step": 1328 }, { "epoch": 0.3248594475678318, "grad_norm": 2.90238618850708, "learning_rate": 1.5777484895913462e-05, "loss": 0.9205, "step": 1329 }, { "epoch": 0.3251038865802982, "grad_norm": 3.016227960586548, "learning_rate": 1.5771020851487893e-05, "loss": 0.9908, "step": 1330 }, { "epoch": 0.3253483255927646, "grad_norm": 3.5532524585723877, "learning_rate": 1.576455318955816e-05, "loss": 0.9425, "step": 1331 }, { "epoch": 0.325592764605231, "grad_norm": 3.1646616458892822, "learning_rate": 1.5758081914178457e-05, "loss": 0.8908, "step": 1332 }, { "epoch": 0.32583720361769736, "grad_norm": 3.4908621311187744, "learning_rate": 1.5751607029405226e-05, "loss": 0.985, "step": 1333 }, { "epoch": 0.32608164263016376, "grad_norm": 3.6697299480438232, "learning_rate": 1.574512853929719e-05, "loss": 1.0036, "step": 1334 }, { "epoch": 0.32632608164263016, "grad_norm": 4.173372268676758, "learning_rate": 1.573864644791532e-05, "loss": 0.9331, "step": 1335 }, { "epoch": 0.32657052065509656, "grad_norm": 2.790220260620117, "learning_rate": 1.5732160759322847e-05, "loss": 1.9576, "step": 1336 }, { "epoch": 0.32681495966756297, "grad_norm": 4.081017971038818, "learning_rate": 1.5725671477585255e-05, "loss": 1.003, "step": 1337 }, { "epoch": 0.3270593986800293, "grad_norm": 4.018110275268555, "learning_rate": 1.5719178606770283e-05, "loss": 1.0492, "step": 1338 }, { "epoch": 0.3273038376924957, "grad_norm": 3.4949355125427246, "learning_rate": 1.5712682150947926e-05, "loss": 0.9217, "step": 1339 }, { "epoch": 0.3275482767049621, "grad_norm": 3.6490862369537354, "learning_rate": 1.5706182114190412e-05, "loss": 0.8558, "step": 1340 }, { "epoch": 0.3277927157174285, "grad_norm": 3.302030563354492, "learning_rate": 1.569967850057222e-05, "loss": 0.9701, "step": 1341 }, { "epoch": 0.32803715472989486, "grad_norm": 3.829453706741333, "learning_rate": 1.5693171314170078e-05, "loss": 0.9252, "step": 1342 }, { "epoch": 0.32828159374236127, "grad_norm": 2.87931227684021, "learning_rate": 1.568666055906295e-05, "loss": 0.9877, "step": 1343 }, { "epoch": 0.32852603275482767, "grad_norm": 2.9058079719543457, "learning_rate": 1.5680146239332024e-05, "loss": 2.1484, "step": 1344 }, { "epoch": 0.32877047176729407, "grad_norm": 3.9817657470703125, "learning_rate": 1.567362835906074e-05, "loss": 0.9291, "step": 1345 }, { "epoch": 0.32901491077976047, "grad_norm": 3.298549175262451, "learning_rate": 1.5667106922334763e-05, "loss": 0.986, "step": 1346 }, { "epoch": 0.3292593497922268, "grad_norm": 3.3551080226898193, "learning_rate": 1.5660581933241994e-05, "loss": 0.8962, "step": 1347 }, { "epoch": 0.3295037888046932, "grad_norm": 2.3159947395324707, "learning_rate": 1.565405339587254e-05, "loss": 1.9325, "step": 1348 }, { "epoch": 0.3297482278171596, "grad_norm": 3.5716497898101807, "learning_rate": 1.564752131431876e-05, "loss": 0.9268, "step": 1349 }, { "epoch": 0.329992666829626, "grad_norm": 3.3860905170440674, "learning_rate": 1.5640985692675213e-05, "loss": 0.9784, "step": 1350 }, { "epoch": 0.33023710584209237, "grad_norm": 3.5350513458251953, "learning_rate": 1.5634446535038688e-05, "loss": 1.0282, "step": 1351 }, { "epoch": 0.3304815448545588, "grad_norm": 3.6100711822509766, "learning_rate": 1.5627903845508184e-05, "loss": 0.9305, "step": 1352 }, { "epoch": 0.3307259838670252, "grad_norm": 3.686244487762451, "learning_rate": 1.5621357628184924e-05, "loss": 0.8673, "step": 1353 }, { "epoch": 0.3309704228794916, "grad_norm": 2.6438686847686768, "learning_rate": 1.5614807887172332e-05, "loss": 2.0463, "step": 1354 }, { "epoch": 0.331214861891958, "grad_norm": 3.5230236053466797, "learning_rate": 1.5608254626576046e-05, "loss": 0.8949, "step": 1355 }, { "epoch": 0.3314593009044243, "grad_norm": 3.89335036277771, "learning_rate": 1.5601697850503907e-05, "loss": 0.8568, "step": 1356 }, { "epoch": 0.3317037399168907, "grad_norm": 3.775359630584717, "learning_rate": 1.5595137563065965e-05, "loss": 1.0217, "step": 1357 }, { "epoch": 0.33194817892935713, "grad_norm": 4.173966407775879, "learning_rate": 1.5588573768374464e-05, "loss": 0.9904, "step": 1358 }, { "epoch": 0.33219261794182353, "grad_norm": 4.213604927062988, "learning_rate": 1.5582006470543855e-05, "loss": 1.0392, "step": 1359 }, { "epoch": 0.3324370569542899, "grad_norm": 3.8599865436553955, "learning_rate": 1.5575435673690774e-05, "loss": 1.0709, "step": 1360 }, { "epoch": 0.3326814959667563, "grad_norm": 3.6756699085235596, "learning_rate": 1.556886138193406e-05, "loss": 0.8283, "step": 1361 }, { "epoch": 0.3329259349792227, "grad_norm": 3.2679405212402344, "learning_rate": 1.556228359939474e-05, "loss": 1.0301, "step": 1362 }, { "epoch": 0.3331703739916891, "grad_norm": 3.217700719833374, "learning_rate": 1.5555702330196024e-05, "loss": 0.9494, "step": 1363 }, { "epoch": 0.3334148130041555, "grad_norm": 3.4018821716308594, "learning_rate": 1.5549117578463316e-05, "loss": 0.8486, "step": 1364 }, { "epoch": 0.33365925201662183, "grad_norm": 3.5787065029144287, "learning_rate": 1.55425293483242e-05, "loss": 1.0495, "step": 1365 }, { "epoch": 0.33390369102908823, "grad_norm": 3.589322566986084, "learning_rate": 1.5535937643908435e-05, "loss": 0.948, "step": 1366 }, { "epoch": 0.33414813004155464, "grad_norm": 3.325261116027832, "learning_rate": 1.5529342469347958e-05, "loss": 0.9936, "step": 1367 }, { "epoch": 0.33439256905402104, "grad_norm": 3.6978461742401123, "learning_rate": 1.5522743828776898e-05, "loss": 0.9901, "step": 1368 }, { "epoch": 0.3346370080664874, "grad_norm": 3.383035182952881, "learning_rate": 1.551614172633154e-05, "loss": 1.0426, "step": 1369 }, { "epoch": 0.3348814470789538, "grad_norm": 3.634697198867798, "learning_rate": 1.5509536166150334e-05, "loss": 0.9485, "step": 1370 }, { "epoch": 0.3351258860914202, "grad_norm": 3.330472230911255, "learning_rate": 1.5502927152373913e-05, "loss": 0.9106, "step": 1371 }, { "epoch": 0.3353703251038866, "grad_norm": 2.659522771835327, "learning_rate": 1.5496314689145072e-05, "loss": 0.8889, "step": 1372 }, { "epoch": 0.335614764116353, "grad_norm": 3.5730326175689697, "learning_rate": 1.548969878060876e-05, "loss": 0.9087, "step": 1373 }, { "epoch": 0.33585920312881934, "grad_norm": 3.6332874298095703, "learning_rate": 1.5483079430912088e-05, "loss": 0.9751, "step": 1374 }, { "epoch": 0.33610364214128574, "grad_norm": 3.5821280479431152, "learning_rate": 1.5476456644204328e-05, "loss": 0.9511, "step": 1375 }, { "epoch": 0.33634808115375214, "grad_norm": 3.1708550453186035, "learning_rate": 1.5469830424636906e-05, "loss": 0.9146, "step": 1376 }, { "epoch": 0.33659252016621855, "grad_norm": 3.0782763957977295, "learning_rate": 1.54632007763634e-05, "loss": 2.0277, "step": 1377 }, { "epoch": 0.3368369591786849, "grad_norm": 3.970358371734619, "learning_rate": 1.5456567703539532e-05, "loss": 1.0644, "step": 1378 }, { "epoch": 0.3370813981911513, "grad_norm": 4.540128231048584, "learning_rate": 1.544993121032318e-05, "loss": 0.8518, "step": 1379 }, { "epoch": 0.3373258372036177, "grad_norm": 3.16619873046875, "learning_rate": 1.5443291300874353e-05, "loss": 0.9677, "step": 1380 }, { "epoch": 0.3375702762160841, "grad_norm": 4.074929237365723, "learning_rate": 1.5436647979355214e-05, "loss": 0.9551, "step": 1381 }, { "epoch": 0.3378147152285505, "grad_norm": 3.1865148544311523, "learning_rate": 1.5430001249930057e-05, "loss": 0.9697, "step": 1382 }, { "epoch": 0.33805915424101685, "grad_norm": 3.514235019683838, "learning_rate": 1.5423351116765316e-05, "loss": 1.0022, "step": 1383 }, { "epoch": 0.33830359325348325, "grad_norm": 3.3248844146728516, "learning_rate": 1.541669758402956e-05, "loss": 1.0536, "step": 1384 }, { "epoch": 0.33854803226594965, "grad_norm": 3.2770636081695557, "learning_rate": 1.5410040655893484e-05, "loss": 1.0274, "step": 1385 }, { "epoch": 0.33879247127841605, "grad_norm": 3.009883403778076, "learning_rate": 1.5403380336529914e-05, "loss": 0.8489, "step": 1386 }, { "epoch": 0.3390369102908824, "grad_norm": 3.4994874000549316, "learning_rate": 1.53967166301138e-05, "loss": 1.0033, "step": 1387 }, { "epoch": 0.3392813493033488, "grad_norm": 3.02510666847229, "learning_rate": 1.5390049540822223e-05, "loss": 0.9407, "step": 1388 }, { "epoch": 0.3395257883158152, "grad_norm": 3.872352361679077, "learning_rate": 1.5383379072834373e-05, "loss": 0.9573, "step": 1389 }, { "epoch": 0.3397702273282816, "grad_norm": 3.6080379486083984, "learning_rate": 1.537670523033156e-05, "loss": 0.9771, "step": 1390 }, { "epoch": 0.340014666340748, "grad_norm": 3.544776678085327, "learning_rate": 1.5370028017497217e-05, "loss": 1.0244, "step": 1391 }, { "epoch": 0.34025910535321435, "grad_norm": 3.672668933868408, "learning_rate": 1.536334743851689e-05, "loss": 0.8461, "step": 1392 }, { "epoch": 0.34050354436568075, "grad_norm": 3.6929850578308105, "learning_rate": 1.5356663497578223e-05, "loss": 0.9417, "step": 1393 }, { "epoch": 0.34074798337814716, "grad_norm": 3.087480306625366, "learning_rate": 1.5349976198870974e-05, "loss": 0.9396, "step": 1394 }, { "epoch": 0.34099242239061356, "grad_norm": 3.1820108890533447, "learning_rate": 1.534328554658701e-05, "loss": 0.9691, "step": 1395 }, { "epoch": 0.3412368614030799, "grad_norm": 4.0018630027771, "learning_rate": 1.5336591544920297e-05, "loss": 0.9179, "step": 1396 }, { "epoch": 0.3414813004155463, "grad_norm": 3.438683271408081, "learning_rate": 1.53298941980669e-05, "loss": 1.0616, "step": 1397 }, { "epoch": 0.3417257394280127, "grad_norm": 3.162782669067383, "learning_rate": 1.5323193510224973e-05, "loss": 0.9266, "step": 1398 }, { "epoch": 0.3419701784404791, "grad_norm": 3.303999185562134, "learning_rate": 1.5316489485594785e-05, "loss": 0.9208, "step": 1399 }, { "epoch": 0.3422146174529455, "grad_norm": 3.0878005027770996, "learning_rate": 1.5309782128378676e-05, "loss": 0.9152, "step": 1400 }, { "epoch": 0.34245905646541186, "grad_norm": 3.1952340602874756, "learning_rate": 1.5303071442781083e-05, "loss": 0.9222, "step": 1401 }, { "epoch": 0.34270349547787826, "grad_norm": 3.550318479537964, "learning_rate": 1.5296357433008532e-05, "loss": 0.8696, "step": 1402 }, { "epoch": 0.34294793449034466, "grad_norm": 3.1809489727020264, "learning_rate": 1.5289640103269626e-05, "loss": 1.0164, "step": 1403 }, { "epoch": 0.34319237350281107, "grad_norm": 3.1642727851867676, "learning_rate": 1.5282919457775058e-05, "loss": 0.8859, "step": 1404 }, { "epoch": 0.3434368125152774, "grad_norm": 3.2544121742248535, "learning_rate": 1.5276195500737592e-05, "loss": 0.8058, "step": 1405 }, { "epoch": 0.3436812515277438, "grad_norm": 2.817103862762451, "learning_rate": 1.5269468236372068e-05, "loss": 2.0819, "step": 1406 }, { "epoch": 0.3439256905402102, "grad_norm": 3.803536891937256, "learning_rate": 1.5262737668895403e-05, "loss": 0.9362, "step": 1407 }, { "epoch": 0.3441701295526766, "grad_norm": 3.5028889179229736, "learning_rate": 1.5256003802526586e-05, "loss": 0.8609, "step": 1408 }, { "epoch": 0.344414568565143, "grad_norm": 3.3524909019470215, "learning_rate": 1.5249266641486666e-05, "loss": 0.9216, "step": 1409 }, { "epoch": 0.34465900757760937, "grad_norm": 3.9344277381896973, "learning_rate": 1.5242526189998768e-05, "loss": 0.9408, "step": 1410 }, { "epoch": 0.34490344659007577, "grad_norm": 4.254050254821777, "learning_rate": 1.5235782452288068e-05, "loss": 1.0183, "step": 1411 }, { "epoch": 0.34514788560254217, "grad_norm": 2.2866740226745605, "learning_rate": 1.5229035432581812e-05, "loss": 1.9729, "step": 1412 }, { "epoch": 0.3453923246150086, "grad_norm": 3.434361696243286, "learning_rate": 1.5222285135109291e-05, "loss": 0.9577, "step": 1413 }, { "epoch": 0.3456367636274749, "grad_norm": 3.364114284515381, "learning_rate": 1.5215531564101866e-05, "loss": 0.9309, "step": 1414 }, { "epoch": 0.3458812026399413, "grad_norm": 3.732811212539673, "learning_rate": 1.5208774723792943e-05, "loss": 0.8709, "step": 1415 }, { "epoch": 0.3461256416524077, "grad_norm": 3.326068878173828, "learning_rate": 1.5202014618417973e-05, "loss": 0.9526, "step": 1416 }, { "epoch": 0.3463700806648741, "grad_norm": 3.0876729488372803, "learning_rate": 1.5195251252214461e-05, "loss": 0.9559, "step": 1417 }, { "epoch": 0.3466145196773405, "grad_norm": 3.168135404586792, "learning_rate": 1.5188484629421946e-05, "loss": 0.9279, "step": 1418 }, { "epoch": 0.3468589586898069, "grad_norm": 3.341728925704956, "learning_rate": 1.518171475428202e-05, "loss": 1.0006, "step": 1419 }, { "epoch": 0.3471033977022733, "grad_norm": 3.315312385559082, "learning_rate": 1.5174941631038307e-05, "loss": 1.0684, "step": 1420 }, { "epoch": 0.3473478367147397, "grad_norm": 4.451924800872803, "learning_rate": 1.5168165263936472e-05, "loss": 0.873, "step": 1421 }, { "epoch": 0.3475922757272061, "grad_norm": 3.256350517272949, "learning_rate": 1.5161385657224203e-05, "loss": 1.0304, "step": 1422 }, { "epoch": 0.3478367147396724, "grad_norm": 3.359121084213257, "learning_rate": 1.5154602815151233e-05, "loss": 1.0106, "step": 1423 }, { "epoch": 0.3480811537521388, "grad_norm": 3.1137757301330566, "learning_rate": 1.514781674196931e-05, "loss": 1.0017, "step": 1424 }, { "epoch": 0.34832559276460523, "grad_norm": 3.178013324737549, "learning_rate": 1.5141027441932217e-05, "loss": 0.8749, "step": 1425 }, { "epoch": 0.34857003177707163, "grad_norm": 4.293848037719727, "learning_rate": 1.5134234919295757e-05, "loss": 0.9248, "step": 1426 }, { "epoch": 0.34881447078953803, "grad_norm": 3.428175449371338, "learning_rate": 1.5127439178317747e-05, "loss": 1.0673, "step": 1427 }, { "epoch": 0.3490589098020044, "grad_norm": 3.1072657108306885, "learning_rate": 1.5120640223258027e-05, "loss": 0.9458, "step": 1428 }, { "epoch": 0.3493033488144708, "grad_norm": 3.426907777786255, "learning_rate": 1.5113838058378454e-05, "loss": 0.9146, "step": 1429 }, { "epoch": 0.3495477878269372, "grad_norm": 3.247302293777466, "learning_rate": 1.5107032687942891e-05, "loss": 0.8704, "step": 1430 }, { "epoch": 0.3497922268394036, "grad_norm": 3.252124786376953, "learning_rate": 1.5100224116217217e-05, "loss": 0.9551, "step": 1431 }, { "epoch": 0.35003666585186993, "grad_norm": 3.9852957725524902, "learning_rate": 1.5093412347469312e-05, "loss": 0.9488, "step": 1432 }, { "epoch": 0.35028110486433633, "grad_norm": 3.5851423740386963, "learning_rate": 1.5086597385969062e-05, "loss": 1.0187, "step": 1433 }, { "epoch": 0.35052554387680274, "grad_norm": 3.946223735809326, "learning_rate": 1.5079779235988355e-05, "loss": 0.973, "step": 1434 }, { "epoch": 0.35076998288926914, "grad_norm": 3.861586093902588, "learning_rate": 1.5072957901801075e-05, "loss": 0.911, "step": 1435 }, { "epoch": 0.35101442190173554, "grad_norm": 3.4354403018951416, "learning_rate": 1.5066133387683106e-05, "loss": 0.8669, "step": 1436 }, { "epoch": 0.3512588609142019, "grad_norm": 3.3271026611328125, "learning_rate": 1.5059305697912323e-05, "loss": 0.8781, "step": 1437 }, { "epoch": 0.3515032999266683, "grad_norm": 3.2607266902923584, "learning_rate": 1.5052474836768594e-05, "loss": 0.9524, "step": 1438 }, { "epoch": 0.3517477389391347, "grad_norm": 3.6708407402038574, "learning_rate": 1.5045640808533772e-05, "loss": 1.0256, "step": 1439 }, { "epoch": 0.3519921779516011, "grad_norm": 3.570585250854492, "learning_rate": 1.5038803617491695e-05, "loss": 0.9217, "step": 1440 }, { "epoch": 0.35223661696406744, "grad_norm": 3.3812408447265625, "learning_rate": 1.5031963267928185e-05, "loss": 0.8215, "step": 1441 }, { "epoch": 0.35248105597653384, "grad_norm": 3.565835952758789, "learning_rate": 1.5025119764131044e-05, "loss": 0.9133, "step": 1442 }, { "epoch": 0.35272549498900024, "grad_norm": 4.318790912628174, "learning_rate": 1.501827311039005e-05, "loss": 0.9073, "step": 1443 }, { "epoch": 0.35296993400146665, "grad_norm": 3.476274251937866, "learning_rate": 1.501142331099696e-05, "loss": 0.9068, "step": 1444 }, { "epoch": 0.35321437301393305, "grad_norm": 3.3230934143066406, "learning_rate": 1.5004570370245492e-05, "loss": 0.988, "step": 1445 }, { "epoch": 0.3534588120263994, "grad_norm": 3.812248706817627, "learning_rate": 1.4997714292431347e-05, "loss": 1.0304, "step": 1446 }, { "epoch": 0.3537032510388658, "grad_norm": 3.7100956439971924, "learning_rate": 1.4990855081852185e-05, "loss": 0.9584, "step": 1447 }, { "epoch": 0.3539476900513322, "grad_norm": 2.9337382316589355, "learning_rate": 1.4983992742807627e-05, "loss": 0.8695, "step": 1448 }, { "epoch": 0.3541921290637986, "grad_norm": 3.947782039642334, "learning_rate": 1.4977127279599264e-05, "loss": 1.0827, "step": 1449 }, { "epoch": 0.35443656807626495, "grad_norm": 3.16684889793396, "learning_rate": 1.4970258696530635e-05, "loss": 0.8774, "step": 1450 }, { "epoch": 0.35468100708873135, "grad_norm": 4.034370422363281, "learning_rate": 1.4963386997907242e-05, "loss": 0.949, "step": 1451 }, { "epoch": 0.35492544610119775, "grad_norm": 3.9608681201934814, "learning_rate": 1.4956512188036536e-05, "loss": 0.9403, "step": 1452 }, { "epoch": 0.35516988511366415, "grad_norm": 3.7968034744262695, "learning_rate": 1.4949634271227918e-05, "loss": 0.9495, "step": 1453 }, { "epoch": 0.35541432412613055, "grad_norm": 3.5002903938293457, "learning_rate": 1.4942753251792746e-05, "loss": 0.9865, "step": 1454 }, { "epoch": 0.3556587631385969, "grad_norm": 3.36322021484375, "learning_rate": 1.4935869134044308e-05, "loss": 0.8611, "step": 1455 }, { "epoch": 0.3559032021510633, "grad_norm": 4.71673583984375, "learning_rate": 1.4928981922297842e-05, "loss": 0.8905, "step": 1456 }, { "epoch": 0.3561476411635297, "grad_norm": 3.0979466438293457, "learning_rate": 1.4922091620870527e-05, "loss": 0.9622, "step": 1457 }, { "epoch": 0.3563920801759961, "grad_norm": 3.6254913806915283, "learning_rate": 1.4915198234081479e-05, "loss": 1.0715, "step": 1458 }, { "epoch": 0.35663651918846245, "grad_norm": 3.1062965393066406, "learning_rate": 1.4908301766251739e-05, "loss": 0.8824, "step": 1459 }, { "epoch": 0.35688095820092886, "grad_norm": 3.2139036655426025, "learning_rate": 1.4901402221704287e-05, "loss": 0.9883, "step": 1460 }, { "epoch": 0.35712539721339526, "grad_norm": 3.6782898902893066, "learning_rate": 1.4894499604764035e-05, "loss": 0.9123, "step": 1461 }, { "epoch": 0.35736983622586166, "grad_norm": 2.6064562797546387, "learning_rate": 1.4887593919757814e-05, "loss": 2.121, "step": 1462 }, { "epoch": 0.35761427523832806, "grad_norm": 4.038392066955566, "learning_rate": 1.4880685171014382e-05, "loss": 0.9432, "step": 1463 }, { "epoch": 0.3578587142507944, "grad_norm": 4.281519412994385, "learning_rate": 1.4873773362864413e-05, "loss": 0.9428, "step": 1464 }, { "epoch": 0.3581031532632608, "grad_norm": 4.620762825012207, "learning_rate": 1.4866858499640504e-05, "loss": 0.9179, "step": 1465 }, { "epoch": 0.3583475922757272, "grad_norm": 3.9565887451171875, "learning_rate": 1.4859940585677165e-05, "loss": 0.9347, "step": 1466 }, { "epoch": 0.3585920312881936, "grad_norm": 3.9693334102630615, "learning_rate": 1.4853019625310813e-05, "loss": 0.9412, "step": 1467 }, { "epoch": 0.35883647030065996, "grad_norm": 3.7491869926452637, "learning_rate": 1.4846095622879784e-05, "loss": 0.921, "step": 1468 }, { "epoch": 0.35908090931312636, "grad_norm": 2.97538161277771, "learning_rate": 1.483916858272432e-05, "loss": 0.8116, "step": 1469 }, { "epoch": 0.35932534832559276, "grad_norm": 3.518476724624634, "learning_rate": 1.4832238509186559e-05, "loss": 0.957, "step": 1470 }, { "epoch": 0.35956978733805917, "grad_norm": 3.6826512813568115, "learning_rate": 1.4825305406610547e-05, "loss": 0.8955, "step": 1471 }, { "epoch": 0.35981422635052557, "grad_norm": 3.420494794845581, "learning_rate": 1.4818369279342227e-05, "loss": 0.957, "step": 1472 }, { "epoch": 0.3600586653629919, "grad_norm": 3.42960262298584, "learning_rate": 1.4811430131729436e-05, "loss": 0.918, "step": 1473 }, { "epoch": 0.3603031043754583, "grad_norm": 2.91977596282959, "learning_rate": 1.480448796812191e-05, "loss": 0.9817, "step": 1474 }, { "epoch": 0.3605475433879247, "grad_norm": 5.7857489585876465, "learning_rate": 1.4797542792871267e-05, "loss": 0.8361, "step": 1475 }, { "epoch": 0.3607919824003911, "grad_norm": 4.280693531036377, "learning_rate": 1.4790594610331019e-05, "loss": 0.8621, "step": 1476 }, { "epoch": 0.36103642141285747, "grad_norm": 3.5429012775421143, "learning_rate": 1.4783643424856564e-05, "loss": 0.9482, "step": 1477 }, { "epoch": 0.36128086042532387, "grad_norm": 2.4663522243499756, "learning_rate": 1.477668924080518e-05, "loss": 2.0677, "step": 1478 }, { "epoch": 0.36152529943779027, "grad_norm": 2.56671404838562, "learning_rate": 1.4769732062536022e-05, "loss": 2.0516, "step": 1479 }, { "epoch": 0.3617697384502567, "grad_norm": 5.004128932952881, "learning_rate": 1.4762771894410129e-05, "loss": 0.9523, "step": 1480 }, { "epoch": 0.3620141774627231, "grad_norm": 3.212191104888916, "learning_rate": 1.4755808740790403e-05, "loss": 0.8761, "step": 1481 }, { "epoch": 0.3622586164751894, "grad_norm": 4.195330619812012, "learning_rate": 1.4748842606041626e-05, "loss": 0.9086, "step": 1482 }, { "epoch": 0.3625030554876558, "grad_norm": 3.0373518466949463, "learning_rate": 1.4741873494530452e-05, "loss": 0.8122, "step": 1483 }, { "epoch": 0.3627474945001222, "grad_norm": 2.4052069187164307, "learning_rate": 1.4734901410625392e-05, "loss": 2.074, "step": 1484 }, { "epoch": 0.3629919335125886, "grad_norm": 3.4554340839385986, "learning_rate": 1.4727926358696825e-05, "loss": 1.0138, "step": 1485 }, { "epoch": 0.363236372525055, "grad_norm": 2.9953768253326416, "learning_rate": 1.4720948343116987e-05, "loss": 0.9359, "step": 1486 }, { "epoch": 0.3634808115375214, "grad_norm": 3.070591926574707, "learning_rate": 1.4713967368259981e-05, "loss": 1.0384, "step": 1487 }, { "epoch": 0.3637252505499878, "grad_norm": 3.7442331314086914, "learning_rate": 1.4706983438501745e-05, "loss": 0.9505, "step": 1488 }, { "epoch": 0.3639696895624542, "grad_norm": 3.5531771183013916, "learning_rate": 1.4699996558220095e-05, "loss": 1.0305, "step": 1489 }, { "epoch": 0.3642141285749206, "grad_norm": 3.0107712745666504, "learning_rate": 1.4693006731794682e-05, "loss": 1.0072, "step": 1490 }, { "epoch": 0.36445856758738693, "grad_norm": 3.122729539871216, "learning_rate": 1.4686013963607e-05, "loss": 0.9986, "step": 1491 }, { "epoch": 0.36470300659985333, "grad_norm": 2.9052510261535645, "learning_rate": 1.4679018258040399e-05, "loss": 0.8143, "step": 1492 }, { "epoch": 0.36494744561231973, "grad_norm": 4.365767478942871, "learning_rate": 1.4672019619480056e-05, "loss": 1.0574, "step": 1493 }, { "epoch": 0.36519188462478613, "grad_norm": 3.220187187194824, "learning_rate": 1.4665018052313004e-05, "loss": 0.8821, "step": 1494 }, { "epoch": 0.3654363236372525, "grad_norm": 3.0632996559143066, "learning_rate": 1.4658013560928098e-05, "loss": 0.9838, "step": 1495 }, { "epoch": 0.3656807626497189, "grad_norm": 3.8072338104248047, "learning_rate": 1.4651006149716027e-05, "loss": 1.0137, "step": 1496 }, { "epoch": 0.3659252016621853, "grad_norm": 3.2164018154144287, "learning_rate": 1.4643995823069318e-05, "loss": 1.0438, "step": 1497 }, { "epoch": 0.3661696406746517, "grad_norm": 3.714524269104004, "learning_rate": 1.4636982585382321e-05, "loss": 1.0036, "step": 1498 }, { "epoch": 0.3664140796871181, "grad_norm": 3.291309118270874, "learning_rate": 1.4629966441051208e-05, "loss": 0.977, "step": 1499 }, { "epoch": 0.36665851869958443, "grad_norm": 4.559157848358154, "learning_rate": 1.462294739447398e-05, "loss": 0.9867, "step": 1500 }, { "epoch": 0.36690295771205084, "grad_norm": 3.373988389968872, "learning_rate": 1.4615925450050448e-05, "loss": 0.9148, "step": 1501 }, { "epoch": 0.36714739672451724, "grad_norm": 3.6921863555908203, "learning_rate": 1.4608900612182246e-05, "loss": 0.8028, "step": 1502 }, { "epoch": 0.36739183573698364, "grad_norm": 3.494563579559326, "learning_rate": 1.4601872885272824e-05, "loss": 0.8157, "step": 1503 }, { "epoch": 0.36763627474945, "grad_norm": 3.2356550693511963, "learning_rate": 1.4594842273727434e-05, "loss": 0.9164, "step": 1504 }, { "epoch": 0.3678807137619164, "grad_norm": 3.729564666748047, "learning_rate": 1.4587808781953146e-05, "loss": 0.8828, "step": 1505 }, { "epoch": 0.3681251527743828, "grad_norm": 2.970707416534424, "learning_rate": 1.4580772414358829e-05, "loss": 0.8681, "step": 1506 }, { "epoch": 0.3683695917868492, "grad_norm": 3.6039421558380127, "learning_rate": 1.4573733175355152e-05, "loss": 0.9996, "step": 1507 }, { "epoch": 0.3686140307993156, "grad_norm": 4.286959171295166, "learning_rate": 1.4566691069354595e-05, "loss": 0.9506, "step": 1508 }, { "epoch": 0.36885846981178194, "grad_norm": 3.340758800506592, "learning_rate": 1.4559646100771428e-05, "loss": 0.8819, "step": 1509 }, { "epoch": 0.36910290882424834, "grad_norm": 4.261744022369385, "learning_rate": 1.4552598274021713e-05, "loss": 1.0301, "step": 1510 }, { "epoch": 0.36934734783671475, "grad_norm": 3.2351458072662354, "learning_rate": 1.4545547593523308e-05, "loss": 0.9546, "step": 1511 }, { "epoch": 0.36959178684918115, "grad_norm": 3.675766706466675, "learning_rate": 1.4538494063695861e-05, "loss": 0.9193, "step": 1512 }, { "epoch": 0.3698362258616475, "grad_norm": 4.477265357971191, "learning_rate": 1.4531437688960796e-05, "loss": 0.989, "step": 1513 }, { "epoch": 0.3700806648741139, "grad_norm": 4.235387325286865, "learning_rate": 1.4524378473741335e-05, "loss": 0.9246, "step": 1514 }, { "epoch": 0.3703251038865803, "grad_norm": 4.069041728973389, "learning_rate": 1.4517316422462471e-05, "loss": 0.8349, "step": 1515 }, { "epoch": 0.3705695428990467, "grad_norm": 4.676631450653076, "learning_rate": 1.4510251539550978e-05, "loss": 0.9418, "step": 1516 }, { "epoch": 0.3708139819115131, "grad_norm": 3.579474687576294, "learning_rate": 1.4503183829435403e-05, "loss": 0.8662, "step": 1517 }, { "epoch": 0.37105842092397945, "grad_norm": 3.939366340637207, "learning_rate": 1.4496113296546068e-05, "loss": 0.9567, "step": 1518 }, { "epoch": 0.37130285993644585, "grad_norm": 3.652257204055786, "learning_rate": 1.4489039945315059e-05, "loss": 0.826, "step": 1519 }, { "epoch": 0.37154729894891225, "grad_norm": 3.49208664894104, "learning_rate": 1.4481963780176235e-05, "loss": 0.815, "step": 1520 }, { "epoch": 0.37179173796137865, "grad_norm": 4.103240013122559, "learning_rate": 1.4474884805565217e-05, "loss": 0.8812, "step": 1521 }, { "epoch": 0.372036176973845, "grad_norm": 3.390845537185669, "learning_rate": 1.4467803025919383e-05, "loss": 0.9073, "step": 1522 }, { "epoch": 0.3722806159863114, "grad_norm": 4.154443740844727, "learning_rate": 1.4460718445677877e-05, "loss": 0.9916, "step": 1523 }, { "epoch": 0.3725250549987778, "grad_norm": 3.4157791137695312, "learning_rate": 1.4453631069281593e-05, "loss": 0.9102, "step": 1524 }, { "epoch": 0.3727694940112442, "grad_norm": 3.5285110473632812, "learning_rate": 1.4446540901173176e-05, "loss": 0.8537, "step": 1525 }, { "epoch": 0.3730139330237106, "grad_norm": 4.130982875823975, "learning_rate": 1.4439447945797026e-05, "loss": 0.8885, "step": 1526 }, { "epoch": 0.37325837203617696, "grad_norm": 3.806924819946289, "learning_rate": 1.4432352207599288e-05, "loss": 0.9369, "step": 1527 }, { "epoch": 0.37350281104864336, "grad_norm": 3.0304319858551025, "learning_rate": 1.442525369102785e-05, "loss": 0.884, "step": 1528 }, { "epoch": 0.37374725006110976, "grad_norm": 3.4081881046295166, "learning_rate": 1.4418152400532344e-05, "loss": 0.9057, "step": 1529 }, { "epoch": 0.37399168907357616, "grad_norm": 3.4484832286834717, "learning_rate": 1.441104834056414e-05, "loss": 0.9393, "step": 1530 }, { "epoch": 0.3742361280860425, "grad_norm": 3.4003255367279053, "learning_rate": 1.4403941515576344e-05, "loss": 0.8726, "step": 1531 }, { "epoch": 0.3744805670985089, "grad_norm": 4.160417079925537, "learning_rate": 1.4396831930023796e-05, "loss": 0.9071, "step": 1532 }, { "epoch": 0.3747250061109753, "grad_norm": 3.3831288814544678, "learning_rate": 1.4389719588363062e-05, "loss": 0.8404, "step": 1533 }, { "epoch": 0.3749694451234417, "grad_norm": 3.3936948776245117, "learning_rate": 1.4382604495052444e-05, "loss": 0.9059, "step": 1534 }, { "epoch": 0.3752138841359081, "grad_norm": 4.137432098388672, "learning_rate": 1.437548665455196e-05, "loss": 0.7942, "step": 1535 }, { "epoch": 0.37545832314837446, "grad_norm": 4.356401443481445, "learning_rate": 1.4368366071323354e-05, "loss": 0.9009, "step": 1536 }, { "epoch": 0.37570276216084086, "grad_norm": 3.207725763320923, "learning_rate": 1.4361242749830093e-05, "loss": 0.8208, "step": 1537 }, { "epoch": 0.37594720117330727, "grad_norm": 3.3889639377593994, "learning_rate": 1.435411669453735e-05, "loss": 0.9573, "step": 1538 }, { "epoch": 0.37619164018577367, "grad_norm": 4.115054130554199, "learning_rate": 1.4346987909912024e-05, "loss": 0.92, "step": 1539 }, { "epoch": 0.37643607919824, "grad_norm": 3.934098720550537, "learning_rate": 1.4339856400422721e-05, "loss": 0.8831, "step": 1540 }, { "epoch": 0.3766805182107064, "grad_norm": 4.6525559425354, "learning_rate": 1.4332722170539748e-05, "loss": 0.843, "step": 1541 }, { "epoch": 0.3769249572231728, "grad_norm": 3.4838201999664307, "learning_rate": 1.4325585224735125e-05, "loss": 0.9346, "step": 1542 }, { "epoch": 0.3771693962356392, "grad_norm": 3.2306668758392334, "learning_rate": 1.431844556748257e-05, "loss": 0.8231, "step": 1543 }, { "epoch": 0.3774138352481056, "grad_norm": 3.608369827270508, "learning_rate": 1.4311303203257505e-05, "loss": 0.9087, "step": 1544 }, { "epoch": 0.37765827426057197, "grad_norm": 3.370871067047119, "learning_rate": 1.4304158136537047e-05, "loss": 0.9316, "step": 1545 }, { "epoch": 0.37790271327303837, "grad_norm": 3.6224355697631836, "learning_rate": 1.4297010371800002e-05, "loss": 0.8198, "step": 1546 }, { "epoch": 0.3781471522855048, "grad_norm": 2.598627805709839, "learning_rate": 1.4289859913526876e-05, "loss": 2.0515, "step": 1547 }, { "epoch": 0.3783915912979712, "grad_norm": 3.645094394683838, "learning_rate": 1.4282706766199859e-05, "loss": 0.9776, "step": 1548 }, { "epoch": 0.3786360303104375, "grad_norm": 3.49668288230896, "learning_rate": 1.4275550934302822e-05, "loss": 0.9527, "step": 1549 }, { "epoch": 0.3788804693229039, "grad_norm": 3.822556972503662, "learning_rate": 1.426839242232133e-05, "loss": 0.9844, "step": 1550 }, { "epoch": 0.3791249083353703, "grad_norm": 3.3121814727783203, "learning_rate": 1.4261231234742618e-05, "loss": 0.8436, "step": 1551 }, { "epoch": 0.3793693473478367, "grad_norm": 3.4897799491882324, "learning_rate": 1.4254067376055602e-05, "loss": 0.9729, "step": 1552 }, { "epoch": 0.37961378636030313, "grad_norm": 3.6050734519958496, "learning_rate": 1.424690085075087e-05, "loss": 0.6889, "step": 1553 }, { "epoch": 0.3798582253727695, "grad_norm": 3.686805248260498, "learning_rate": 1.4239731663320688e-05, "loss": 0.8947, "step": 1554 }, { "epoch": 0.3801026643852359, "grad_norm": 3.1977193355560303, "learning_rate": 1.4232559818258983e-05, "loss": 0.9087, "step": 1555 }, { "epoch": 0.3803471033977023, "grad_norm": 5.1845526695251465, "learning_rate": 1.4225385320061353e-05, "loss": 0.8625, "step": 1556 }, { "epoch": 0.3805915424101687, "grad_norm": 3.603829860687256, "learning_rate": 1.4218208173225057e-05, "loss": 0.9319, "step": 1557 }, { "epoch": 0.38083598142263503, "grad_norm": 3.646444797515869, "learning_rate": 1.4211028382249013e-05, "loss": 0.8586, "step": 1558 }, { "epoch": 0.38108042043510143, "grad_norm": 4.110007286071777, "learning_rate": 1.4203845951633799e-05, "loss": 0.9234, "step": 1559 }, { "epoch": 0.38132485944756783, "grad_norm": 4.331524848937988, "learning_rate": 1.4196660885881646e-05, "loss": 0.9518, "step": 1560 }, { "epoch": 0.38156929846003423, "grad_norm": 3.5555384159088135, "learning_rate": 1.4189473189496437e-05, "loss": 0.8741, "step": 1561 }, { "epoch": 0.38181373747250064, "grad_norm": 3.843526840209961, "learning_rate": 1.4182282866983708e-05, "loss": 0.8401, "step": 1562 }, { "epoch": 0.382058176484967, "grad_norm": 3.4764273166656494, "learning_rate": 1.4175089922850633e-05, "loss": 0.9026, "step": 1563 }, { "epoch": 0.3823026154974334, "grad_norm": 3.949537754058838, "learning_rate": 1.4167894361606038e-05, "loss": 0.8916, "step": 1564 }, { "epoch": 0.3825470545098998, "grad_norm": 3.3268473148345947, "learning_rate": 1.4160696187760383e-05, "loss": 0.8919, "step": 1565 }, { "epoch": 0.3827914935223662, "grad_norm": 4.385059356689453, "learning_rate": 1.4153495405825768e-05, "loss": 1.0173, "step": 1566 }, { "epoch": 0.38303593253483254, "grad_norm": 4.0300445556640625, "learning_rate": 1.4146292020315926e-05, "loss": 0.9705, "step": 1567 }, { "epoch": 0.38328037154729894, "grad_norm": 3.9647634029388428, "learning_rate": 1.413908603574623e-05, "loss": 0.9537, "step": 1568 }, { "epoch": 0.38352481055976534, "grad_norm": 3.9260358810424805, "learning_rate": 1.413187745663367e-05, "loss": 0.8883, "step": 1569 }, { "epoch": 0.38376924957223174, "grad_norm": 3.1854300498962402, "learning_rate": 1.4124666287496873e-05, "loss": 0.9824, "step": 1570 }, { "epoch": 0.38401368858469814, "grad_norm": 3.3261890411376953, "learning_rate": 1.4117452532856084e-05, "loss": 0.9699, "step": 1571 }, { "epoch": 0.3842581275971645, "grad_norm": 3.499786615371704, "learning_rate": 1.4110236197233164e-05, "loss": 0.8683, "step": 1572 }, { "epoch": 0.3845025666096309, "grad_norm": 3.24682879447937, "learning_rate": 1.4103017285151607e-05, "loss": 0.9277, "step": 1573 }, { "epoch": 0.3847470056220973, "grad_norm": 6.736312389373779, "learning_rate": 1.4095795801136509e-05, "loss": 0.9788, "step": 1574 }, { "epoch": 0.3849914446345637, "grad_norm": 3.485842227935791, "learning_rate": 1.408857174971458e-05, "loss": 0.818, "step": 1575 }, { "epoch": 0.38523588364703004, "grad_norm": 3.612020969390869, "learning_rate": 1.4081345135414144e-05, "loss": 1.0562, "step": 1576 }, { "epoch": 0.38548032265949644, "grad_norm": 3.5550265312194824, "learning_rate": 1.4074115962765125e-05, "loss": 0.9049, "step": 1577 }, { "epoch": 0.38572476167196285, "grad_norm": 3.0933775901794434, "learning_rate": 1.4066884236299057e-05, "loss": 0.8222, "step": 1578 }, { "epoch": 0.38596920068442925, "grad_norm": 3.570155143737793, "learning_rate": 1.4059649960549071e-05, "loss": 0.8068, "step": 1579 }, { "epoch": 0.38621363969689565, "grad_norm": 3.3882038593292236, "learning_rate": 1.4052413140049898e-05, "loss": 0.874, "step": 1580 }, { "epoch": 0.386458078709362, "grad_norm": 3.5525708198547363, "learning_rate": 1.4045173779337866e-05, "loss": 0.9939, "step": 1581 }, { "epoch": 0.3867025177218284, "grad_norm": 4.007676124572754, "learning_rate": 1.4037931882950888e-05, "loss": 0.8346, "step": 1582 }, { "epoch": 0.3869469567342948, "grad_norm": 3.0331296920776367, "learning_rate": 1.4030687455428473e-05, "loss": 0.9635, "step": 1583 }, { "epoch": 0.3871913957467612, "grad_norm": 15.911253929138184, "learning_rate": 1.4023440501311713e-05, "loss": 0.8254, "step": 1584 }, { "epoch": 0.38743583475922755, "grad_norm": 3.9656739234924316, "learning_rate": 1.4016191025143289e-05, "loss": 0.9775, "step": 1585 }, { "epoch": 0.38768027377169395, "grad_norm": 3.3272018432617188, "learning_rate": 1.400893903146746e-05, "loss": 0.9645, "step": 1586 }, { "epoch": 0.38792471278416035, "grad_norm": 4.041081428527832, "learning_rate": 1.4001684524830057e-05, "loss": 0.8346, "step": 1587 }, { "epoch": 0.38816915179662675, "grad_norm": 3.2119650840759277, "learning_rate": 1.3994427509778499e-05, "loss": 0.8845, "step": 1588 }, { "epoch": 0.38841359080909316, "grad_norm": 2.7611374855041504, "learning_rate": 1.3987167990861763e-05, "loss": 0.8931, "step": 1589 }, { "epoch": 0.3886580298215595, "grad_norm": 3.178654670715332, "learning_rate": 1.3979905972630405e-05, "loss": 0.9309, "step": 1590 }, { "epoch": 0.3889024688340259, "grad_norm": 4.22986364364624, "learning_rate": 1.3972641459636548e-05, "loss": 0.7984, "step": 1591 }, { "epoch": 0.3891469078464923, "grad_norm": 3.732235908508301, "learning_rate": 1.396537445643387e-05, "loss": 0.8091, "step": 1592 }, { "epoch": 0.3893913468589587, "grad_norm": 3.6782584190368652, "learning_rate": 1.3958104967577627e-05, "loss": 1.053, "step": 1593 }, { "epoch": 0.38963578587142506, "grad_norm": 3.923552989959717, "learning_rate": 1.3950832997624612e-05, "loss": 0.8426, "step": 1594 }, { "epoch": 0.38988022488389146, "grad_norm": 3.2947030067443848, "learning_rate": 1.3943558551133186e-05, "loss": 0.9391, "step": 1595 }, { "epoch": 0.39012466389635786, "grad_norm": 3.272634506225586, "learning_rate": 1.3936281632663263e-05, "loss": 0.886, "step": 1596 }, { "epoch": 0.39036910290882426, "grad_norm": 4.148128986358643, "learning_rate": 1.39290022467763e-05, "loss": 1.045, "step": 1597 }, { "epoch": 0.39061354192129066, "grad_norm": 3.9834187030792236, "learning_rate": 1.3921720398035306e-05, "loss": 0.8731, "step": 1598 }, { "epoch": 0.390857980933757, "grad_norm": 3.3036372661590576, "learning_rate": 1.3914436091004829e-05, "loss": 0.8053, "step": 1599 }, { "epoch": 0.3911024199462234, "grad_norm": 3.2304863929748535, "learning_rate": 1.3907149330250962e-05, "loss": 0.8049, "step": 1600 }, { "epoch": 0.3913468589586898, "grad_norm": 3.1071078777313232, "learning_rate": 1.3899860120341338e-05, "loss": 0.7691, "step": 1601 }, { "epoch": 0.3915912979711562, "grad_norm": 3.505664110183716, "learning_rate": 1.3892568465845118e-05, "loss": 0.8412, "step": 1602 }, { "epoch": 0.39183573698362256, "grad_norm": 3.2916007041931152, "learning_rate": 1.3885274371333001e-05, "loss": 0.8779, "step": 1603 }, { "epoch": 0.39208017599608896, "grad_norm": 3.703036069869995, "learning_rate": 1.3877977841377217e-05, "loss": 0.8388, "step": 1604 }, { "epoch": 0.39232461500855537, "grad_norm": 3.327202320098877, "learning_rate": 1.3870678880551516e-05, "loss": 0.9072, "step": 1605 }, { "epoch": 0.39256905402102177, "grad_norm": 5.386071681976318, "learning_rate": 1.3863377493431177e-05, "loss": 0.975, "step": 1606 }, { "epoch": 0.39281349303348817, "grad_norm": 3.7295799255371094, "learning_rate": 1.3856073684592994e-05, "loss": 0.7813, "step": 1607 }, { "epoch": 0.3930579320459545, "grad_norm": 3.692769765853882, "learning_rate": 1.3848767458615286e-05, "loss": 0.8892, "step": 1608 }, { "epoch": 0.3933023710584209, "grad_norm": 3.072274923324585, "learning_rate": 1.3841458820077889e-05, "loss": 0.8592, "step": 1609 }, { "epoch": 0.3935468100708873, "grad_norm": 3.234300136566162, "learning_rate": 1.3834147773562143e-05, "loss": 1.0064, "step": 1610 }, { "epoch": 0.3937912490833537, "grad_norm": 4.414983749389648, "learning_rate": 1.3826834323650899e-05, "loss": 0.8768, "step": 1611 }, { "epoch": 0.39403568809582007, "grad_norm": 3.7046587467193604, "learning_rate": 1.3819518474928519e-05, "loss": 0.8014, "step": 1612 }, { "epoch": 0.39428012710828647, "grad_norm": 3.0449559688568115, "learning_rate": 1.3812200231980869e-05, "loss": 0.9361, "step": 1613 }, { "epoch": 0.3945245661207529, "grad_norm": 3.53568696975708, "learning_rate": 1.380487959939531e-05, "loss": 0.9733, "step": 1614 }, { "epoch": 0.3947690051332193, "grad_norm": 3.7599360942840576, "learning_rate": 1.37975565817607e-05, "loss": 0.9341, "step": 1615 }, { "epoch": 0.3950134441456857, "grad_norm": 3.2119531631469727, "learning_rate": 1.3790231183667403e-05, "loss": 0.926, "step": 1616 }, { "epoch": 0.395257883158152, "grad_norm": 3.0126593112945557, "learning_rate": 1.3782903409707266e-05, "loss": 0.8564, "step": 1617 }, { "epoch": 0.3955023221706184, "grad_norm": 3.4946446418762207, "learning_rate": 1.3775573264473629e-05, "loss": 0.8601, "step": 1618 }, { "epoch": 0.3957467611830848, "grad_norm": 3.624980926513672, "learning_rate": 1.3768240752561315e-05, "loss": 0.9043, "step": 1619 }, { "epoch": 0.39599120019555123, "grad_norm": 3.868273973464966, "learning_rate": 1.3760905878566633e-05, "loss": 0.9251, "step": 1620 }, { "epoch": 0.3962356392080176, "grad_norm": 3.1690731048583984, "learning_rate": 1.3753568647087372e-05, "loss": 0.8974, "step": 1621 }, { "epoch": 0.396480078220484, "grad_norm": 4.424831867218018, "learning_rate": 1.37462290627228e-05, "loss": 0.8481, "step": 1622 }, { "epoch": 0.3967245172329504, "grad_norm": 3.5461859703063965, "learning_rate": 1.3738887130073655e-05, "loss": 0.8403, "step": 1623 }, { "epoch": 0.3969689562454168, "grad_norm": 3.7361207008361816, "learning_rate": 1.3731542853742155e-05, "loss": 0.8354, "step": 1624 }, { "epoch": 0.3972133952578832, "grad_norm": 2.770488977432251, "learning_rate": 1.3724196238331983e-05, "loss": 2.013, "step": 1625 }, { "epoch": 0.39745783427034953, "grad_norm": 2.9239935874938965, "learning_rate": 1.3716847288448287e-05, "loss": 0.7827, "step": 1626 }, { "epoch": 0.39770227328281593, "grad_norm": 3.373499870300293, "learning_rate": 1.370949600869768e-05, "loss": 0.9397, "step": 1627 }, { "epoch": 0.39794671229528233, "grad_norm": 3.6640830039978027, "learning_rate": 1.3702142403688234e-05, "loss": 0.7888, "step": 1628 }, { "epoch": 0.39819115130774874, "grad_norm": 3.442188024520874, "learning_rate": 1.3694786478029478e-05, "loss": 0.8488, "step": 1629 }, { "epoch": 0.3984355903202151, "grad_norm": 3.294236183166504, "learning_rate": 1.3687428236332401e-05, "loss": 0.8168, "step": 1630 }, { "epoch": 0.3986800293326815, "grad_norm": 3.681321620941162, "learning_rate": 1.3680067683209438e-05, "loss": 0.8385, "step": 1631 }, { "epoch": 0.3989244683451479, "grad_norm": 5.7639384269714355, "learning_rate": 1.3672704823274472e-05, "loss": 0.8752, "step": 1632 }, { "epoch": 0.3991689073576143, "grad_norm": 4.043473720550537, "learning_rate": 1.366533966114284e-05, "loss": 0.8396, "step": 1633 }, { "epoch": 0.3994133463700807, "grad_norm": 3.78446626663208, "learning_rate": 1.3657972201431315e-05, "loss": 0.9433, "step": 1634 }, { "epoch": 0.39965778538254704, "grad_norm": 3.113582134246826, "learning_rate": 1.3650602448758113e-05, "loss": 0.8771, "step": 1635 }, { "epoch": 0.39990222439501344, "grad_norm": 3.1159162521362305, "learning_rate": 1.3643230407742883e-05, "loss": 0.7404, "step": 1636 }, { "epoch": 0.40014666340747984, "grad_norm": 2.6976401805877686, "learning_rate": 1.3635856083006715e-05, "loss": 2.1143, "step": 1637 }, { "epoch": 0.40039110241994624, "grad_norm": 3.4253532886505127, "learning_rate": 1.3628479479172128e-05, "loss": 0.6934, "step": 1638 }, { "epoch": 0.4006355414324126, "grad_norm": 3.560701847076416, "learning_rate": 1.3621100600863066e-05, "loss": 0.9006, "step": 1639 }, { "epoch": 0.400879980444879, "grad_norm": 3.8821518421173096, "learning_rate": 1.3613719452704906e-05, "loss": 0.7986, "step": 1640 }, { "epoch": 0.4011244194573454, "grad_norm": 3.5182454586029053, "learning_rate": 1.3606336039324439e-05, "loss": 0.8107, "step": 1641 }, { "epoch": 0.4013688584698118, "grad_norm": 3.8955798149108887, "learning_rate": 1.3598950365349884e-05, "loss": 0.885, "step": 1642 }, { "epoch": 0.4016132974822782, "grad_norm": 3.7581288814544678, "learning_rate": 1.3591562435410873e-05, "loss": 0.7875, "step": 1643 }, { "epoch": 0.40185773649474454, "grad_norm": 4.369556427001953, "learning_rate": 1.3584172254138452e-05, "loss": 0.9033, "step": 1644 }, { "epoch": 0.40210217550721095, "grad_norm": 5.720822811126709, "learning_rate": 1.357677982616508e-05, "loss": 0.9166, "step": 1645 }, { "epoch": 0.40234661451967735, "grad_norm": 3.371997117996216, "learning_rate": 1.356938515612462e-05, "loss": 0.9141, "step": 1646 }, { "epoch": 0.40259105353214375, "grad_norm": 4.52249002456665, "learning_rate": 1.3561988248652346e-05, "loss": 0.9192, "step": 1647 }, { "epoch": 0.4028354925446101, "grad_norm": 3.647069215774536, "learning_rate": 1.3554589108384937e-05, "loss": 0.9729, "step": 1648 }, { "epoch": 0.4030799315570765, "grad_norm": 3.2838327884674072, "learning_rate": 1.3547187739960458e-05, "loss": 0.8376, "step": 1649 }, { "epoch": 0.4033243705695429, "grad_norm": 4.263354778289795, "learning_rate": 1.3539784148018387e-05, "loss": 0.9311, "step": 1650 }, { "epoch": 0.4035688095820093, "grad_norm": 3.7330758571624756, "learning_rate": 1.353237833719958e-05, "loss": 0.8295, "step": 1651 }, { "epoch": 0.4038132485944757, "grad_norm": 3.026618003845215, "learning_rate": 1.3524970312146305e-05, "loss": 0.9112, "step": 1652 }, { "epoch": 0.40405768760694205, "grad_norm": 3.332324266433716, "learning_rate": 1.3517560077502191e-05, "loss": 0.8801, "step": 1653 }, { "epoch": 0.40430212661940845, "grad_norm": 3.2940218448638916, "learning_rate": 1.3510147637912276e-05, "loss": 0.9387, "step": 1654 }, { "epoch": 0.40454656563187485, "grad_norm": 3.100508213043213, "learning_rate": 1.3502732998022965e-05, "loss": 0.8721, "step": 1655 }, { "epoch": 0.40479100464434126, "grad_norm": 3.2187600135803223, "learning_rate": 1.3495316162482051e-05, "loss": 0.773, "step": 1656 }, { "epoch": 0.4050354436568076, "grad_norm": 3.792860984802246, "learning_rate": 1.3487897135938697e-05, "loss": 0.7794, "step": 1657 }, { "epoch": 0.405279882669274, "grad_norm": 4.043926239013672, "learning_rate": 1.3480475923043445e-05, "loss": 0.845, "step": 1658 }, { "epoch": 0.4055243216817404, "grad_norm": 3.7164793014526367, "learning_rate": 1.3473052528448203e-05, "loss": 0.9005, "step": 1659 }, { "epoch": 0.4057687606942068, "grad_norm": 4.479531288146973, "learning_rate": 1.346562695680625e-05, "loss": 0.8135, "step": 1660 }, { "epoch": 0.4060131997066732, "grad_norm": 3.599146604537964, "learning_rate": 1.3458199212772227e-05, "loss": 0.7886, "step": 1661 }, { "epoch": 0.40625763871913956, "grad_norm": 3.1297788619995117, "learning_rate": 1.3450769301002135e-05, "loss": 0.8802, "step": 1662 }, { "epoch": 0.40650207773160596, "grad_norm": 3.232038974761963, "learning_rate": 1.3443337226153343e-05, "loss": 0.888, "step": 1663 }, { "epoch": 0.40674651674407236, "grad_norm": 3.5411183834075928, "learning_rate": 1.3435902992884567e-05, "loss": 0.88, "step": 1664 }, { "epoch": 0.40699095575653876, "grad_norm": 4.1356682777404785, "learning_rate": 1.3428466605855874e-05, "loss": 0.8242, "step": 1665 }, { "epoch": 0.4072353947690051, "grad_norm": 4.684967517852783, "learning_rate": 1.342102806972869e-05, "loss": 0.837, "step": 1666 }, { "epoch": 0.4074798337814715, "grad_norm": 4.071098327636719, "learning_rate": 1.3413587389165783e-05, "loss": 0.9001, "step": 1667 }, { "epoch": 0.4077242727939379, "grad_norm": 3.038487434387207, "learning_rate": 1.3406144568831265e-05, "loss": 0.8962, "step": 1668 }, { "epoch": 0.4079687118064043, "grad_norm": 3.531371593475342, "learning_rate": 1.3398699613390593e-05, "loss": 0.89, "step": 1669 }, { "epoch": 0.4082131508188707, "grad_norm": 3.3554468154907227, "learning_rate": 1.3391252527510554e-05, "loss": 0.9256, "step": 1670 }, { "epoch": 0.40845758983133706, "grad_norm": 4.093055725097656, "learning_rate": 1.3383803315859281e-05, "loss": 0.8241, "step": 1671 }, { "epoch": 0.40870202884380347, "grad_norm": 3.222688913345337, "learning_rate": 1.3376351983106233e-05, "loss": 0.9334, "step": 1672 }, { "epoch": 0.40894646785626987, "grad_norm": 3.072899341583252, "learning_rate": 1.3368898533922202e-05, "loss": 0.7888, "step": 1673 }, { "epoch": 0.40919090686873627, "grad_norm": 3.434206008911133, "learning_rate": 1.3361442972979301e-05, "loss": 0.9011, "step": 1674 }, { "epoch": 0.4094353458812026, "grad_norm": 3.513918161392212, "learning_rate": 1.3353985304950974e-05, "loss": 0.8411, "step": 1675 }, { "epoch": 0.409679784893669, "grad_norm": 3.3258190155029297, "learning_rate": 1.3346525534511978e-05, "loss": 0.846, "step": 1676 }, { "epoch": 0.4099242239061354, "grad_norm": 4.432790756225586, "learning_rate": 1.3339063666338396e-05, "loss": 0.8694, "step": 1677 }, { "epoch": 0.4101686629186018, "grad_norm": 3.975558042526245, "learning_rate": 1.333159970510762e-05, "loss": 0.9027, "step": 1678 }, { "epoch": 0.4104131019310682, "grad_norm": 3.4570248126983643, "learning_rate": 1.3324133655498361e-05, "loss": 0.8495, "step": 1679 }, { "epoch": 0.41065754094353457, "grad_norm": 3.605822801589966, "learning_rate": 1.3316665522190628e-05, "loss": 0.8228, "step": 1680 }, { "epoch": 0.410901979956001, "grad_norm": 3.1482040882110596, "learning_rate": 1.3309195309865746e-05, "loss": 0.7748, "step": 1681 }, { "epoch": 0.4111464189684674, "grad_norm": 3.5796356201171875, "learning_rate": 1.3301723023206336e-05, "loss": 0.8505, "step": 1682 }, { "epoch": 0.4113908579809338, "grad_norm": 3.5911896228790283, "learning_rate": 1.329424866689633e-05, "loss": 0.8742, "step": 1683 }, { "epoch": 0.4116352969934001, "grad_norm": 2.871198892593384, "learning_rate": 1.3286772245620942e-05, "loss": 0.8436, "step": 1684 }, { "epoch": 0.4118797360058665, "grad_norm": 3.2517623901367188, "learning_rate": 1.3279293764066693e-05, "loss": 0.8314, "step": 1685 }, { "epoch": 0.41212417501833293, "grad_norm": 3.427194833755493, "learning_rate": 1.3271813226921388e-05, "loss": 0.8339, "step": 1686 }, { "epoch": 0.41236861403079933, "grad_norm": 4.1440582275390625, "learning_rate": 1.3264330638874128e-05, "loss": 0.9002, "step": 1687 }, { "epoch": 0.41261305304326573, "grad_norm": 3.1286559104919434, "learning_rate": 1.3256846004615293e-05, "loss": 0.7998, "step": 1688 }, { "epoch": 0.4128574920557321, "grad_norm": 3.3335187435150146, "learning_rate": 1.3249359328836549e-05, "loss": 0.8403, "step": 1689 }, { "epoch": 0.4131019310681985, "grad_norm": 3.547224998474121, "learning_rate": 1.3241870616230839e-05, "loss": 0.9001, "step": 1690 }, { "epoch": 0.4133463700806649, "grad_norm": 3.3161604404449463, "learning_rate": 1.3234379871492381e-05, "loss": 0.7957, "step": 1691 }, { "epoch": 0.4135908090931313, "grad_norm": 4.421625137329102, "learning_rate": 1.3226887099316678e-05, "loss": 0.8809, "step": 1692 }, { "epoch": 0.41383524810559763, "grad_norm": 3.8033032417297363, "learning_rate": 1.3219392304400489e-05, "loss": 0.8683, "step": 1693 }, { "epoch": 0.41407968711806403, "grad_norm": 3.9016315937042236, "learning_rate": 1.3211895491441853e-05, "loss": 0.8793, "step": 1694 }, { "epoch": 0.41432412613053043, "grad_norm": 3.785693407058716, "learning_rate": 1.3204396665140066e-05, "loss": 0.895, "step": 1695 }, { "epoch": 0.41456856514299684, "grad_norm": 3.5152273178100586, "learning_rate": 1.3196895830195691e-05, "loss": 0.7666, "step": 1696 }, { "epoch": 0.41481300415546324, "grad_norm": 3.0873780250549316, "learning_rate": 1.3189392991310545e-05, "loss": 0.8339, "step": 1697 }, { "epoch": 0.4150574431679296, "grad_norm": 4.051246643066406, "learning_rate": 1.3181888153187705e-05, "loss": 0.9735, "step": 1698 }, { "epoch": 0.415301882180396, "grad_norm": 3.4806783199310303, "learning_rate": 1.3174381320531504e-05, "loss": 0.7755, "step": 1699 }, { "epoch": 0.4155463211928624, "grad_norm": 3.6442301273345947, "learning_rate": 1.3166872498047515e-05, "loss": 0.8943, "step": 1700 }, { "epoch": 0.4157907602053288, "grad_norm": 4.0153374671936035, "learning_rate": 1.315936169044257e-05, "loss": 0.884, "step": 1701 }, { "epoch": 0.41603519921779514, "grad_norm": 3.542607545852661, "learning_rate": 1.3151848902424743e-05, "loss": 0.9262, "step": 1702 }, { "epoch": 0.41627963823026154, "grad_norm": 3.4306719303131104, "learning_rate": 1.3144334138703341e-05, "loss": 0.9948, "step": 1703 }, { "epoch": 0.41652407724272794, "grad_norm": 3.551039934158325, "learning_rate": 1.3136817403988918e-05, "loss": 0.7998, "step": 1704 }, { "epoch": 0.41676851625519434, "grad_norm": 3.1957240104675293, "learning_rate": 1.3129298702993256e-05, "loss": 0.929, "step": 1705 }, { "epoch": 0.41701295526766075, "grad_norm": 3.0402817726135254, "learning_rate": 1.3121778040429382e-05, "loss": 0.8153, "step": 1706 }, { "epoch": 0.4172573942801271, "grad_norm": 2.92474102973938, "learning_rate": 1.311425542101154e-05, "loss": 0.8377, "step": 1707 }, { "epoch": 0.4175018332925935, "grad_norm": 3.0902931690216064, "learning_rate": 1.3106730849455205e-05, "loss": 0.8092, "step": 1708 }, { "epoch": 0.4177462723050599, "grad_norm": 2.998044967651367, "learning_rate": 1.3099204330477078e-05, "loss": 0.8138, "step": 1709 }, { "epoch": 0.4179907113175263, "grad_norm": 3.3012495040893555, "learning_rate": 1.3091675868795077e-05, "loss": 0.697, "step": 1710 }, { "epoch": 0.41823515032999264, "grad_norm": 3.38061261177063, "learning_rate": 1.3084145469128343e-05, "loss": 0.8677, "step": 1711 }, { "epoch": 0.41847958934245905, "grad_norm": 3.9767093658447266, "learning_rate": 1.3076613136197225e-05, "loss": 0.8228, "step": 1712 }, { "epoch": 0.41872402835492545, "grad_norm": 3.446218252182007, "learning_rate": 1.306907887472329e-05, "loss": 0.7189, "step": 1713 }, { "epoch": 0.41896846736739185, "grad_norm": 4.28263521194458, "learning_rate": 1.3061542689429308e-05, "loss": 0.7953, "step": 1714 }, { "epoch": 0.41921290637985825, "grad_norm": 3.218669891357422, "learning_rate": 1.305400458503926e-05, "loss": 0.7847, "step": 1715 }, { "epoch": 0.4194573453923246, "grad_norm": 4.574066638946533, "learning_rate": 1.3046464566278325e-05, "loss": 0.7932, "step": 1716 }, { "epoch": 0.419701784404791, "grad_norm": 3.627200126647949, "learning_rate": 1.303892263787289e-05, "loss": 0.9059, "step": 1717 }, { "epoch": 0.4199462234172574, "grad_norm": 3.5440592765808105, "learning_rate": 1.3031378804550533e-05, "loss": 0.7798, "step": 1718 }, { "epoch": 0.4201906624297238, "grad_norm": 3.743830442428589, "learning_rate": 1.3023833071040026e-05, "loss": 0.8675, "step": 1719 }, { "epoch": 0.42043510144219015, "grad_norm": 4.3647356033325195, "learning_rate": 1.3016285442071332e-05, "loss": 0.8428, "step": 1720 }, { "epoch": 0.42067954045465655, "grad_norm": 3.2686777114868164, "learning_rate": 1.3008735922375607e-05, "loss": 0.7906, "step": 1721 }, { "epoch": 0.42092397946712296, "grad_norm": 4.280836582183838, "learning_rate": 1.3001184516685181e-05, "loss": 0.8492, "step": 1722 }, { "epoch": 0.42116841847958936, "grad_norm": 3.6485018730163574, "learning_rate": 1.2993631229733584e-05, "loss": 0.845, "step": 1723 }, { "epoch": 0.42141285749205576, "grad_norm": 3.2848522663116455, "learning_rate": 1.2986076066255504e-05, "loss": 0.7116, "step": 1724 }, { "epoch": 0.4216572965045221, "grad_norm": 3.6487104892730713, "learning_rate": 1.2978519030986827e-05, "loss": 0.8676, "step": 1725 }, { "epoch": 0.4219017355169885, "grad_norm": 3.091217279434204, "learning_rate": 1.2970960128664597e-05, "loss": 0.8372, "step": 1726 }, { "epoch": 0.4221461745294549, "grad_norm": 2.7542967796325684, "learning_rate": 1.2963399364027031e-05, "loss": 1.9989, "step": 1727 }, { "epoch": 0.4223906135419213, "grad_norm": 3.3845884799957275, "learning_rate": 1.2955836741813519e-05, "loss": 0.7785, "step": 1728 }, { "epoch": 0.42263505255438766, "grad_norm": 3.99977970123291, "learning_rate": 1.294827226676461e-05, "loss": 0.8772, "step": 1729 }, { "epoch": 0.42287949156685406, "grad_norm": 3.4157869815826416, "learning_rate": 1.2940705943622013e-05, "loss": 0.8963, "step": 1730 }, { "epoch": 0.42312393057932046, "grad_norm": 4.183679580688477, "learning_rate": 1.2933137777128607e-05, "loss": 0.8847, "step": 1731 }, { "epoch": 0.42336836959178686, "grad_norm": 3.5591185092926025, "learning_rate": 1.2925567772028412e-05, "loss": 0.7772, "step": 1732 }, { "epoch": 0.42361280860425327, "grad_norm": 4.990171909332275, "learning_rate": 1.2917995933066604e-05, "loss": 0.901, "step": 1733 }, { "epoch": 0.4238572476167196, "grad_norm": 2.8711509704589844, "learning_rate": 1.2910422264989519e-05, "loss": 0.6467, "step": 1734 }, { "epoch": 0.424101686629186, "grad_norm": 3.38092041015625, "learning_rate": 1.2902846772544625e-05, "loss": 0.9127, "step": 1735 }, { "epoch": 0.4243461256416524, "grad_norm": 3.018601894378662, "learning_rate": 1.2895269460480544e-05, "loss": 0.8054, "step": 1736 }, { "epoch": 0.4245905646541188, "grad_norm": 3.373168706893921, "learning_rate": 1.2887690333547034e-05, "loss": 0.8284, "step": 1737 }, { "epoch": 0.42483500366658516, "grad_norm": 3.377891778945923, "learning_rate": 1.2880109396494993e-05, "loss": 0.8189, "step": 1738 }, { "epoch": 0.42507944267905157, "grad_norm": 4.429141044616699, "learning_rate": 1.287252665407645e-05, "loss": 0.7497, "step": 1739 }, { "epoch": 0.42532388169151797, "grad_norm": 3.5740861892700195, "learning_rate": 1.2864942111044567e-05, "loss": 0.7699, "step": 1740 }, { "epoch": 0.42556832070398437, "grad_norm": 3.383955478668213, "learning_rate": 1.2857355772153637e-05, "loss": 0.747, "step": 1741 }, { "epoch": 0.4258127597164508, "grad_norm": 3.5648624897003174, "learning_rate": 1.2849767642159079e-05, "loss": 0.9059, "step": 1742 }, { "epoch": 0.4260571987289171, "grad_norm": 3.2955615520477295, "learning_rate": 1.2842177725817433e-05, "loss": 0.8256, "step": 1743 }, { "epoch": 0.4263016377413835, "grad_norm": 3.646454334259033, "learning_rate": 1.2834586027886355e-05, "loss": 0.8469, "step": 1744 }, { "epoch": 0.4265460767538499, "grad_norm": 2.9300601482391357, "learning_rate": 1.2826992553124628e-05, "loss": 2.0021, "step": 1745 }, { "epoch": 0.4267905157663163, "grad_norm": 2.608389139175415, "learning_rate": 1.2819397306292136e-05, "loss": 2.051, "step": 1746 }, { "epoch": 0.42703495477878267, "grad_norm": 3.829514265060425, "learning_rate": 1.2811800292149881e-05, "loss": 0.786, "step": 1747 }, { "epoch": 0.4272793937912491, "grad_norm": 3.398343563079834, "learning_rate": 1.2804201515459975e-05, "loss": 0.9001, "step": 1748 }, { "epoch": 0.4275238328037155, "grad_norm": 3.9880435466766357, "learning_rate": 1.279660098098563e-05, "loss": 0.8368, "step": 1749 }, { "epoch": 0.4277682718161819, "grad_norm": 3.6510822772979736, "learning_rate": 1.2788998693491163e-05, "loss": 0.7814, "step": 1750 }, { "epoch": 0.4280127108286483, "grad_norm": 5.199392795562744, "learning_rate": 1.2781394657741988e-05, "loss": 0.7836, "step": 1751 }, { "epoch": 0.4282571498411146, "grad_norm": 3.309724807739258, "learning_rate": 1.2773788878504614e-05, "loss": 0.8544, "step": 1752 }, { "epoch": 0.42850158885358103, "grad_norm": 4.230286121368408, "learning_rate": 1.2766181360546646e-05, "loss": 0.8382, "step": 1753 }, { "epoch": 0.42874602786604743, "grad_norm": 3.4013397693634033, "learning_rate": 1.2758572108636775e-05, "loss": 0.7617, "step": 1754 }, { "epoch": 0.42899046687851383, "grad_norm": 3.3231332302093506, "learning_rate": 1.2750961127544782e-05, "loss": 0.8446, "step": 1755 }, { "epoch": 0.4292349058909802, "grad_norm": 3.5105698108673096, "learning_rate": 1.2743348422041532e-05, "loss": 0.7687, "step": 1756 }, { "epoch": 0.4294793449034466, "grad_norm": 3.3357815742492676, "learning_rate": 1.2735733996898972e-05, "loss": 0.8778, "step": 1757 }, { "epoch": 0.429723783915913, "grad_norm": 3.030113935470581, "learning_rate": 1.2728117856890121e-05, "loss": 2.0554, "step": 1758 }, { "epoch": 0.4299682229283794, "grad_norm": 4.042571067810059, "learning_rate": 1.2720500006789079e-05, "loss": 0.7558, "step": 1759 }, { "epoch": 0.4302126619408458, "grad_norm": 4.2148237228393555, "learning_rate": 1.2712880451371015e-05, "loss": 0.8378, "step": 1760 }, { "epoch": 0.43045710095331213, "grad_norm": 3.0944976806640625, "learning_rate": 1.2705259195412168e-05, "loss": 0.9255, "step": 1761 }, { "epoch": 0.43070153996577853, "grad_norm": 3.7010090351104736, "learning_rate": 1.269763624368984e-05, "loss": 0.8578, "step": 1762 }, { "epoch": 0.43094597897824494, "grad_norm": 3.246903419494629, "learning_rate": 1.2690011600982401e-05, "loss": 0.8243, "step": 1763 }, { "epoch": 0.43119041799071134, "grad_norm": 3.145646333694458, "learning_rate": 1.2682385272069281e-05, "loss": 0.8352, "step": 1764 }, { "epoch": 0.4314348570031777, "grad_norm": 4.118160247802734, "learning_rate": 1.2674757261730963e-05, "loss": 0.7928, "step": 1765 }, { "epoch": 0.4316792960156441, "grad_norm": 4.461591720581055, "learning_rate": 1.2667127574748985e-05, "loss": 0.8024, "step": 1766 }, { "epoch": 0.4319237350281105, "grad_norm": 3.4589874744415283, "learning_rate": 1.2659496215905937e-05, "loss": 0.7367, "step": 1767 }, { "epoch": 0.4321681740405769, "grad_norm": 3.478851556777954, "learning_rate": 1.2651863189985455e-05, "loss": 0.8295, "step": 1768 }, { "epoch": 0.4324126130530433, "grad_norm": 2.5998263359069824, "learning_rate": 1.2644228501772225e-05, "loss": 1.8908, "step": 1769 }, { "epoch": 0.43265705206550964, "grad_norm": 3.7075815200805664, "learning_rate": 1.2636592156051965e-05, "loss": 0.8881, "step": 1770 }, { "epoch": 0.43290149107797604, "grad_norm": 3.9388482570648193, "learning_rate": 1.2628954157611449e-05, "loss": 0.8284, "step": 1771 }, { "epoch": 0.43314593009044244, "grad_norm": 6.109549045562744, "learning_rate": 1.2621314511238469e-05, "loss": 0.9496, "step": 1772 }, { "epoch": 0.43339036910290885, "grad_norm": 2.9666237831115723, "learning_rate": 1.2613673221721859e-05, "loss": 0.7682, "step": 1773 }, { "epoch": 0.4336348081153752, "grad_norm": 3.7829387187957764, "learning_rate": 1.260603029385148e-05, "loss": 0.9357, "step": 1774 }, { "epoch": 0.4338792471278416, "grad_norm": 3.2909271717071533, "learning_rate": 1.2598385732418226e-05, "loss": 0.8327, "step": 1775 }, { "epoch": 0.434123686140308, "grad_norm": 3.5076942443847656, "learning_rate": 1.2590739542214007e-05, "loss": 0.82, "step": 1776 }, { "epoch": 0.4343681251527744, "grad_norm": 3.365180015563965, "learning_rate": 1.2583091728031757e-05, "loss": 0.7442, "step": 1777 }, { "epoch": 0.4346125641652408, "grad_norm": 14.855374336242676, "learning_rate": 1.257544229466543e-05, "loss": 0.7882, "step": 1778 }, { "epoch": 0.43485700317770715, "grad_norm": 3.5813956260681152, "learning_rate": 1.2567791246909995e-05, "loss": 1.0035, "step": 1779 }, { "epoch": 0.43510144219017355, "grad_norm": 3.7195217609405518, "learning_rate": 1.2560138589561427e-05, "loss": 0.7808, "step": 1780 }, { "epoch": 0.43534588120263995, "grad_norm": 3.236727714538574, "learning_rate": 1.255248432741672e-05, "loss": 0.8204, "step": 1781 }, { "epoch": 0.43559032021510635, "grad_norm": 3.5511107444763184, "learning_rate": 1.2544828465273864e-05, "loss": 0.8246, "step": 1782 }, { "epoch": 0.4358347592275727, "grad_norm": 3.321305990219116, "learning_rate": 1.2537171007931859e-05, "loss": 0.7877, "step": 1783 }, { "epoch": 0.4360791982400391, "grad_norm": 3.3074100017547607, "learning_rate": 1.2529511960190699e-05, "loss": 0.8372, "step": 1784 }, { "epoch": 0.4363236372525055, "grad_norm": 3.4279422760009766, "learning_rate": 1.252185132685138e-05, "loss": 0.8411, "step": 1785 }, { "epoch": 0.4365680762649719, "grad_norm": 3.2827301025390625, "learning_rate": 1.2514189112715888e-05, "loss": 0.7474, "step": 1786 }, { "epoch": 0.4368125152774383, "grad_norm": 3.407957077026367, "learning_rate": 1.2506525322587207e-05, "loss": 0.7737, "step": 1787 }, { "epoch": 0.43705695428990465, "grad_norm": 4.449306011199951, "learning_rate": 1.2498859961269299e-05, "loss": 0.802, "step": 1788 }, { "epoch": 0.43730139330237106, "grad_norm": 3.125731945037842, "learning_rate": 1.2491193033567117e-05, "loss": 0.8348, "step": 1789 }, { "epoch": 0.43754583231483746, "grad_norm": 4.144693374633789, "learning_rate": 1.2483524544286596e-05, "loss": 0.7007, "step": 1790 }, { "epoch": 0.43779027132730386, "grad_norm": 4.416677951812744, "learning_rate": 1.2475854498234647e-05, "loss": 0.8256, "step": 1791 }, { "epoch": 0.4380347103397702, "grad_norm": 4.807706356048584, "learning_rate": 1.2468182900219158e-05, "loss": 0.7557, "step": 1792 }, { "epoch": 0.4382791493522366, "grad_norm": 3.2665696144104004, "learning_rate": 1.2460509755048991e-05, "loss": 0.6903, "step": 1793 }, { "epoch": 0.438523588364703, "grad_norm": 4.200536251068115, "learning_rate": 1.2452835067533975e-05, "loss": 0.785, "step": 1794 }, { "epoch": 0.4387680273771694, "grad_norm": 3.9921391010284424, "learning_rate": 1.2445158842484913e-05, "loss": 0.9047, "step": 1795 }, { "epoch": 0.43901246638963576, "grad_norm": 5.3824849128723145, "learning_rate": 1.2437481084713558e-05, "loss": 0.8092, "step": 1796 }, { "epoch": 0.43925690540210216, "grad_norm": 4.081234455108643, "learning_rate": 1.242980179903264e-05, "loss": 0.858, "step": 1797 }, { "epoch": 0.43950134441456856, "grad_norm": 4.144099712371826, "learning_rate": 1.2422120990255836e-05, "loss": 0.8211, "step": 1798 }, { "epoch": 0.43974578342703496, "grad_norm": 2.637526273727417, "learning_rate": 1.2414438663197782e-05, "loss": 1.9785, "step": 1799 }, { "epoch": 0.43999022243950137, "grad_norm": 3.5604233741760254, "learning_rate": 1.2406754822674062e-05, "loss": 0.7446, "step": 1800 }, { "epoch": 0.4402346614519677, "grad_norm": 3.7411844730377197, "learning_rate": 1.239906947350121e-05, "loss": 1.0626, "step": 1801 }, { "epoch": 0.4404791004644341, "grad_norm": 3.6301281452178955, "learning_rate": 1.2391382620496715e-05, "loss": 0.8541, "step": 1802 }, { "epoch": 0.4407235394769005, "grad_norm": 3.982541799545288, "learning_rate": 1.2383694268478992e-05, "loss": 0.8327, "step": 1803 }, { "epoch": 0.4409679784893669, "grad_norm": 4.221193790435791, "learning_rate": 1.2376004422267407e-05, "loss": 0.9032, "step": 1804 }, { "epoch": 0.44121241750183327, "grad_norm": 3.3037431240081787, "learning_rate": 1.2368313086682262e-05, "loss": 0.8737, "step": 1805 }, { "epoch": 0.44145685651429967, "grad_norm": 3.748661756515503, "learning_rate": 1.2360620266544787e-05, "loss": 0.9294, "step": 1806 }, { "epoch": 0.44170129552676607, "grad_norm": 3.5204269886016846, "learning_rate": 1.2352925966677147e-05, "loss": 0.772, "step": 1807 }, { "epoch": 0.44194573453923247, "grad_norm": 3.786822557449341, "learning_rate": 1.2345230191902435e-05, "loss": 0.7596, "step": 1808 }, { "epoch": 0.4421901735516989, "grad_norm": 4.062258720397949, "learning_rate": 1.2337532947044664e-05, "loss": 0.853, "step": 1809 }, { "epoch": 0.4424346125641652, "grad_norm": 3.43574857711792, "learning_rate": 1.2329834236928774e-05, "loss": 0.6149, "step": 1810 }, { "epoch": 0.4426790515766316, "grad_norm": 5.103590965270996, "learning_rate": 1.2322134066380622e-05, "loss": 0.746, "step": 1811 }, { "epoch": 0.442923490589098, "grad_norm": 4.332845211029053, "learning_rate": 1.231443244022698e-05, "loss": 0.75, "step": 1812 }, { "epoch": 0.4431679296015644, "grad_norm": 3.292994737625122, "learning_rate": 1.2306729363295529e-05, "loss": 0.9762, "step": 1813 }, { "epoch": 0.44341236861403077, "grad_norm": 3.4701523780822754, "learning_rate": 1.2299024840414861e-05, "loss": 0.7983, "step": 1814 }, { "epoch": 0.4436568076264972, "grad_norm": 4.202922821044922, "learning_rate": 1.2291318876414477e-05, "loss": 0.8687, "step": 1815 }, { "epoch": 0.4439012466389636, "grad_norm": 3.164367437362671, "learning_rate": 1.2283611476124784e-05, "loss": 0.8103, "step": 1816 }, { "epoch": 0.44414568565143, "grad_norm": 3.6044301986694336, "learning_rate": 1.2275902644377082e-05, "loss": 0.8641, "step": 1817 }, { "epoch": 0.4443901246638964, "grad_norm": 3.4861881732940674, "learning_rate": 1.2268192386003572e-05, "loss": 0.9032, "step": 1818 }, { "epoch": 0.4446345636763627, "grad_norm": 3.0395853519439697, "learning_rate": 1.226048070583735e-05, "loss": 0.73, "step": 1819 }, { "epoch": 0.44487900268882913, "grad_norm": 3.807711124420166, "learning_rate": 1.22527676087124e-05, "loss": 0.8286, "step": 1820 }, { "epoch": 0.44512344170129553, "grad_norm": 3.1775054931640625, "learning_rate": 1.22450530994636e-05, "loss": 0.773, "step": 1821 }, { "epoch": 0.44536788071376193, "grad_norm": 11.516867637634277, "learning_rate": 1.2237337182926706e-05, "loss": 0.8356, "step": 1822 }, { "epoch": 0.4456123197262283, "grad_norm": 3.5438990592956543, "learning_rate": 1.222961986393836e-05, "loss": 0.9018, "step": 1823 }, { "epoch": 0.4458567587386947, "grad_norm": 3.439138412475586, "learning_rate": 1.2221901147336086e-05, "loss": 0.759, "step": 1824 }, { "epoch": 0.4461011977511611, "grad_norm": 3.017598867416382, "learning_rate": 1.2214181037958274e-05, "loss": 0.8366, "step": 1825 }, { "epoch": 0.4463456367636275, "grad_norm": 3.530103921890259, "learning_rate": 1.2206459540644205e-05, "loss": 0.7565, "step": 1826 }, { "epoch": 0.4465900757760939, "grad_norm": 3.458371639251709, "learning_rate": 1.2198736660234009e-05, "loss": 0.766, "step": 1827 }, { "epoch": 0.44683451478856023, "grad_norm": 3.8791873455047607, "learning_rate": 1.2191012401568698e-05, "loss": 0.8749, "step": 1828 }, { "epoch": 0.44707895380102664, "grad_norm": 3.9143688678741455, "learning_rate": 1.2183286769490143e-05, "loss": 0.833, "step": 1829 }, { "epoch": 0.44732339281349304, "grad_norm": 3.308911085128784, "learning_rate": 1.2175559768841071e-05, "loss": 0.7712, "step": 1830 }, { "epoch": 0.44756783182595944, "grad_norm": 4.858282089233398, "learning_rate": 1.2167831404465078e-05, "loss": 0.9879, "step": 1831 }, { "epoch": 0.4478122708384258, "grad_norm": 3.2932534217834473, "learning_rate": 1.2160101681206602e-05, "loss": 0.9242, "step": 1832 }, { "epoch": 0.4480567098508922, "grad_norm": 3.069490671157837, "learning_rate": 1.2152370603910946e-05, "loss": 0.817, "step": 1833 }, { "epoch": 0.4483011488633586, "grad_norm": 3.297757863998413, "learning_rate": 1.214463817742425e-05, "loss": 0.7234, "step": 1834 }, { "epoch": 0.448545587875825, "grad_norm": 3.113157272338867, "learning_rate": 1.2136904406593507e-05, "loss": 0.8785, "step": 1835 }, { "epoch": 0.4487900268882914, "grad_norm": 2.865623950958252, "learning_rate": 1.2129169296266552e-05, "loss": 2.0621, "step": 1836 }, { "epoch": 0.44903446590075774, "grad_norm": 3.2112600803375244, "learning_rate": 1.2121432851292057e-05, "loss": 0.7704, "step": 1837 }, { "epoch": 0.44927890491322414, "grad_norm": 2.9245505332946777, "learning_rate": 1.2113695076519529e-05, "loss": 0.848, "step": 1838 }, { "epoch": 0.44952334392569054, "grad_norm": 3.3004794120788574, "learning_rate": 1.2105955976799316e-05, "loss": 0.7734, "step": 1839 }, { "epoch": 0.44976778293815695, "grad_norm": 3.367562770843506, "learning_rate": 1.2098215556982583e-05, "loss": 0.8719, "step": 1840 }, { "epoch": 0.4500122219506233, "grad_norm": 2.2853074073791504, "learning_rate": 1.2090473821921343e-05, "loss": 2.0348, "step": 1841 }, { "epoch": 0.4502566609630897, "grad_norm": 4.120794773101807, "learning_rate": 1.2082730776468414e-05, "loss": 0.8574, "step": 1842 }, { "epoch": 0.4505010999755561, "grad_norm": 2.4619553089141846, "learning_rate": 1.2074986425477447e-05, "loss": 2.0985, "step": 1843 }, { "epoch": 0.4507455389880225, "grad_norm": 3.2559115886688232, "learning_rate": 1.2067240773802907e-05, "loss": 0.7199, "step": 1844 }, { "epoch": 0.4509899780004889, "grad_norm": 3.8532941341400146, "learning_rate": 1.205949382630007e-05, "loss": 0.8172, "step": 1845 }, { "epoch": 0.45123441701295525, "grad_norm": 3.294890880584717, "learning_rate": 1.2051745587825036e-05, "loss": 0.8261, "step": 1846 }, { "epoch": 0.45147885602542165, "grad_norm": 2.884343147277832, "learning_rate": 1.2043996063234707e-05, "loss": 0.8833, "step": 1847 }, { "epoch": 0.45172329503788805, "grad_norm": 3.017421245574951, "learning_rate": 1.2036245257386783e-05, "loss": 0.7939, "step": 1848 }, { "epoch": 0.45196773405035445, "grad_norm": 3.6880812644958496, "learning_rate": 1.2028493175139784e-05, "loss": 0.7836, "step": 1849 }, { "epoch": 0.4522121730628208, "grad_norm": 3.3114538192749023, "learning_rate": 1.2020739821353022e-05, "loss": 0.815, "step": 1850 }, { "epoch": 0.4524566120752872, "grad_norm": 3.5703442096710205, "learning_rate": 1.2012985200886602e-05, "loss": 0.8905, "step": 1851 }, { "epoch": 0.4527010510877536, "grad_norm": 2.905712604522705, "learning_rate": 1.2005229318601429e-05, "loss": 0.8272, "step": 1852 }, { "epoch": 0.45294549010022, "grad_norm": 2.9451382160186768, "learning_rate": 1.1997472179359196e-05, "loss": 0.8074, "step": 1853 }, { "epoch": 0.4531899291126864, "grad_norm": 3.0270562171936035, "learning_rate": 1.1989713788022383e-05, "loss": 0.7081, "step": 1854 }, { "epoch": 0.45343436812515275, "grad_norm": 3.377389907836914, "learning_rate": 1.1981954149454263e-05, "loss": 0.7165, "step": 1855 }, { "epoch": 0.45367880713761916, "grad_norm": 3.476201295852661, "learning_rate": 1.1974193268518874e-05, "loss": 0.8254, "step": 1856 }, { "epoch": 0.45392324615008556, "grad_norm": 3.7356038093566895, "learning_rate": 1.1966431150081053e-05, "loss": 0.7156, "step": 1857 }, { "epoch": 0.45416768516255196, "grad_norm": 2.9921445846557617, "learning_rate": 1.1958667799006395e-05, "loss": 0.8301, "step": 1858 }, { "epoch": 0.4544121241750183, "grad_norm": 3.6033852100372314, "learning_rate": 1.1950903220161286e-05, "loss": 0.7017, "step": 1859 }, { "epoch": 0.4546565631874847, "grad_norm": 3.0926873683929443, "learning_rate": 1.194313741841286e-05, "loss": 0.8492, "step": 1860 }, { "epoch": 0.4549010021999511, "grad_norm": 4.124162197113037, "learning_rate": 1.1935370398629033e-05, "loss": 0.8546, "step": 1861 }, { "epoch": 0.4551454412124175, "grad_norm": 3.167569398880005, "learning_rate": 1.1927602165678485e-05, "loss": 0.7296, "step": 1862 }, { "epoch": 0.4553898802248839, "grad_norm": 3.3567657470703125, "learning_rate": 1.1919832724430642e-05, "loss": 0.8011, "step": 1863 }, { "epoch": 0.45563431923735026, "grad_norm": 4.348156452178955, "learning_rate": 1.1912062079755704e-05, "loss": 0.797, "step": 1864 }, { "epoch": 0.45587875824981666, "grad_norm": 3.06976056098938, "learning_rate": 1.1904290236524618e-05, "loss": 0.8518, "step": 1865 }, { "epoch": 0.45612319726228306, "grad_norm": 3.9246575832366943, "learning_rate": 1.1896517199609083e-05, "loss": 0.7095, "step": 1866 }, { "epoch": 0.45636763627474947, "grad_norm": 3.9467039108276367, "learning_rate": 1.1888742973881544e-05, "loss": 0.8455, "step": 1867 }, { "epoch": 0.4566120752872158, "grad_norm": 3.6636672019958496, "learning_rate": 1.1880967564215194e-05, "loss": 0.865, "step": 1868 }, { "epoch": 0.4568565142996822, "grad_norm": 3.345817804336548, "learning_rate": 1.1873190975483965e-05, "loss": 0.7664, "step": 1869 }, { "epoch": 0.4571009533121486, "grad_norm": 3.5963964462280273, "learning_rate": 1.1865413212562537e-05, "loss": 0.7425, "step": 1870 }, { "epoch": 0.457345392324615, "grad_norm": 3.2574336528778076, "learning_rate": 1.185763428032631e-05, "loss": 0.78, "step": 1871 }, { "epoch": 0.4575898313370814, "grad_norm": 3.143759250640869, "learning_rate": 1.1849854183651435e-05, "loss": 2.088, "step": 1872 }, { "epoch": 0.45783427034954777, "grad_norm": 3.23978590965271, "learning_rate": 1.1842072927414781e-05, "loss": 0.763, "step": 1873 }, { "epoch": 0.45807870936201417, "grad_norm": 3.4633617401123047, "learning_rate": 1.183429051649395e-05, "loss": 0.8203, "step": 1874 }, { "epoch": 0.45832314837448057, "grad_norm": 3.5833747386932373, "learning_rate": 1.1826506955767259e-05, "loss": 0.5899, "step": 1875 }, { "epoch": 0.458567587386947, "grad_norm": 2.921069860458374, "learning_rate": 1.1818722250113755e-05, "loss": 0.6378, "step": 1876 }, { "epoch": 0.4588120263994133, "grad_norm": 3.160431385040283, "learning_rate": 1.18109364044132e-05, "loss": 0.7744, "step": 1877 }, { "epoch": 0.4590564654118797, "grad_norm": 2.5235769748687744, "learning_rate": 1.180314942354607e-05, "loss": 2.0303, "step": 1878 }, { "epoch": 0.4593009044243461, "grad_norm": 3.5756592750549316, "learning_rate": 1.179536131239355e-05, "loss": 0.8258, "step": 1879 }, { "epoch": 0.4595453434368125, "grad_norm": 4.374162673950195, "learning_rate": 1.1787572075837539e-05, "loss": 0.8376, "step": 1880 }, { "epoch": 0.4597897824492789, "grad_norm": 4.300260066986084, "learning_rate": 1.1779781718760641e-05, "loss": 0.867, "step": 1881 }, { "epoch": 0.4600342214617453, "grad_norm": 3.3397669792175293, "learning_rate": 1.1771990246046154e-05, "loss": 0.7503, "step": 1882 }, { "epoch": 0.4602786604742117, "grad_norm": 3.846641778945923, "learning_rate": 1.1764197662578087e-05, "loss": 0.7351, "step": 1883 }, { "epoch": 0.4605230994866781, "grad_norm": 4.7137980461120605, "learning_rate": 1.1756403973241136e-05, "loss": 0.7328, "step": 1884 }, { "epoch": 0.4607675384991445, "grad_norm": 3.4643044471740723, "learning_rate": 1.1748609182920694e-05, "loss": 0.7494, "step": 1885 }, { "epoch": 0.4610119775116108, "grad_norm": 4.772416591644287, "learning_rate": 1.1740813296502845e-05, "loss": 0.8827, "step": 1886 }, { "epoch": 0.46125641652407723, "grad_norm": 3.249156951904297, "learning_rate": 1.1733016318874357e-05, "loss": 0.9389, "step": 1887 }, { "epoch": 0.46150085553654363, "grad_norm": 3.579883575439453, "learning_rate": 1.1725218254922685e-05, "loss": 0.7083, "step": 1888 }, { "epoch": 0.46174529454901003, "grad_norm": 3.744875192642212, "learning_rate": 1.1717419109535967e-05, "loss": 0.7988, "step": 1889 }, { "epoch": 0.46198973356147643, "grad_norm": 3.765515089035034, "learning_rate": 1.1709618887603013e-05, "loss": 0.8993, "step": 1890 }, { "epoch": 0.4622341725739428, "grad_norm": 4.36591100692749, "learning_rate": 1.1701817594013312e-05, "loss": 0.6889, "step": 1891 }, { "epoch": 0.4624786115864092, "grad_norm": 2.7193212509155273, "learning_rate": 1.1694015233657023e-05, "loss": 0.8109, "step": 1892 }, { "epoch": 0.4627230505988756, "grad_norm": 4.089737415313721, "learning_rate": 1.1686211811424971e-05, "loss": 0.8103, "step": 1893 }, { "epoch": 0.462967489611342, "grad_norm": 4.676167964935303, "learning_rate": 1.1678407332208652e-05, "loss": 0.7724, "step": 1894 }, { "epoch": 0.46321192862380833, "grad_norm": 2.849379062652588, "learning_rate": 1.1670601800900225e-05, "loss": 0.8262, "step": 1895 }, { "epoch": 0.46345636763627474, "grad_norm": 3.5339317321777344, "learning_rate": 1.1662795222392503e-05, "loss": 0.7615, "step": 1896 }, { "epoch": 0.46370080664874114, "grad_norm": 3.3167636394500732, "learning_rate": 1.1654987601578958e-05, "loss": 0.8335, "step": 1897 }, { "epoch": 0.46394524566120754, "grad_norm": 4.1246771812438965, "learning_rate": 1.1647178943353716e-05, "loss": 0.7546, "step": 1898 }, { "epoch": 0.46418968467367394, "grad_norm": 3.220127582550049, "learning_rate": 1.1639369252611552e-05, "loss": 0.7727, "step": 1899 }, { "epoch": 0.4644341236861403, "grad_norm": 3.6973886489868164, "learning_rate": 1.1631558534247894e-05, "loss": 0.8367, "step": 1900 }, { "epoch": 0.4646785626986067, "grad_norm": 3.421121835708618, "learning_rate": 1.1623746793158803e-05, "loss": 0.7907, "step": 1901 }, { "epoch": 0.4649230017110731, "grad_norm": 5.688394546508789, "learning_rate": 1.161593403424099e-05, "loss": 0.7756, "step": 1902 }, { "epoch": 0.4651674407235395, "grad_norm": 3.4840052127838135, "learning_rate": 1.1608120262391803e-05, "loss": 0.7949, "step": 1903 }, { "epoch": 0.46541187973600584, "grad_norm": 3.7205164432525635, "learning_rate": 1.1600305482509215e-05, "loss": 0.8698, "step": 1904 }, { "epoch": 0.46565631874847224, "grad_norm": 3.2122421264648438, "learning_rate": 1.159248969949185e-05, "loss": 0.9021, "step": 1905 }, { "epoch": 0.46590075776093864, "grad_norm": 3.5042636394500732, "learning_rate": 1.1584672918238948e-05, "loss": 0.9091, "step": 1906 }, { "epoch": 0.46614519677340505, "grad_norm": 3.6216835975646973, "learning_rate": 1.1576855143650372e-05, "loss": 0.7081, "step": 1907 }, { "epoch": 0.46638963578587145, "grad_norm": 3.4164111614227295, "learning_rate": 1.1569036380626615e-05, "loss": 0.869, "step": 1908 }, { "epoch": 0.4666340747983378, "grad_norm": 3.370806932449341, "learning_rate": 1.1561216634068785e-05, "loss": 0.7781, "step": 1909 }, { "epoch": 0.4668785138108042, "grad_norm": 3.592364549636841, "learning_rate": 1.155339590887861e-05, "loss": 0.7527, "step": 1910 }, { "epoch": 0.4671229528232706, "grad_norm": 3.2882182598114014, "learning_rate": 1.1545574209958433e-05, "loss": 0.8143, "step": 1911 }, { "epoch": 0.467367391835737, "grad_norm": 2.8944430351257324, "learning_rate": 1.15377515422112e-05, "loss": 0.7112, "step": 1912 }, { "epoch": 0.46761183084820335, "grad_norm": 3.555100202560425, "learning_rate": 1.152992791054047e-05, "loss": 0.6917, "step": 1913 }, { "epoch": 0.46785626986066975, "grad_norm": 2.828277349472046, "learning_rate": 1.1522103319850406e-05, "loss": 0.7013, "step": 1914 }, { "epoch": 0.46810070887313615, "grad_norm": 3.50907039642334, "learning_rate": 1.1514277775045768e-05, "loss": 0.7952, "step": 1915 }, { "epoch": 0.46834514788560255, "grad_norm": 6.130204677581787, "learning_rate": 1.1506451281031918e-05, "loss": 0.804, "step": 1916 }, { "epoch": 0.46858958689806895, "grad_norm": 3.4739468097686768, "learning_rate": 1.1498623842714816e-05, "loss": 0.8219, "step": 1917 }, { "epoch": 0.4688340259105353, "grad_norm": 4.35533332824707, "learning_rate": 1.1490795465001001e-05, "loss": 0.73, "step": 1918 }, { "epoch": 0.4690784649230017, "grad_norm": 3.3043723106384277, "learning_rate": 1.148296615279762e-05, "loss": 0.8029, "step": 1919 }, { "epoch": 0.4693229039354681, "grad_norm": 3.4134469032287598, "learning_rate": 1.1475135911012392e-05, "loss": 0.787, "step": 1920 }, { "epoch": 0.4695673429479345, "grad_norm": 3.1891961097717285, "learning_rate": 1.1467304744553618e-05, "loss": 0.6445, "step": 1921 }, { "epoch": 0.46981178196040085, "grad_norm": 3.495227098464966, "learning_rate": 1.1459472658330188e-05, "loss": 0.7557, "step": 1922 }, { "epoch": 0.47005622097286726, "grad_norm": 2.7130374908447266, "learning_rate": 1.1451639657251564e-05, "loss": 2.031, "step": 1923 }, { "epoch": 0.47030065998533366, "grad_norm": 3.4243030548095703, "learning_rate": 1.1443805746227776e-05, "loss": 0.8233, "step": 1924 }, { "epoch": 0.47054509899780006, "grad_norm": 3.749325752258301, "learning_rate": 1.143597093016943e-05, "loss": 0.7684, "step": 1925 }, { "epoch": 0.47078953801026646, "grad_norm": 3.5366694927215576, "learning_rate": 1.1428135213987706e-05, "loss": 0.8376, "step": 1926 }, { "epoch": 0.4710339770227328, "grad_norm": 3.4410207271575928, "learning_rate": 1.1420298602594334e-05, "loss": 0.8641, "step": 1927 }, { "epoch": 0.4712784160351992, "grad_norm": 3.6780996322631836, "learning_rate": 1.1412461100901614e-05, "loss": 0.7465, "step": 1928 }, { "epoch": 0.4715228550476656, "grad_norm": 3.1814169883728027, "learning_rate": 1.14046227138224e-05, "loss": 0.7965, "step": 1929 }, { "epoch": 0.471767294060132, "grad_norm": 3.3467376232147217, "learning_rate": 1.1396783446270105e-05, "loss": 0.7431, "step": 1930 }, { "epoch": 0.47201173307259836, "grad_norm": 2.4791579246520996, "learning_rate": 1.1388943303158692e-05, "loss": 2.0101, "step": 1931 }, { "epoch": 0.47225617208506476, "grad_norm": 3.1137938499450684, "learning_rate": 1.1381102289402675e-05, "loss": 0.9033, "step": 1932 }, { "epoch": 0.47250061109753116, "grad_norm": 3.520462989807129, "learning_rate": 1.1373260409917104e-05, "loss": 0.8201, "step": 1933 }, { "epoch": 0.47274505010999757, "grad_norm": 2.4430835247039795, "learning_rate": 1.136541766961759e-05, "loss": 2.0582, "step": 1934 }, { "epoch": 0.47298948912246397, "grad_norm": 3.2520508766174316, "learning_rate": 1.1357574073420265e-05, "loss": 0.7676, "step": 1935 }, { "epoch": 0.4732339281349303, "grad_norm": 3.3971803188323975, "learning_rate": 1.1349729626241807e-05, "loss": 0.8358, "step": 1936 }, { "epoch": 0.4734783671473967, "grad_norm": 3.3537838459014893, "learning_rate": 1.1341884332999431e-05, "loss": 0.923, "step": 1937 }, { "epoch": 0.4737228061598631, "grad_norm": 3.855825424194336, "learning_rate": 1.1334038198610868e-05, "loss": 0.8022, "step": 1938 }, { "epoch": 0.4739672451723295, "grad_norm": 3.2089316844940186, "learning_rate": 1.1326191227994392e-05, "loss": 0.7227, "step": 1939 }, { "epoch": 0.47421168418479587, "grad_norm": 3.6445322036743164, "learning_rate": 1.1318343426068795e-05, "loss": 0.8884, "step": 1940 }, { "epoch": 0.47445612319726227, "grad_norm": 3.0271670818328857, "learning_rate": 1.1310494797753382e-05, "loss": 0.618, "step": 1941 }, { "epoch": 0.47470056220972867, "grad_norm": 4.566411018371582, "learning_rate": 1.1302645347967993e-05, "loss": 0.776, "step": 1942 }, { "epoch": 0.4749450012221951, "grad_norm": 3.470848321914673, "learning_rate": 1.1294795081632968e-05, "loss": 0.8234, "step": 1943 }, { "epoch": 0.4751894402346615, "grad_norm": 3.8001720905303955, "learning_rate": 1.128694400366917e-05, "loss": 0.7269, "step": 1944 }, { "epoch": 0.4754338792471278, "grad_norm": 3.591465950012207, "learning_rate": 1.1279092118997958e-05, "loss": 0.7565, "step": 1945 }, { "epoch": 0.4756783182595942, "grad_norm": 3.489041328430176, "learning_rate": 1.1271239432541208e-05, "loss": 0.7692, "step": 1946 }, { "epoch": 0.4759227572720606, "grad_norm": 2.8316056728363037, "learning_rate": 1.1263385949221294e-05, "loss": 0.8811, "step": 1947 }, { "epoch": 0.47616719628452703, "grad_norm": 3.651979684829712, "learning_rate": 1.1255531673961091e-05, "loss": 0.7406, "step": 1948 }, { "epoch": 0.4764116352969934, "grad_norm": 3.323721170425415, "learning_rate": 1.1247676611683964e-05, "loss": 0.7923, "step": 1949 }, { "epoch": 0.4766560743094598, "grad_norm": 4.642440319061279, "learning_rate": 1.1239820767313788e-05, "loss": 0.7638, "step": 1950 }, { "epoch": 0.4769005133219262, "grad_norm": 3.692936897277832, "learning_rate": 1.1231964145774906e-05, "loss": 0.7265, "step": 1951 }, { "epoch": 0.4771449523343926, "grad_norm": 4.496062755584717, "learning_rate": 1.1224106751992164e-05, "loss": 0.7691, "step": 1952 }, { "epoch": 0.477389391346859, "grad_norm": 3.4846792221069336, "learning_rate": 1.1216248590890886e-05, "loss": 0.7858, "step": 1953 }, { "epoch": 0.47763383035932533, "grad_norm": 4.358302116394043, "learning_rate": 1.1208389667396879e-05, "loss": 0.8424, "step": 1954 }, { "epoch": 0.47787826937179173, "grad_norm": 3.3236751556396484, "learning_rate": 1.120052998643643e-05, "loss": 0.8013, "step": 1955 }, { "epoch": 0.47812270838425813, "grad_norm": 4.438330173492432, "learning_rate": 1.119266955293629e-05, "loss": 0.7891, "step": 1956 }, { "epoch": 0.47836714739672453, "grad_norm": 4.703066825866699, "learning_rate": 1.1184808371823696e-05, "loss": 0.7928, "step": 1957 }, { "epoch": 0.4786115864091909, "grad_norm": 4.193386554718018, "learning_rate": 1.1176946448026348e-05, "loss": 0.8049, "step": 1958 }, { "epoch": 0.4788560254216573, "grad_norm": 3.0352625846862793, "learning_rate": 1.1169083786472407e-05, "loss": 0.7952, "step": 1959 }, { "epoch": 0.4791004644341237, "grad_norm": 3.2521796226501465, "learning_rate": 1.1161220392090503e-05, "loss": 0.6706, "step": 1960 }, { "epoch": 0.4793449034465901, "grad_norm": 3.6830103397369385, "learning_rate": 1.1153356269809721e-05, "loss": 0.7813, "step": 1961 }, { "epoch": 0.4795893424590565, "grad_norm": 4.002415180206299, "learning_rate": 1.1145491424559607e-05, "loss": 0.7248, "step": 1962 }, { "epoch": 0.47983378147152284, "grad_norm": 3.116703987121582, "learning_rate": 1.1137625861270151e-05, "loss": 0.7807, "step": 1963 }, { "epoch": 0.48007822048398924, "grad_norm": 2.9933393001556396, "learning_rate": 1.1129759584871802e-05, "loss": 2.0964, "step": 1964 }, { "epoch": 0.48032265949645564, "grad_norm": 3.455176830291748, "learning_rate": 1.1121892600295456e-05, "loss": 0.7856, "step": 1965 }, { "epoch": 0.48056709850892204, "grad_norm": 2.866671085357666, "learning_rate": 1.1114024912472449e-05, "loss": 2.049, "step": 1966 }, { "epoch": 0.4808115375213884, "grad_norm": 3.603726625442505, "learning_rate": 1.1106156526334559e-05, "loss": 0.7538, "step": 1967 }, { "epoch": 0.4810559765338548, "grad_norm": 2.406770706176758, "learning_rate": 1.1098287446814001e-05, "loss": 1.944, "step": 1968 }, { "epoch": 0.4813004155463212, "grad_norm": 4.4387712478637695, "learning_rate": 1.1090417678843423e-05, "loss": 0.9268, "step": 1969 }, { "epoch": 0.4815448545587876, "grad_norm": 3.8196306228637695, "learning_rate": 1.1082547227355911e-05, "loss": 0.8084, "step": 1970 }, { "epoch": 0.481789293571254, "grad_norm": 3.659147024154663, "learning_rate": 1.1074676097284973e-05, "loss": 0.7973, "step": 1971 }, { "epoch": 0.48203373258372034, "grad_norm": 4.428442001342773, "learning_rate": 1.1066804293564541e-05, "loss": 0.8604, "step": 1972 }, { "epoch": 0.48227817159618674, "grad_norm": 4.519818305969238, "learning_rate": 1.1058931821128982e-05, "loss": 0.6759, "step": 1973 }, { "epoch": 0.48252261060865315, "grad_norm": 3.2999141216278076, "learning_rate": 1.1051058684913067e-05, "loss": 0.7477, "step": 1974 }, { "epoch": 0.48276704962111955, "grad_norm": 3.5467281341552734, "learning_rate": 1.1043184889851992e-05, "loss": 0.7753, "step": 1975 }, { "epoch": 0.4830114886335859, "grad_norm": 3.5430102348327637, "learning_rate": 1.1035310440881359e-05, "loss": 0.8758, "step": 1976 }, { "epoch": 0.4832559276460523, "grad_norm": 4.608526229858398, "learning_rate": 1.1027435342937189e-05, "loss": 0.7839, "step": 1977 }, { "epoch": 0.4835003666585187, "grad_norm": 3.1568756103515625, "learning_rate": 1.10195596009559e-05, "loss": 0.6923, "step": 1978 }, { "epoch": 0.4837448056709851, "grad_norm": 3.5793933868408203, "learning_rate": 1.1011683219874324e-05, "loss": 0.8686, "step": 1979 }, { "epoch": 0.4839892446834515, "grad_norm": 4.421467304229736, "learning_rate": 1.1003806204629683e-05, "loss": 0.7134, "step": 1980 }, { "epoch": 0.48423368369591785, "grad_norm": 4.978440284729004, "learning_rate": 1.0995928560159608e-05, "loss": 0.6484, "step": 1981 }, { "epoch": 0.48447812270838425, "grad_norm": 3.6223433017730713, "learning_rate": 1.0988050291402113e-05, "loss": 0.7283, "step": 1982 }, { "epoch": 0.48472256172085065, "grad_norm": 3.2677972316741943, "learning_rate": 1.098017140329561e-05, "loss": 0.7855, "step": 1983 }, { "epoch": 0.48496700073331706, "grad_norm": 3.3966641426086426, "learning_rate": 1.0972291900778892e-05, "loss": 0.8409, "step": 1984 }, { "epoch": 0.4852114397457834, "grad_norm": 2.9886679649353027, "learning_rate": 1.0964411788791156e-05, "loss": 0.7301, "step": 1985 }, { "epoch": 0.4854558787582498, "grad_norm": 3.46282958984375, "learning_rate": 1.0956531072271959e-05, "loss": 0.7322, "step": 1986 }, { "epoch": 0.4857003177707162, "grad_norm": 3.8131372928619385, "learning_rate": 1.0948649756161246e-05, "loss": 0.7173, "step": 1987 }, { "epoch": 0.4859447567831826, "grad_norm": 3.8756089210510254, "learning_rate": 1.0940767845399341e-05, "loss": 0.786, "step": 1988 }, { "epoch": 0.486189195795649, "grad_norm": 5.3649725914001465, "learning_rate": 1.0932885344926931e-05, "loss": 0.8385, "step": 1989 }, { "epoch": 0.48643363480811536, "grad_norm": 3.4769184589385986, "learning_rate": 1.0925002259685089e-05, "loss": 0.7209, "step": 1990 }, { "epoch": 0.48667807382058176, "grad_norm": 4.807114601135254, "learning_rate": 1.0917118594615237e-05, "loss": 0.7383, "step": 1991 }, { "epoch": 0.48692251283304816, "grad_norm": 2.774432897567749, "learning_rate": 1.0909234354659173e-05, "loss": 0.7314, "step": 1992 }, { "epoch": 0.48716695184551456, "grad_norm": 3.699162006378174, "learning_rate": 1.0901349544759046e-05, "loss": 0.8145, "step": 1993 }, { "epoch": 0.4874113908579809, "grad_norm": 3.1336519718170166, "learning_rate": 1.0893464169857368e-05, "loss": 0.7637, "step": 1994 }, { "epoch": 0.4876558298704473, "grad_norm": 3.9432170391082764, "learning_rate": 1.0885578234897003e-05, "loss": 0.883, "step": 1995 }, { "epoch": 0.4879002688829137, "grad_norm": 2.927171230316162, "learning_rate": 1.0877691744821171e-05, "loss": 0.719, "step": 1996 }, { "epoch": 0.4881447078953801, "grad_norm": 3.5491185188293457, "learning_rate": 1.0869804704573435e-05, "loss": 0.7115, "step": 1997 }, { "epoch": 0.4883891469078465, "grad_norm": 3.6908175945281982, "learning_rate": 1.08619171190977e-05, "loss": 0.8634, "step": 1998 }, { "epoch": 0.48863358592031286, "grad_norm": 3.3738112449645996, "learning_rate": 1.0854028993338222e-05, "loss": 0.664, "step": 1999 }, { "epoch": 0.48887802493277926, "grad_norm": 3.688183307647705, "learning_rate": 1.0846140332239584e-05, "loss": 0.8448, "step": 2000 }, { "epoch": 0.48912246394524567, "grad_norm": 3.9575772285461426, "learning_rate": 1.0838251140746717e-05, "loss": 0.8176, "step": 2001 }, { "epoch": 0.48936690295771207, "grad_norm": 3.4544336795806885, "learning_rate": 1.0830361423804872e-05, "loss": 0.7331, "step": 2002 }, { "epoch": 0.4896113419701784, "grad_norm": 2.768550157546997, "learning_rate": 1.082247118635964e-05, "loss": 0.8275, "step": 2003 }, { "epoch": 0.4898557809826448, "grad_norm": 3.2085041999816895, "learning_rate": 1.0814580433356933e-05, "loss": 0.717, "step": 2004 }, { "epoch": 0.4901002199951112, "grad_norm": 3.5018882751464844, "learning_rate": 1.0806689169742988e-05, "loss": 0.6366, "step": 2005 }, { "epoch": 0.4903446590075776, "grad_norm": 3.169766426086426, "learning_rate": 1.0798797400464361e-05, "loss": 0.7642, "step": 2006 }, { "epoch": 0.490589098020044, "grad_norm": 3.3302865028381348, "learning_rate": 1.0790905130467923e-05, "loss": 0.801, "step": 2007 }, { "epoch": 0.49083353703251037, "grad_norm": 4.264266490936279, "learning_rate": 1.0783012364700865e-05, "loss": 0.7877, "step": 2008 }, { "epoch": 0.49107797604497677, "grad_norm": 2.8109889030456543, "learning_rate": 1.077511910811068e-05, "loss": 0.7575, "step": 2009 }, { "epoch": 0.4913224150574432, "grad_norm": 3.112183094024658, "learning_rate": 1.0767225365645178e-05, "loss": 0.7362, "step": 2010 }, { "epoch": 0.4915668540699096, "grad_norm": 3.1281116008758545, "learning_rate": 1.0759331142252463e-05, "loss": 0.6488, "step": 2011 }, { "epoch": 0.4918112930823759, "grad_norm": 3.025792121887207, "learning_rate": 1.0751436442880953e-05, "loss": 0.7768, "step": 2012 }, { "epoch": 0.4920557320948423, "grad_norm": 3.433082342147827, "learning_rate": 1.0743541272479356e-05, "loss": 0.7959, "step": 2013 }, { "epoch": 0.4923001711073087, "grad_norm": 3.600435495376587, "learning_rate": 1.0735645635996676e-05, "loss": 0.7833, "step": 2014 }, { "epoch": 0.49254461011977513, "grad_norm": 3.4164845943450928, "learning_rate": 1.072774953838221e-05, "loss": 0.6478, "step": 2015 }, { "epoch": 0.49278904913224153, "grad_norm": 3.6529622077941895, "learning_rate": 1.0719852984585546e-05, "loss": 0.766, "step": 2016 }, { "epoch": 0.4930334881447079, "grad_norm": 4.316786289215088, "learning_rate": 1.0711955979556555e-05, "loss": 0.8255, "step": 2017 }, { "epoch": 0.4932779271571743, "grad_norm": 2.940098762512207, "learning_rate": 1.0704058528245389e-05, "loss": 0.7522, "step": 2018 }, { "epoch": 0.4935223661696407, "grad_norm": 4.512490272521973, "learning_rate": 1.0696160635602488e-05, "loss": 0.8231, "step": 2019 }, { "epoch": 0.4937668051821071, "grad_norm": 3.141284227371216, "learning_rate": 1.0688262306578566e-05, "loss": 0.9031, "step": 2020 }, { "epoch": 0.49401124419457343, "grad_norm": 4.085526466369629, "learning_rate": 1.0680363546124599e-05, "loss": 0.7519, "step": 2021 }, { "epoch": 0.49425568320703983, "grad_norm": 2.9810550212860107, "learning_rate": 1.067246435919185e-05, "loss": 0.6218, "step": 2022 }, { "epoch": 0.49450012221950623, "grad_norm": 2.8658111095428467, "learning_rate": 1.0664564750731838e-05, "loss": 0.6835, "step": 2023 }, { "epoch": 0.49474456123197263, "grad_norm": 3.3811020851135254, "learning_rate": 1.0656664725696352e-05, "loss": 0.822, "step": 2024 }, { "epoch": 0.49498900024443904, "grad_norm": 3.2734344005584717, "learning_rate": 1.064876428903744e-05, "loss": 0.835, "step": 2025 }, { "epoch": 0.4952334392569054, "grad_norm": 3.8637382984161377, "learning_rate": 1.0640863445707408e-05, "loss": 0.7283, "step": 2026 }, { "epoch": 0.4954778782693718, "grad_norm": 3.770451307296753, "learning_rate": 1.0632962200658816e-05, "loss": 0.7497, "step": 2027 }, { "epoch": 0.4957223172818382, "grad_norm": 3.022958755493164, "learning_rate": 1.0625060558844478e-05, "loss": 2.0316, "step": 2028 }, { "epoch": 0.4959667562943046, "grad_norm": 3.7786545753479004, "learning_rate": 1.0617158525217457e-05, "loss": 0.762, "step": 2029 }, { "epoch": 0.49621119530677094, "grad_norm": 3.3836238384246826, "learning_rate": 1.0609256104731057e-05, "loss": 0.7447, "step": 2030 }, { "epoch": 0.49645563431923734, "grad_norm": 3.8169052600860596, "learning_rate": 1.060135330233883e-05, "loss": 0.6994, "step": 2031 }, { "epoch": 0.49670007333170374, "grad_norm": 3.290356397628784, "learning_rate": 1.0593450122994567e-05, "loss": 0.6643, "step": 2032 }, { "epoch": 0.49694451234417014, "grad_norm": 2.974830150604248, "learning_rate": 1.0585546571652288e-05, "loss": 0.7194, "step": 2033 }, { "epoch": 0.49718895135663654, "grad_norm": 2.969827651977539, "learning_rate": 1.0577642653266253e-05, "loss": 0.8106, "step": 2034 }, { "epoch": 0.4974333903691029, "grad_norm": 3.1830921173095703, "learning_rate": 1.0569738372790956e-05, "loss": 0.7061, "step": 2035 }, { "epoch": 0.4976778293815693, "grad_norm": 2.936081886291504, "learning_rate": 1.0561833735181108e-05, "loss": 2.0002, "step": 2036 }, { "epoch": 0.4979222683940357, "grad_norm": 2.954249143600464, "learning_rate": 1.055392874539165e-05, "loss": 0.7313, "step": 2037 }, { "epoch": 0.4981667074065021, "grad_norm": 3.6443800926208496, "learning_rate": 1.0546023408377743e-05, "loss": 0.8397, "step": 2038 }, { "epoch": 0.49841114641896844, "grad_norm": 3.4366455078125, "learning_rate": 1.0538117729094766e-05, "loss": 0.7084, "step": 2039 }, { "epoch": 0.49865558543143484, "grad_norm": 2.2596499919891357, "learning_rate": 1.0530211712498306e-05, "loss": 1.9601, "step": 2040 }, { "epoch": 0.49890002444390125, "grad_norm": 3.0489089488983154, "learning_rate": 1.0522305363544172e-05, "loss": 0.7073, "step": 2041 }, { "epoch": 0.49914446345636765, "grad_norm": 4.950798511505127, "learning_rate": 1.0514398687188372e-05, "loss": 0.8156, "step": 2042 }, { "epoch": 0.49938890246883405, "grad_norm": 3.044152021408081, "learning_rate": 1.0506491688387128e-05, "loss": 0.6842, "step": 2043 }, { "epoch": 0.4996333414813004, "grad_norm": 3.6759657859802246, "learning_rate": 1.0498584372096858e-05, "loss": 0.7377, "step": 2044 }, { "epoch": 0.4998777804937668, "grad_norm": 2.9382901191711426, "learning_rate": 1.0490676743274181e-05, "loss": 0.749, "step": 2045 }, { "epoch": 0.5001222195062331, "grad_norm": 3.570497989654541, "learning_rate": 1.0482768806875911e-05, "loss": 0.8168, "step": 2046 }, { "epoch": 0.5003666585186995, "grad_norm": 3.9341888427734375, "learning_rate": 1.0474860567859058e-05, "loss": 0.7277, "step": 2047 }, { "epoch": 0.500611097531166, "grad_norm": 7.264518737792969, "learning_rate": 1.0466952031180813e-05, "loss": 0.6585, "step": 2048 }, { "epoch": 0.5008555365436324, "grad_norm": 3.4198923110961914, "learning_rate": 1.0459043201798563e-05, "loss": 0.8649, "step": 2049 }, { "epoch": 0.5010999755560988, "grad_norm": 3.1877851486206055, "learning_rate": 1.0451134084669876e-05, "loss": 0.6722, "step": 2050 }, { "epoch": 0.5013444145685652, "grad_norm": 3.139477014541626, "learning_rate": 1.04432246847525e-05, "loss": 0.761, "step": 2051 }, { "epoch": 0.5015888535810316, "grad_norm": 3.5756216049194336, "learning_rate": 1.0435315007004357e-05, "loss": 0.7999, "step": 2052 }, { "epoch": 0.501833292593498, "grad_norm": 2.963740587234497, "learning_rate": 1.0427405056383548e-05, "loss": 0.7494, "step": 2053 }, { "epoch": 0.5020777316059644, "grad_norm": 2.5787642002105713, "learning_rate": 1.0419494837848343e-05, "loss": 1.9864, "step": 2054 }, { "epoch": 0.5023221706184307, "grad_norm": 3.3031718730926514, "learning_rate": 1.0411584356357178e-05, "loss": 0.7257, "step": 2055 }, { "epoch": 0.502566609630897, "grad_norm": 3.0101404190063477, "learning_rate": 1.0403673616868659e-05, "loss": 0.8019, "step": 2056 }, { "epoch": 0.5028110486433635, "grad_norm": 2.8920490741729736, "learning_rate": 1.0395762624341544e-05, "loss": 0.7606, "step": 2057 }, { "epoch": 0.5030554876558299, "grad_norm": 3.88023042678833, "learning_rate": 1.0387851383734763e-05, "loss": 0.7647, "step": 2058 }, { "epoch": 0.5032999266682963, "grad_norm": 3.9406485557556152, "learning_rate": 1.0379939900007394e-05, "loss": 0.7223, "step": 2059 }, { "epoch": 0.5035443656807627, "grad_norm": 3.2017197608947754, "learning_rate": 1.0372028178118665e-05, "loss": 0.7909, "step": 2060 }, { "epoch": 0.5037888046932291, "grad_norm": 3.2319176197052, "learning_rate": 1.0364116223027956e-05, "loss": 0.6608, "step": 2061 }, { "epoch": 0.5040332437056955, "grad_norm": 3.2559704780578613, "learning_rate": 1.0356204039694795e-05, "loss": 0.7812, "step": 2062 }, { "epoch": 0.5042776827181619, "grad_norm": 3.1422622203826904, "learning_rate": 1.0348291633078851e-05, "loss": 0.7641, "step": 2063 }, { "epoch": 0.5045221217306282, "grad_norm": 3.148240804672241, "learning_rate": 1.0340379008139933e-05, "loss": 0.7373, "step": 2064 }, { "epoch": 0.5047665607430946, "grad_norm": 3.469862699508667, "learning_rate": 1.0332466169837992e-05, "loss": 0.811, "step": 2065 }, { "epoch": 0.505010999755561, "grad_norm": 3.4699959754943848, "learning_rate": 1.03245531231331e-05, "loss": 0.6805, "step": 2066 }, { "epoch": 0.5052554387680274, "grad_norm": 4.243104934692383, "learning_rate": 1.0316639872985471e-05, "loss": 0.7452, "step": 2067 }, { "epoch": 0.5054998777804938, "grad_norm": 3.567861557006836, "learning_rate": 1.0308726424355446e-05, "loss": 0.7222, "step": 2068 }, { "epoch": 0.5057443167929602, "grad_norm": 3.9209423065185547, "learning_rate": 1.0300812782203486e-05, "loss": 0.7033, "step": 2069 }, { "epoch": 0.5059887558054266, "grad_norm": 4.002355098724365, "learning_rate": 1.0292898951490174e-05, "loss": 0.6573, "step": 2070 }, { "epoch": 0.506233194817893, "grad_norm": 3.7021231651306152, "learning_rate": 1.0284984937176213e-05, "loss": 0.7672, "step": 2071 }, { "epoch": 0.5064776338303594, "grad_norm": 2.4699697494506836, "learning_rate": 1.0277070744222419e-05, "loss": 2.0226, "step": 2072 }, { "epoch": 0.5067220728428257, "grad_norm": 4.644677639007568, "learning_rate": 1.0269156377589722e-05, "loss": 0.7015, "step": 2073 }, { "epoch": 0.5069665118552921, "grad_norm": 3.0466513633728027, "learning_rate": 1.026124184223916e-05, "loss": 0.8128, "step": 2074 }, { "epoch": 0.5072109508677585, "grad_norm": 4.4984354972839355, "learning_rate": 1.025332714313188e-05, "loss": 0.7525, "step": 2075 }, { "epoch": 0.5074553898802249, "grad_norm": 4.234147071838379, "learning_rate": 1.0245412285229124e-05, "loss": 0.7366, "step": 2076 }, { "epoch": 0.5076998288926913, "grad_norm": 3.7217657566070557, "learning_rate": 1.023749727349224e-05, "loss": 0.7647, "step": 2077 }, { "epoch": 0.5079442679051577, "grad_norm": 3.2942984104156494, "learning_rate": 1.0229582112882668e-05, "loss": 0.8248, "step": 2078 }, { "epoch": 0.5081887069176241, "grad_norm": 3.738570213317871, "learning_rate": 1.0221666808361947e-05, "loss": 0.5877, "step": 2079 }, { "epoch": 0.5084331459300905, "grad_norm": 3.184098720550537, "learning_rate": 1.0213751364891702e-05, "loss": 0.6752, "step": 2080 }, { "epoch": 0.5086775849425569, "grad_norm": 3.1130599975585938, "learning_rate": 1.0205835787433645e-05, "loss": 0.6944, "step": 2081 }, { "epoch": 0.5089220239550232, "grad_norm": 2.9968113899230957, "learning_rate": 1.0197920080949574e-05, "loss": 0.8001, "step": 2082 }, { "epoch": 0.5091664629674896, "grad_norm": 3.2675774097442627, "learning_rate": 1.0190004250401369e-05, "loss": 0.7529, "step": 2083 }, { "epoch": 0.509410901979956, "grad_norm": 2.6524808406829834, "learning_rate": 1.0182088300750983e-05, "loss": 1.9075, "step": 2084 }, { "epoch": 0.5096553409924224, "grad_norm": 3.6236817836761475, "learning_rate": 1.0174172236960447e-05, "loss": 0.8717, "step": 2085 }, { "epoch": 0.5098997800048888, "grad_norm": 3.115124225616455, "learning_rate": 1.0166256063991861e-05, "loss": 0.6924, "step": 2086 }, { "epoch": 0.5101442190173552, "grad_norm": 3.547389030456543, "learning_rate": 1.0158339786807401e-05, "loss": 0.6893, "step": 2087 }, { "epoch": 0.5103886580298216, "grad_norm": 3.550128221511841, "learning_rate": 1.0150423410369294e-05, "loss": 0.6903, "step": 2088 }, { "epoch": 0.510633097042288, "grad_norm": 3.768012523651123, "learning_rate": 1.0142506939639847e-05, "loss": 0.8392, "step": 2089 }, { "epoch": 0.5108775360547544, "grad_norm": 2.932687759399414, "learning_rate": 1.013459037958141e-05, "loss": 0.7111, "step": 2090 }, { "epoch": 0.5111219750672207, "grad_norm": 3.274658441543579, "learning_rate": 1.0126673735156402e-05, "loss": 0.8295, "step": 2091 }, { "epoch": 0.5113664140796871, "grad_norm": 4.2612624168396, "learning_rate": 1.0118757011327285e-05, "loss": 0.8573, "step": 2092 }, { "epoch": 0.5116108530921535, "grad_norm": 2.732764959335327, "learning_rate": 1.0110840213056575e-05, "loss": 0.6972, "step": 2093 }, { "epoch": 0.5118552921046199, "grad_norm": 5.253904342651367, "learning_rate": 1.0102923345306834e-05, "loss": 0.7623, "step": 2094 }, { "epoch": 0.5120997311170863, "grad_norm": 2.9598002433776855, "learning_rate": 1.009500641304067e-05, "loss": 0.5901, "step": 2095 }, { "epoch": 0.5123441701295527, "grad_norm": 7.8216776847839355, "learning_rate": 1.0087089421220727e-05, "loss": 0.8268, "step": 2096 }, { "epoch": 0.5125886091420191, "grad_norm": 3.346170663833618, "learning_rate": 1.0079172374809689e-05, "loss": 0.6633, "step": 2097 }, { "epoch": 0.5128330481544855, "grad_norm": 3.266075372695923, "learning_rate": 1.0071255278770276e-05, "loss": 0.8366, "step": 2098 }, { "epoch": 0.5130774871669519, "grad_norm": 3.055297613143921, "learning_rate": 1.0063338138065235e-05, "loss": 0.6402, "step": 2099 }, { "epoch": 0.5133219261794182, "grad_norm": 5.6008830070495605, "learning_rate": 1.0055420957657344e-05, "loss": 0.7894, "step": 2100 }, { "epoch": 0.5135663651918846, "grad_norm": 2.906796455383301, "learning_rate": 1.0047503742509405e-05, "loss": 0.7169, "step": 2101 }, { "epoch": 0.513810804204351, "grad_norm": 3.6960554122924805, "learning_rate": 1.0039586497584241e-05, "loss": 0.7333, "step": 2102 }, { "epoch": 0.5140552432168174, "grad_norm": 3.835498094558716, "learning_rate": 1.0031669227844693e-05, "loss": 0.7082, "step": 2103 }, { "epoch": 0.5142996822292838, "grad_norm": 3.1887574195861816, "learning_rate": 1.0023751938253618e-05, "loss": 0.8432, "step": 2104 }, { "epoch": 0.5145441212417502, "grad_norm": 3.2100727558135986, "learning_rate": 1.0015834633773893e-05, "loss": 0.6839, "step": 2105 }, { "epoch": 0.5147885602542166, "grad_norm": 4.217765808105469, "learning_rate": 1.0007917319368392e-05, "loss": 0.7569, "step": 2106 }, { "epoch": 0.515032999266683, "grad_norm": 4.070542812347412, "learning_rate": 1e-05, "loss": 0.753, "step": 2107 }, { "epoch": 0.5152774382791494, "grad_norm": 3.809004783630371, "learning_rate": 9.99208268063161e-06, "loss": 0.7754, "step": 2108 }, { "epoch": 0.5155218772916157, "grad_norm": 2.945784568786621, "learning_rate": 9.98416536622611e-06, "loss": 0.7516, "step": 2109 }, { "epoch": 0.5157663163040821, "grad_norm": 3.0451085567474365, "learning_rate": 9.976248061746383e-06, "loss": 0.8051, "step": 2110 }, { "epoch": 0.5160107553165485, "grad_norm": 4.2851786613464355, "learning_rate": 9.968330772155312e-06, "loss": 0.8682, "step": 2111 }, { "epoch": 0.5162551943290149, "grad_norm": 3.8674421310424805, "learning_rate": 9.960413502415764e-06, "loss": 0.7483, "step": 2112 }, { "epoch": 0.5164996333414813, "grad_norm": 3.2832260131835938, "learning_rate": 9.9524962574906e-06, "loss": 0.7464, "step": 2113 }, { "epoch": 0.5167440723539477, "grad_norm": 3.5348575115203857, "learning_rate": 9.944579042342661e-06, "loss": 0.886, "step": 2114 }, { "epoch": 0.5169885113664141, "grad_norm": 3.5706822872161865, "learning_rate": 9.936661861934765e-06, "loss": 0.6823, "step": 2115 }, { "epoch": 0.5172329503788805, "grad_norm": 3.5740959644317627, "learning_rate": 9.928744721229726e-06, "loss": 0.7477, "step": 2116 }, { "epoch": 0.5174773893913469, "grad_norm": 3.688131093978882, "learning_rate": 9.920827625190311e-06, "loss": 0.8044, "step": 2117 }, { "epoch": 0.5177218284038132, "grad_norm": 2.408571481704712, "learning_rate": 9.912910578779275e-06, "loss": 1.9138, "step": 2118 }, { "epoch": 0.5179662674162796, "grad_norm": 3.5460824966430664, "learning_rate": 9.904993586959333e-06, "loss": 0.858, "step": 2119 }, { "epoch": 0.518210706428746, "grad_norm": 2.603878974914551, "learning_rate": 9.89707665469317e-06, "loss": 1.9577, "step": 2120 }, { "epoch": 0.5184551454412124, "grad_norm": 4.377145290374756, "learning_rate": 9.889159786943428e-06, "loss": 0.7671, "step": 2121 }, { "epoch": 0.5186995844536788, "grad_norm": 3.42803692817688, "learning_rate": 9.881242988672719e-06, "loss": 0.7304, "step": 2122 }, { "epoch": 0.5189440234661452, "grad_norm": 2.933485269546509, "learning_rate": 9.873326264843601e-06, "loss": 0.7568, "step": 2123 }, { "epoch": 0.5191884624786116, "grad_norm": 3.03100323677063, "learning_rate": 9.865409620418593e-06, "loss": 0.6751, "step": 2124 }, { "epoch": 0.519432901491078, "grad_norm": 3.503622055053711, "learning_rate": 9.857493060360157e-06, "loss": 0.8314, "step": 2125 }, { "epoch": 0.5196773405035444, "grad_norm": 3.3902294635772705, "learning_rate": 9.84957658963071e-06, "loss": 0.7963, "step": 2126 }, { "epoch": 0.5199217795160107, "grad_norm": 3.721219062805176, "learning_rate": 9.841660213192606e-06, "loss": 0.7359, "step": 2127 }, { "epoch": 0.5201662185284771, "grad_norm": 3.8676271438598633, "learning_rate": 9.833743936008144e-06, "loss": 0.7679, "step": 2128 }, { "epoch": 0.5204106575409435, "grad_norm": 2.977606773376465, "learning_rate": 9.825827763039558e-06, "loss": 0.7072, "step": 2129 }, { "epoch": 0.5206550965534099, "grad_norm": 4.037961006164551, "learning_rate": 9.817911699249022e-06, "loss": 0.7505, "step": 2130 }, { "epoch": 0.5208995355658763, "grad_norm": 3.039395332336426, "learning_rate": 9.809995749598633e-06, "loss": 0.6771, "step": 2131 }, { "epoch": 0.5211439745783427, "grad_norm": 4.734142303466797, "learning_rate": 9.802079919050426e-06, "loss": 0.6907, "step": 2132 }, { "epoch": 0.5213884135908091, "grad_norm": 3.565530300140381, "learning_rate": 9.794164212566355e-06, "loss": 0.5967, "step": 2133 }, { "epoch": 0.5216328526032755, "grad_norm": 4.300550937652588, "learning_rate": 9.7862486351083e-06, "loss": 0.7463, "step": 2134 }, { "epoch": 0.5218772916157419, "grad_norm": 4.79160737991333, "learning_rate": 9.778333191638055e-06, "loss": 0.7749, "step": 2135 }, { "epoch": 0.5221217306282082, "grad_norm": 5.3668951988220215, "learning_rate": 9.770417887117333e-06, "loss": 0.7261, "step": 2136 }, { "epoch": 0.5223661696406746, "grad_norm": 2.6563315391540527, "learning_rate": 9.762502726507764e-06, "loss": 2.0795, "step": 2137 }, { "epoch": 0.522610608653141, "grad_norm": 3.5295982360839844, "learning_rate": 9.75458771477088e-06, "loss": 0.6785, "step": 2138 }, { "epoch": 0.5228550476656074, "grad_norm": 3.7802164554595947, "learning_rate": 9.746672856868124e-06, "loss": 0.9156, "step": 2139 }, { "epoch": 0.5230994866780738, "grad_norm": 3.2856857776641846, "learning_rate": 9.738758157760841e-06, "loss": 0.791, "step": 2140 }, { "epoch": 0.5233439256905402, "grad_norm": 3.8294010162353516, "learning_rate": 9.730843622410282e-06, "loss": 0.7145, "step": 2141 }, { "epoch": 0.5235883647030066, "grad_norm": 3.379720687866211, "learning_rate": 9.722929255777583e-06, "loss": 0.7279, "step": 2142 }, { "epoch": 0.523832803715473, "grad_norm": 4.032857894897461, "learning_rate": 9.715015062823789e-06, "loss": 0.7665, "step": 2143 }, { "epoch": 0.5240772427279394, "grad_norm": 3.1897528171539307, "learning_rate": 9.707101048509828e-06, "loss": 0.8118, "step": 2144 }, { "epoch": 0.5243216817404057, "grad_norm": 2.891777992248535, "learning_rate": 9.699187217796518e-06, "loss": 0.7128, "step": 2145 }, { "epoch": 0.5245661207528721, "grad_norm": 2.616403818130493, "learning_rate": 9.691273575644554e-06, "loss": 0.702, "step": 2146 }, { "epoch": 0.5248105597653385, "grad_norm": 3.5809245109558105, "learning_rate": 9.68336012701453e-06, "loss": 0.833, "step": 2147 }, { "epoch": 0.5250549987778049, "grad_norm": 4.2926025390625, "learning_rate": 9.675446876866903e-06, "loss": 0.6003, "step": 2148 }, { "epoch": 0.5252994377902713, "grad_norm": 3.5323801040649414, "learning_rate": 9.667533830162013e-06, "loss": 0.6544, "step": 2149 }, { "epoch": 0.5255438768027377, "grad_norm": 2.839102029800415, "learning_rate": 9.65962099186007e-06, "loss": 0.753, "step": 2150 }, { "epoch": 0.5257883158152041, "grad_norm": 4.1956706047058105, "learning_rate": 9.651708366921152e-06, "loss": 0.909, "step": 2151 }, { "epoch": 0.5260327548276705, "grad_norm": 3.9112610816955566, "learning_rate": 9.643795960305207e-06, "loss": 0.7354, "step": 2152 }, { "epoch": 0.5262771938401369, "grad_norm": 3.4937121868133545, "learning_rate": 9.635883776972046e-06, "loss": 0.7002, "step": 2153 }, { "epoch": 0.5265216328526032, "grad_norm": 3.4174258708953857, "learning_rate": 9.62797182188134e-06, "loss": 0.679, "step": 2154 }, { "epoch": 0.5267660718650696, "grad_norm": 3.6268627643585205, "learning_rate": 9.620060099992609e-06, "loss": 0.5634, "step": 2155 }, { "epoch": 0.527010510877536, "grad_norm": 3.2625319957733154, "learning_rate": 9.612148616265238e-06, "loss": 0.7009, "step": 2156 }, { "epoch": 0.5272549498900024, "grad_norm": 3.6257452964782715, "learning_rate": 9.604237375658458e-06, "loss": 0.8636, "step": 2157 }, { "epoch": 0.5274993889024688, "grad_norm": 3.004514694213867, "learning_rate": 9.596326383131346e-06, "loss": 0.6685, "step": 2158 }, { "epoch": 0.5277438279149352, "grad_norm": 3.018902063369751, "learning_rate": 9.588415643642827e-06, "loss": 0.7041, "step": 2159 }, { "epoch": 0.5279882669274016, "grad_norm": 4.501762390136719, "learning_rate": 9.580505162151662e-06, "loss": 0.8137, "step": 2160 }, { "epoch": 0.528232705939868, "grad_norm": 3.087576150894165, "learning_rate": 9.572594943616457e-06, "loss": 0.7622, "step": 2161 }, { "epoch": 0.5284771449523344, "grad_norm": 4.186478614807129, "learning_rate": 9.564684992995645e-06, "loss": 0.7919, "step": 2162 }, { "epoch": 0.5287215839648007, "grad_norm": 4.6504693031311035, "learning_rate": 9.556775315247502e-06, "loss": 0.7596, "step": 2163 }, { "epoch": 0.5289660229772671, "grad_norm": 3.2655417919158936, "learning_rate": 9.548865915330124e-06, "loss": 0.618, "step": 2164 }, { "epoch": 0.5292104619897335, "grad_norm": 3.9398417472839355, "learning_rate": 9.540956798201439e-06, "loss": 0.6942, "step": 2165 }, { "epoch": 0.5294549010021999, "grad_norm": 3.884324550628662, "learning_rate": 9.53304796881919e-06, "loss": 0.7633, "step": 2166 }, { "epoch": 0.5296993400146663, "grad_norm": 3.5051238536834717, "learning_rate": 9.525139432140946e-06, "loss": 0.6823, "step": 2167 }, { "epoch": 0.5299437790271327, "grad_norm": 2.708531379699707, "learning_rate": 9.51723119312409e-06, "loss": 2.0782, "step": 2168 }, { "epoch": 0.5301882180395991, "grad_norm": 3.4217796325683594, "learning_rate": 9.50932325672582e-06, "loss": 0.7089, "step": 2169 }, { "epoch": 0.5304326570520655, "grad_norm": 3.7905852794647217, "learning_rate": 9.501415627903143e-06, "loss": 0.8027, "step": 2170 }, { "epoch": 0.530677096064532, "grad_norm": 3.344539165496826, "learning_rate": 9.493508311612874e-06, "loss": 0.6425, "step": 2171 }, { "epoch": 0.5309215350769982, "grad_norm": 3.495182752609253, "learning_rate": 9.485601312811631e-06, "loss": 0.7988, "step": 2172 }, { "epoch": 0.5311659740894646, "grad_norm": 3.6051738262176514, "learning_rate": 9.477694636455833e-06, "loss": 0.7847, "step": 2173 }, { "epoch": 0.531410413101931, "grad_norm": 3.1293435096740723, "learning_rate": 9.4697882875017e-06, "loss": 0.7358, "step": 2174 }, { "epoch": 0.5316548521143974, "grad_norm": 3.203247308731079, "learning_rate": 9.46188227090524e-06, "loss": 0.75, "step": 2175 }, { "epoch": 0.5318992911268638, "grad_norm": 3.34112548828125, "learning_rate": 9.45397659162226e-06, "loss": 0.6814, "step": 2176 }, { "epoch": 0.5321437301393303, "grad_norm": 3.279837131500244, "learning_rate": 9.44607125460835e-06, "loss": 0.6884, "step": 2177 }, { "epoch": 0.5323881691517967, "grad_norm": 4.138041019439697, "learning_rate": 9.438166264818892e-06, "loss": 0.797, "step": 2178 }, { "epoch": 0.532632608164263, "grad_norm": 2.98604416847229, "learning_rate": 9.430261627209044e-06, "loss": 0.6787, "step": 2179 }, { "epoch": 0.5328770471767295, "grad_norm": 3.883176565170288, "learning_rate": 9.422357346733747e-06, "loss": 0.8277, "step": 2180 }, { "epoch": 0.5331214861891957, "grad_norm": 4.455817699432373, "learning_rate": 9.414453428347715e-06, "loss": 0.8231, "step": 2181 }, { "epoch": 0.5333659252016621, "grad_norm": 3.1506121158599854, "learning_rate": 9.406549877005438e-06, "loss": 0.6412, "step": 2182 }, { "epoch": 0.5336103642141286, "grad_norm": 3.284975528717041, "learning_rate": 9.398646697661173e-06, "loss": 0.7799, "step": 2183 }, { "epoch": 0.533854803226595, "grad_norm": 3.78600811958313, "learning_rate": 9.390743895268945e-06, "loss": 0.6762, "step": 2184 }, { "epoch": 0.5340992422390614, "grad_norm": 3.901350975036621, "learning_rate": 9.382841474782546e-06, "loss": 0.7135, "step": 2185 }, { "epoch": 0.5343436812515278, "grad_norm": 4.45625114440918, "learning_rate": 9.374939441155525e-06, "loss": 0.8253, "step": 2186 }, { "epoch": 0.5345881202639942, "grad_norm": 3.3413562774658203, "learning_rate": 9.367037799341187e-06, "loss": 0.7585, "step": 2187 }, { "epoch": 0.5348325592764606, "grad_norm": 4.1224517822265625, "learning_rate": 9.359136554292596e-06, "loss": 0.8461, "step": 2188 }, { "epoch": 0.535076998288927, "grad_norm": 3.8622636795043945, "learning_rate": 9.351235710962565e-06, "loss": 0.7622, "step": 2189 }, { "epoch": 0.5353214373013933, "grad_norm": 3.3691866397857666, "learning_rate": 9.343335274303651e-06, "loss": 0.7139, "step": 2190 }, { "epoch": 0.5355658763138597, "grad_norm": 3.879357099533081, "learning_rate": 9.335435249268165e-06, "loss": 0.6645, "step": 2191 }, { "epoch": 0.5358103153263261, "grad_norm": 3.245859146118164, "learning_rate": 9.327535640808154e-06, "loss": 0.8594, "step": 2192 }, { "epoch": 0.5360547543387925, "grad_norm": 3.7075939178466797, "learning_rate": 9.3196364538754e-06, "loss": 0.8228, "step": 2193 }, { "epoch": 0.5362991933512589, "grad_norm": 3.3815085887908936, "learning_rate": 9.311737693421436e-06, "loss": 0.7622, "step": 2194 }, { "epoch": 0.5365436323637253, "grad_norm": 3.013307571411133, "learning_rate": 9.303839364397512e-06, "loss": 0.6551, "step": 2195 }, { "epoch": 0.5367880713761917, "grad_norm": 3.158292770385742, "learning_rate": 9.295941471754611e-06, "loss": 0.6397, "step": 2196 }, { "epoch": 0.5370325103886581, "grad_norm": 3.4648733139038086, "learning_rate": 9.288044020443449e-06, "loss": 0.6763, "step": 2197 }, { "epoch": 0.5372769494011245, "grad_norm": 4.360973834991455, "learning_rate": 9.280147015414458e-06, "loss": 0.7493, "step": 2198 }, { "epoch": 0.5375213884135908, "grad_norm": 3.1984996795654297, "learning_rate": 9.272250461617794e-06, "loss": 0.7879, "step": 2199 }, { "epoch": 0.5377658274260572, "grad_norm": 4.430309295654297, "learning_rate": 9.264354364003327e-06, "loss": 0.6783, "step": 2200 }, { "epoch": 0.5380102664385236, "grad_norm": 3.2827811241149902, "learning_rate": 9.256458727520648e-06, "loss": 0.6597, "step": 2201 }, { "epoch": 0.53825470545099, "grad_norm": 3.689610719680786, "learning_rate": 9.24856355711905e-06, "loss": 0.7053, "step": 2202 }, { "epoch": 0.5384991444634564, "grad_norm": 3.9610626697540283, "learning_rate": 9.24066885774754e-06, "loss": 0.5845, "step": 2203 }, { "epoch": 0.5387435834759228, "grad_norm": 3.752939462661743, "learning_rate": 9.232774634354827e-06, "loss": 0.8217, "step": 2204 }, { "epoch": 0.5389880224883892, "grad_norm": 3.2424635887145996, "learning_rate": 9.224880891889324e-06, "loss": 0.7209, "step": 2205 }, { "epoch": 0.5392324615008556, "grad_norm": 3.151031017303467, "learning_rate": 9.21698763529914e-06, "loss": 0.6904, "step": 2206 }, { "epoch": 0.539476900513322, "grad_norm": 3.540400981903076, "learning_rate": 9.209094869532082e-06, "loss": 0.5895, "step": 2207 }, { "epoch": 0.5397213395257883, "grad_norm": 2.9011690616607666, "learning_rate": 9.201202599535639e-06, "loss": 0.7095, "step": 2208 }, { "epoch": 0.5399657785382547, "grad_norm": 3.394333600997925, "learning_rate": 9.193310830257012e-06, "loss": 0.7982, "step": 2209 }, { "epoch": 0.5402102175507211, "grad_norm": 2.996051788330078, "learning_rate": 9.185419566643068e-06, "loss": 0.6493, "step": 2210 }, { "epoch": 0.5404546565631875, "grad_norm": 3.2841668128967285, "learning_rate": 9.177528813640362e-06, "loss": 0.6623, "step": 2211 }, { "epoch": 0.5406990955756539, "grad_norm": 3.1686737537384033, "learning_rate": 9.16963857619513e-06, "loss": 0.6757, "step": 2212 }, { "epoch": 0.5409435345881203, "grad_norm": 3.0455615520477295, "learning_rate": 9.161748859253288e-06, "loss": 0.7552, "step": 2213 }, { "epoch": 0.5411879736005867, "grad_norm": 3.6713058948516846, "learning_rate": 9.153859667760418e-06, "loss": 0.6814, "step": 2214 }, { "epoch": 0.5414324126130531, "grad_norm": 3.267914295196533, "learning_rate": 9.145971006661783e-06, "loss": 0.6431, "step": 2215 }, { "epoch": 0.5416768516255195, "grad_norm": 3.4913010597229004, "learning_rate": 9.138082880902302e-06, "loss": 0.7237, "step": 2216 }, { "epoch": 0.5419212906379858, "grad_norm": 3.234109878540039, "learning_rate": 9.13019529542657e-06, "loss": 0.6613, "step": 2217 }, { "epoch": 0.5421657296504522, "grad_norm": 3.133761405944824, "learning_rate": 9.12230825517883e-06, "loss": 0.7178, "step": 2218 }, { "epoch": 0.5424101686629186, "grad_norm": 4.280647277832031, "learning_rate": 9.114421765103e-06, "loss": 0.7539, "step": 2219 }, { "epoch": 0.542654607675385, "grad_norm": 3.2164087295532227, "learning_rate": 9.106535830142638e-06, "loss": 0.6519, "step": 2220 }, { "epoch": 0.5428990466878514, "grad_norm": 3.8115792274475098, "learning_rate": 9.098650455240959e-06, "loss": 0.9071, "step": 2221 }, { "epoch": 0.5431434857003178, "grad_norm": 3.463438034057617, "learning_rate": 9.090765645340832e-06, "loss": 0.768, "step": 2222 }, { "epoch": 0.5433879247127842, "grad_norm": 3.009864330291748, "learning_rate": 9.082881405384767e-06, "loss": 0.7107, "step": 2223 }, { "epoch": 0.5436323637252506, "grad_norm": 3.3293442726135254, "learning_rate": 9.074997740314913e-06, "loss": 0.7244, "step": 2224 }, { "epoch": 0.543876802737717, "grad_norm": 3.3943569660186768, "learning_rate": 9.067114655073069e-06, "loss": 0.6464, "step": 2225 }, { "epoch": 0.5441212417501833, "grad_norm": 3.474684000015259, "learning_rate": 9.059232154600662e-06, "loss": 0.7722, "step": 2226 }, { "epoch": 0.5443656807626497, "grad_norm": 2.908451795578003, "learning_rate": 9.051350243838757e-06, "loss": 1.9873, "step": 2227 }, { "epoch": 0.5446101197751161, "grad_norm": 3.428748846054077, "learning_rate": 9.043468927728045e-06, "loss": 0.6635, "step": 2228 }, { "epoch": 0.5448545587875825, "grad_norm": 3.169649124145508, "learning_rate": 9.035588211208847e-06, "loss": 0.6984, "step": 2229 }, { "epoch": 0.5450989978000489, "grad_norm": 2.6431148052215576, "learning_rate": 9.02770809922111e-06, "loss": 0.6158, "step": 2230 }, { "epoch": 0.5453434368125153, "grad_norm": 3.05291748046875, "learning_rate": 9.019828596704394e-06, "loss": 0.6906, "step": 2231 }, { "epoch": 0.5455878758249817, "grad_norm": 3.4386649131774902, "learning_rate": 9.011949708597892e-06, "loss": 0.607, "step": 2232 }, { "epoch": 0.5458323148374481, "grad_norm": 3.2652945518493652, "learning_rate": 9.004071439840396e-06, "loss": 0.7593, "step": 2233 }, { "epoch": 0.5460767538499145, "grad_norm": 3.2680203914642334, "learning_rate": 8.996193795370319e-06, "loss": 0.7699, "step": 2234 }, { "epoch": 0.5463211928623808, "grad_norm": 3.967542886734009, "learning_rate": 8.98831678012568e-06, "loss": 0.7487, "step": 2235 }, { "epoch": 0.5465656318748472, "grad_norm": 2.661180257797241, "learning_rate": 8.980440399044103e-06, "loss": 1.9339, "step": 2236 }, { "epoch": 0.5468100708873136, "grad_norm": 2.546757698059082, "learning_rate": 8.972564657062816e-06, "loss": 2.0937, "step": 2237 }, { "epoch": 0.54705450989978, "grad_norm": 3.3631649017333984, "learning_rate": 8.964689559118646e-06, "loss": 0.7624, "step": 2238 }, { "epoch": 0.5472989489122464, "grad_norm": 2.922140121459961, "learning_rate": 8.956815110148012e-06, "loss": 0.6668, "step": 2239 }, { "epoch": 0.5475433879247128, "grad_norm": 2.4277820587158203, "learning_rate": 8.948941315086936e-06, "loss": 1.9969, "step": 2240 }, { "epoch": 0.5477878269371792, "grad_norm": 3.292299747467041, "learning_rate": 8.941068178871021e-06, "loss": 0.6945, "step": 2241 }, { "epoch": 0.5480322659496456, "grad_norm": 3.2512614727020264, "learning_rate": 8.933195706435459e-06, "loss": 0.6505, "step": 2242 }, { "epoch": 0.548276704962112, "grad_norm": 3.8018834590911865, "learning_rate": 8.92532390271503e-06, "loss": 0.7485, "step": 2243 }, { "epoch": 0.5485211439745783, "grad_norm": 3.6477644443511963, "learning_rate": 8.917452772644092e-06, "loss": 0.6455, "step": 2244 }, { "epoch": 0.5487655829870447, "grad_norm": 3.821967124938965, "learning_rate": 8.90958232115658e-06, "loss": 0.6256, "step": 2245 }, { "epoch": 0.5490100219995111, "grad_norm": 3.82509708404541, "learning_rate": 8.901712553186002e-06, "loss": 0.8019, "step": 2246 }, { "epoch": 0.5492544610119775, "grad_norm": 3.373044729232788, "learning_rate": 8.893843473665443e-06, "loss": 0.6713, "step": 2247 }, { "epoch": 0.5494989000244439, "grad_norm": 3.1013381481170654, "learning_rate": 8.885975087527553e-06, "loss": 0.7236, "step": 2248 }, { "epoch": 0.5497433390369103, "grad_norm": 3.8983516693115234, "learning_rate": 8.878107399704546e-06, "loss": 0.7146, "step": 2249 }, { "epoch": 0.5499877780493767, "grad_norm": 3.556209087371826, "learning_rate": 8.870240415128201e-06, "loss": 0.6534, "step": 2250 }, { "epoch": 0.5502322170618431, "grad_norm": 3.2743749618530273, "learning_rate": 8.862374138729854e-06, "loss": 0.7089, "step": 2251 }, { "epoch": 0.5504766560743095, "grad_norm": 3.252896308898926, "learning_rate": 8.8545085754404e-06, "loss": 0.7129, "step": 2252 }, { "epoch": 0.5507210950867758, "grad_norm": 3.328610897064209, "learning_rate": 8.846643730190284e-06, "loss": 0.6465, "step": 2253 }, { "epoch": 0.5509655340992422, "grad_norm": 3.4283292293548584, "learning_rate": 8.838779607909502e-06, "loss": 0.6214, "step": 2254 }, { "epoch": 0.5512099731117086, "grad_norm": 2.920253038406372, "learning_rate": 8.830916213527593e-06, "loss": 0.6444, "step": 2255 }, { "epoch": 0.551454412124175, "grad_norm": 3.2998158931732178, "learning_rate": 8.823053551973653e-06, "loss": 0.6866, "step": 2256 }, { "epoch": 0.5516988511366414, "grad_norm": 3.3970022201538086, "learning_rate": 8.815191628176304e-06, "loss": 0.6624, "step": 2257 }, { "epoch": 0.5519432901491078, "grad_norm": 3.831857919692993, "learning_rate": 8.807330447063712e-06, "loss": 0.7256, "step": 2258 }, { "epoch": 0.5521877291615742, "grad_norm": 4.044900417327881, "learning_rate": 8.799470013563573e-06, "loss": 0.639, "step": 2259 }, { "epoch": 0.5524321681740406, "grad_norm": 3.4633517265319824, "learning_rate": 8.791610332603123e-06, "loss": 0.5559, "step": 2260 }, { "epoch": 0.552676607186507, "grad_norm": 4.289843559265137, "learning_rate": 8.783751409109116e-06, "loss": 0.7395, "step": 2261 }, { "epoch": 0.5529210461989733, "grad_norm": 3.948843002319336, "learning_rate": 8.77589324800784e-06, "loss": 0.8218, "step": 2262 }, { "epoch": 0.5531654852114397, "grad_norm": 4.675017356872559, "learning_rate": 8.768035854225098e-06, "loss": 0.8025, "step": 2263 }, { "epoch": 0.5534099242239061, "grad_norm": 3.471562385559082, "learning_rate": 8.760179232686217e-06, "loss": 0.7199, "step": 2264 }, { "epoch": 0.5536543632363725, "grad_norm": 3.258531093597412, "learning_rate": 8.752323388316038e-06, "loss": 0.725, "step": 2265 }, { "epoch": 0.5538988022488389, "grad_norm": 3.5345466136932373, "learning_rate": 8.744468326038914e-06, "loss": 0.6563, "step": 2266 }, { "epoch": 0.5541432412613053, "grad_norm": 3.502413272857666, "learning_rate": 8.73661405077871e-06, "loss": 0.5881, "step": 2267 }, { "epoch": 0.5543876802737717, "grad_norm": 5.4487810134887695, "learning_rate": 8.728760567458797e-06, "loss": 0.7555, "step": 2268 }, { "epoch": 0.5546321192862381, "grad_norm": 3.4690463542938232, "learning_rate": 8.720907881002047e-06, "loss": 0.8153, "step": 2269 }, { "epoch": 0.5548765582987045, "grad_norm": 3.3505942821502686, "learning_rate": 8.713055996330832e-06, "loss": 0.6743, "step": 2270 }, { "epoch": 0.5551209973111708, "grad_norm": 3.491990089416504, "learning_rate": 8.705204918367032e-06, "loss": 0.6761, "step": 2271 }, { "epoch": 0.5553654363236372, "grad_norm": 4.659776210784912, "learning_rate": 8.697354652032009e-06, "loss": 0.6884, "step": 2272 }, { "epoch": 0.5556098753361036, "grad_norm": 3.3480889797210693, "learning_rate": 8.689505202246618e-06, "loss": 0.6282, "step": 2273 }, { "epoch": 0.55585431434857, "grad_norm": 3.3003580570220947, "learning_rate": 8.681656573931209e-06, "loss": 0.6511, "step": 2274 }, { "epoch": 0.5560987533610364, "grad_norm": 2.958965539932251, "learning_rate": 8.67380877200561e-06, "loss": 0.5956, "step": 2275 }, { "epoch": 0.5563431923735028, "grad_norm": 3.247917890548706, "learning_rate": 8.665961801389133e-06, "loss": 0.6243, "step": 2276 }, { "epoch": 0.5565876313859692, "grad_norm": 3.247648000717163, "learning_rate": 8.658115667000572e-06, "loss": 0.7174, "step": 2277 }, { "epoch": 0.5568320703984356, "grad_norm": 3.5899839401245117, "learning_rate": 8.650270373758194e-06, "loss": 0.7944, "step": 2278 }, { "epoch": 0.557076509410902, "grad_norm": 3.4315006732940674, "learning_rate": 8.642425926579738e-06, "loss": 0.6849, "step": 2279 }, { "epoch": 0.5573209484233683, "grad_norm": 4.766323089599609, "learning_rate": 8.634582330382414e-06, "loss": 0.8128, "step": 2280 }, { "epoch": 0.5575653874358347, "grad_norm": 3.2827932834625244, "learning_rate": 8.626739590082897e-06, "loss": 0.6284, "step": 2281 }, { "epoch": 0.5578098264483011, "grad_norm": 3.4646546840667725, "learning_rate": 8.61889771059733e-06, "loss": 0.578, "step": 2282 }, { "epoch": 0.5580542654607675, "grad_norm": 3.7662670612335205, "learning_rate": 8.611056696841313e-06, "loss": 0.7601, "step": 2283 }, { "epoch": 0.5582987044732339, "grad_norm": 2.9055752754211426, "learning_rate": 8.6032165537299e-06, "loss": 0.7847, "step": 2284 }, { "epoch": 0.5585431434857003, "grad_norm": 4.044349670410156, "learning_rate": 8.595377286177606e-06, "loss": 0.7565, "step": 2285 }, { "epoch": 0.5587875824981667, "grad_norm": 2.8701465129852295, "learning_rate": 8.587538899098388e-06, "loss": 1.9964, "step": 2286 }, { "epoch": 0.5590320215106331, "grad_norm": 3.0263373851776123, "learning_rate": 8.579701397405668e-06, "loss": 0.635, "step": 2287 }, { "epoch": 0.5592764605230995, "grad_norm": 2.837200403213501, "learning_rate": 8.571864786012296e-06, "loss": 2.005, "step": 2288 }, { "epoch": 0.5595208995355658, "grad_norm": 5.424970626831055, "learning_rate": 8.564029069830569e-06, "loss": 0.6735, "step": 2289 }, { "epoch": 0.5597653385480322, "grad_norm": 2.9481616020202637, "learning_rate": 8.556194253772226e-06, "loss": 0.5457, "step": 2290 }, { "epoch": 0.5600097775604986, "grad_norm": 4.5754241943359375, "learning_rate": 8.54836034274844e-06, "loss": 0.7518, "step": 2291 }, { "epoch": 0.560254216572965, "grad_norm": 3.5060579776763916, "learning_rate": 8.540527341669814e-06, "loss": 0.6221, "step": 2292 }, { "epoch": 0.5604986555854314, "grad_norm": 3.9225127696990967, "learning_rate": 8.532695255446384e-06, "loss": 0.8294, "step": 2293 }, { "epoch": 0.5607430945978978, "grad_norm": 4.013699054718018, "learning_rate": 8.524864088987613e-06, "loss": 0.6485, "step": 2294 }, { "epoch": 0.5609875336103642, "grad_norm": 3.0791900157928467, "learning_rate": 8.517033847202383e-06, "loss": 0.6977, "step": 2295 }, { "epoch": 0.5612319726228306, "grad_norm": 3.0367400646209717, "learning_rate": 8.509204534999002e-06, "loss": 0.6768, "step": 2296 }, { "epoch": 0.561476411635297, "grad_norm": 3.0239598751068115, "learning_rate": 8.501376157285191e-06, "loss": 0.6619, "step": 2297 }, { "epoch": 0.5617208506477633, "grad_norm": 4.120974063873291, "learning_rate": 8.493548718968085e-06, "loss": 0.6598, "step": 2298 }, { "epoch": 0.5619652896602297, "grad_norm": 4.155196666717529, "learning_rate": 8.485722224954237e-06, "loss": 0.6761, "step": 2299 }, { "epoch": 0.5622097286726961, "grad_norm": 3.427410364151001, "learning_rate": 8.477896680149597e-06, "loss": 0.5893, "step": 2300 }, { "epoch": 0.5624541676851625, "grad_norm": 2.950836658477783, "learning_rate": 8.47007208945953e-06, "loss": 0.7799, "step": 2301 }, { "epoch": 0.562698606697629, "grad_norm": 3.4842567443847656, "learning_rate": 8.4622484577888e-06, "loss": 0.6295, "step": 2302 }, { "epoch": 0.5629430457100953, "grad_norm": 3.7331862449645996, "learning_rate": 8.454425790041567e-06, "loss": 0.7276, "step": 2303 }, { "epoch": 0.5631874847225617, "grad_norm": 3.904407501220703, "learning_rate": 8.44660409112139e-06, "loss": 0.7983, "step": 2304 }, { "epoch": 0.5634319237350282, "grad_norm": 3.3740532398223877, "learning_rate": 8.438783365931216e-06, "loss": 0.7281, "step": 2305 }, { "epoch": 0.5636763627474946, "grad_norm": 3.833692789077759, "learning_rate": 8.430963619373388e-06, "loss": 0.8742, "step": 2306 }, { "epoch": 0.5639208017599608, "grad_norm": 3.350278377532959, "learning_rate": 8.423144856349631e-06, "loss": 0.6639, "step": 2307 }, { "epoch": 0.5641652407724272, "grad_norm": 2.7863752841949463, "learning_rate": 8.415327081761056e-06, "loss": 0.4901, "step": 2308 }, { "epoch": 0.5644096797848936, "grad_norm": 3.5297372341156006, "learning_rate": 8.407510300508151e-06, "loss": 0.6328, "step": 2309 }, { "epoch": 0.56465411879736, "grad_norm": 3.285212278366089, "learning_rate": 8.399694517490787e-06, "loss": 0.7632, "step": 2310 }, { "epoch": 0.5648985578098265, "grad_norm": 3.537651777267456, "learning_rate": 8.391879737608202e-06, "loss": 0.6741, "step": 2311 }, { "epoch": 0.5651429968222929, "grad_norm": 3.2823824882507324, "learning_rate": 8.384065965759014e-06, "loss": 0.7633, "step": 2312 }, { "epoch": 0.5653874358347593, "grad_norm": 3.2051095962524414, "learning_rate": 8.3762532068412e-06, "loss": 0.6253, "step": 2313 }, { "epoch": 0.5656318748472257, "grad_norm": 3.437474489212036, "learning_rate": 8.368441465752111e-06, "loss": 0.7761, "step": 2314 }, { "epoch": 0.5658763138596921, "grad_norm": 3.0659470558166504, "learning_rate": 8.36063074738845e-06, "loss": 0.6317, "step": 2315 }, { "epoch": 0.5661207528721584, "grad_norm": 6.433424472808838, "learning_rate": 8.352821056646289e-06, "loss": 0.7217, "step": 2316 }, { "epoch": 0.5663651918846248, "grad_norm": 3.3285651206970215, "learning_rate": 8.345012398421046e-06, "loss": 0.7531, "step": 2317 }, { "epoch": 0.5666096308970912, "grad_norm": 3.584637403488159, "learning_rate": 8.3372047776075e-06, "loss": 0.7354, "step": 2318 }, { "epoch": 0.5668540699095576, "grad_norm": 3.155935764312744, "learning_rate": 8.329398199099778e-06, "loss": 0.7217, "step": 2319 }, { "epoch": 0.567098508922024, "grad_norm": 3.0519256591796875, "learning_rate": 8.321592667791351e-06, "loss": 0.6979, "step": 2320 }, { "epoch": 0.5673429479344904, "grad_norm": 2.8886446952819824, "learning_rate": 8.313788188575032e-06, "loss": 0.7433, "step": 2321 }, { "epoch": 0.5675873869469568, "grad_norm": 3.074049472808838, "learning_rate": 8.30598476634298e-06, "loss": 0.743, "step": 2322 }, { "epoch": 0.5678318259594232, "grad_norm": 3.1361405849456787, "learning_rate": 8.29818240598669e-06, "loss": 0.7218, "step": 2323 }, { "epoch": 0.5680762649718896, "grad_norm": 3.6520795822143555, "learning_rate": 8.290381112396989e-06, "loss": 0.8034, "step": 2324 }, { "epoch": 0.5683207039843559, "grad_norm": 3.4125802516937256, "learning_rate": 8.282580890464034e-06, "loss": 0.8232, "step": 2325 }, { "epoch": 0.5685651429968223, "grad_norm": 4.007071018218994, "learning_rate": 8.274781745077316e-06, "loss": 0.5776, "step": 2326 }, { "epoch": 0.5688095820092887, "grad_norm": 3.6633801460266113, "learning_rate": 8.266983681125647e-06, "loss": 0.6866, "step": 2327 }, { "epoch": 0.5690540210217551, "grad_norm": 2.8978915214538574, "learning_rate": 8.25918670349716e-06, "loss": 0.5959, "step": 2328 }, { "epoch": 0.5692984600342215, "grad_norm": 3.001453399658203, "learning_rate": 8.25139081707931e-06, "loss": 0.694, "step": 2329 }, { "epoch": 0.5695428990466879, "grad_norm": 2.951714038848877, "learning_rate": 8.24359602675887e-06, "loss": 0.7599, "step": 2330 }, { "epoch": 0.5697873380591543, "grad_norm": 6.370380878448486, "learning_rate": 8.23580233742192e-06, "loss": 0.6479, "step": 2331 }, { "epoch": 0.5700317770716207, "grad_norm": 3.31969952583313, "learning_rate": 8.228009753953845e-06, "loss": 0.7751, "step": 2332 }, { "epoch": 0.5702762160840871, "grad_norm": 3.5400986671447754, "learning_rate": 8.22021828123936e-06, "loss": 0.6758, "step": 2333 }, { "epoch": 0.5705206550965534, "grad_norm": 3.052528142929077, "learning_rate": 8.21242792416246e-06, "loss": 0.5639, "step": 2334 }, { "epoch": 0.5707650941090198, "grad_norm": 3.4429566860198975, "learning_rate": 8.20463868760645e-06, "loss": 0.7557, "step": 2335 }, { "epoch": 0.5710095331214862, "grad_norm": 3.9706220626831055, "learning_rate": 8.196850576453933e-06, "loss": 0.7045, "step": 2336 }, { "epoch": 0.5712539721339526, "grad_norm": 3.4687039852142334, "learning_rate": 8.189063595586804e-06, "loss": 0.6398, "step": 2337 }, { "epoch": 0.571498411146419, "grad_norm": 3.567106246948242, "learning_rate": 8.181277749886248e-06, "loss": 0.639, "step": 2338 }, { "epoch": 0.5717428501588854, "grad_norm": 4.738640785217285, "learning_rate": 8.173493044232745e-06, "loss": 0.7687, "step": 2339 }, { "epoch": 0.5719872891713518, "grad_norm": 3.808171510696411, "learning_rate": 8.165709483506054e-06, "loss": 0.7441, "step": 2340 }, { "epoch": 0.5722317281838182, "grad_norm": 3.3038058280944824, "learning_rate": 8.15792707258522e-06, "loss": 0.6759, "step": 2341 }, { "epoch": 0.5724761671962846, "grad_norm": 3.4772489070892334, "learning_rate": 8.150145816348567e-06, "loss": 0.6799, "step": 2342 }, { "epoch": 0.5727206062087509, "grad_norm": 3.959152936935425, "learning_rate": 8.142365719673693e-06, "loss": 0.7665, "step": 2343 }, { "epoch": 0.5729650452212173, "grad_norm": 4.0073137283325195, "learning_rate": 8.13458678743747e-06, "loss": 0.7992, "step": 2344 }, { "epoch": 0.5732094842336837, "grad_norm": 4.013628005981445, "learning_rate": 8.126809024516038e-06, "loss": 0.6077, "step": 2345 }, { "epoch": 0.5734539232461501, "grad_norm": 3.3716201782226562, "learning_rate": 8.119032435784811e-06, "loss": 0.7101, "step": 2346 }, { "epoch": 0.5736983622586165, "grad_norm": 4.247351169586182, "learning_rate": 8.111257026118461e-06, "loss": 0.6254, "step": 2347 }, { "epoch": 0.5739428012710829, "grad_norm": 3.6460940837860107, "learning_rate": 8.103482800390919e-06, "loss": 0.695, "step": 2348 }, { "epoch": 0.5741872402835493, "grad_norm": 2.9894654750823975, "learning_rate": 8.09570976347538e-06, "loss": 0.6567, "step": 2349 }, { "epoch": 0.5744316792960157, "grad_norm": 3.431727409362793, "learning_rate": 8.087937920244296e-06, "loss": 0.5698, "step": 2350 }, { "epoch": 0.5746761183084821, "grad_norm": 3.010704755783081, "learning_rate": 8.08016727556936e-06, "loss": 0.7224, "step": 2351 }, { "epoch": 0.5749205573209484, "grad_norm": 2.5566749572753906, "learning_rate": 8.07239783432152e-06, "loss": 0.5189, "step": 2352 }, { "epoch": 0.5751649963334148, "grad_norm": 4.0545220375061035, "learning_rate": 8.064629601370968e-06, "loss": 0.6559, "step": 2353 }, { "epoch": 0.5754094353458812, "grad_norm": 3.776262044906616, "learning_rate": 8.056862581587143e-06, "loss": 0.63, "step": 2354 }, { "epoch": 0.5756538743583476, "grad_norm": 2.8388962745666504, "learning_rate": 8.04909677983872e-06, "loss": 0.7652, "step": 2355 }, { "epoch": 0.575898313370814, "grad_norm": 3.862236976623535, "learning_rate": 8.041332200993606e-06, "loss": 0.6503, "step": 2356 }, { "epoch": 0.5761427523832804, "grad_norm": 3.48097825050354, "learning_rate": 8.03356884991895e-06, "loss": 0.7618, "step": 2357 }, { "epoch": 0.5763871913957468, "grad_norm": 2.9636313915252686, "learning_rate": 8.02580673148113e-06, "loss": 0.6997, "step": 2358 }, { "epoch": 0.5766316304082132, "grad_norm": 7.425884246826172, "learning_rate": 8.018045850545744e-06, "loss": 0.6347, "step": 2359 }, { "epoch": 0.5768760694206796, "grad_norm": 2.9083340167999268, "learning_rate": 8.01028621197762e-06, "loss": 0.533, "step": 2360 }, { "epoch": 0.5771205084331459, "grad_norm": 3.161489725112915, "learning_rate": 8.002527820640809e-06, "loss": 0.7049, "step": 2361 }, { "epoch": 0.5773649474456123, "grad_norm": 4.079695701599121, "learning_rate": 7.994770681398576e-06, "loss": 0.766, "step": 2362 }, { "epoch": 0.5776093864580787, "grad_norm": 3.2374866008758545, "learning_rate": 7.987014799113398e-06, "loss": 0.7038, "step": 2363 }, { "epoch": 0.5778538254705451, "grad_norm": 3.604962110519409, "learning_rate": 7.97926017864698e-06, "loss": 0.7046, "step": 2364 }, { "epoch": 0.5780982644830115, "grad_norm": 4.064101219177246, "learning_rate": 7.971506824860216e-06, "loss": 0.5358, "step": 2365 }, { "epoch": 0.5783427034954779, "grad_norm": 3.2200396060943604, "learning_rate": 7.963754742613218e-06, "loss": 0.7982, "step": 2366 }, { "epoch": 0.5785871425079443, "grad_norm": 3.961225986480713, "learning_rate": 7.956003936765298e-06, "loss": 0.7388, "step": 2367 }, { "epoch": 0.5788315815204107, "grad_norm": 3.4508423805236816, "learning_rate": 7.948254412174966e-06, "loss": 2.0356, "step": 2368 }, { "epoch": 0.5790760205328771, "grad_norm": 4.107931613922119, "learning_rate": 7.940506173699933e-06, "loss": 0.6808, "step": 2369 }, { "epoch": 0.5793204595453434, "grad_norm": 3.6741394996643066, "learning_rate": 7.932759226197098e-06, "loss": 0.6325, "step": 2370 }, { "epoch": 0.5795648985578098, "grad_norm": 3.243452548980713, "learning_rate": 7.925013574522556e-06, "loss": 0.5824, "step": 2371 }, { "epoch": 0.5798093375702762, "grad_norm": 3.2412142753601074, "learning_rate": 7.91726922353159e-06, "loss": 0.6886, "step": 2372 }, { "epoch": 0.5800537765827426, "grad_norm": 3.364396333694458, "learning_rate": 7.90952617807866e-06, "loss": 0.6811, "step": 2373 }, { "epoch": 0.580298215595209, "grad_norm": 3.435480833053589, "learning_rate": 7.901784443017419e-06, "loss": 0.6123, "step": 2374 }, { "epoch": 0.5805426546076754, "grad_norm": 3.247061252593994, "learning_rate": 7.894044023200691e-06, "loss": 0.652, "step": 2375 }, { "epoch": 0.5807870936201418, "grad_norm": 4.547329425811768, "learning_rate": 7.886304923480476e-06, "loss": 0.7178, "step": 2376 }, { "epoch": 0.5810315326326082, "grad_norm": 2.973055601119995, "learning_rate": 7.878567148707948e-06, "loss": 0.761, "step": 2377 }, { "epoch": 0.5812759716450746, "grad_norm": 4.504457473754883, "learning_rate": 7.870830703733451e-06, "loss": 0.6513, "step": 2378 }, { "epoch": 0.5815204106575409, "grad_norm": 3.269958734512329, "learning_rate": 7.86309559340649e-06, "loss": 0.6604, "step": 2379 }, { "epoch": 0.5817648496700073, "grad_norm": 2.929678201675415, "learning_rate": 7.85536182257575e-06, "loss": 0.6358, "step": 2380 }, { "epoch": 0.5820092886824737, "grad_norm": 4.424942970275879, "learning_rate": 7.847629396089054e-06, "loss": 0.6125, "step": 2381 }, { "epoch": 0.5822537276949401, "grad_norm": 3.2533557415008545, "learning_rate": 7.839898318793398e-06, "loss": 0.6891, "step": 2382 }, { "epoch": 0.5824981667074065, "grad_norm": 5.905025959014893, "learning_rate": 7.832168595534925e-06, "loss": 0.6394, "step": 2383 }, { "epoch": 0.5827426057198729, "grad_norm": 3.163419723510742, "learning_rate": 7.82444023115893e-06, "loss": 0.5843, "step": 2384 }, { "epoch": 0.5829870447323393, "grad_norm": 4.014256954193115, "learning_rate": 7.81671323050986e-06, "loss": 0.5822, "step": 2385 }, { "epoch": 0.5832314837448057, "grad_norm": 3.2915475368499756, "learning_rate": 7.808987598431303e-06, "loss": 0.6776, "step": 2386 }, { "epoch": 0.5834759227572721, "grad_norm": 3.713676929473877, "learning_rate": 7.801263339765992e-06, "loss": 0.8284, "step": 2387 }, { "epoch": 0.5837203617697384, "grad_norm": 3.4072582721710205, "learning_rate": 7.793540459355798e-06, "loss": 0.7618, "step": 2388 }, { "epoch": 0.5839648007822048, "grad_norm": 2.850114583969116, "learning_rate": 7.785818962041727e-06, "loss": 0.7429, "step": 2389 }, { "epoch": 0.5842092397946712, "grad_norm": 3.4230334758758545, "learning_rate": 7.778098852663918e-06, "loss": 0.6766, "step": 2390 }, { "epoch": 0.5844536788071376, "grad_norm": 2.526958465576172, "learning_rate": 7.770380136061643e-06, "loss": 2.0334, "step": 2391 }, { "epoch": 0.584698117819604, "grad_norm": 5.009162902832031, "learning_rate": 7.762662817073297e-06, "loss": 0.6014, "step": 2392 }, { "epoch": 0.5849425568320704, "grad_norm": 3.332958221435547, "learning_rate": 7.754946900536403e-06, "loss": 0.6578, "step": 2393 }, { "epoch": 0.5851869958445368, "grad_norm": 3.8514368534088135, "learning_rate": 7.747232391287602e-06, "loss": 0.7524, "step": 2394 }, { "epoch": 0.5854314348570032, "grad_norm": 5.737468719482422, "learning_rate": 7.739519294162652e-06, "loss": 0.6417, "step": 2395 }, { "epoch": 0.5856758738694696, "grad_norm": 3.6034488677978516, "learning_rate": 7.73180761399643e-06, "loss": 0.7302, "step": 2396 }, { "epoch": 0.5859203128819359, "grad_norm": 3.2893946170806885, "learning_rate": 7.72409735562292e-06, "loss": 0.6442, "step": 2397 }, { "epoch": 0.5861647518944023, "grad_norm": 5.023566246032715, "learning_rate": 7.716388523875219e-06, "loss": 0.7578, "step": 2398 }, { "epoch": 0.5864091909068687, "grad_norm": 3.2663068771362305, "learning_rate": 7.708681123585525e-06, "loss": 0.6888, "step": 2399 }, { "epoch": 0.5866536299193351, "grad_norm": 3.55702805519104, "learning_rate": 7.700975159585142e-06, "loss": 0.7193, "step": 2400 }, { "epoch": 0.5868980689318015, "grad_norm": 3.774721622467041, "learning_rate": 7.693270636704476e-06, "loss": 0.7504, "step": 2401 }, { "epoch": 0.5871425079442679, "grad_norm": 5.629326820373535, "learning_rate": 7.685567559773024e-06, "loss": 0.7078, "step": 2402 }, { "epoch": 0.5873869469567343, "grad_norm": 3.1027450561523438, "learning_rate": 7.67786593361938e-06, "loss": 0.7157, "step": 2403 }, { "epoch": 0.5876313859692007, "grad_norm": 3.616337537765503, "learning_rate": 7.670165763071227e-06, "loss": 0.6647, "step": 2404 }, { "epoch": 0.5878758249816671, "grad_norm": 3.252295732498169, "learning_rate": 7.662467052955337e-06, "loss": 0.7219, "step": 2405 }, { "epoch": 0.5881202639941334, "grad_norm": 3.382500171661377, "learning_rate": 7.65476980809757e-06, "loss": 0.5509, "step": 2406 }, { "epoch": 0.5883647030065998, "grad_norm": 3.338959217071533, "learning_rate": 7.647074033322856e-06, "loss": 0.6593, "step": 2407 }, { "epoch": 0.5886091420190662, "grad_norm": 2.3664376735687256, "learning_rate": 7.639379733455219e-06, "loss": 1.9463, "step": 2408 }, { "epoch": 0.5888535810315326, "grad_norm": 2.9018778800964355, "learning_rate": 7.631686913317743e-06, "loss": 0.6245, "step": 2409 }, { "epoch": 0.589098020043999, "grad_norm": 3.5737507343292236, "learning_rate": 7.623995577732594e-06, "loss": 0.6687, "step": 2410 }, { "epoch": 0.5893424590564654, "grad_norm": 2.874079704284668, "learning_rate": 7.616305731521009e-06, "loss": 0.7341, "step": 2411 }, { "epoch": 0.5895868980689318, "grad_norm": 4.047252655029297, "learning_rate": 7.6086173795032866e-06, "loss": 0.6604, "step": 2412 }, { "epoch": 0.5898313370813982, "grad_norm": 3.585425615310669, "learning_rate": 7.600930526498788e-06, "loss": 0.7895, "step": 2413 }, { "epoch": 0.5900757760938646, "grad_norm": 2.9749667644500732, "learning_rate": 7.593245177325941e-06, "loss": 0.5783, "step": 2414 }, { "epoch": 0.5903202151063309, "grad_norm": 2.7515406608581543, "learning_rate": 7.585561336802221e-06, "loss": 0.6703, "step": 2415 }, { "epoch": 0.5905646541187973, "grad_norm": 3.3146510124206543, "learning_rate": 7.577879009744165e-06, "loss": 0.6058, "step": 2416 }, { "epoch": 0.5908090931312637, "grad_norm": 3.073467969894409, "learning_rate": 7.570198200967363e-06, "loss": 0.5668, "step": 2417 }, { "epoch": 0.5910535321437301, "grad_norm": 3.7749462127685547, "learning_rate": 7.562518915286443e-06, "loss": 0.7409, "step": 2418 }, { "epoch": 0.5912979711561965, "grad_norm": 3.177457571029663, "learning_rate": 7.554841157515091e-06, "loss": 0.7827, "step": 2419 }, { "epoch": 0.5915424101686629, "grad_norm": 2.8870596885681152, "learning_rate": 7.547164932466028e-06, "loss": 0.7256, "step": 2420 }, { "epoch": 0.5917868491811293, "grad_norm": 3.333930253982544, "learning_rate": 7.539490244951013e-06, "loss": 0.6556, "step": 2421 }, { "epoch": 0.5920312881935957, "grad_norm": 6.634274482727051, "learning_rate": 7.531817099780846e-06, "loss": 0.69, "step": 2422 }, { "epoch": 0.5922757272060621, "grad_norm": 3.813068151473999, "learning_rate": 7.524145501765357e-06, "loss": 0.6719, "step": 2423 }, { "epoch": 0.5925201662185284, "grad_norm": 3.0771398544311523, "learning_rate": 7.516475455713409e-06, "loss": 0.6068, "step": 2424 }, { "epoch": 0.5927646052309948, "grad_norm": 3.213043689727783, "learning_rate": 7.508806966432882e-06, "loss": 0.6483, "step": 2425 }, { "epoch": 0.5930090442434612, "grad_norm": 4.564669132232666, "learning_rate": 7.501140038730701e-06, "loss": 0.5449, "step": 2426 }, { "epoch": 0.5932534832559276, "grad_norm": 3.1842901706695557, "learning_rate": 7.493474677412795e-06, "loss": 0.6448, "step": 2427 }, { "epoch": 0.593497922268394, "grad_norm": 2.83215594291687, "learning_rate": 7.4858108872841115e-06, "loss": 0.5718, "step": 2428 }, { "epoch": 0.5937423612808604, "grad_norm": 3.178879499435425, "learning_rate": 7.4781486731486215e-06, "loss": 0.6779, "step": 2429 }, { "epoch": 0.5939868002933268, "grad_norm": 3.813783884048462, "learning_rate": 7.470488039809304e-06, "loss": 0.7745, "step": 2430 }, { "epoch": 0.5942312393057932, "grad_norm": 4.4895195960998535, "learning_rate": 7.462828992068144e-06, "loss": 0.6709, "step": 2431 }, { "epoch": 0.5944756783182596, "grad_norm": 3.431575059890747, "learning_rate": 7.455171534726138e-06, "loss": 0.8067, "step": 2432 }, { "epoch": 0.5947201173307259, "grad_norm": 3.2519891262054443, "learning_rate": 7.447515672583282e-06, "loss": 0.7389, "step": 2433 }, { "epoch": 0.5949645563431923, "grad_norm": 3.4478063583374023, "learning_rate": 7.4398614104385734e-06, "loss": 0.7747, "step": 2434 }, { "epoch": 0.5952089953556587, "grad_norm": 3.40944504737854, "learning_rate": 7.432208753090009e-06, "loss": 0.7563, "step": 2435 }, { "epoch": 0.5954534343681251, "grad_norm": 3.2093870639801025, "learning_rate": 7.4245577053345715e-06, "loss": 0.7171, "step": 2436 }, { "epoch": 0.5956978733805915, "grad_norm": 3.222412347793579, "learning_rate": 7.416908271968246e-06, "loss": 0.6656, "step": 2437 }, { "epoch": 0.595942312393058, "grad_norm": 2.9758498668670654, "learning_rate": 7.409260457785999e-06, "loss": 0.7187, "step": 2438 }, { "epoch": 0.5961867514055244, "grad_norm": 2.8184404373168945, "learning_rate": 7.401614267581779e-06, "loss": 0.6505, "step": 2439 }, { "epoch": 0.5964311904179908, "grad_norm": 3.585016965866089, "learning_rate": 7.393969706148524e-06, "loss": 0.7019, "step": 2440 }, { "epoch": 0.5966756294304572, "grad_norm": 4.0438995361328125, "learning_rate": 7.386326778278142e-06, "loss": 0.6532, "step": 2441 }, { "epoch": 0.5969200684429234, "grad_norm": 3.1334328651428223, "learning_rate": 7.378685488761533e-06, "loss": 0.7686, "step": 2442 }, { "epoch": 0.5971645074553898, "grad_norm": 3.9417247772216797, "learning_rate": 7.371045842388552e-06, "loss": 0.8383, "step": 2443 }, { "epoch": 0.5974089464678562, "grad_norm": 4.557409763336182, "learning_rate": 7.363407843948034e-06, "loss": 0.6166, "step": 2444 }, { "epoch": 0.5976533854803227, "grad_norm": 3.150182008743286, "learning_rate": 7.355771498227777e-06, "loss": 0.6707, "step": 2445 }, { "epoch": 0.597897824492789, "grad_norm": 5.719577312469482, "learning_rate": 7.3481368100145464e-06, "loss": 0.7095, "step": 2446 }, { "epoch": 0.5981422635052555, "grad_norm": 3.6504716873168945, "learning_rate": 7.340503784094066e-06, "loss": 0.6926, "step": 2447 }, { "epoch": 0.5983867025177219, "grad_norm": 4.071256160736084, "learning_rate": 7.332872425251017e-06, "loss": 0.6978, "step": 2448 }, { "epoch": 0.5986311415301883, "grad_norm": 3.5719997882843018, "learning_rate": 7.325242738269039e-06, "loss": 0.5985, "step": 2449 }, { "epoch": 0.5988755805426547, "grad_norm": 2.8013484477996826, "learning_rate": 7.317614727930722e-06, "loss": 0.5605, "step": 2450 }, { "epoch": 0.599120019555121, "grad_norm": 2.548478126525879, "learning_rate": 7.3099883990176025e-06, "loss": 1.9373, "step": 2451 }, { "epoch": 0.5993644585675874, "grad_norm": 2.671673059463501, "learning_rate": 7.302363756310165e-06, "loss": 1.9074, "step": 2452 }, { "epoch": 0.5996088975800538, "grad_norm": 2.897136926651001, "learning_rate": 7.294740804587838e-06, "loss": 0.6139, "step": 2453 }, { "epoch": 0.5998533365925202, "grad_norm": 3.2317192554473877, "learning_rate": 7.287119548628992e-06, "loss": 0.7463, "step": 2454 }, { "epoch": 0.6000977756049866, "grad_norm": 2.8740546703338623, "learning_rate": 7.279499993210928e-06, "loss": 0.6704, "step": 2455 }, { "epoch": 0.600342214617453, "grad_norm": 3.8139729499816895, "learning_rate": 7.27188214310988e-06, "loss": 0.6726, "step": 2456 }, { "epoch": 0.6005866536299194, "grad_norm": 2.5996851921081543, "learning_rate": 7.26426600310103e-06, "loss": 0.561, "step": 2457 }, { "epoch": 0.6008310926423858, "grad_norm": 3.3971922397613525, "learning_rate": 7.256651577958469e-06, "loss": 0.7901, "step": 2458 }, { "epoch": 0.6010755316548522, "grad_norm": 3.3375139236450195, "learning_rate": 7.24903887245522e-06, "loss": 0.596, "step": 2459 }, { "epoch": 0.6013199706673185, "grad_norm": 3.5682437419891357, "learning_rate": 7.241427891363228e-06, "loss": 0.7292, "step": 2460 }, { "epoch": 0.6015644096797849, "grad_norm": 2.866892099380493, "learning_rate": 7.233818639453358e-06, "loss": 0.6654, "step": 2461 }, { "epoch": 0.6018088486922513, "grad_norm": 2.9490864276885986, "learning_rate": 7.22621112149539e-06, "loss": 0.5401, "step": 2462 }, { "epoch": 0.6020532877047177, "grad_norm": 2.8294754028320312, "learning_rate": 7.218605342258016e-06, "loss": 0.6467, "step": 2463 }, { "epoch": 0.6022977267171841, "grad_norm": 3.224865436553955, "learning_rate": 7.21100130650884e-06, "loss": 0.7148, "step": 2464 }, { "epoch": 0.6025421657296505, "grad_norm": 4.785635471343994, "learning_rate": 7.203399019014374e-06, "loss": 0.6656, "step": 2465 }, { "epoch": 0.6027866047421169, "grad_norm": 3.1482491493225098, "learning_rate": 7.195798484540028e-06, "loss": 0.6414, "step": 2466 }, { "epoch": 0.6030310437545833, "grad_norm": 4.730430603027344, "learning_rate": 7.188199707850123e-06, "loss": 0.7205, "step": 2467 }, { "epoch": 0.6032754827670497, "grad_norm": 3.990285873413086, "learning_rate": 7.1806026937078695e-06, "loss": 0.6635, "step": 2468 }, { "epoch": 0.603519921779516, "grad_norm": 3.3099493980407715, "learning_rate": 7.173007446875375e-06, "loss": 0.7286, "step": 2469 }, { "epoch": 0.6037643607919824, "grad_norm": 3.4665327072143555, "learning_rate": 7.165413972113646e-06, "loss": 0.7668, "step": 2470 }, { "epoch": 0.6040087998044488, "grad_norm": 3.216830015182495, "learning_rate": 7.15782227418257e-06, "loss": 0.6273, "step": 2471 }, { "epoch": 0.6042532388169152, "grad_norm": 2.6301674842834473, "learning_rate": 7.150232357840921e-06, "loss": 0.5984, "step": 2472 }, { "epoch": 0.6044976778293816, "grad_norm": 3.5538887977600098, "learning_rate": 7.1426442278463626e-06, "loss": 0.7371, "step": 2473 }, { "epoch": 0.604742116841848, "grad_norm": 4.289004802703857, "learning_rate": 7.135057888955435e-06, "loss": 0.6264, "step": 2474 }, { "epoch": 0.6049865558543144, "grad_norm": 3.42405366897583, "learning_rate": 7.127473345923553e-06, "loss": 0.6621, "step": 2475 }, { "epoch": 0.6052309948667808, "grad_norm": 3.0243327617645264, "learning_rate": 7.1198906035050105e-06, "loss": 0.7344, "step": 2476 }, { "epoch": 0.6054754338792472, "grad_norm": 2.5014827251434326, "learning_rate": 7.112309666452968e-06, "loss": 2.0079, "step": 2477 }, { "epoch": 0.6057198728917135, "grad_norm": 2.786853313446045, "learning_rate": 7.10473053951946e-06, "loss": 0.7084, "step": 2478 }, { "epoch": 0.6059643119041799, "grad_norm": 3.303546667098999, "learning_rate": 7.097153227455379e-06, "loss": 0.7626, "step": 2479 }, { "epoch": 0.6062087509166463, "grad_norm": 3.2983829975128174, "learning_rate": 7.089577735010485e-06, "loss": 0.7953, "step": 2480 }, { "epoch": 0.6064531899291127, "grad_norm": 3.149487257003784, "learning_rate": 7.0820040669333975e-06, "loss": 0.7116, "step": 2481 }, { "epoch": 0.6066976289415791, "grad_norm": 3.133228063583374, "learning_rate": 7.074432227971593e-06, "loss": 0.6555, "step": 2482 }, { "epoch": 0.6069420679540455, "grad_norm": 3.6981680393218994, "learning_rate": 7.066862222871397e-06, "loss": 0.5982, "step": 2483 }, { "epoch": 0.6071865069665119, "grad_norm": 4.039612770080566, "learning_rate": 7.059294056377988e-06, "loss": 0.5946, "step": 2484 }, { "epoch": 0.6074309459789783, "grad_norm": 3.019282579421997, "learning_rate": 7.051727733235395e-06, "loss": 0.6346, "step": 2485 }, { "epoch": 0.6076753849914447, "grad_norm": 2.6989247798919678, "learning_rate": 7.044163258186485e-06, "loss": 0.6808, "step": 2486 }, { "epoch": 0.607919824003911, "grad_norm": 3.9518232345581055, "learning_rate": 7.036600635972971e-06, "loss": 0.7237, "step": 2487 }, { "epoch": 0.6081642630163774, "grad_norm": 4.522764682769775, "learning_rate": 7.029039871335406e-06, "loss": 0.6655, "step": 2488 }, { "epoch": 0.6084087020288438, "grad_norm": 3.1290531158447266, "learning_rate": 7.021480969013176e-06, "loss": 0.745, "step": 2489 }, { "epoch": 0.6086531410413102, "grad_norm": 2.9655075073242188, "learning_rate": 7.0139239337444956e-06, "loss": 0.7403, "step": 2490 }, { "epoch": 0.6088975800537766, "grad_norm": 3.183637857437134, "learning_rate": 7.006368770266421e-06, "loss": 0.6196, "step": 2491 }, { "epoch": 0.609142019066243, "grad_norm": 3.102365016937256, "learning_rate": 6.99881548331482e-06, "loss": 0.6792, "step": 2492 }, { "epoch": 0.6093864580787094, "grad_norm": 3.733527660369873, "learning_rate": 6.9912640776243974e-06, "loss": 0.6448, "step": 2493 }, { "epoch": 0.6096308970911758, "grad_norm": 2.7660675048828125, "learning_rate": 6.983714557928672e-06, "loss": 0.6567, "step": 2494 }, { "epoch": 0.6098753361036422, "grad_norm": 4.2180938720703125, "learning_rate": 6.976166928959978e-06, "loss": 0.7289, "step": 2495 }, { "epoch": 0.6101197751161085, "grad_norm": 2.9590981006622314, "learning_rate": 6.96862119544947e-06, "loss": 0.6344, "step": 2496 }, { "epoch": 0.6103642141285749, "grad_norm": 5.102367401123047, "learning_rate": 6.961077362127111e-06, "loss": 0.6535, "step": 2497 }, { "epoch": 0.6106086531410413, "grad_norm": 2.679868221282959, "learning_rate": 6.953535433721677e-06, "loss": 0.6962, "step": 2498 }, { "epoch": 0.6108530921535077, "grad_norm": 4.007815837860107, "learning_rate": 6.945995414960744e-06, "loss": 0.8283, "step": 2499 }, { "epoch": 0.6110975311659741, "grad_norm": 4.890908241271973, "learning_rate": 6.938457310570696e-06, "loss": 0.6869, "step": 2500 }, { "epoch": 0.6113419701784405, "grad_norm": 3.021857261657715, "learning_rate": 6.930921125276715e-06, "loss": 0.6152, "step": 2501 }, { "epoch": 0.6115864091909069, "grad_norm": 2.963789701461792, "learning_rate": 6.923386863802779e-06, "loss": 0.5795, "step": 2502 }, { "epoch": 0.6118308482033733, "grad_norm": 3.034787178039551, "learning_rate": 6.915854530871657e-06, "loss": 0.5507, "step": 2503 }, { "epoch": 0.6120752872158397, "grad_norm": 2.8483729362487793, "learning_rate": 6.9083241312049235e-06, "loss": 0.7486, "step": 2504 }, { "epoch": 0.612319726228306, "grad_norm": 3.7511532306671143, "learning_rate": 6.900795669522924e-06, "loss": 0.8166, "step": 2505 }, { "epoch": 0.6125641652407724, "grad_norm": 3.2344181537628174, "learning_rate": 6.893269150544797e-06, "loss": 0.6897, "step": 2506 }, { "epoch": 0.6128086042532388, "grad_norm": 3.738460063934326, "learning_rate": 6.885744578988463e-06, "loss": 0.7133, "step": 2507 }, { "epoch": 0.6130530432657052, "grad_norm": 5.127227783203125, "learning_rate": 6.878221959570621e-06, "loss": 0.595, "step": 2508 }, { "epoch": 0.6132974822781716, "grad_norm": 4.383499622344971, "learning_rate": 6.870701297006746e-06, "loss": 0.6411, "step": 2509 }, { "epoch": 0.613541921290638, "grad_norm": 3.2642929553985596, "learning_rate": 6.8631825960110866e-06, "loss": 0.7823, "step": 2510 }, { "epoch": 0.6137863603031044, "grad_norm": 3.017092704772949, "learning_rate": 6.855665861296662e-06, "loss": 0.6362, "step": 2511 }, { "epoch": 0.6140307993155708, "grad_norm": 3.371415138244629, "learning_rate": 6.84815109757526e-06, "loss": 0.6656, "step": 2512 }, { "epoch": 0.6142752383280372, "grad_norm": 3.385425329208374, "learning_rate": 6.84063830955743e-06, "loss": 0.6934, "step": 2513 }, { "epoch": 0.6145196773405035, "grad_norm": 3.7038681507110596, "learning_rate": 6.833127501952488e-06, "loss": 0.5947, "step": 2514 }, { "epoch": 0.6147641163529699, "grad_norm": 2.9765119552612305, "learning_rate": 6.8256186794685016e-06, "loss": 0.5329, "step": 2515 }, { "epoch": 0.6150085553654363, "grad_norm": 2.7302298545837402, "learning_rate": 6.818111846812299e-06, "loss": 0.6509, "step": 2516 }, { "epoch": 0.6152529943779027, "grad_norm": 3.5046465396881104, "learning_rate": 6.810607008689461e-06, "loss": 0.6428, "step": 2517 }, { "epoch": 0.6154974333903691, "grad_norm": 3.45648455619812, "learning_rate": 6.803104169804311e-06, "loss": 0.707, "step": 2518 }, { "epoch": 0.6157418724028355, "grad_norm": 3.033381938934326, "learning_rate": 6.795603334859935e-06, "loss": 0.6832, "step": 2519 }, { "epoch": 0.6159863114153019, "grad_norm": 3.4825875759124756, "learning_rate": 6.788104508558147e-06, "loss": 0.6363, "step": 2520 }, { "epoch": 0.6162307504277683, "grad_norm": 3.1465296745300293, "learning_rate": 6.78060769559951e-06, "loss": 0.6199, "step": 2521 }, { "epoch": 0.6164751894402347, "grad_norm": 3.8383195400238037, "learning_rate": 6.773112900683323e-06, "loss": 0.6372, "step": 2522 }, { "epoch": 0.616719628452701, "grad_norm": 3.987447738647461, "learning_rate": 6.7656201285076195e-06, "loss": 0.7747, "step": 2523 }, { "epoch": 0.6169640674651674, "grad_norm": 2.686070680618286, "learning_rate": 6.758129383769165e-06, "loss": 0.6127, "step": 2524 }, { "epoch": 0.6172085064776338, "grad_norm": 2.738743543624878, "learning_rate": 6.7506406711634545e-06, "loss": 0.6823, "step": 2525 }, { "epoch": 0.6174529454901002, "grad_norm": 3.0114638805389404, "learning_rate": 6.743153995384709e-06, "loss": 0.7053, "step": 2526 }, { "epoch": 0.6176973845025666, "grad_norm": 3.4111175537109375, "learning_rate": 6.735669361125874e-06, "loss": 0.64, "step": 2527 }, { "epoch": 0.617941823515033, "grad_norm": 2.8889999389648438, "learning_rate": 6.728186773078614e-06, "loss": 0.7055, "step": 2528 }, { "epoch": 0.6181862625274994, "grad_norm": 3.439328670501709, "learning_rate": 6.720706235933312e-06, "loss": 0.5602, "step": 2529 }, { "epoch": 0.6184307015399658, "grad_norm": 2.9301159381866455, "learning_rate": 6.713227754379063e-06, "loss": 0.7021, "step": 2530 }, { "epoch": 0.6186751405524322, "grad_norm": 3.2925527095794678, "learning_rate": 6.705751333103676e-06, "loss": 0.6667, "step": 2531 }, { "epoch": 0.6189195795648985, "grad_norm": 3.848553419113159, "learning_rate": 6.6982769767936676e-06, "loss": 0.7464, "step": 2532 }, { "epoch": 0.6191640185773649, "grad_norm": 3.1005401611328125, "learning_rate": 6.690804690134259e-06, "loss": 0.6674, "step": 2533 }, { "epoch": 0.6194084575898313, "grad_norm": 3.32075834274292, "learning_rate": 6.6833344778093735e-06, "loss": 0.6011, "step": 2534 }, { "epoch": 0.6196528966022977, "grad_norm": 3.6397931575775146, "learning_rate": 6.675866344501641e-06, "loss": 0.727, "step": 2535 }, { "epoch": 0.6198973356147641, "grad_norm": 3.8641738891601562, "learning_rate": 6.668400294892379e-06, "loss": 0.7457, "step": 2536 }, { "epoch": 0.6201417746272305, "grad_norm": 2.848632335662842, "learning_rate": 6.660936333661605e-06, "loss": 0.667, "step": 2537 }, { "epoch": 0.6203862136396969, "grad_norm": 2.7623350620269775, "learning_rate": 6.653474465488023e-06, "loss": 0.6282, "step": 2538 }, { "epoch": 0.6206306526521633, "grad_norm": 3.0257644653320312, "learning_rate": 6.64601469504903e-06, "loss": 0.5238, "step": 2539 }, { "epoch": 0.6208750916646297, "grad_norm": 2.996584892272949, "learning_rate": 6.638557027020703e-06, "loss": 0.6428, "step": 2540 }, { "epoch": 0.621119530677096, "grad_norm": 3.1068155765533447, "learning_rate": 6.631101466077801e-06, "loss": 0.6324, "step": 2541 }, { "epoch": 0.6213639696895624, "grad_norm": 4.016943454742432, "learning_rate": 6.623648016893769e-06, "loss": 0.6182, "step": 2542 }, { "epoch": 0.6216084087020288, "grad_norm": 3.138049602508545, "learning_rate": 6.616196684140721e-06, "loss": 0.5747, "step": 2543 }, { "epoch": 0.6218528477144952, "grad_norm": 2.8710036277770996, "learning_rate": 6.60874747248945e-06, "loss": 0.6195, "step": 2544 }, { "epoch": 0.6220972867269616, "grad_norm": 4.811293601989746, "learning_rate": 6.601300386609412e-06, "loss": 0.6909, "step": 2545 }, { "epoch": 0.622341725739428, "grad_norm": 4.38634729385376, "learning_rate": 6.593855431168739e-06, "loss": 0.593, "step": 2546 }, { "epoch": 0.6225861647518944, "grad_norm": 3.7089486122131348, "learning_rate": 6.5864126108342215e-06, "loss": 0.6747, "step": 2547 }, { "epoch": 0.6228306037643608, "grad_norm": 2.638113260269165, "learning_rate": 6.578971930271313e-06, "loss": 0.5938, "step": 2548 }, { "epoch": 0.6230750427768272, "grad_norm": 3.1624960899353027, "learning_rate": 6.571533394144127e-06, "loss": 0.6043, "step": 2549 }, { "epoch": 0.6233194817892935, "grad_norm": 3.2632787227630615, "learning_rate": 6.5640970071154356e-06, "loss": 0.6845, "step": 2550 }, { "epoch": 0.6235639208017599, "grad_norm": 3.290696382522583, "learning_rate": 6.556662773846658e-06, "loss": 0.7496, "step": 2551 }, { "epoch": 0.6238083598142263, "grad_norm": 3.2277283668518066, "learning_rate": 6.549230698997865e-06, "loss": 0.7165, "step": 2552 }, { "epoch": 0.6240527988266927, "grad_norm": 3.4584851264953613, "learning_rate": 6.541800787227776e-06, "loss": 0.7162, "step": 2553 }, { "epoch": 0.6242972378391591, "grad_norm": 2.970109462738037, "learning_rate": 6.534373043193751e-06, "loss": 0.661, "step": 2554 }, { "epoch": 0.6245416768516255, "grad_norm": 4.127514362335205, "learning_rate": 6.526947471551799e-06, "loss": 0.6906, "step": 2555 }, { "epoch": 0.6247861158640919, "grad_norm": 3.361604690551758, "learning_rate": 6.519524076956558e-06, "loss": 0.5777, "step": 2556 }, { "epoch": 0.6250305548765583, "grad_norm": 2.528740167617798, "learning_rate": 6.512102864061305e-06, "loss": 1.9661, "step": 2557 }, { "epoch": 0.6252749938890246, "grad_norm": 2.5687520503997803, "learning_rate": 6.504683837517953e-06, "loss": 0.5021, "step": 2558 }, { "epoch": 0.625519432901491, "grad_norm": 3.5585079193115234, "learning_rate": 6.497267001977037e-06, "loss": 0.6798, "step": 2559 }, { "epoch": 0.6257638719139574, "grad_norm": 3.9767425060272217, "learning_rate": 6.489852362087727e-06, "loss": 0.6911, "step": 2560 }, { "epoch": 0.6260083109264238, "grad_norm": 3.0334603786468506, "learning_rate": 6.48243992249781e-06, "loss": 0.5789, "step": 2561 }, { "epoch": 0.6262527499388902, "grad_norm": 2.5939929485321045, "learning_rate": 6.4750296878537004e-06, "loss": 0.5324, "step": 2562 }, { "epoch": 0.6264971889513566, "grad_norm": 2.6828372478485107, "learning_rate": 6.4676216628004204e-06, "loss": 0.5582, "step": 2563 }, { "epoch": 0.626741627963823, "grad_norm": 2.956088066101074, "learning_rate": 6.460215851981617e-06, "loss": 0.6962, "step": 2564 }, { "epoch": 0.6269860669762894, "grad_norm": 3.4469640254974365, "learning_rate": 6.452812260039542e-06, "loss": 0.8145, "step": 2565 }, { "epoch": 0.6272305059887558, "grad_norm": 2.3670003414154053, "learning_rate": 6.445410891615067e-06, "loss": 1.9115, "step": 2566 }, { "epoch": 0.6274749450012221, "grad_norm": 2.997681140899658, "learning_rate": 6.4380117513476545e-06, "loss": 0.5838, "step": 2567 }, { "epoch": 0.6277193840136885, "grad_norm": 2.9319982528686523, "learning_rate": 6.430614843875384e-06, "loss": 0.6775, "step": 2568 }, { "epoch": 0.6279638230261549, "grad_norm": 2.865135431289673, "learning_rate": 6.423220173834923e-06, "loss": 0.6572, "step": 2569 }, { "epoch": 0.6282082620386213, "grad_norm": 2.836655616760254, "learning_rate": 6.415827745861552e-06, "loss": 0.6656, "step": 2570 }, { "epoch": 0.6284527010510877, "grad_norm": 3.49206805229187, "learning_rate": 6.40843756458913e-06, "loss": 0.6767, "step": 2571 }, { "epoch": 0.6286971400635541, "grad_norm": 2.854118585586548, "learning_rate": 6.401049634650119e-06, "loss": 0.6471, "step": 2572 }, { "epoch": 0.6289415790760206, "grad_norm": 3.155752420425415, "learning_rate": 6.393663960675565e-06, "loss": 0.7015, "step": 2573 }, { "epoch": 0.629186018088487, "grad_norm": 2.570833206176758, "learning_rate": 6.386280547295099e-06, "loss": 0.5743, "step": 2574 }, { "epoch": 0.6294304571009534, "grad_norm": 2.7275097370147705, "learning_rate": 6.378899399136937e-06, "loss": 0.5756, "step": 2575 }, { "epoch": 0.6296748961134196, "grad_norm": 3.417999267578125, "learning_rate": 6.371520520827876e-06, "loss": 0.6533, "step": 2576 }, { "epoch": 0.629919335125886, "grad_norm": 2.962523937225342, "learning_rate": 6.3641439169932895e-06, "loss": 0.7356, "step": 2577 }, { "epoch": 0.6301637741383525, "grad_norm": 7.3755269050598145, "learning_rate": 6.356769592257122e-06, "loss": 0.7784, "step": 2578 }, { "epoch": 0.6304082131508189, "grad_norm": 2.9884088039398193, "learning_rate": 6.349397551241894e-06, "loss": 0.6326, "step": 2579 }, { "epoch": 0.6306526521632853, "grad_norm": 3.609339952468872, "learning_rate": 6.342027798568686e-06, "loss": 0.6522, "step": 2580 }, { "epoch": 0.6308970911757517, "grad_norm": 3.33388090133667, "learning_rate": 6.3346603388571605e-06, "loss": 0.7649, "step": 2581 }, { "epoch": 0.6311415301882181, "grad_norm": 2.7813782691955566, "learning_rate": 6.3272951767255284e-06, "loss": 0.731, "step": 2582 }, { "epoch": 0.6313859692006845, "grad_norm": 3.0105741024017334, "learning_rate": 6.319932316790565e-06, "loss": 0.8428, "step": 2583 }, { "epoch": 0.6316304082131509, "grad_norm": 3.0146050453186035, "learning_rate": 6.3125717636676005e-06, "loss": 0.6955, "step": 2584 }, { "epoch": 0.6318748472256172, "grad_norm": 2.816695213317871, "learning_rate": 6.305213521970523e-06, "loss": 0.7129, "step": 2585 }, { "epoch": 0.6321192862380836, "grad_norm": 2.3536274433135986, "learning_rate": 6.297857596311769e-06, "loss": 1.9621, "step": 2586 }, { "epoch": 0.63236372525055, "grad_norm": 3.4241769313812256, "learning_rate": 6.290503991302324e-06, "loss": 0.6107, "step": 2587 }, { "epoch": 0.6326081642630164, "grad_norm": 2.7386038303375244, "learning_rate": 6.283152711551714e-06, "loss": 0.6496, "step": 2588 }, { "epoch": 0.6328526032754828, "grad_norm": 5.59810209274292, "learning_rate": 6.275803761668019e-06, "loss": 0.6817, "step": 2589 }, { "epoch": 0.6330970422879492, "grad_norm": 2.6066410541534424, "learning_rate": 6.2684571462578465e-06, "loss": 0.703, "step": 2590 }, { "epoch": 0.6333414813004156, "grad_norm": 3.150466203689575, "learning_rate": 6.261112869926348e-06, "loss": 0.7259, "step": 2591 }, { "epoch": 0.633585920312882, "grad_norm": 2.94368052482605, "learning_rate": 6.253770937277206e-06, "loss": 0.6505, "step": 2592 }, { "epoch": 0.6338303593253484, "grad_norm": 3.3030643463134766, "learning_rate": 6.246431352912632e-06, "loss": 0.6825, "step": 2593 }, { "epoch": 0.6340747983378147, "grad_norm": 2.6119279861450195, "learning_rate": 6.239094121433371e-06, "loss": 0.6397, "step": 2594 }, { "epoch": 0.6343192373502811, "grad_norm": 3.283381700515747, "learning_rate": 6.231759247438689e-06, "loss": 0.7639, "step": 2595 }, { "epoch": 0.6345636763627475, "grad_norm": 2.4869697093963623, "learning_rate": 6.224426735526372e-06, "loss": 0.5081, "step": 2596 }, { "epoch": 0.6348081153752139, "grad_norm": 3.226597785949707, "learning_rate": 6.217096590292734e-06, "loss": 0.59, "step": 2597 }, { "epoch": 0.6350525543876803, "grad_norm": 2.7441442012786865, "learning_rate": 6.2097688163325974e-06, "loss": 0.5933, "step": 2598 }, { "epoch": 0.6352969934001467, "grad_norm": 2.381624937057495, "learning_rate": 6.202443418239301e-06, "loss": 1.9185, "step": 2599 }, { "epoch": 0.6355414324126131, "grad_norm": 2.9569995403289795, "learning_rate": 6.195120400604695e-06, "loss": 0.7283, "step": 2600 }, { "epoch": 0.6357858714250795, "grad_norm": 3.2580981254577637, "learning_rate": 6.187799768019134e-06, "loss": 0.5857, "step": 2601 }, { "epoch": 0.6360303104375459, "grad_norm": 2.321307897567749, "learning_rate": 6.180481525071482e-06, "loss": 1.9926, "step": 2602 }, { "epoch": 0.6362747494500122, "grad_norm": 3.270001173019409, "learning_rate": 6.173165676349103e-06, "loss": 0.6934, "step": 2603 }, { "epoch": 0.6365191884624786, "grad_norm": 3.450073003768921, "learning_rate": 6.165852226437861e-06, "loss": 0.6653, "step": 2604 }, { "epoch": 0.636763627474945, "grad_norm": 3.1501264572143555, "learning_rate": 6.158541179922113e-06, "loss": 0.8334, "step": 2605 }, { "epoch": 0.6370080664874114, "grad_norm": 4.02630615234375, "learning_rate": 6.1512325413847154e-06, "loss": 0.787, "step": 2606 }, { "epoch": 0.6372525054998778, "grad_norm": 2.958813190460205, "learning_rate": 6.143926315407011e-06, "loss": 0.646, "step": 2607 }, { "epoch": 0.6374969445123442, "grad_norm": 3.098252773284912, "learning_rate": 6.13662250656883e-06, "loss": 0.6546, "step": 2608 }, { "epoch": 0.6377413835248106, "grad_norm": 3.045335292816162, "learning_rate": 6.1293211194484895e-06, "loss": 0.6385, "step": 2609 }, { "epoch": 0.637985822537277, "grad_norm": 6.5176568031311035, "learning_rate": 6.122022158622787e-06, "loss": 0.6316, "step": 2610 }, { "epoch": 0.6382302615497434, "grad_norm": 2.843134880065918, "learning_rate": 6.114725628666997e-06, "loss": 0.7969, "step": 2611 }, { "epoch": 0.6384747005622097, "grad_norm": 2.9319586753845215, "learning_rate": 6.107431534154882e-06, "loss": 0.6795, "step": 2612 }, { "epoch": 0.6387191395746761, "grad_norm": 3.498922824859619, "learning_rate": 6.1001398796586644e-06, "loss": 0.668, "step": 2613 }, { "epoch": 0.6389635785871425, "grad_norm": 3.7721149921417236, "learning_rate": 6.092850669749038e-06, "loss": 0.7027, "step": 2614 }, { "epoch": 0.6392080175996089, "grad_norm": 3.661775827407837, "learning_rate": 6.085563908995174e-06, "loss": 0.7541, "step": 2615 }, { "epoch": 0.6394524566120753, "grad_norm": 2.810243606567383, "learning_rate": 6.0782796019646984e-06, "loss": 0.6738, "step": 2616 }, { "epoch": 0.6396968956245417, "grad_norm": 2.9646449089050293, "learning_rate": 6.070997753223703e-06, "loss": 0.5183, "step": 2617 }, { "epoch": 0.6399413346370081, "grad_norm": 3.056952714920044, "learning_rate": 6.063718367336739e-06, "loss": 0.5632, "step": 2618 }, { "epoch": 0.6401857736494745, "grad_norm": 2.522547483444214, "learning_rate": 6.056441448866817e-06, "loss": 0.7386, "step": 2619 }, { "epoch": 0.6404302126619409, "grad_norm": 2.832892894744873, "learning_rate": 6.0491670023753914e-06, "loss": 0.6423, "step": 2620 }, { "epoch": 0.6406746516744072, "grad_norm": 2.9263675212860107, "learning_rate": 6.041895032422377e-06, "loss": 0.638, "step": 2621 }, { "epoch": 0.6409190906868736, "grad_norm": 3.260471820831299, "learning_rate": 6.03462554356613e-06, "loss": 0.6096, "step": 2622 }, { "epoch": 0.64116352969934, "grad_norm": 3.9631731510162354, "learning_rate": 6.027358540363457e-06, "loss": 0.6807, "step": 2623 }, { "epoch": 0.6414079687118064, "grad_norm": 3.1918716430664062, "learning_rate": 6.020094027369598e-06, "loss": 0.7503, "step": 2624 }, { "epoch": 0.6416524077242728, "grad_norm": 2.8288021087646484, "learning_rate": 6.012832009138243e-06, "loss": 0.6821, "step": 2625 }, { "epoch": 0.6418968467367392, "grad_norm": 2.9716744422912598, "learning_rate": 6.005572490221507e-06, "loss": 0.6767, "step": 2626 }, { "epoch": 0.6421412857492056, "grad_norm": 4.20635986328125, "learning_rate": 5.998315475169942e-06, "loss": 0.6081, "step": 2627 }, { "epoch": 0.642385724761672, "grad_norm": 2.957223415374756, "learning_rate": 5.9910609685325415e-06, "loss": 0.7475, "step": 2628 }, { "epoch": 0.6426301637741384, "grad_norm": 2.925553798675537, "learning_rate": 5.983808974856711e-06, "loss": 0.6628, "step": 2629 }, { "epoch": 0.6428746027866047, "grad_norm": 3.361935615539551, "learning_rate": 5.976559498688288e-06, "loss": 0.6694, "step": 2630 }, { "epoch": 0.6431190417990711, "grad_norm": 2.64375376701355, "learning_rate": 5.969312544571529e-06, "loss": 0.5882, "step": 2631 }, { "epoch": 0.6433634808115375, "grad_norm": 3.661879777908325, "learning_rate": 5.962068117049115e-06, "loss": 0.5921, "step": 2632 }, { "epoch": 0.6436079198240039, "grad_norm": 3.00673770904541, "learning_rate": 5.9548262206621375e-06, "loss": 0.7094, "step": 2633 }, { "epoch": 0.6438523588364703, "grad_norm": 3.405564546585083, "learning_rate": 5.947586859950103e-06, "loss": 0.7097, "step": 2634 }, { "epoch": 0.6440967978489367, "grad_norm": 3.2848403453826904, "learning_rate": 5.94035003945093e-06, "loss": 0.7331, "step": 2635 }, { "epoch": 0.6443412368614031, "grad_norm": 3.589048385620117, "learning_rate": 5.9331157637009466e-06, "loss": 0.6206, "step": 2636 }, { "epoch": 0.6445856758738695, "grad_norm": 3.4497900009155273, "learning_rate": 5.925884037234879e-06, "loss": 0.6693, "step": 2637 }, { "epoch": 0.6448301148863359, "grad_norm": 3.1013028621673584, "learning_rate": 5.918654864585859e-06, "loss": 0.7106, "step": 2638 }, { "epoch": 0.6450745538988022, "grad_norm": 2.397865056991577, "learning_rate": 5.911428250285422e-06, "loss": 1.8904, "step": 2639 }, { "epoch": 0.6453189929112686, "grad_norm": 4.107816219329834, "learning_rate": 5.904204198863492e-06, "loss": 0.645, "step": 2640 }, { "epoch": 0.645563431923735, "grad_norm": 3.6357409954071045, "learning_rate": 5.8969827148483935e-06, "loss": 0.6926, "step": 2641 }, { "epoch": 0.6458078709362014, "grad_norm": 3.5307352542877197, "learning_rate": 5.889763802766835e-06, "loss": 0.6974, "step": 2642 }, { "epoch": 0.6460523099486678, "grad_norm": 2.94061279296875, "learning_rate": 5.88254746714392e-06, "loss": 0.6028, "step": 2643 }, { "epoch": 0.6462967489611342, "grad_norm": 2.9789211750030518, "learning_rate": 5.87533371250313e-06, "loss": 0.6147, "step": 2644 }, { "epoch": 0.6465411879736006, "grad_norm": 3.805901050567627, "learning_rate": 5.868122543366332e-06, "loss": 0.7588, "step": 2645 }, { "epoch": 0.646785626986067, "grad_norm": 2.642486095428467, "learning_rate": 5.860913964253774e-06, "loss": 0.6221, "step": 2646 }, { "epoch": 0.6470300659985334, "grad_norm": 2.385053873062134, "learning_rate": 5.853707979684077e-06, "loss": 1.9012, "step": 2647 }, { "epoch": 0.6472745050109997, "grad_norm": 3.222817897796631, "learning_rate": 5.846504594174236e-06, "loss": 0.6914, "step": 2648 }, { "epoch": 0.6475189440234661, "grad_norm": 2.6899101734161377, "learning_rate": 5.8393038122396205e-06, "loss": 0.6335, "step": 2649 }, { "epoch": 0.6477633830359325, "grad_norm": 2.827629327774048, "learning_rate": 5.832105638393968e-06, "loss": 0.4106, "step": 2650 }, { "epoch": 0.6480078220483989, "grad_norm": 3.3581056594848633, "learning_rate": 5.824910077149372e-06, "loss": 0.6776, "step": 2651 }, { "epoch": 0.6482522610608653, "grad_norm": 3.319446325302124, "learning_rate": 5.817717133016298e-06, "loss": 0.6217, "step": 2652 }, { "epoch": 0.6484967000733317, "grad_norm": 3.3250958919525146, "learning_rate": 5.810526810503567e-06, "loss": 0.638, "step": 2653 }, { "epoch": 0.6487411390857981, "grad_norm": 3.6977198123931885, "learning_rate": 5.80333911411836e-06, "loss": 0.6203, "step": 2654 }, { "epoch": 0.6489855780982645, "grad_norm": 2.956408739089966, "learning_rate": 5.796154048366209e-06, "loss": 0.5279, "step": 2655 }, { "epoch": 0.6492300171107309, "grad_norm": 3.3112738132476807, "learning_rate": 5.7889716177509945e-06, "loss": 0.6822, "step": 2656 }, { "epoch": 0.6494744561231972, "grad_norm": 3.304767370223999, "learning_rate": 5.7817918267749505e-06, "loss": 0.552, "step": 2657 }, { "epoch": 0.6497188951356636, "grad_norm": 3.368194103240967, "learning_rate": 5.774614679938651e-06, "loss": 0.7749, "step": 2658 }, { "epoch": 0.64996333414813, "grad_norm": 3.636293411254883, "learning_rate": 5.76744018174102e-06, "loss": 0.7058, "step": 2659 }, { "epoch": 0.6502077731605964, "grad_norm": 2.703385829925537, "learning_rate": 5.760268336679315e-06, "loss": 0.6515, "step": 2660 }, { "epoch": 0.6504522121730628, "grad_norm": 9.689297676086426, "learning_rate": 5.753099149249133e-06, "loss": 0.6111, "step": 2661 }, { "epoch": 0.6506966511855292, "grad_norm": 3.1031436920166016, "learning_rate": 5.7459326239444025e-06, "loss": 0.6971, "step": 2662 }, { "epoch": 0.6509410901979956, "grad_norm": 3.6608095169067383, "learning_rate": 5.738768765257386e-06, "loss": 0.6315, "step": 2663 }, { "epoch": 0.651185529210462, "grad_norm": 4.4993743896484375, "learning_rate": 5.731607577678675e-06, "loss": 0.6462, "step": 2664 }, { "epoch": 0.6514299682229284, "grad_norm": 3.131669282913208, "learning_rate": 5.724449065697182e-06, "loss": 0.6861, "step": 2665 }, { "epoch": 0.6516744072353947, "grad_norm": 4.2920756340026855, "learning_rate": 5.717293233800148e-06, "loss": 0.6796, "step": 2666 }, { "epoch": 0.6519188462478611, "grad_norm": 3.0195488929748535, "learning_rate": 5.7101400864731294e-06, "loss": 0.6626, "step": 2667 }, { "epoch": 0.6521632852603275, "grad_norm": 3.632920742034912, "learning_rate": 5.702989628200004e-06, "loss": 0.6894, "step": 2668 }, { "epoch": 0.6524077242727939, "grad_norm": 4.062261581420898, "learning_rate": 5.69584186346296e-06, "loss": 0.7573, "step": 2669 }, { "epoch": 0.6526521632852603, "grad_norm": 3.829195499420166, "learning_rate": 5.688696796742501e-06, "loss": 0.7282, "step": 2670 }, { "epoch": 0.6528966022977267, "grad_norm": 3.4981908798217773, "learning_rate": 5.681554432517435e-06, "loss": 0.5847, "step": 2671 }, { "epoch": 0.6531410413101931, "grad_norm": 2.9106080532073975, "learning_rate": 5.674414775264878e-06, "loss": 0.4843, "step": 2672 }, { "epoch": 0.6533854803226595, "grad_norm": 2.600187301635742, "learning_rate": 5.667277829460254e-06, "loss": 1.965, "step": 2673 }, { "epoch": 0.6536299193351259, "grad_norm": 2.7137207984924316, "learning_rate": 5.6601435995772814e-06, "loss": 0.6175, "step": 2674 }, { "epoch": 0.6538743583475922, "grad_norm": 3.5779716968536377, "learning_rate": 5.6530120900879774e-06, "loss": 0.6038, "step": 2675 }, { "epoch": 0.6541187973600586, "grad_norm": 3.0512759685516357, "learning_rate": 5.6458833054626514e-06, "loss": 0.6629, "step": 2676 }, { "epoch": 0.654363236372525, "grad_norm": 2.890873908996582, "learning_rate": 5.638757250169913e-06, "loss": 0.5274, "step": 2677 }, { "epoch": 0.6546076753849914, "grad_norm": 2.889688491821289, "learning_rate": 5.6316339286766496e-06, "loss": 0.609, "step": 2678 }, { "epoch": 0.6548521143974578, "grad_norm": 4.647675037384033, "learning_rate": 5.6245133454480446e-06, "loss": 0.6843, "step": 2679 }, { "epoch": 0.6550965534099242, "grad_norm": 3.1891238689422607, "learning_rate": 5.61739550494756e-06, "loss": 0.6044, "step": 2680 }, { "epoch": 0.6553409924223906, "grad_norm": 2.9441967010498047, "learning_rate": 5.610280411636941e-06, "loss": 0.6735, "step": 2681 }, { "epoch": 0.655585431434857, "grad_norm": 3.5355992317199707, "learning_rate": 5.603168069976209e-06, "loss": 0.6165, "step": 2682 }, { "epoch": 0.6558298704473234, "grad_norm": 3.233180522918701, "learning_rate": 5.5960584844236565e-06, "loss": 0.6039, "step": 2683 }, { "epoch": 0.6560743094597897, "grad_norm": 2.8662421703338623, "learning_rate": 5.588951659435861e-06, "loss": 0.6255, "step": 2684 }, { "epoch": 0.6563187484722561, "grad_norm": 3.419118642807007, "learning_rate": 5.581847599467658e-06, "loss": 0.7289, "step": 2685 }, { "epoch": 0.6565631874847225, "grad_norm": 3.948251485824585, "learning_rate": 5.574746308972153e-06, "loss": 0.5889, "step": 2686 }, { "epoch": 0.6568076264971889, "grad_norm": 2.250866174697876, "learning_rate": 5.5676477924007165e-06, "loss": 1.9108, "step": 2687 }, { "epoch": 0.6570520655096553, "grad_norm": 3.220052719116211, "learning_rate": 5.5605520542029776e-06, "loss": 0.6972, "step": 2688 }, { "epoch": 0.6572965045221217, "grad_norm": 3.2889387607574463, "learning_rate": 5.553459098826828e-06, "loss": 0.5567, "step": 2689 }, { "epoch": 0.6575409435345881, "grad_norm": 3.8833532333374023, "learning_rate": 5.546368930718411e-06, "loss": 0.6254, "step": 2690 }, { "epoch": 0.6577853825470545, "grad_norm": 4.197533130645752, "learning_rate": 5.539281554322126e-06, "loss": 0.7118, "step": 2691 }, { "epoch": 0.6580298215595209, "grad_norm": 3.916963815689087, "learning_rate": 5.532196974080619e-06, "loss": 0.6149, "step": 2692 }, { "epoch": 0.6582742605719872, "grad_norm": 2.631883144378662, "learning_rate": 5.5251151944347825e-06, "loss": 1.8987, "step": 2693 }, { "epoch": 0.6585186995844536, "grad_norm": 3.2310924530029297, "learning_rate": 5.518036219823765e-06, "loss": 0.6472, "step": 2694 }, { "epoch": 0.65876313859692, "grad_norm": 3.2843449115753174, "learning_rate": 5.510960054684942e-06, "loss": 0.7493, "step": 2695 }, { "epoch": 0.6590075776093864, "grad_norm": 3.5885109901428223, "learning_rate": 5.503886703453933e-06, "loss": 0.5283, "step": 2696 }, { "epoch": 0.6592520166218528, "grad_norm": 4.44291877746582, "learning_rate": 5.496816170564597e-06, "loss": 0.6773, "step": 2697 }, { "epoch": 0.6594964556343192, "grad_norm": 2.9985857009887695, "learning_rate": 5.4897484604490225e-06, "loss": 0.7804, "step": 2698 }, { "epoch": 0.6597408946467856, "grad_norm": 2.802274227142334, "learning_rate": 5.482683577537529e-06, "loss": 0.6248, "step": 2699 }, { "epoch": 0.659985333659252, "grad_norm": 3.206056594848633, "learning_rate": 5.475621526258665e-06, "loss": 0.6502, "step": 2700 }, { "epoch": 0.6602297726717185, "grad_norm": 3.777484655380249, "learning_rate": 5.468562311039205e-06, "loss": 0.6926, "step": 2701 }, { "epoch": 0.6604742116841847, "grad_norm": 3.5440683364868164, "learning_rate": 5.461505936304144e-06, "loss": 0.5646, "step": 2702 }, { "epoch": 0.6607186506966511, "grad_norm": 3.01408314704895, "learning_rate": 5.4544524064766944e-06, "loss": 0.5474, "step": 2703 }, { "epoch": 0.6609630897091175, "grad_norm": 4.224313735961914, "learning_rate": 5.4474017259782866e-06, "loss": 0.6911, "step": 2704 }, { "epoch": 0.661207528721584, "grad_norm": 3.4431660175323486, "learning_rate": 5.440353899228572e-06, "loss": 0.6834, "step": 2705 }, { "epoch": 0.6614519677340504, "grad_norm": 3.106137990951538, "learning_rate": 5.4333089306454045e-06, "loss": 0.5567, "step": 2706 }, { "epoch": 0.6616964067465168, "grad_norm": 3.4715335369110107, "learning_rate": 5.426266824644848e-06, "loss": 0.7525, "step": 2707 }, { "epoch": 0.6619408457589832, "grad_norm": 2.917834758758545, "learning_rate": 5.419227585641174e-06, "loss": 0.462, "step": 2708 }, { "epoch": 0.6621852847714496, "grad_norm": 4.073702812194824, "learning_rate": 5.412191218046856e-06, "loss": 0.5448, "step": 2709 }, { "epoch": 0.662429723783916, "grad_norm": 2.8708159923553467, "learning_rate": 5.405157726272567e-06, "loss": 0.6345, "step": 2710 }, { "epoch": 0.6626741627963822, "grad_norm": 2.726679563522339, "learning_rate": 5.3981271147271786e-06, "loss": 0.7013, "step": 2711 }, { "epoch": 0.6629186018088487, "grad_norm": 2.372623920440674, "learning_rate": 5.391099387817756e-06, "loss": 0.6221, "step": 2712 }, { "epoch": 0.663163040821315, "grad_norm": 2.783414363861084, "learning_rate": 5.384074549949554e-06, "loss": 0.6144, "step": 2713 }, { "epoch": 0.6634074798337815, "grad_norm": 3.0188844203948975, "learning_rate": 5.377052605526023e-06, "loss": 0.73, "step": 2714 }, { "epoch": 0.6636519188462479, "grad_norm": 2.9153764247894287, "learning_rate": 5.370033558948793e-06, "loss": 0.5721, "step": 2715 }, { "epoch": 0.6638963578587143, "grad_norm": 2.798769235610962, "learning_rate": 5.363017414617682e-06, "loss": 0.6387, "step": 2716 }, { "epoch": 0.6641407968711807, "grad_norm": 2.9460997581481934, "learning_rate": 5.356004176930683e-06, "loss": 0.6611, "step": 2717 }, { "epoch": 0.6643852358836471, "grad_norm": 2.4899098873138428, "learning_rate": 5.348993850283974e-06, "loss": 1.9264, "step": 2718 }, { "epoch": 0.6646296748961135, "grad_norm": 3.263526439666748, "learning_rate": 5.341986439071907e-06, "loss": 0.6537, "step": 2719 }, { "epoch": 0.6648741139085798, "grad_norm": 3.4318578243255615, "learning_rate": 5.334981947686995e-06, "loss": 0.6641, "step": 2720 }, { "epoch": 0.6651185529210462, "grad_norm": 3.2904953956604004, "learning_rate": 5.327980380519942e-06, "loss": 0.7114, "step": 2721 }, { "epoch": 0.6653629919335126, "grad_norm": 2.738497734069824, "learning_rate": 5.320981741959603e-06, "loss": 0.53, "step": 2722 }, { "epoch": 0.665607430945979, "grad_norm": 6.3192267417907715, "learning_rate": 5.313986036393e-06, "loss": 0.6404, "step": 2723 }, { "epoch": 0.6658518699584454, "grad_norm": 2.9573323726654053, "learning_rate": 5.306993268205319e-06, "loss": 0.5863, "step": 2724 }, { "epoch": 0.6660963089709118, "grad_norm": 3.060486078262329, "learning_rate": 5.300003441779904e-06, "loss": 0.6007, "step": 2725 }, { "epoch": 0.6663407479833782, "grad_norm": 3.0728816986083984, "learning_rate": 5.293016561498255e-06, "loss": 0.7026, "step": 2726 }, { "epoch": 0.6665851869958446, "grad_norm": 2.4286727905273438, "learning_rate": 5.286032631740023e-06, "loss": 1.9812, "step": 2727 }, { "epoch": 0.666829626008311, "grad_norm": 2.3266477584838867, "learning_rate": 5.279051656883014e-06, "loss": 1.9323, "step": 2728 }, { "epoch": 0.6670740650207773, "grad_norm": 4.493717670440674, "learning_rate": 5.272073641303177e-06, "loss": 0.7852, "step": 2729 }, { "epoch": 0.6673185040332437, "grad_norm": 5.459487438201904, "learning_rate": 5.26509858937461e-06, "loss": 0.7235, "step": 2730 }, { "epoch": 0.6675629430457101, "grad_norm": 3.1351940631866455, "learning_rate": 5.25812650546955e-06, "loss": 0.5807, "step": 2731 }, { "epoch": 0.6678073820581765, "grad_norm": 3.1465470790863037, "learning_rate": 5.251157393958376e-06, "loss": 0.6398, "step": 2732 }, { "epoch": 0.6680518210706429, "grad_norm": 2.907792329788208, "learning_rate": 5.244191259209602e-06, "loss": 0.5835, "step": 2733 }, { "epoch": 0.6682962600831093, "grad_norm": 2.9785854816436768, "learning_rate": 5.237228105589877e-06, "loss": 0.6426, "step": 2734 }, { "epoch": 0.6685406990955757, "grad_norm": 3.418694257736206, "learning_rate": 5.230267937463977e-06, "loss": 0.6216, "step": 2735 }, { "epoch": 0.6687851381080421, "grad_norm": 2.9922711849212646, "learning_rate": 5.223310759194822e-06, "loss": 0.7244, "step": 2736 }, { "epoch": 0.6690295771205085, "grad_norm": 2.6697285175323486, "learning_rate": 5.216356575143436e-06, "loss": 0.5847, "step": 2737 }, { "epoch": 0.6692740161329748, "grad_norm": 2.942906141281128, "learning_rate": 5.209405389668982e-06, "loss": 0.6909, "step": 2738 }, { "epoch": 0.6695184551454412, "grad_norm": 3.1385350227355957, "learning_rate": 5.202457207128736e-06, "loss": 0.5378, "step": 2739 }, { "epoch": 0.6697628941579076, "grad_norm": 3.10185170173645, "learning_rate": 5.195512031878093e-06, "loss": 0.5174, "step": 2740 }, { "epoch": 0.670007333170374, "grad_norm": 3.0209953784942627, "learning_rate": 5.188569868270566e-06, "loss": 0.6357, "step": 2741 }, { "epoch": 0.6702517721828404, "grad_norm": 3.6313841342926025, "learning_rate": 5.181630720657776e-06, "loss": 0.6426, "step": 2742 }, { "epoch": 0.6704962111953068, "grad_norm": 3.3282594680786133, "learning_rate": 5.1746945933894555e-06, "loss": 0.6618, "step": 2743 }, { "epoch": 0.6707406502077732, "grad_norm": 2.451758623123169, "learning_rate": 5.167761490813444e-06, "loss": 0.5011, "step": 2744 }, { "epoch": 0.6709850892202396, "grad_norm": 3.1366803646087646, "learning_rate": 5.160831417275683e-06, "loss": 0.636, "step": 2745 }, { "epoch": 0.671229528232706, "grad_norm": 2.826927661895752, "learning_rate": 5.153904377120218e-06, "loss": 0.5909, "step": 2746 }, { "epoch": 0.6714739672451723, "grad_norm": 2.7017712593078613, "learning_rate": 5.146980374689192e-06, "loss": 0.5149, "step": 2747 }, { "epoch": 0.6717184062576387, "grad_norm": 2.9147655963897705, "learning_rate": 5.140059414322842e-06, "loss": 0.6561, "step": 2748 }, { "epoch": 0.6719628452701051, "grad_norm": 3.0260162353515625, "learning_rate": 5.133141500359502e-06, "loss": 0.672, "step": 2749 }, { "epoch": 0.6722072842825715, "grad_norm": 3.405888557434082, "learning_rate": 5.1262266371355914e-06, "loss": 0.6075, "step": 2750 }, { "epoch": 0.6724517232950379, "grad_norm": 3.4034671783447266, "learning_rate": 5.11931482898562e-06, "loss": 0.7385, "step": 2751 }, { "epoch": 0.6726961623075043, "grad_norm": 3.760180950164795, "learning_rate": 5.112406080242186e-06, "loss": 0.6619, "step": 2752 }, { "epoch": 0.6729406013199707, "grad_norm": 3.1953327655792236, "learning_rate": 5.105500395235965e-06, "loss": 0.6422, "step": 2753 }, { "epoch": 0.6731850403324371, "grad_norm": 3.0213558673858643, "learning_rate": 5.098597778295713e-06, "loss": 0.51, "step": 2754 }, { "epoch": 0.6734294793449035, "grad_norm": 3.2140281200408936, "learning_rate": 5.091698233748264e-06, "loss": 0.6254, "step": 2755 }, { "epoch": 0.6736739183573698, "grad_norm": 2.8674018383026123, "learning_rate": 5.0848017659185254e-06, "loss": 0.6284, "step": 2756 }, { "epoch": 0.6739183573698362, "grad_norm": 3.3768415451049805, "learning_rate": 5.077908379129475e-06, "loss": 0.6839, "step": 2757 }, { "epoch": 0.6741627963823026, "grad_norm": 3.359588861465454, "learning_rate": 5.071018077702161e-06, "loss": 0.658, "step": 2758 }, { "epoch": 0.674407235394769, "grad_norm": 2.669429063796997, "learning_rate": 5.064130865955696e-06, "loss": 0.5982, "step": 2759 }, { "epoch": 0.6746516744072354, "grad_norm": 2.9091298580169678, "learning_rate": 5.057246748207258e-06, "loss": 0.5159, "step": 2760 }, { "epoch": 0.6748961134197018, "grad_norm": 3.143890857696533, "learning_rate": 5.050365728772084e-06, "loss": 0.6629, "step": 2761 }, { "epoch": 0.6751405524321682, "grad_norm": 3.617204189300537, "learning_rate": 5.043487811963469e-06, "loss": 0.7287, "step": 2762 }, { "epoch": 0.6753849914446346, "grad_norm": 3.3309438228607178, "learning_rate": 5.0366130020927625e-06, "loss": 0.6314, "step": 2763 }, { "epoch": 0.675629430457101, "grad_norm": 4.037367343902588, "learning_rate": 5.02974130346937e-06, "loss": 0.6331, "step": 2764 }, { "epoch": 0.6758738694695673, "grad_norm": 2.586611747741699, "learning_rate": 5.022872720400741e-06, "loss": 0.5245, "step": 2765 }, { "epoch": 0.6761183084820337, "grad_norm": 3.3242573738098145, "learning_rate": 5.0160072571923725e-06, "loss": 0.6345, "step": 2766 }, { "epoch": 0.6763627474945001, "grad_norm": 2.9784135818481445, "learning_rate": 5.009144918147817e-06, "loss": 0.6177, "step": 2767 }, { "epoch": 0.6766071865069665, "grad_norm": 3.950590133666992, "learning_rate": 5.002285707568653e-06, "loss": 0.5661, "step": 2768 }, { "epoch": 0.6768516255194329, "grad_norm": 3.056154251098633, "learning_rate": 4.995429629754509e-06, "loss": 0.5672, "step": 2769 }, { "epoch": 0.6770960645318993, "grad_norm": 2.899587392807007, "learning_rate": 4.988576689003044e-06, "loss": 0.5638, "step": 2770 }, { "epoch": 0.6773405035443657, "grad_norm": 3.3658998012542725, "learning_rate": 4.981726889609952e-06, "loss": 0.6199, "step": 2771 }, { "epoch": 0.6775849425568321, "grad_norm": 3.157362699508667, "learning_rate": 4.974880235868959e-06, "loss": 0.6285, "step": 2772 }, { "epoch": 0.6778293815692985, "grad_norm": 3.2186131477355957, "learning_rate": 4.968036732071818e-06, "loss": 0.6394, "step": 2773 }, { "epoch": 0.6780738205817648, "grad_norm": 3.415682077407837, "learning_rate": 4.9611963825083085e-06, "loss": 0.5819, "step": 2774 }, { "epoch": 0.6783182595942312, "grad_norm": 3.1678550243377686, "learning_rate": 4.954359191466232e-06, "loss": 0.6095, "step": 2775 }, { "epoch": 0.6785626986066976, "grad_norm": 3.0142929553985596, "learning_rate": 4.947525163231408e-06, "loss": 0.615, "step": 2776 }, { "epoch": 0.678807137619164, "grad_norm": 2.398094415664673, "learning_rate": 4.940694302087679e-06, "loss": 1.9171, "step": 2777 }, { "epoch": 0.6790515766316304, "grad_norm": 2.8858699798583984, "learning_rate": 4.933866612316898e-06, "loss": 0.6783, "step": 2778 }, { "epoch": 0.6792960156440968, "grad_norm": 3.315502166748047, "learning_rate": 4.9270420981989295e-06, "loss": 0.6088, "step": 2779 }, { "epoch": 0.6795404546565632, "grad_norm": 2.7693212032318115, "learning_rate": 4.92022076401165e-06, "loss": 0.6648, "step": 2780 }, { "epoch": 0.6797848936690296, "grad_norm": 3.115082263946533, "learning_rate": 4.913402614030944e-06, "loss": 0.5976, "step": 2781 }, { "epoch": 0.680029332681496, "grad_norm": 3.340075969696045, "learning_rate": 4.9065876525306884e-06, "loss": 0.7053, "step": 2782 }, { "epoch": 0.6802737716939623, "grad_norm": 2.9971578121185303, "learning_rate": 4.899775883782784e-06, "loss": 0.6259, "step": 2783 }, { "epoch": 0.6805182107064287, "grad_norm": 2.8590915203094482, "learning_rate": 4.89296731205711e-06, "loss": 0.6496, "step": 2784 }, { "epoch": 0.6807626497188951, "grad_norm": 3.2005624771118164, "learning_rate": 4.8861619416215475e-06, "loss": 0.5618, "step": 2785 }, { "epoch": 0.6810070887313615, "grad_norm": 5.067270278930664, "learning_rate": 4.8793597767419755e-06, "loss": 0.5195, "step": 2786 }, { "epoch": 0.6812515277438279, "grad_norm": 2.8245551586151123, "learning_rate": 4.872560821682256e-06, "loss": 0.5953, "step": 2787 }, { "epoch": 0.6814959667562943, "grad_norm": 2.979853630065918, "learning_rate": 4.8657650807042465e-06, "loss": 0.6692, "step": 2788 }, { "epoch": 0.6817404057687607, "grad_norm": 3.309493064880371, "learning_rate": 4.858972558067784e-06, "loss": 0.5598, "step": 2789 }, { "epoch": 0.6819848447812271, "grad_norm": 3.0199952125549316, "learning_rate": 4.852183258030691e-06, "loss": 0.6, "step": 2790 }, { "epoch": 0.6822292837936935, "grad_norm": 3.136849880218506, "learning_rate": 4.84539718484877e-06, "loss": 0.5988, "step": 2791 }, { "epoch": 0.6824737228061598, "grad_norm": 3.2955336570739746, "learning_rate": 4.8386143427757995e-06, "loss": 0.6135, "step": 2792 }, { "epoch": 0.6827181618186262, "grad_norm": 4.007350921630859, "learning_rate": 4.831834736063533e-06, "loss": 0.4945, "step": 2793 }, { "epoch": 0.6829626008310926, "grad_norm": 3.997565507888794, "learning_rate": 4.825058368961696e-06, "loss": 0.7764, "step": 2794 }, { "epoch": 0.683207039843559, "grad_norm": 3.0956525802612305, "learning_rate": 4.818285245717984e-06, "loss": 0.6211, "step": 2795 }, { "epoch": 0.6834514788560254, "grad_norm": 3.000917911529541, "learning_rate": 4.811515370578059e-06, "loss": 0.5148, "step": 2796 }, { "epoch": 0.6836959178684918, "grad_norm": 3.562018394470215, "learning_rate": 4.804748747785542e-06, "loss": 0.5863, "step": 2797 }, { "epoch": 0.6839403568809582, "grad_norm": 3.1512999534606934, "learning_rate": 4.7979853815820275e-06, "loss": 0.6411, "step": 2798 }, { "epoch": 0.6841847958934246, "grad_norm": 3.354006290435791, "learning_rate": 4.791225276207058e-06, "loss": 0.5797, "step": 2799 }, { "epoch": 0.684429234905891, "grad_norm": 2.866476535797119, "learning_rate": 4.784468435898134e-06, "loss": 0.6366, "step": 2800 }, { "epoch": 0.6846736739183573, "grad_norm": 2.913511037826538, "learning_rate": 4.77771486489071e-06, "loss": 0.6318, "step": 2801 }, { "epoch": 0.6849181129308237, "grad_norm": 2.909332036972046, "learning_rate": 4.7709645674181925e-06, "loss": 0.6117, "step": 2802 }, { "epoch": 0.6851625519432901, "grad_norm": 2.971118450164795, "learning_rate": 4.764217547711935e-06, "loss": 0.6834, "step": 2803 }, { "epoch": 0.6854069909557565, "grad_norm": 2.961076021194458, "learning_rate": 4.757473810001235e-06, "loss": 0.6081, "step": 2804 }, { "epoch": 0.6856514299682229, "grad_norm": 3.154080867767334, "learning_rate": 4.750733358513335e-06, "loss": 0.6245, "step": 2805 }, { "epoch": 0.6858958689806893, "grad_norm": 3.5119946002960205, "learning_rate": 4.743996197473417e-06, "loss": 0.6834, "step": 2806 }, { "epoch": 0.6861403079931557, "grad_norm": 3.775698661804199, "learning_rate": 4.7372623311045995e-06, "loss": 0.6074, "step": 2807 }, { "epoch": 0.6863847470056221, "grad_norm": 2.907153606414795, "learning_rate": 4.730531763627937e-06, "loss": 0.6108, "step": 2808 }, { "epoch": 0.6866291860180885, "grad_norm": 2.9769084453582764, "learning_rate": 4.723804499262414e-06, "loss": 0.5833, "step": 2809 }, { "epoch": 0.6868736250305548, "grad_norm": 4.168307781219482, "learning_rate": 4.7170805422249474e-06, "loss": 0.6263, "step": 2810 }, { "epoch": 0.6871180640430212, "grad_norm": 3.363607883453369, "learning_rate": 4.710359896730379e-06, "loss": 0.4936, "step": 2811 }, { "epoch": 0.6873625030554876, "grad_norm": 3.9064033031463623, "learning_rate": 4.7036425669914756e-06, "loss": 0.6862, "step": 2812 }, { "epoch": 0.687606942067954, "grad_norm": 2.661390542984009, "learning_rate": 4.696928557218919e-06, "loss": 0.5281, "step": 2813 }, { "epoch": 0.6878513810804204, "grad_norm": 3.045010566711426, "learning_rate": 4.690217871621327e-06, "loss": 0.5213, "step": 2814 }, { "epoch": 0.6880958200928868, "grad_norm": 3.699244976043701, "learning_rate": 4.683510514405219e-06, "loss": 0.6995, "step": 2815 }, { "epoch": 0.6883402591053532, "grad_norm": 3.2344930171966553, "learning_rate": 4.676806489775028e-06, "loss": 0.7577, "step": 2816 }, { "epoch": 0.6885846981178196, "grad_norm": 2.8136589527130127, "learning_rate": 4.670105801933106e-06, "loss": 0.6066, "step": 2817 }, { "epoch": 0.688829137130286, "grad_norm": 3.5094377994537354, "learning_rate": 4.663408455079707e-06, "loss": 0.6551, "step": 2818 }, { "epoch": 0.6890735761427523, "grad_norm": 3.238950490951538, "learning_rate": 4.6567144534129926e-06, "loss": 0.655, "step": 2819 }, { "epoch": 0.6893180151552187, "grad_norm": 2.555016040802002, "learning_rate": 4.65002380112903e-06, "loss": 1.9487, "step": 2820 }, { "epoch": 0.6895624541676851, "grad_norm": 3.152003049850464, "learning_rate": 4.643336502421783e-06, "loss": 0.5787, "step": 2821 }, { "epoch": 0.6898068931801515, "grad_norm": 2.7996885776519775, "learning_rate": 4.636652561483116e-06, "loss": 0.6433, "step": 2822 }, { "epoch": 0.6900513321926179, "grad_norm": 4.192671775817871, "learning_rate": 4.629971982502786e-06, "loss": 0.5745, "step": 2823 }, { "epoch": 0.6902957712050843, "grad_norm": 3.3025522232055664, "learning_rate": 4.623294769668445e-06, "loss": 0.6802, "step": 2824 }, { "epoch": 0.6905402102175507, "grad_norm": 3.519629955291748, "learning_rate": 4.616620927165635e-06, "loss": 0.647, "step": 2825 }, { "epoch": 0.6907846492300171, "grad_norm": 3.209341049194336, "learning_rate": 4.6099504591777845e-06, "loss": 0.6831, "step": 2826 }, { "epoch": 0.6910290882424835, "grad_norm": 2.624837636947632, "learning_rate": 4.603283369886205e-06, "loss": 0.5011, "step": 2827 }, { "epoch": 0.6912735272549498, "grad_norm": 3.348609685897827, "learning_rate": 4.596619663470089e-06, "loss": 0.6547, "step": 2828 }, { "epoch": 0.6915179662674162, "grad_norm": 4.262418746948242, "learning_rate": 4.5899593441065195e-06, "loss": 0.6114, "step": 2829 }, { "epoch": 0.6917624052798826, "grad_norm": 3.3939056396484375, "learning_rate": 4.583302415970443e-06, "loss": 0.6676, "step": 2830 }, { "epoch": 0.692006844292349, "grad_norm": 2.8519091606140137, "learning_rate": 4.576648883234686e-06, "loss": 0.5483, "step": 2831 }, { "epoch": 0.6922512833048154, "grad_norm": 3.068394184112549, "learning_rate": 4.5699987500699465e-06, "loss": 0.7529, "step": 2832 }, { "epoch": 0.6924957223172818, "grad_norm": 3.2849552631378174, "learning_rate": 4.563352020644791e-06, "loss": 0.7754, "step": 2833 }, { "epoch": 0.6927401613297482, "grad_norm": 3.2050352096557617, "learning_rate": 4.556708699125652e-06, "loss": 0.5716, "step": 2834 }, { "epoch": 0.6929846003422147, "grad_norm": 2.885378360748291, "learning_rate": 4.550068789676826e-06, "loss": 0.6766, "step": 2835 }, { "epoch": 0.693229039354681, "grad_norm": 3.097829818725586, "learning_rate": 4.5434322964604715e-06, "loss": 0.6388, "step": 2836 }, { "epoch": 0.6934734783671473, "grad_norm": 3.3307607173919678, "learning_rate": 4.536799223636604e-06, "loss": 0.6683, "step": 2837 }, { "epoch": 0.6937179173796137, "grad_norm": 3.108682155609131, "learning_rate": 4.530169575363098e-06, "loss": 0.5499, "step": 2838 }, { "epoch": 0.6939623563920801, "grad_norm": 2.598700761795044, "learning_rate": 4.523543355795676e-06, "loss": 0.5555, "step": 2839 }, { "epoch": 0.6942067954045466, "grad_norm": 2.4821794033050537, "learning_rate": 4.516920569087918e-06, "loss": 0.535, "step": 2840 }, { "epoch": 0.694451234417013, "grad_norm": 2.8600430488586426, "learning_rate": 4.510301219391245e-06, "loss": 0.6832, "step": 2841 }, { "epoch": 0.6946956734294794, "grad_norm": 3.905129909515381, "learning_rate": 4.503685310854931e-06, "loss": 0.7206, "step": 2842 }, { "epoch": 0.6949401124419458, "grad_norm": 2.983491897583008, "learning_rate": 4.497072847626087e-06, "loss": 0.6094, "step": 2843 }, { "epoch": 0.6951845514544122, "grad_norm": 2.955796003341675, "learning_rate": 4.490463833849669e-06, "loss": 0.5527, "step": 2844 }, { "epoch": 0.6954289904668786, "grad_norm": 3.2205801010131836, "learning_rate": 4.483858273668464e-06, "loss": 0.6138, "step": 2845 }, { "epoch": 0.6956734294793449, "grad_norm": 2.5980865955352783, "learning_rate": 4.477256171223103e-06, "loss": 1.8764, "step": 2846 }, { "epoch": 0.6959178684918113, "grad_norm": 3.737715482711792, "learning_rate": 4.470657530652042e-06, "loss": 0.6366, "step": 2847 }, { "epoch": 0.6961623075042777, "grad_norm": 2.8147330284118652, "learning_rate": 4.464062356091571e-06, "loss": 0.4609, "step": 2848 }, { "epoch": 0.6964067465167441, "grad_norm": 2.979423999786377, "learning_rate": 4.457470651675806e-06, "loss": 0.5153, "step": 2849 }, { "epoch": 0.6966511855292105, "grad_norm": 3.7111847400665283, "learning_rate": 4.450882421536688e-06, "loss": 0.7031, "step": 2850 }, { "epoch": 0.6968956245416769, "grad_norm": 3.0900397300720215, "learning_rate": 4.444297669803981e-06, "loss": 0.7322, "step": 2851 }, { "epoch": 0.6971400635541433, "grad_norm": 3.054860830307007, "learning_rate": 4.437716400605263e-06, "loss": 0.5775, "step": 2852 }, { "epoch": 0.6973845025666097, "grad_norm": 3.381854772567749, "learning_rate": 4.431138618065942e-06, "loss": 0.7313, "step": 2853 }, { "epoch": 0.6976289415790761, "grad_norm": 7.616814613342285, "learning_rate": 4.424564326309228e-06, "loss": 0.5338, "step": 2854 }, { "epoch": 0.6978733805915424, "grad_norm": 2.912473440170288, "learning_rate": 4.417993529456148e-06, "loss": 0.536, "step": 2855 }, { "epoch": 0.6981178196040088, "grad_norm": 2.8949167728424072, "learning_rate": 4.411426231625537e-06, "loss": 0.6394, "step": 2856 }, { "epoch": 0.6983622586164752, "grad_norm": 3.7534406185150146, "learning_rate": 4.404862436934037e-06, "loss": 0.6242, "step": 2857 }, { "epoch": 0.6986066976289416, "grad_norm": 2.8737173080444336, "learning_rate": 4.3983021494960944e-06, "loss": 0.6676, "step": 2858 }, { "epoch": 0.698851136641408, "grad_norm": 2.802374839782715, "learning_rate": 4.391745373423957e-06, "loss": 0.6221, "step": 2859 }, { "epoch": 0.6990955756538744, "grad_norm": 2.5841798782348633, "learning_rate": 4.38519211282767e-06, "loss": 0.5938, "step": 2860 }, { "epoch": 0.6993400146663408, "grad_norm": 3.130535364151001, "learning_rate": 4.378642371815078e-06, "loss": 0.5083, "step": 2861 }, { "epoch": 0.6995844536788072, "grad_norm": 2.743894338607788, "learning_rate": 4.372096154491814e-06, "loss": 0.6127, "step": 2862 }, { "epoch": 0.6998288926912736, "grad_norm": 4.366552352905273, "learning_rate": 4.365553464961313e-06, "loss": 0.6796, "step": 2863 }, { "epoch": 0.7000733317037399, "grad_norm": 3.4443304538726807, "learning_rate": 4.359014307324787e-06, "loss": 0.5831, "step": 2864 }, { "epoch": 0.7003177707162063, "grad_norm": 3.2634642124176025, "learning_rate": 4.3524786856812406e-06, "loss": 0.6987, "step": 2865 }, { "epoch": 0.7005622097286727, "grad_norm": 5.027645587921143, "learning_rate": 4.3459466041274575e-06, "loss": 0.6459, "step": 2866 }, { "epoch": 0.7008066487411391, "grad_norm": 3.121417284011841, "learning_rate": 4.339418066758008e-06, "loss": 0.6671, "step": 2867 }, { "epoch": 0.7010510877536055, "grad_norm": 3.2320797443389893, "learning_rate": 4.332893077665234e-06, "loss": 0.6006, "step": 2868 }, { "epoch": 0.7012955267660719, "grad_norm": 3.52091646194458, "learning_rate": 4.32637164093926e-06, "loss": 0.7388, "step": 2869 }, { "epoch": 0.7015399657785383, "grad_norm": 2.5063316822052, "learning_rate": 4.319853760667978e-06, "loss": 1.8042, "step": 2870 }, { "epoch": 0.7017844047910047, "grad_norm": 2.862105131149292, "learning_rate": 4.313339440937055e-06, "loss": 0.51, "step": 2871 }, { "epoch": 0.7020288438034711, "grad_norm": 2.7071707248687744, "learning_rate": 4.3068286858299235e-06, "loss": 0.609, "step": 2872 }, { "epoch": 0.7022732828159374, "grad_norm": 3.4979820251464844, "learning_rate": 4.300321499427782e-06, "loss": 0.5811, "step": 2873 }, { "epoch": 0.7025177218284038, "grad_norm": 2.958627223968506, "learning_rate": 4.293817885809594e-06, "loss": 0.5751, "step": 2874 }, { "epoch": 0.7027621608408702, "grad_norm": 3.80709171295166, "learning_rate": 4.287317849052075e-06, "loss": 0.6346, "step": 2875 }, { "epoch": 0.7030065998533366, "grad_norm": 3.485647678375244, "learning_rate": 4.280821393229716e-06, "loss": 0.5961, "step": 2876 }, { "epoch": 0.703251038865803, "grad_norm": 3.2261605262756348, "learning_rate": 4.274328522414747e-06, "loss": 0.5644, "step": 2877 }, { "epoch": 0.7034954778782694, "grad_norm": 2.664689779281616, "learning_rate": 4.2678392406771555e-06, "loss": 0.4838, "step": 2878 }, { "epoch": 0.7037399168907358, "grad_norm": 3.4998104572296143, "learning_rate": 4.261353552084682e-06, "loss": 0.5976, "step": 2879 }, { "epoch": 0.7039843559032022, "grad_norm": 3.702908515930176, "learning_rate": 4.254871460702811e-06, "loss": 0.6867, "step": 2880 }, { "epoch": 0.7042287949156686, "grad_norm": 3.084543466567993, "learning_rate": 4.248392970594774e-06, "loss": 0.6221, "step": 2881 }, { "epoch": 0.7044732339281349, "grad_norm": 3.0470752716064453, "learning_rate": 4.241918085821547e-06, "loss": 0.6252, "step": 2882 }, { "epoch": 0.7047176729406013, "grad_norm": 2.600609540939331, "learning_rate": 4.2354468104418415e-06, "loss": 1.8867, "step": 2883 }, { "epoch": 0.7049621119530677, "grad_norm": 4.214003562927246, "learning_rate": 4.228979148512109e-06, "loss": 0.7683, "step": 2884 }, { "epoch": 0.7052065509655341, "grad_norm": 6.833065032958984, "learning_rate": 4.2225151040865375e-06, "loss": 0.659, "step": 2885 }, { "epoch": 0.7054509899780005, "grad_norm": 3.0870535373687744, "learning_rate": 4.216054681217044e-06, "loss": 0.5151, "step": 2886 }, { "epoch": 0.7056954289904669, "grad_norm": 2.8286168575286865, "learning_rate": 4.209597883953278e-06, "loss": 0.4955, "step": 2887 }, { "epoch": 0.7059398680029333, "grad_norm": 2.5923023223876953, "learning_rate": 4.203144716342617e-06, "loss": 1.9329, "step": 2888 }, { "epoch": 0.7061843070153997, "grad_norm": 3.287935733795166, "learning_rate": 4.196695182430157e-06, "loss": 0.5925, "step": 2889 }, { "epoch": 0.7064287460278661, "grad_norm": 3.587629556655884, "learning_rate": 4.190249286258723e-06, "loss": 0.7135, "step": 2890 }, { "epoch": 0.7066731850403324, "grad_norm": 2.9305241107940674, "learning_rate": 4.18380703186886e-06, "loss": 0.5792, "step": 2891 }, { "epoch": 0.7069176240527988, "grad_norm": 3.006047248840332, "learning_rate": 4.1773684232988285e-06, "loss": 0.5638, "step": 2892 }, { "epoch": 0.7071620630652652, "grad_norm": 2.8415372371673584, "learning_rate": 4.170933464584601e-06, "loss": 0.5779, "step": 2893 }, { "epoch": 0.7074065020777316, "grad_norm": 3.2072343826293945, "learning_rate": 4.1645021597598635e-06, "loss": 0.6026, "step": 2894 }, { "epoch": 0.707650941090198, "grad_norm": 2.8006460666656494, "learning_rate": 4.158074512856014e-06, "loss": 0.5016, "step": 2895 }, { "epoch": 0.7078953801026644, "grad_norm": 3.4516077041625977, "learning_rate": 4.151650527902155e-06, "loss": 0.6544, "step": 2896 }, { "epoch": 0.7081398191151308, "grad_norm": 2.9812870025634766, "learning_rate": 4.145230208925094e-06, "loss": 0.6984, "step": 2897 }, { "epoch": 0.7083842581275972, "grad_norm": 11.79186725616455, "learning_rate": 4.138813559949341e-06, "loss": 0.7448, "step": 2898 }, { "epoch": 0.7086286971400636, "grad_norm": 3.4640963077545166, "learning_rate": 4.132400584997106e-06, "loss": 0.6699, "step": 2899 }, { "epoch": 0.7088731361525299, "grad_norm": 2.5082526206970215, "learning_rate": 4.125991288088294e-06, "loss": 0.5293, "step": 2900 }, { "epoch": 0.7091175751649963, "grad_norm": 2.9462575912475586, "learning_rate": 4.1195856732405094e-06, "loss": 0.6258, "step": 2901 }, { "epoch": 0.7093620141774627, "grad_norm": 3.4112930297851562, "learning_rate": 4.1131837444690404e-06, "loss": 0.6114, "step": 2902 }, { "epoch": 0.7096064531899291, "grad_norm": 2.7927472591400146, "learning_rate": 4.106785505786872e-06, "loss": 0.4668, "step": 2903 }, { "epoch": 0.7098508922023955, "grad_norm": 3.160207748413086, "learning_rate": 4.100390961204672e-06, "loss": 0.5719, "step": 2904 }, { "epoch": 0.7100953312148619, "grad_norm": 2.6186273097991943, "learning_rate": 4.094000114730795e-06, "loss": 0.5201, "step": 2905 }, { "epoch": 0.7103397702273283, "grad_norm": 3.6046929359436035, "learning_rate": 4.087612970371273e-06, "loss": 0.6772, "step": 2906 }, { "epoch": 0.7105842092397947, "grad_norm": 3.1790590286254883, "learning_rate": 4.081229532129826e-06, "loss": 0.7094, "step": 2907 }, { "epoch": 0.7108286482522611, "grad_norm": 3.196411371231079, "learning_rate": 4.0748498040078446e-06, "loss": 0.6551, "step": 2908 }, { "epoch": 0.7110730872647274, "grad_norm": 3.279223918914795, "learning_rate": 4.068473790004393e-06, "loss": 0.6332, "step": 2909 }, { "epoch": 0.7113175262771938, "grad_norm": 3.0042917728424072, "learning_rate": 4.062101494116209e-06, "loss": 0.5151, "step": 2910 }, { "epoch": 0.7115619652896602, "grad_norm": 2.7973415851593018, "learning_rate": 4.055732920337699e-06, "loss": 0.5038, "step": 2911 }, { "epoch": 0.7118064043021266, "grad_norm": 3.4121341705322266, "learning_rate": 4.04936807266094e-06, "loss": 0.5919, "step": 2912 }, { "epoch": 0.712050843314593, "grad_norm": 3.2722136974334717, "learning_rate": 4.043006955075667e-06, "loss": 0.69, "step": 2913 }, { "epoch": 0.7122952823270594, "grad_norm": 3.4652504920959473, "learning_rate": 4.036649571569282e-06, "loss": 0.6469, "step": 2914 }, { "epoch": 0.7125397213395258, "grad_norm": 3.436591863632202, "learning_rate": 4.030295926126845e-06, "loss": 0.6366, "step": 2915 }, { "epoch": 0.7127841603519922, "grad_norm": 3.649648666381836, "learning_rate": 4.02394602273107e-06, "loss": 0.5092, "step": 2916 }, { "epoch": 0.7130285993644586, "grad_norm": 3.596709966659546, "learning_rate": 4.0175998653623295e-06, "loss": 0.5792, "step": 2917 }, { "epoch": 0.7132730383769249, "grad_norm": 3.119182586669922, "learning_rate": 4.011257457998645e-06, "loss": 0.5226, "step": 2918 }, { "epoch": 0.7135174773893913, "grad_norm": 3.1880149841308594, "learning_rate": 4.004918804615689e-06, "loss": 0.5484, "step": 2919 }, { "epoch": 0.7137619164018577, "grad_norm": 2.373765230178833, "learning_rate": 3.998583909186781e-06, "loss": 1.9726, "step": 2920 }, { "epoch": 0.7140063554143241, "grad_norm": 2.5934154987335205, "learning_rate": 3.992252775682877e-06, "loss": 0.5649, "step": 2921 }, { "epoch": 0.7142507944267905, "grad_norm": 3.5194554328918457, "learning_rate": 3.985925408072594e-06, "loss": 0.6557, "step": 2922 }, { "epoch": 0.7144952334392569, "grad_norm": 3.8668274879455566, "learning_rate": 3.979601810322169e-06, "loss": 0.5971, "step": 2923 }, { "epoch": 0.7147396724517233, "grad_norm": 3.060781240463257, "learning_rate": 3.973281986395485e-06, "loss": 0.7926, "step": 2924 }, { "epoch": 0.7149841114641897, "grad_norm": 3.3815793991088867, "learning_rate": 3.966965940254057e-06, "loss": 0.6712, "step": 2925 }, { "epoch": 0.7152285504766561, "grad_norm": 3.3370070457458496, "learning_rate": 3.960653675857033e-06, "loss": 0.6374, "step": 2926 }, { "epoch": 0.7154729894891224, "grad_norm": 3.052048444747925, "learning_rate": 3.954345197161188e-06, "loss": 0.6032, "step": 2927 }, { "epoch": 0.7157174285015888, "grad_norm": 2.9592669010162354, "learning_rate": 3.948040508120929e-06, "loss": 0.5775, "step": 2928 }, { "epoch": 0.7159618675140552, "grad_norm": 2.4977145195007324, "learning_rate": 3.941739612688284e-06, "loss": 0.5427, "step": 2929 }, { "epoch": 0.7162063065265216, "grad_norm": 2.9116251468658447, "learning_rate": 3.935442514812901e-06, "loss": 0.6394, "step": 2930 }, { "epoch": 0.716450745538988, "grad_norm": 2.811952590942383, "learning_rate": 3.929149218442052e-06, "loss": 0.5343, "step": 2931 }, { "epoch": 0.7166951845514544, "grad_norm": 3.4379796981811523, "learning_rate": 3.922859727520625e-06, "loss": 0.6813, "step": 2932 }, { "epoch": 0.7169396235639208, "grad_norm": 3.4804670810699463, "learning_rate": 3.91657404599112e-06, "loss": 0.624, "step": 2933 }, { "epoch": 0.7171840625763872, "grad_norm": 2.6202759742736816, "learning_rate": 3.910292177793651e-06, "loss": 0.5678, "step": 2934 }, { "epoch": 0.7174285015888536, "grad_norm": 3.9853694438934326, "learning_rate": 3.904014126865943e-06, "loss": 0.6223, "step": 2935 }, { "epoch": 0.7176729406013199, "grad_norm": 3.015868902206421, "learning_rate": 3.897739897143324e-06, "loss": 0.5786, "step": 2936 }, { "epoch": 0.7179173796137863, "grad_norm": 2.8636083602905273, "learning_rate": 3.891469492558728e-06, "loss": 0.6114, "step": 2937 }, { "epoch": 0.7181618186262527, "grad_norm": 3.039456844329834, "learning_rate": 3.8852029170427e-06, "loss": 0.6155, "step": 2938 }, { "epoch": 0.7184062576387191, "grad_norm": 2.8267924785614014, "learning_rate": 3.878940174523371e-06, "loss": 0.6414, "step": 2939 }, { "epoch": 0.7186506966511855, "grad_norm": 2.8125078678131104, "learning_rate": 3.872681268926479e-06, "loss": 0.6549, "step": 2940 }, { "epoch": 0.7188951356636519, "grad_norm": 3.337916135787964, "learning_rate": 3.866426204175353e-06, "loss": 0.5881, "step": 2941 }, { "epoch": 0.7191395746761183, "grad_norm": 2.848170757293701, "learning_rate": 3.860174984190914e-06, "loss": 0.6312, "step": 2942 }, { "epoch": 0.7193840136885847, "grad_norm": 2.944838047027588, "learning_rate": 3.8539276128916736e-06, "loss": 0.6264, "step": 2943 }, { "epoch": 0.7196284527010511, "grad_norm": 3.2724449634552, "learning_rate": 3.847684094193733e-06, "loss": 0.6204, "step": 2944 }, { "epoch": 0.7198728917135174, "grad_norm": 3.8080027103424072, "learning_rate": 3.841444432010775e-06, "loss": 0.625, "step": 2945 }, { "epoch": 0.7201173307259838, "grad_norm": 3.054201602935791, "learning_rate": 3.835208630254068e-06, "loss": 0.7294, "step": 2946 }, { "epoch": 0.7203617697384502, "grad_norm": 3.6399037837982178, "learning_rate": 3.828976692832458e-06, "loss": 0.5939, "step": 2947 }, { "epoch": 0.7206062087509166, "grad_norm": 3.149604320526123, "learning_rate": 3.822748623652372e-06, "loss": 0.6319, "step": 2948 }, { "epoch": 0.720850647763383, "grad_norm": 3.125087261199951, "learning_rate": 3.816524426617807e-06, "loss": 0.5891, "step": 2949 }, { "epoch": 0.7210950867758494, "grad_norm": 3.2292184829711914, "learning_rate": 3.810304105630338e-06, "loss": 0.561, "step": 2950 }, { "epoch": 0.7213395257883158, "grad_norm": 2.8233754634857178, "learning_rate": 3.804087664589108e-06, "loss": 0.4755, "step": 2951 }, { "epoch": 0.7215839648007822, "grad_norm": 3.1551756858825684, "learning_rate": 3.7978751073908237e-06, "loss": 0.6226, "step": 2952 }, { "epoch": 0.7218284038132486, "grad_norm": 4.270998001098633, "learning_rate": 3.791666437929771e-06, "loss": 0.5689, "step": 2953 }, { "epoch": 0.7220728428257149, "grad_norm": 3.086409568786621, "learning_rate": 3.7854616600977834e-06, "loss": 0.669, "step": 2954 }, { "epoch": 0.7223172818381813, "grad_norm": 3.408801555633545, "learning_rate": 3.7792607777842626e-06, "loss": 0.6059, "step": 2955 }, { "epoch": 0.7225617208506477, "grad_norm": 3.2890207767486572, "learning_rate": 3.773063794876167e-06, "loss": 0.7136, "step": 2956 }, { "epoch": 0.7228061598631141, "grad_norm": 3.042243242263794, "learning_rate": 3.7668707152580097e-06, "loss": 0.5738, "step": 2957 }, { "epoch": 0.7230505988755805, "grad_norm": 2.7547099590301514, "learning_rate": 3.760681542811859e-06, "loss": 0.5578, "step": 2958 }, { "epoch": 0.7232950378880469, "grad_norm": 3.3390941619873047, "learning_rate": 3.754496281417331e-06, "loss": 0.6282, "step": 2959 }, { "epoch": 0.7235394769005133, "grad_norm": 3.3198680877685547, "learning_rate": 3.7483149349515957e-06, "loss": 0.5808, "step": 2960 }, { "epoch": 0.7237839159129797, "grad_norm": 2.6300973892211914, "learning_rate": 3.742137507289363e-06, "loss": 0.538, "step": 2961 }, { "epoch": 0.7240283549254461, "grad_norm": 3.129347085952759, "learning_rate": 3.7359640023028888e-06, "loss": 0.6722, "step": 2962 }, { "epoch": 0.7242727939379124, "grad_norm": 2.856234550476074, "learning_rate": 3.729794423861971e-06, "loss": 0.6394, "step": 2963 }, { "epoch": 0.7245172329503788, "grad_norm": 3.1484363079071045, "learning_rate": 3.7236287758339464e-06, "loss": 0.6045, "step": 2964 }, { "epoch": 0.7247616719628452, "grad_norm": 3.6585779190063477, "learning_rate": 3.717467062083685e-06, "loss": 0.5096, "step": 2965 }, { "epoch": 0.7250061109753116, "grad_norm": 3.5617711544036865, "learning_rate": 3.7113092864735943e-06, "loss": 0.6079, "step": 2966 }, { "epoch": 0.725250549987778, "grad_norm": 3.1096692085266113, "learning_rate": 3.7051554528636114e-06, "loss": 0.6274, "step": 2967 }, { "epoch": 0.7254949890002445, "grad_norm": 3.7432024478912354, "learning_rate": 3.699005565111198e-06, "loss": 0.6443, "step": 2968 }, { "epoch": 0.7257394280127109, "grad_norm": 3.277007579803467, "learning_rate": 3.6928596270713558e-06, "loss": 0.6341, "step": 2969 }, { "epoch": 0.7259838670251773, "grad_norm": 2.8448712825775146, "learning_rate": 3.6867176425965967e-06, "loss": 0.6601, "step": 2970 }, { "epoch": 0.7262283060376437, "grad_norm": 2.983934164047241, "learning_rate": 3.680579615536961e-06, "loss": 0.6822, "step": 2971 }, { "epoch": 0.72647274505011, "grad_norm": 2.7468302249908447, "learning_rate": 3.6744455497400055e-06, "loss": 0.5229, "step": 2972 }, { "epoch": 0.7267171840625763, "grad_norm": 3.0838937759399414, "learning_rate": 3.6683154490508064e-06, "loss": 0.7091, "step": 2973 }, { "epoch": 0.7269616230750428, "grad_norm": 2.580299139022827, "learning_rate": 3.662189317311953e-06, "loss": 0.6071, "step": 2974 }, { "epoch": 0.7272060620875092, "grad_norm": 3.2007927894592285, "learning_rate": 3.6560671583635467e-06, "loss": 0.6529, "step": 2975 }, { "epoch": 0.7274505010999756, "grad_norm": 2.868992805480957, "learning_rate": 3.6499489760431993e-06, "loss": 1.8465, "step": 2976 }, { "epoch": 0.727694940112442, "grad_norm": 3.0807933807373047, "learning_rate": 3.643834774186029e-06, "loss": 0.6774, "step": 2977 }, { "epoch": 0.7279393791249084, "grad_norm": 3.0739243030548096, "learning_rate": 3.6377245566246597e-06, "loss": 0.5611, "step": 2978 }, { "epoch": 0.7281838181373748, "grad_norm": 3.160921812057495, "learning_rate": 3.631618327189218e-06, "loss": 0.5293, "step": 2979 }, { "epoch": 0.7284282571498412, "grad_norm": 7.900955677032471, "learning_rate": 3.6255160897073293e-06, "loss": 0.6552, "step": 2980 }, { "epoch": 0.7286726961623075, "grad_norm": 2.859971046447754, "learning_rate": 3.6194178480041174e-06, "loss": 0.5905, "step": 2981 }, { "epoch": 0.7289171351747739, "grad_norm": 3.280907154083252, "learning_rate": 3.6133236059022016e-06, "loss": 0.6404, "step": 2982 }, { "epoch": 0.7291615741872403, "grad_norm": 4.005970478057861, "learning_rate": 3.607233367221691e-06, "loss": 0.6535, "step": 2983 }, { "epoch": 0.7294060131997067, "grad_norm": 2.936605453491211, "learning_rate": 3.601147135780194e-06, "loss": 0.5508, "step": 2984 }, { "epoch": 0.7296504522121731, "grad_norm": 2.9298524856567383, "learning_rate": 3.5950649153927997e-06, "loss": 0.5346, "step": 2985 }, { "epoch": 0.7298948912246395, "grad_norm": 3.077277421951294, "learning_rate": 3.588986709872082e-06, "loss": 0.5944, "step": 2986 }, { "epoch": 0.7301393302371059, "grad_norm": 3.2504830360412598, "learning_rate": 3.5829125230281014e-06, "loss": 0.6482, "step": 2987 }, { "epoch": 0.7303837692495723, "grad_norm": 2.85105037689209, "learning_rate": 3.5768423586684e-06, "loss": 0.6188, "step": 2988 }, { "epoch": 0.7306282082620387, "grad_norm": 2.931598663330078, "learning_rate": 3.5707762205979946e-06, "loss": 0.6613, "step": 2989 }, { "epoch": 0.730872647274505, "grad_norm": 3.099318265914917, "learning_rate": 3.564714112619383e-06, "loss": 0.5855, "step": 2990 }, { "epoch": 0.7311170862869714, "grad_norm": 2.8876891136169434, "learning_rate": 3.558656038532532e-06, "loss": 0.533, "step": 2991 }, { "epoch": 0.7313615252994378, "grad_norm": 2.955615997314453, "learning_rate": 3.5526020021348827e-06, "loss": 0.5758, "step": 2992 }, { "epoch": 0.7316059643119042, "grad_norm": 3.9695706367492676, "learning_rate": 3.5465520072213455e-06, "loss": 0.6492, "step": 2993 }, { "epoch": 0.7318504033243706, "grad_norm": 3.853733777999878, "learning_rate": 3.5405060575842953e-06, "loss": 0.5827, "step": 2994 }, { "epoch": 0.732094842336837, "grad_norm": 3.007627487182617, "learning_rate": 3.5344641570135742e-06, "loss": 0.672, "step": 2995 }, { "epoch": 0.7323392813493034, "grad_norm": 3.5044028759002686, "learning_rate": 3.5284263092964823e-06, "loss": 0.668, "step": 2996 }, { "epoch": 0.7325837203617698, "grad_norm": 3.6388285160064697, "learning_rate": 3.5223925182177833e-06, "loss": 0.6515, "step": 2997 }, { "epoch": 0.7328281593742362, "grad_norm": 3.2376761436462402, "learning_rate": 3.516362787559695e-06, "loss": 0.5645, "step": 2998 }, { "epoch": 0.7330725983867025, "grad_norm": 4.141454219818115, "learning_rate": 3.510337121101888e-06, "loss": 0.4661, "step": 2999 }, { "epoch": 0.7333170373991689, "grad_norm": 3.248368740081787, "learning_rate": 3.5043155226214954e-06, "loss": 0.6578, "step": 3000 }, { "epoch": 0.7335614764116353, "grad_norm": 2.848895788192749, "learning_rate": 3.4982979958930896e-06, "loss": 0.5574, "step": 3001 }, { "epoch": 0.7338059154241017, "grad_norm": 3.017939805984497, "learning_rate": 3.4922845446886944e-06, "loss": 0.5379, "step": 3002 }, { "epoch": 0.7340503544365681, "grad_norm": 3.1488208770751953, "learning_rate": 3.48627517277778e-06, "loss": 0.5726, "step": 3003 }, { "epoch": 0.7342947934490345, "grad_norm": 3.541123151779175, "learning_rate": 3.4802698839272564e-06, "loss": 0.5333, "step": 3004 }, { "epoch": 0.7345392324615009, "grad_norm": 3.4302146434783936, "learning_rate": 3.4742686819014782e-06, "loss": 0.5916, "step": 3005 }, { "epoch": 0.7347836714739673, "grad_norm": 3.442814350128174, "learning_rate": 3.468271570462235e-06, "loss": 0.7797, "step": 3006 }, { "epoch": 0.7350281104864337, "grad_norm": 3.1694416999816895, "learning_rate": 3.462278553368753e-06, "loss": 0.5926, "step": 3007 }, { "epoch": 0.7352725494989, "grad_norm": 5.13250207901001, "learning_rate": 3.4562896343776943e-06, "loss": 0.7076, "step": 3008 }, { "epoch": 0.7355169885113664, "grad_norm": 3.201740264892578, "learning_rate": 3.4503048172431495e-06, "loss": 0.5783, "step": 3009 }, { "epoch": 0.7357614275238328, "grad_norm": 3.3980233669281006, "learning_rate": 3.4443241057166342e-06, "loss": 0.6619, "step": 3010 }, { "epoch": 0.7360058665362992, "grad_norm": 3.5091662406921387, "learning_rate": 3.4383475035471026e-06, "loss": 0.5663, "step": 3011 }, { "epoch": 0.7362503055487656, "grad_norm": 3.1590633392333984, "learning_rate": 3.432375014480922e-06, "loss": 0.5789, "step": 3012 }, { "epoch": 0.736494744561232, "grad_norm": 2.7551000118255615, "learning_rate": 3.426406642261885e-06, "loss": 1.8702, "step": 3013 }, { "epoch": 0.7367391835736984, "grad_norm": 2.7853903770446777, "learning_rate": 3.4204423906312023e-06, "loss": 0.5757, "step": 3014 }, { "epoch": 0.7369836225861648, "grad_norm": 2.52852201461792, "learning_rate": 3.4144822633275042e-06, "loss": 1.845, "step": 3015 }, { "epoch": 0.7372280615986312, "grad_norm": 3.035095691680908, "learning_rate": 3.408526264086833e-06, "loss": 0.4472, "step": 3016 }, { "epoch": 0.7374725006110975, "grad_norm": 3.2289512157440186, "learning_rate": 3.402574396642646e-06, "loss": 0.5751, "step": 3017 }, { "epoch": 0.7377169396235639, "grad_norm": 5.053186893463135, "learning_rate": 3.3966266647258085e-06, "loss": 0.5418, "step": 3018 }, { "epoch": 0.7379613786360303, "grad_norm": 3.862535238265991, "learning_rate": 3.3906830720645943e-06, "loss": 0.452, "step": 3019 }, { "epoch": 0.7382058176484967, "grad_norm": 6.835580348968506, "learning_rate": 3.3847436223846788e-06, "loss": 0.605, "step": 3020 }, { "epoch": 0.7384502566609631, "grad_norm": 2.5224106311798096, "learning_rate": 3.378808319409149e-06, "loss": 0.5702, "step": 3021 }, { "epoch": 0.7386946956734295, "grad_norm": 2.9877500534057617, "learning_rate": 3.372877166858488e-06, "loss": 0.6121, "step": 3022 }, { "epoch": 0.7389391346858959, "grad_norm": 2.6402807235717773, "learning_rate": 3.3669501684505734e-06, "loss": 2.0057, "step": 3023 }, { "epoch": 0.7391835736983623, "grad_norm": 3.599905490875244, "learning_rate": 3.3610273279006842e-06, "loss": 0.5901, "step": 3024 }, { "epoch": 0.7394280127108287, "grad_norm": 3.2533516883850098, "learning_rate": 3.3551086489214913e-06, "loss": 0.5949, "step": 3025 }, { "epoch": 0.739672451723295, "grad_norm": 3.3682477474212646, "learning_rate": 3.349194135223056e-06, "loss": 0.5901, "step": 3026 }, { "epoch": 0.7399168907357614, "grad_norm": 2.728986978530884, "learning_rate": 3.343283790512829e-06, "loss": 0.699, "step": 3027 }, { "epoch": 0.7401613297482278, "grad_norm": 2.769089460372925, "learning_rate": 3.3373776184956485e-06, "loss": 0.7134, "step": 3028 }, { "epoch": 0.7404057687606942, "grad_norm": 2.7086477279663086, "learning_rate": 3.331475622873738e-06, "loss": 0.5526, "step": 3029 }, { "epoch": 0.7406502077731606, "grad_norm": 3.8077518939971924, "learning_rate": 3.3255778073467004e-06, "loss": 0.5683, "step": 3030 }, { "epoch": 0.740894646785627, "grad_norm": 3.039224147796631, "learning_rate": 3.319684175611517e-06, "loss": 0.5612, "step": 3031 }, { "epoch": 0.7411390857980934, "grad_norm": 3.579586982727051, "learning_rate": 3.3137947313625563e-06, "loss": 0.6918, "step": 3032 }, { "epoch": 0.7413835248105598, "grad_norm": 3.2064623832702637, "learning_rate": 3.30790947829155e-06, "loss": 0.5335, "step": 3033 }, { "epoch": 0.7416279638230262, "grad_norm": 6.851931571960449, "learning_rate": 3.3020284200876095e-06, "loss": 0.5767, "step": 3034 }, { "epoch": 0.7418724028354925, "grad_norm": 3.6696505546569824, "learning_rate": 3.296151560437214e-06, "loss": 0.6169, "step": 3035 }, { "epoch": 0.7421168418479589, "grad_norm": 3.404021978378296, "learning_rate": 3.290278903024211e-06, "loss": 0.6475, "step": 3036 }, { "epoch": 0.7423612808604253, "grad_norm": 3.687025308609009, "learning_rate": 3.284410451529816e-06, "loss": 0.6712, "step": 3037 }, { "epoch": 0.7426057198728917, "grad_norm": 3.5287423133850098, "learning_rate": 3.2785462096326047e-06, "loss": 1.9579, "step": 3038 }, { "epoch": 0.7428501588853581, "grad_norm": 2.978653907775879, "learning_rate": 3.2726861810085163e-06, "loss": 0.514, "step": 3039 }, { "epoch": 0.7430945978978245, "grad_norm": 3.3357956409454346, "learning_rate": 3.26683036933085e-06, "loss": 0.5666, "step": 3040 }, { "epoch": 0.7433390369102909, "grad_norm": 3.517115354537964, "learning_rate": 3.2609787782702595e-06, "loss": 0.6104, "step": 3041 }, { "epoch": 0.7435834759227573, "grad_norm": 2.289231538772583, "learning_rate": 3.255131411494753e-06, "loss": 2.043, "step": 3042 }, { "epoch": 0.7438279149352237, "grad_norm": 2.9872188568115234, "learning_rate": 3.2492882726696907e-06, "loss": 0.6355, "step": 3043 }, { "epoch": 0.74407235394769, "grad_norm": 8.818699836730957, "learning_rate": 3.243449365457785e-06, "loss": 0.5859, "step": 3044 }, { "epoch": 0.7443167929601564, "grad_norm": 2.9520797729492188, "learning_rate": 3.237614693519089e-06, "loss": 0.6683, "step": 3045 }, { "epoch": 0.7445612319726228, "grad_norm": 3.3904762268066406, "learning_rate": 3.2317842605110126e-06, "loss": 0.7299, "step": 3046 }, { "epoch": 0.7448056709850892, "grad_norm": 2.8226022720336914, "learning_rate": 3.225958070088301e-06, "loss": 0.5884, "step": 3047 }, { "epoch": 0.7450501099975556, "grad_norm": 2.8809454441070557, "learning_rate": 3.2201361259030374e-06, "loss": 0.5376, "step": 3048 }, { "epoch": 0.745294549010022, "grad_norm": 2.6132972240448, "learning_rate": 3.2143184316046485e-06, "loss": 0.5241, "step": 3049 }, { "epoch": 0.7455389880224884, "grad_norm": 3.0091755390167236, "learning_rate": 3.2085049908398945e-06, "loss": 0.5732, "step": 3050 }, { "epoch": 0.7457834270349548, "grad_norm": 3.3569703102111816, "learning_rate": 3.2026958072528715e-06, "loss": 0.5624, "step": 3051 }, { "epoch": 0.7460278660474212, "grad_norm": 3.4056758880615234, "learning_rate": 3.1968908844850045e-06, "loss": 0.5958, "step": 3052 }, { "epoch": 0.7462723050598875, "grad_norm": 3.654890298843384, "learning_rate": 3.191090226175049e-06, "loss": 0.663, "step": 3053 }, { "epoch": 0.7465167440723539, "grad_norm": 4.088568210601807, "learning_rate": 3.1852938359590857e-06, "loss": 0.665, "step": 3054 }, { "epoch": 0.7467611830848203, "grad_norm": 2.4330692291259766, "learning_rate": 3.179501717470522e-06, "loss": 2.0035, "step": 3055 }, { "epoch": 0.7470056220972867, "grad_norm": 3.3720357418060303, "learning_rate": 3.1737138743400896e-06, "loss": 0.6575, "step": 3056 }, { "epoch": 0.7472500611097531, "grad_norm": 2.899686336517334, "learning_rate": 3.167930310195834e-06, "loss": 0.6562, "step": 3057 }, { "epoch": 0.7474945001222195, "grad_norm": 3.022397518157959, "learning_rate": 3.162151028663123e-06, "loss": 0.6167, "step": 3058 }, { "epoch": 0.7477389391346859, "grad_norm": 5.976170063018799, "learning_rate": 3.15637603336464e-06, "loss": 0.5957, "step": 3059 }, { "epoch": 0.7479833781471523, "grad_norm": 3.8271026611328125, "learning_rate": 3.1506053279203787e-06, "loss": 0.5749, "step": 3060 }, { "epoch": 0.7482278171596187, "grad_norm": 3.036123275756836, "learning_rate": 3.1448389159476433e-06, "loss": 0.5961, "step": 3061 }, { "epoch": 0.748472256172085, "grad_norm": 3.2765238285064697, "learning_rate": 3.139076801061056e-06, "loss": 0.6276, "step": 3062 }, { "epoch": 0.7487166951845514, "grad_norm": 8.13082218170166, "learning_rate": 3.1333189868725343e-06, "loss": 0.5269, "step": 3063 }, { "epoch": 0.7489611341970178, "grad_norm": 2.5719962120056152, "learning_rate": 3.1275654769913043e-06, "loss": 1.9888, "step": 3064 }, { "epoch": 0.7492055732094842, "grad_norm": 2.6839728355407715, "learning_rate": 3.1218162750238933e-06, "loss": 0.482, "step": 3065 }, { "epoch": 0.7494500122219506, "grad_norm": 3.614717960357666, "learning_rate": 3.1160713845741287e-06, "loss": 0.5976, "step": 3066 }, { "epoch": 0.749694451234417, "grad_norm": 3.5247299671173096, "learning_rate": 3.110330809243134e-06, "loss": 0.6967, "step": 3067 }, { "epoch": 0.7499388902468834, "grad_norm": 3.888062000274658, "learning_rate": 3.1045945526293307e-06, "loss": 0.5281, "step": 3068 }, { "epoch": 0.7501833292593498, "grad_norm": 2.962456464767456, "learning_rate": 3.098862618328431e-06, "loss": 1.8584, "step": 3069 }, { "epoch": 0.7504277682718162, "grad_norm": 2.798931121826172, "learning_rate": 3.0931350099334366e-06, "loss": 0.5392, "step": 3070 }, { "epoch": 0.7506722072842825, "grad_norm": 3.3497190475463867, "learning_rate": 3.087411731034641e-06, "loss": 0.6611, "step": 3071 }, { "epoch": 0.7509166462967489, "grad_norm": 3.342142343521118, "learning_rate": 3.08169278521962e-06, "loss": 0.6632, "step": 3072 }, { "epoch": 0.7511610853092153, "grad_norm": 2.9381072521209717, "learning_rate": 3.0759781760732365e-06, "loss": 0.6726, "step": 3073 }, { "epoch": 0.7514055243216817, "grad_norm": 3.3888957500457764, "learning_rate": 3.070267907177633e-06, "loss": 0.6559, "step": 3074 }, { "epoch": 0.7516499633341481, "grad_norm": 2.966024875640869, "learning_rate": 3.064561982112232e-06, "loss": 0.6128, "step": 3075 }, { "epoch": 0.7518944023466145, "grad_norm": 3.589439630508423, "learning_rate": 3.0588604044537297e-06, "loss": 0.6941, "step": 3076 }, { "epoch": 0.7521388413590809, "grad_norm": 2.973039388656616, "learning_rate": 3.053163177776106e-06, "loss": 0.6058, "step": 3077 }, { "epoch": 0.7523832803715473, "grad_norm": 3.746828079223633, "learning_rate": 3.047470305650607e-06, "loss": 0.5882, "step": 3078 }, { "epoch": 0.7526277193840137, "grad_norm": 3.47501277923584, "learning_rate": 3.0417817916457482e-06, "loss": 0.5496, "step": 3079 }, { "epoch": 0.75287215839648, "grad_norm": 3.0161173343658447, "learning_rate": 3.036097639327317e-06, "loss": 0.6409, "step": 3080 }, { "epoch": 0.7531165974089464, "grad_norm": 2.6342554092407227, "learning_rate": 3.0304178522583626e-06, "loss": 0.5368, "step": 3081 }, { "epoch": 0.7533610364214128, "grad_norm": 3.435028553009033, "learning_rate": 3.0247424339992017e-06, "loss": 0.6462, "step": 3082 }, { "epoch": 0.7536054754338792, "grad_norm": 3.231520175933838, "learning_rate": 3.0190713881074106e-06, "loss": 0.5733, "step": 3083 }, { "epoch": 0.7538499144463456, "grad_norm": 3.6398332118988037, "learning_rate": 3.0134047181378256e-06, "loss": 0.6683, "step": 3084 }, { "epoch": 0.754094353458812, "grad_norm": 3.8633766174316406, "learning_rate": 3.007742427642539e-06, "loss": 0.5601, "step": 3085 }, { "epoch": 0.7543387924712784, "grad_norm": 3.340566635131836, "learning_rate": 3.0020845201708993e-06, "loss": 0.4947, "step": 3086 }, { "epoch": 0.7545832314837448, "grad_norm": 3.265387773513794, "learning_rate": 2.9964309992695053e-06, "loss": 0.5801, "step": 3087 }, { "epoch": 0.7548276704962112, "grad_norm": 2.477468490600586, "learning_rate": 2.9907818684822088e-06, "loss": 1.9359, "step": 3088 }, { "epoch": 0.7550721095086775, "grad_norm": 2.758582592010498, "learning_rate": 2.985137131350109e-06, "loss": 0.432, "step": 3089 }, { "epoch": 0.7553165485211439, "grad_norm": 3.1395211219787598, "learning_rate": 2.979496791411548e-06, "loss": 0.5712, "step": 3090 }, { "epoch": 0.7555609875336103, "grad_norm": 3.1856343746185303, "learning_rate": 2.973860852202117e-06, "loss": 0.7134, "step": 3091 }, { "epoch": 0.7558054265460767, "grad_norm": 3.088649034500122, "learning_rate": 2.9682293172546407e-06, "loss": 0.6315, "step": 3092 }, { "epoch": 0.7560498655585431, "grad_norm": 3.3430135250091553, "learning_rate": 2.962602190099193e-06, "loss": 0.629, "step": 3093 }, { "epoch": 0.7562943045710095, "grad_norm": 4.353748321533203, "learning_rate": 2.9569794742630796e-06, "loss": 0.6365, "step": 3094 }, { "epoch": 0.756538743583476, "grad_norm": 3.6114869117736816, "learning_rate": 2.9513611732708404e-06, "loss": 0.7775, "step": 3095 }, { "epoch": 0.7567831825959423, "grad_norm": 2.5576424598693848, "learning_rate": 2.9457472906442476e-06, "loss": 1.9289, "step": 3096 }, { "epoch": 0.7570276216084088, "grad_norm": 3.31587553024292, "learning_rate": 2.9401378299023043e-06, "loss": 0.5658, "step": 3097 }, { "epoch": 0.757272060620875, "grad_norm": 3.4183433055877686, "learning_rate": 2.9345327945612446e-06, "loss": 0.512, "step": 3098 }, { "epoch": 0.7575164996333414, "grad_norm": 3.2171711921691895, "learning_rate": 2.9289321881345257e-06, "loss": 0.7227, "step": 3099 }, { "epoch": 0.7577609386458078, "grad_norm": 5.540745258331299, "learning_rate": 2.9233360141328283e-06, "loss": 0.6218, "step": 3100 }, { "epoch": 0.7580053776582742, "grad_norm": 3.9040966033935547, "learning_rate": 2.917744276064056e-06, "loss": 0.4521, "step": 3101 }, { "epoch": 0.7582498166707407, "grad_norm": 3.151887893676758, "learning_rate": 2.9121569774333327e-06, "loss": 0.6599, "step": 3102 }, { "epoch": 0.758494255683207, "grad_norm": 3.7008299827575684, "learning_rate": 2.9065741217429964e-06, "loss": 0.593, "step": 3103 }, { "epoch": 0.7587386946956735, "grad_norm": 4.508377552032471, "learning_rate": 2.9009957124926026e-06, "loss": 0.5348, "step": 3104 }, { "epoch": 0.7589831337081399, "grad_norm": 2.980142116546631, "learning_rate": 2.8954217531789186e-06, "loss": 0.5423, "step": 3105 }, { "epoch": 0.7592275727206063, "grad_norm": 2.6276297569274902, "learning_rate": 2.8898522472959234e-06, "loss": 0.581, "step": 3106 }, { "epoch": 0.7594720117330725, "grad_norm": 3.8187081813812256, "learning_rate": 2.8842871983348e-06, "loss": 0.6095, "step": 3107 }, { "epoch": 0.759716450745539, "grad_norm": 3.7661819458007812, "learning_rate": 2.878726609783947e-06, "loss": 0.5473, "step": 3108 }, { "epoch": 0.7599608897580054, "grad_norm": 5.643442153930664, "learning_rate": 2.8731704851289598e-06, "loss": 0.5268, "step": 3109 }, { "epoch": 0.7602053287704718, "grad_norm": 3.5928304195404053, "learning_rate": 2.8676188278526353e-06, "loss": 0.5524, "step": 3110 }, { "epoch": 0.7604497677829382, "grad_norm": 3.473266124725342, "learning_rate": 2.8620716414349714e-06, "loss": 0.4903, "step": 3111 }, { "epoch": 0.7606942067954046, "grad_norm": 3.6228229999542236, "learning_rate": 2.856528929353166e-06, "loss": 0.6172, "step": 3112 }, { "epoch": 0.760938645807871, "grad_norm": 3.694929361343384, "learning_rate": 2.85099069508161e-06, "loss": 0.584, "step": 3113 }, { "epoch": 0.7611830848203374, "grad_norm": 4.231311798095703, "learning_rate": 2.8454569420918865e-06, "loss": 0.7133, "step": 3114 }, { "epoch": 0.7614275238328038, "grad_norm": 4.651334285736084, "learning_rate": 2.8399276738527716e-06, "loss": 0.5553, "step": 3115 }, { "epoch": 0.7616719628452701, "grad_norm": 3.3225133419036865, "learning_rate": 2.8344028938302292e-06, "loss": 0.5635, "step": 3116 }, { "epoch": 0.7619164018577365, "grad_norm": 2.9594101905822754, "learning_rate": 2.8288826054874095e-06, "loss": 0.6375, "step": 3117 }, { "epoch": 0.7621608408702029, "grad_norm": 3.052440643310547, "learning_rate": 2.823366812284649e-06, "loss": 0.5714, "step": 3118 }, { "epoch": 0.7624052798826693, "grad_norm": 4.923966884613037, "learning_rate": 2.817855517679464e-06, "loss": 0.6374, "step": 3119 }, { "epoch": 0.7626497188951357, "grad_norm": 3.6699352264404297, "learning_rate": 2.812348725126551e-06, "loss": 0.5905, "step": 3120 }, { "epoch": 0.7628941579076021, "grad_norm": 3.297264575958252, "learning_rate": 2.806846438077787e-06, "loss": 0.6301, "step": 3121 }, { "epoch": 0.7631385969200685, "grad_norm": 2.899012327194214, "learning_rate": 2.801348659982224e-06, "loss": 0.4651, "step": 3122 }, { "epoch": 0.7633830359325349, "grad_norm": 3.3183135986328125, "learning_rate": 2.795855394286081e-06, "loss": 0.5737, "step": 3123 }, { "epoch": 0.7636274749450013, "grad_norm": 3.0558860301971436, "learning_rate": 2.7903666444327615e-06, "loss": 0.5875, "step": 3124 }, { "epoch": 0.7638719139574676, "grad_norm": 5.014186859130859, "learning_rate": 2.7848824138628295e-06, "loss": 0.595, "step": 3125 }, { "epoch": 0.764116352969934, "grad_norm": 3.420767307281494, "learning_rate": 2.779402706014016e-06, "loss": 0.5561, "step": 3126 }, { "epoch": 0.7643607919824004, "grad_norm": 3.7397305965423584, "learning_rate": 2.7739275243212183e-06, "loss": 0.5283, "step": 3127 }, { "epoch": 0.7646052309948668, "grad_norm": 2.853084087371826, "learning_rate": 2.7684568722164983e-06, "loss": 0.6438, "step": 3128 }, { "epoch": 0.7648496700073332, "grad_norm": 3.4784562587738037, "learning_rate": 2.7629907531290767e-06, "loss": 0.6024, "step": 3129 }, { "epoch": 0.7650941090197996, "grad_norm": 3.2651922702789307, "learning_rate": 2.7575291704853325e-06, "loss": 0.6737, "step": 3130 }, { "epoch": 0.765338548032266, "grad_norm": 3.167628526687622, "learning_rate": 2.7520721277088023e-06, "loss": 0.6024, "step": 3131 }, { "epoch": 0.7655829870447324, "grad_norm": 3.006815195083618, "learning_rate": 2.746619628220176e-06, "loss": 0.5909, "step": 3132 }, { "epoch": 0.7658274260571988, "grad_norm": 3.065413236618042, "learning_rate": 2.7411716754372964e-06, "loss": 0.6586, "step": 3133 }, { "epoch": 0.7660718650696651, "grad_norm": 2.7193782329559326, "learning_rate": 2.7357282727751567e-06, "loss": 2.0259, "step": 3134 }, { "epoch": 0.7663163040821315, "grad_norm": 2.904202938079834, "learning_rate": 2.730289423645894e-06, "loss": 0.5203, "step": 3135 }, { "epoch": 0.7665607430945979, "grad_norm": 3.3061468601226807, "learning_rate": 2.724855131458798e-06, "loss": 0.6414, "step": 3136 }, { "epoch": 0.7668051821070643, "grad_norm": 2.839189291000366, "learning_rate": 2.719425399620296e-06, "loss": 0.6055, "step": 3137 }, { "epoch": 0.7670496211195307, "grad_norm": 3.020749807357788, "learning_rate": 2.714000231533954e-06, "loss": 0.4995, "step": 3138 }, { "epoch": 0.7672940601319971, "grad_norm": 3.7392027378082275, "learning_rate": 2.7085796306004907e-06, "loss": 0.5281, "step": 3139 }, { "epoch": 0.7675384991444635, "grad_norm": 2.9909677505493164, "learning_rate": 2.7031636002177496e-06, "loss": 0.5481, "step": 3140 }, { "epoch": 0.7677829381569299, "grad_norm": 2.932992935180664, "learning_rate": 2.697752143780713e-06, "loss": 0.6252, "step": 3141 }, { "epoch": 0.7680273771693963, "grad_norm": 3.954922676086426, "learning_rate": 2.6923452646814942e-06, "loss": 0.6438, "step": 3142 }, { "epoch": 0.7682718161818626, "grad_norm": 3.9304919242858887, "learning_rate": 2.6869429663093405e-06, "loss": 1.8718, "step": 3143 }, { "epoch": 0.768516255194329, "grad_norm": 7.261805057525635, "learning_rate": 2.681545252050626e-06, "loss": 0.56, "step": 3144 }, { "epoch": 0.7687606942067954, "grad_norm": 3.9659855365753174, "learning_rate": 2.676152125288851e-06, "loss": 0.5806, "step": 3145 }, { "epoch": 0.7690051332192618, "grad_norm": 3.0885088443756104, "learning_rate": 2.67076358940464e-06, "loss": 0.6062, "step": 3146 }, { "epoch": 0.7692495722317282, "grad_norm": 3.743597984313965, "learning_rate": 2.665379647775743e-06, "loss": 0.5626, "step": 3147 }, { "epoch": 0.7694940112441946, "grad_norm": 3.111335039138794, "learning_rate": 2.660000303777025e-06, "loss": 0.5573, "step": 3148 }, { "epoch": 0.769738450256661, "grad_norm": 3.303645372390747, "learning_rate": 2.6546255607804726e-06, "loss": 0.6009, "step": 3149 }, { "epoch": 0.7699828892691274, "grad_norm": 3.0187478065490723, "learning_rate": 2.649255422155188e-06, "loss": 0.526, "step": 3150 }, { "epoch": 0.7702273282815938, "grad_norm": 2.943671464920044, "learning_rate": 2.643889891267386e-06, "loss": 0.6326, "step": 3151 }, { "epoch": 0.7704717672940601, "grad_norm": 3.4020495414733887, "learning_rate": 2.638528971480393e-06, "loss": 0.6332, "step": 3152 }, { "epoch": 0.7707162063065265, "grad_norm": 4.747247219085693, "learning_rate": 2.633172666154645e-06, "loss": 0.6612, "step": 3153 }, { "epoch": 0.7709606453189929, "grad_norm": 3.2169346809387207, "learning_rate": 2.627820978647686e-06, "loss": 0.642, "step": 3154 }, { "epoch": 0.7712050843314593, "grad_norm": 2.937433958053589, "learning_rate": 2.6224739123141684e-06, "loss": 0.4554, "step": 3155 }, { "epoch": 0.7714495233439257, "grad_norm": 3.4266116619110107, "learning_rate": 2.6171314705058438e-06, "loss": 0.5838, "step": 3156 }, { "epoch": 0.7716939623563921, "grad_norm": 3.0606844425201416, "learning_rate": 2.611793656571565e-06, "loss": 0.5011, "step": 3157 }, { "epoch": 0.7719384013688585, "grad_norm": 3.9658639430999756, "learning_rate": 2.606460473857285e-06, "loss": 0.6529, "step": 3158 }, { "epoch": 0.7721828403813249, "grad_norm": 3.4948606491088867, "learning_rate": 2.601131925706054e-06, "loss": 0.6081, "step": 3159 }, { "epoch": 0.7724272793937913, "grad_norm": 3.4018239974975586, "learning_rate": 2.5958080154580167e-06, "loss": 0.7337, "step": 3160 }, { "epoch": 0.7726717184062576, "grad_norm": 4.360661029815674, "learning_rate": 2.5904887464504115e-06, "loss": 0.5395, "step": 3161 }, { "epoch": 0.772916157418724, "grad_norm": 2.925231695175171, "learning_rate": 2.5851741220175643e-06, "loss": 0.566, "step": 3162 }, { "epoch": 0.7731605964311904, "grad_norm": 5.550756931304932, "learning_rate": 2.5798641454908945e-06, "loss": 0.6263, "step": 3163 }, { "epoch": 0.7734050354436568, "grad_norm": 3.855212926864624, "learning_rate": 2.5745588201989037e-06, "loss": 0.7732, "step": 3164 }, { "epoch": 0.7736494744561232, "grad_norm": 3.303528070449829, "learning_rate": 2.569258149467181e-06, "loss": 0.4347, "step": 3165 }, { "epoch": 0.7738939134685896, "grad_norm": 2.3593063354492188, "learning_rate": 2.5639621366183953e-06, "loss": 1.8563, "step": 3166 }, { "epoch": 0.774138352481056, "grad_norm": 3.5858919620513916, "learning_rate": 2.558670784972298e-06, "loss": 0.6312, "step": 3167 }, { "epoch": 0.7743827914935224, "grad_norm": 5.1975016593933105, "learning_rate": 2.5533840978457135e-06, "loss": 0.6655, "step": 3168 }, { "epoch": 0.7746272305059888, "grad_norm": 3.6032609939575195, "learning_rate": 2.5481020785525534e-06, "loss": 0.53, "step": 3169 }, { "epoch": 0.7748716695184551, "grad_norm": 3.0232839584350586, "learning_rate": 2.542824730403792e-06, "loss": 0.5679, "step": 3170 }, { "epoch": 0.7751161085309215, "grad_norm": 3.177489995956421, "learning_rate": 2.537552056707483e-06, "loss": 0.6288, "step": 3171 }, { "epoch": 0.7753605475433879, "grad_norm": 2.9517323970794678, "learning_rate": 2.5322840607687426e-06, "loss": 0.6599, "step": 3172 }, { "epoch": 0.7756049865558543, "grad_norm": 3.787276268005371, "learning_rate": 2.5270207458897622e-06, "loss": 0.51, "step": 3173 }, { "epoch": 0.7758494255683207, "grad_norm": 3.143080234527588, "learning_rate": 2.5217621153697947e-06, "loss": 0.585, "step": 3174 }, { "epoch": 0.7760938645807871, "grad_norm": 2.5998079776763916, "learning_rate": 2.5165081725051564e-06, "loss": 1.8594, "step": 3175 }, { "epoch": 0.7763383035932535, "grad_norm": 3.082580327987671, "learning_rate": 2.511258920589229e-06, "loss": 0.5744, "step": 3176 }, { "epoch": 0.7765827426057199, "grad_norm": 4.161632061004639, "learning_rate": 2.50601436291245e-06, "loss": 0.5662, "step": 3177 }, { "epoch": 0.7768271816181863, "grad_norm": 3.5269548892974854, "learning_rate": 2.5007745027623164e-06, "loss": 0.7999, "step": 3178 }, { "epoch": 0.7770716206306526, "grad_norm": 2.985814332962036, "learning_rate": 2.4955393434233756e-06, "loss": 0.4679, "step": 3179 }, { "epoch": 0.777316059643119, "grad_norm": 2.7241947650909424, "learning_rate": 2.4903088881772384e-06, "loss": 0.5001, "step": 3180 }, { "epoch": 0.7775604986555854, "grad_norm": 3.6743762493133545, "learning_rate": 2.4850831403025597e-06, "loss": 0.5835, "step": 3181 }, { "epoch": 0.7778049376680518, "grad_norm": 3.5109803676605225, "learning_rate": 2.4798621030750438e-06, "loss": 0.5953, "step": 3182 }, { "epoch": 0.7780493766805182, "grad_norm": 2.7164785861968994, "learning_rate": 2.4746457797674438e-06, "loss": 1.8747, "step": 3183 }, { "epoch": 0.7782938156929846, "grad_norm": 4.3310723304748535, "learning_rate": 2.4694341736495576e-06, "loss": 0.6783, "step": 3184 }, { "epoch": 0.778538254705451, "grad_norm": 3.274315118789673, "learning_rate": 2.4642272879882245e-06, "loss": 0.6919, "step": 3185 }, { "epoch": 0.7787826937179174, "grad_norm": 2.5512311458587646, "learning_rate": 2.4590251260473288e-06, "loss": 0.5595, "step": 3186 }, { "epoch": 0.7790271327303838, "grad_norm": 3.0792834758758545, "learning_rate": 2.45382769108779e-06, "loss": 0.6065, "step": 3187 }, { "epoch": 0.7792715717428501, "grad_norm": 2.90177321434021, "learning_rate": 2.4486349863675664e-06, "loss": 0.5647, "step": 3188 }, { "epoch": 0.7795160107553165, "grad_norm": 2.9708359241485596, "learning_rate": 2.443447015141648e-06, "loss": 0.4932, "step": 3189 }, { "epoch": 0.7797604497677829, "grad_norm": 3.1886961460113525, "learning_rate": 2.4382637806620646e-06, "loss": 0.5723, "step": 3190 }, { "epoch": 0.7800048887802493, "grad_norm": 3.5020089149475098, "learning_rate": 2.433085286177872e-06, "loss": 0.5639, "step": 3191 }, { "epoch": 0.7802493277927157, "grad_norm": 4.378829002380371, "learning_rate": 2.4279115349351546e-06, "loss": 0.6527, "step": 3192 }, { "epoch": 0.7804937668051821, "grad_norm": 4.593955993652344, "learning_rate": 2.4227425301770236e-06, "loss": 0.7164, "step": 3193 }, { "epoch": 0.7807382058176485, "grad_norm": 2.77014422416687, "learning_rate": 2.417578275143616e-06, "loss": 0.4451, "step": 3194 }, { "epoch": 0.7809826448301149, "grad_norm": 3.2085678577423096, "learning_rate": 2.4124187730720916e-06, "loss": 0.6105, "step": 3195 }, { "epoch": 0.7812270838425813, "grad_norm": 2.853844404220581, "learning_rate": 2.4072640271966297e-06, "loss": 0.5077, "step": 3196 }, { "epoch": 0.7814715228550476, "grad_norm": 2.981623888015747, "learning_rate": 2.402114040748429e-06, "loss": 0.5778, "step": 3197 }, { "epoch": 0.781715961867514, "grad_norm": 3.194746971130371, "learning_rate": 2.396968816955706e-06, "loss": 0.7014, "step": 3198 }, { "epoch": 0.7819604008799804, "grad_norm": 2.9603145122528076, "learning_rate": 2.3918283590436897e-06, "loss": 0.4981, "step": 3199 }, { "epoch": 0.7822048398924468, "grad_norm": 3.691871166229248, "learning_rate": 2.3866926702346182e-06, "loss": 0.495, "step": 3200 }, { "epoch": 0.7824492789049132, "grad_norm": 3.7968385219573975, "learning_rate": 2.381561753747753e-06, "loss": 0.5113, "step": 3201 }, { "epoch": 0.7826937179173796, "grad_norm": 2.7656779289245605, "learning_rate": 2.3764356127993495e-06, "loss": 2.0152, "step": 3202 }, { "epoch": 0.782938156929846, "grad_norm": 3.0234978199005127, "learning_rate": 2.3713142506026786e-06, "loss": 0.6202, "step": 3203 }, { "epoch": 0.7831825959423124, "grad_norm": 3.1162590980529785, "learning_rate": 2.36619767036801e-06, "loss": 0.6806, "step": 3204 }, { "epoch": 0.7834270349547788, "grad_norm": 3.5407638549804688, "learning_rate": 2.361085875302621e-06, "loss": 0.621, "step": 3205 }, { "epoch": 0.7836714739672451, "grad_norm": 4.021968841552734, "learning_rate": 2.3559788686107853e-06, "loss": 0.5917, "step": 3206 }, { "epoch": 0.7839159129797115, "grad_norm": 3.0466601848602295, "learning_rate": 2.3508766534937776e-06, "loss": 0.5675, "step": 3207 }, { "epoch": 0.7841603519921779, "grad_norm": 2.976064920425415, "learning_rate": 2.3457792331498672e-06, "loss": 0.5782, "step": 3208 }, { "epoch": 0.7844047910046443, "grad_norm": 2.7482686042785645, "learning_rate": 2.3406866107743196e-06, "loss": 0.5479, "step": 3209 }, { "epoch": 0.7846492300171107, "grad_norm": 3.396845817565918, "learning_rate": 2.3355987895593923e-06, "loss": 0.556, "step": 3210 }, { "epoch": 0.7848936690295771, "grad_norm": 2.758816957473755, "learning_rate": 2.330515772694333e-06, "loss": 0.5656, "step": 3211 }, { "epoch": 0.7851381080420435, "grad_norm": 2.695256233215332, "learning_rate": 2.325437563365376e-06, "loss": 0.4653, "step": 3212 }, { "epoch": 0.7853825470545099, "grad_norm": 5.168088436126709, "learning_rate": 2.320364164755745e-06, "loss": 0.6637, "step": 3213 }, { "epoch": 0.7856269860669763, "grad_norm": 3.181368350982666, "learning_rate": 2.3152955800456468e-06, "loss": 0.613, "step": 3214 }, { "epoch": 0.7858714250794426, "grad_norm": 6.01630973815918, "learning_rate": 2.3102318124122714e-06, "loss": 0.6279, "step": 3215 }, { "epoch": 0.786115864091909, "grad_norm": 3.028567314147949, "learning_rate": 2.3051728650297856e-06, "loss": 0.5401, "step": 3216 }, { "epoch": 0.7863603031043754, "grad_norm": 3.762362480163574, "learning_rate": 2.3001187410693427e-06, "loss": 0.5725, "step": 3217 }, { "epoch": 0.7866047421168418, "grad_norm": 3.6007094383239746, "learning_rate": 2.295069443699065e-06, "loss": 0.6648, "step": 3218 }, { "epoch": 0.7868491811293082, "grad_norm": 2.9934427738189697, "learning_rate": 2.290024976084052e-06, "loss": 0.5656, "step": 3219 }, { "epoch": 0.7870936201417746, "grad_norm": 3.62497615814209, "learning_rate": 2.2849853413863756e-06, "loss": 0.6877, "step": 3220 }, { "epoch": 0.787338059154241, "grad_norm": 3.790687084197998, "learning_rate": 2.279950542765078e-06, "loss": 0.6179, "step": 3221 }, { "epoch": 0.7875824981667074, "grad_norm": 2.8185369968414307, "learning_rate": 2.274920583376171e-06, "loss": 0.5653, "step": 3222 }, { "epoch": 0.7878269371791738, "grad_norm": 3.2843177318573, "learning_rate": 2.26989546637263e-06, "loss": 0.5257, "step": 3223 }, { "epoch": 0.7880713761916401, "grad_norm": 2.947882890701294, "learning_rate": 2.2648751949044e-06, "loss": 0.6428, "step": 3224 }, { "epoch": 0.7883158152041065, "grad_norm": 3.8019323348999023, "learning_rate": 2.2598597721183834e-06, "loss": 0.5644, "step": 3225 }, { "epoch": 0.7885602542165729, "grad_norm": 4.020700454711914, "learning_rate": 2.2548492011584465e-06, "loss": 0.5394, "step": 3226 }, { "epoch": 0.7888046932290393, "grad_norm": 3.4065825939178467, "learning_rate": 2.2498434851654125e-06, "loss": 0.7011, "step": 3227 }, { "epoch": 0.7890491322415057, "grad_norm": 3.1583659648895264, "learning_rate": 2.244842627277064e-06, "loss": 0.7297, "step": 3228 }, { "epoch": 0.7892935712539721, "grad_norm": 2.93677020072937, "learning_rate": 2.2398466306281364e-06, "loss": 0.6397, "step": 3229 }, { "epoch": 0.7895380102664386, "grad_norm": 3.476424217224121, "learning_rate": 2.2348554983503156e-06, "loss": 0.6933, "step": 3230 }, { "epoch": 0.789782449278905, "grad_norm": 2.351652145385742, "learning_rate": 2.2298692335722403e-06, "loss": 1.98, "step": 3231 }, { "epoch": 0.7900268882913714, "grad_norm": 4.299272537231445, "learning_rate": 2.224887839419503e-06, "loss": 0.6083, "step": 3232 }, { "epoch": 0.7902713273038376, "grad_norm": 3.0552446842193604, "learning_rate": 2.219911319014637e-06, "loss": 0.5848, "step": 3233 }, { "epoch": 0.790515766316304, "grad_norm": 2.289609909057617, "learning_rate": 2.2149396754771203e-06, "loss": 1.9677, "step": 3234 }, { "epoch": 0.7907602053287704, "grad_norm": 3.8874871730804443, "learning_rate": 2.209972911923377e-06, "loss": 0.6953, "step": 3235 }, { "epoch": 0.7910046443412369, "grad_norm": 2.7373220920562744, "learning_rate": 2.205011031466772e-06, "loss": 0.5882, "step": 3236 }, { "epoch": 0.7912490833537033, "grad_norm": 3.414679527282715, "learning_rate": 2.200054037217605e-06, "loss": 0.6221, "step": 3237 }, { "epoch": 0.7914935223661697, "grad_norm": 3.6899285316467285, "learning_rate": 2.1951019322831193e-06, "loss": 0.6806, "step": 3238 }, { "epoch": 0.7917379613786361, "grad_norm": 3.0427722930908203, "learning_rate": 2.190154719767489e-06, "loss": 0.6509, "step": 3239 }, { "epoch": 0.7919824003911025, "grad_norm": 2.772542953491211, "learning_rate": 2.1852124027718212e-06, "loss": 0.4565, "step": 3240 }, { "epoch": 0.7922268394035689, "grad_norm": 3.280536651611328, "learning_rate": 2.1802749843941583e-06, "loss": 0.5821, "step": 3241 }, { "epoch": 0.7924712784160352, "grad_norm": 3.199457883834839, "learning_rate": 2.1753424677294675e-06, "loss": 0.6618, "step": 3242 }, { "epoch": 0.7927157174285016, "grad_norm": 3.418578624725342, "learning_rate": 2.170414855869647e-06, "loss": 0.5916, "step": 3243 }, { "epoch": 0.792960156440968, "grad_norm": 3.6555192470550537, "learning_rate": 2.1654921519035176e-06, "loss": 0.5722, "step": 3244 }, { "epoch": 0.7932045954534344, "grad_norm": 2.9484596252441406, "learning_rate": 2.1605743589168247e-06, "loss": 0.6156, "step": 3245 }, { "epoch": 0.7934490344659008, "grad_norm": 3.5383124351501465, "learning_rate": 2.155661479992236e-06, "loss": 0.5778, "step": 3246 }, { "epoch": 0.7936934734783672, "grad_norm": 5.903968811035156, "learning_rate": 2.1507535182093353e-06, "loss": 0.5943, "step": 3247 }, { "epoch": 0.7939379124908336, "grad_norm": 2.7591583728790283, "learning_rate": 2.145850476644634e-06, "loss": 0.444, "step": 3248 }, { "epoch": 0.7941823515033, "grad_norm": 2.4827725887298584, "learning_rate": 2.1409523583715486e-06, "loss": 0.4999, "step": 3249 }, { "epoch": 0.7944267905157664, "grad_norm": 3.5222556591033936, "learning_rate": 2.1360591664604125e-06, "loss": 0.5139, "step": 3250 }, { "epoch": 0.7946712295282327, "grad_norm": 2.6247849464416504, "learning_rate": 2.1311709039784734e-06, "loss": 0.5401, "step": 3251 }, { "epoch": 0.7949156685406991, "grad_norm": 3.2825536727905273, "learning_rate": 2.126287573989888e-06, "loss": 0.6492, "step": 3252 }, { "epoch": 0.7951601075531655, "grad_norm": 2.7209300994873047, "learning_rate": 2.121409179555719e-06, "loss": 0.5069, "step": 3253 }, { "epoch": 0.7954045465656319, "grad_norm": 3.4337806701660156, "learning_rate": 2.116535723733938e-06, "loss": 0.7743, "step": 3254 }, { "epoch": 0.7956489855780983, "grad_norm": 3.2049851417541504, "learning_rate": 2.11166720957942e-06, "loss": 0.5545, "step": 3255 }, { "epoch": 0.7958934245905647, "grad_norm": 2.9724299907684326, "learning_rate": 2.1068036401439418e-06, "loss": 0.6282, "step": 3256 }, { "epoch": 0.7961378636030311, "grad_norm": 2.6269307136535645, "learning_rate": 2.1019450184761814e-06, "loss": 0.5668, "step": 3257 }, { "epoch": 0.7963823026154975, "grad_norm": 3.181718349456787, "learning_rate": 2.097091347621715e-06, "loss": 0.5737, "step": 3258 }, { "epoch": 0.7966267416279639, "grad_norm": 3.2268974781036377, "learning_rate": 2.092242630623016e-06, "loss": 0.6092, "step": 3259 }, { "epoch": 0.7968711806404302, "grad_norm": 3.338627576828003, "learning_rate": 2.08739887051945e-06, "loss": 0.5365, "step": 3260 }, { "epoch": 0.7971156196528966, "grad_norm": 2.914188861846924, "learning_rate": 2.0825600703472814e-06, "loss": 0.5473, "step": 3261 }, { "epoch": 0.797360058665363, "grad_norm": 3.5348644256591797, "learning_rate": 2.0777262331396543e-06, "loss": 0.6275, "step": 3262 }, { "epoch": 0.7976044976778294, "grad_norm": 3.0596249103546143, "learning_rate": 2.0728973619266178e-06, "loss": 0.5902, "step": 3263 }, { "epoch": 0.7978489366902958, "grad_norm": 2.831759452819824, "learning_rate": 2.0680734597350947e-06, "loss": 0.5543, "step": 3264 }, { "epoch": 0.7980933757027622, "grad_norm": 3.386136054992676, "learning_rate": 2.0632545295888996e-06, "loss": 0.5854, "step": 3265 }, { "epoch": 0.7983378147152286, "grad_norm": 2.5762999057769775, "learning_rate": 2.058440574508727e-06, "loss": 0.4227, "step": 3266 }, { "epoch": 0.798582253727695, "grad_norm": 2.781954288482666, "learning_rate": 2.0536315975121545e-06, "loss": 0.4883, "step": 3267 }, { "epoch": 0.7988266927401614, "grad_norm": 3.280689239501953, "learning_rate": 2.04882760161364e-06, "loss": 0.518, "step": 3268 }, { "epoch": 0.7990711317526277, "grad_norm": 2.670980453491211, "learning_rate": 2.044028589824517e-06, "loss": 0.4825, "step": 3269 }, { "epoch": 0.7993155707650941, "grad_norm": 2.9630038738250732, "learning_rate": 2.0392345651529964e-06, "loss": 0.5242, "step": 3270 }, { "epoch": 0.7995600097775605, "grad_norm": 2.9382364749908447, "learning_rate": 2.0344455306041633e-06, "loss": 0.5933, "step": 3271 }, { "epoch": 0.7998044487900269, "grad_norm": 3.124269723892212, "learning_rate": 2.029661489179974e-06, "loss": 0.5831, "step": 3272 }, { "epoch": 0.8000488878024933, "grad_norm": 3.2077784538269043, "learning_rate": 2.0248824438792526e-06, "loss": 0.6423, "step": 3273 }, { "epoch": 0.8002933268149597, "grad_norm": 2.9261975288391113, "learning_rate": 2.0201083976976966e-06, "loss": 0.5352, "step": 3274 }, { "epoch": 0.8005377658274261, "grad_norm": 2.8065948486328125, "learning_rate": 2.015339353627865e-06, "loss": 0.7017, "step": 3275 }, { "epoch": 0.8007822048398925, "grad_norm": 2.950518846511841, "learning_rate": 2.0105753146591846e-06, "loss": 0.5949, "step": 3276 }, { "epoch": 0.8010266438523589, "grad_norm": 2.621652126312256, "learning_rate": 2.0058162837779437e-06, "loss": 1.9721, "step": 3277 }, { "epoch": 0.8012710828648252, "grad_norm": 2.55247163772583, "learning_rate": 2.001062263967286e-06, "loss": 1.8001, "step": 3278 }, { "epoch": 0.8015155218772916, "grad_norm": 3.304582118988037, "learning_rate": 1.9963132582072277e-06, "loss": 0.4885, "step": 3279 }, { "epoch": 0.801759960889758, "grad_norm": 3.22530198097229, "learning_rate": 1.9915692694746293e-06, "loss": 0.5259, "step": 3280 }, { "epoch": 0.8020043999022244, "grad_norm": 6.259829521179199, "learning_rate": 1.98683030074321e-06, "loss": 0.5966, "step": 3281 }, { "epoch": 0.8022488389146908, "grad_norm": 3.880714178085327, "learning_rate": 1.9820963549835435e-06, "loss": 0.588, "step": 3282 }, { "epoch": 0.8024932779271572, "grad_norm": 3.169184923171997, "learning_rate": 1.9773674351630543e-06, "loss": 0.5565, "step": 3283 }, { "epoch": 0.8027377169396236, "grad_norm": 3.4683644771575928, "learning_rate": 1.972643544246017e-06, "loss": 0.5455, "step": 3284 }, { "epoch": 0.80298215595209, "grad_norm": 2.546109437942505, "learning_rate": 1.967924685193552e-06, "loss": 1.8964, "step": 3285 }, { "epoch": 0.8032265949645564, "grad_norm": 7.747804164886475, "learning_rate": 1.963210860963628e-06, "loss": 0.6601, "step": 3286 }, { "epoch": 0.8034710339770227, "grad_norm": 3.7255680561065674, "learning_rate": 1.958502074511055e-06, "loss": 0.6702, "step": 3287 }, { "epoch": 0.8037154729894891, "grad_norm": 2.965513229370117, "learning_rate": 1.9537983287874883e-06, "loss": 0.6289, "step": 3288 }, { "epoch": 0.8039599120019555, "grad_norm": 2.916699171066284, "learning_rate": 1.949099626741422e-06, "loss": 0.5857, "step": 3289 }, { "epoch": 0.8042043510144219, "grad_norm": 2.9211907386779785, "learning_rate": 1.9444059713181864e-06, "loss": 0.6084, "step": 3290 }, { "epoch": 0.8044487900268883, "grad_norm": 3.1928536891937256, "learning_rate": 1.939717365459952e-06, "loss": 0.5481, "step": 3291 }, { "epoch": 0.8046932290393547, "grad_norm": 3.2966573238372803, "learning_rate": 1.935033812105721e-06, "loss": 0.4839, "step": 3292 }, { "epoch": 0.8049376680518211, "grad_norm": 2.993467330932617, "learning_rate": 1.9303553141913287e-06, "loss": 0.6134, "step": 3293 }, { "epoch": 0.8051821070642875, "grad_norm": 3.431793212890625, "learning_rate": 1.9256818746494487e-06, "loss": 0.5922, "step": 3294 }, { "epoch": 0.8054265460767539, "grad_norm": 2.649465560913086, "learning_rate": 1.9210134964095728e-06, "loss": 0.4745, "step": 3295 }, { "epoch": 0.8056709850892202, "grad_norm": 8.001269340515137, "learning_rate": 1.9163501823980258e-06, "loss": 0.5893, "step": 3296 }, { "epoch": 0.8059154241016866, "grad_norm": 3.9312362670898438, "learning_rate": 1.911691935537958e-06, "loss": 0.5305, "step": 3297 }, { "epoch": 0.806159863114153, "grad_norm": 3.161888360977173, "learning_rate": 1.9070387587493433e-06, "loss": 0.5117, "step": 3298 }, { "epoch": 0.8064043021266194, "grad_norm": 2.8150248527526855, "learning_rate": 1.9023906549489767e-06, "loss": 0.6931, "step": 3299 }, { "epoch": 0.8066487411390858, "grad_norm": 2.817641019821167, "learning_rate": 1.8977476270504725e-06, "loss": 0.6476, "step": 3300 }, { "epoch": 0.8068931801515522, "grad_norm": 3.453606605529785, "learning_rate": 1.8931096779642644e-06, "loss": 0.595, "step": 3301 }, { "epoch": 0.8071376191640186, "grad_norm": 3.230611801147461, "learning_rate": 1.8884768105976037e-06, "loss": 0.5715, "step": 3302 }, { "epoch": 0.807382058176485, "grad_norm": 4.638473033905029, "learning_rate": 1.8838490278545528e-06, "loss": 0.5706, "step": 3303 }, { "epoch": 0.8076264971889514, "grad_norm": 3.2550625801086426, "learning_rate": 1.8792263326359917e-06, "loss": 0.6181, "step": 3304 }, { "epoch": 0.8078709362014177, "grad_norm": 3.597907066345215, "learning_rate": 1.8746087278396074e-06, "loss": 0.6596, "step": 3305 }, { "epoch": 0.8081153752138841, "grad_norm": 3.790290594100952, "learning_rate": 1.8699962163598962e-06, "loss": 0.6337, "step": 3306 }, { "epoch": 0.8083598142263505, "grad_norm": 4.040053844451904, "learning_rate": 1.8653888010881637e-06, "loss": 0.5535, "step": 3307 }, { "epoch": 0.8086042532388169, "grad_norm": 3.4842941761016846, "learning_rate": 1.8607864849125202e-06, "loss": 0.6042, "step": 3308 }, { "epoch": 0.8088486922512833, "grad_norm": 2.6666111946105957, "learning_rate": 1.8561892707178765e-06, "loss": 0.419, "step": 3309 }, { "epoch": 0.8090931312637497, "grad_norm": 2.8186697959899902, "learning_rate": 1.8515971613859541e-06, "loss": 0.5402, "step": 3310 }, { "epoch": 0.8093375702762161, "grad_norm": 3.2007358074188232, "learning_rate": 1.847010159795265e-06, "loss": 2.0194, "step": 3311 }, { "epoch": 0.8095820092886825, "grad_norm": 3.150097608566284, "learning_rate": 1.8424282688211247e-06, "loss": 0.6157, "step": 3312 }, { "epoch": 0.8098264483011489, "grad_norm": 3.6688637733459473, "learning_rate": 1.8378514913356437e-06, "loss": 0.6479, "step": 3313 }, { "epoch": 0.8100708873136152, "grad_norm": 2.9342734813690186, "learning_rate": 1.833279830207726e-06, "loss": 0.5581, "step": 3314 }, { "epoch": 0.8103153263260816, "grad_norm": 2.9184458255767822, "learning_rate": 1.82871328830307e-06, "loss": 0.5614, "step": 3315 }, { "epoch": 0.810559765338548, "grad_norm": 3.491520881652832, "learning_rate": 1.8241518684841642e-06, "loss": 0.5316, "step": 3316 }, { "epoch": 0.8108042043510144, "grad_norm": 3.5354225635528564, "learning_rate": 1.8195955736102888e-06, "loss": 0.5644, "step": 3317 }, { "epoch": 0.8110486433634808, "grad_norm": 3.7676517963409424, "learning_rate": 1.8150444065375084e-06, "loss": 0.5796, "step": 3318 }, { "epoch": 0.8112930823759472, "grad_norm": 3.5893468856811523, "learning_rate": 1.8104983701186751e-06, "loss": 0.5001, "step": 3319 }, { "epoch": 0.8115375213884136, "grad_norm": 2.6374058723449707, "learning_rate": 1.8059574672034242e-06, "loss": 0.5068, "step": 3320 }, { "epoch": 0.81178196040088, "grad_norm": 3.215869188308716, "learning_rate": 1.8014217006381728e-06, "loss": 0.4949, "step": 3321 }, { "epoch": 0.8120263994133464, "grad_norm": 4.275416851043701, "learning_rate": 1.7968910732661193e-06, "loss": 0.5926, "step": 3322 }, { "epoch": 0.8122708384258127, "grad_norm": 3.5583596229553223, "learning_rate": 1.7923655879272395e-06, "loss": 0.5681, "step": 3323 }, { "epoch": 0.8125152774382791, "grad_norm": 2.748497724533081, "learning_rate": 1.7878452474582852e-06, "loss": 0.626, "step": 3324 }, { "epoch": 0.8127597164507455, "grad_norm": 3.1607143878936768, "learning_rate": 1.783330054692789e-06, "loss": 0.4609, "step": 3325 }, { "epoch": 0.8130041554632119, "grad_norm": 3.6746819019317627, "learning_rate": 1.77882001246105e-06, "loss": 0.596, "step": 3326 }, { "epoch": 0.8132485944756783, "grad_norm": 3.008025884628296, "learning_rate": 1.7743151235901401e-06, "loss": 0.5775, "step": 3327 }, { "epoch": 0.8134930334881447, "grad_norm": 3.740696907043457, "learning_rate": 1.769815390903904e-06, "loss": 0.5793, "step": 3328 }, { "epoch": 0.8137374725006111, "grad_norm": 4.715339660644531, "learning_rate": 1.7653208172229497e-06, "loss": 0.7491, "step": 3329 }, { "epoch": 0.8139819115130775, "grad_norm": 3.7908732891082764, "learning_rate": 1.7608314053646558e-06, "loss": 0.5596, "step": 3330 }, { "epoch": 0.8142263505255439, "grad_norm": 3.267061471939087, "learning_rate": 1.7563471581431623e-06, "loss": 0.5367, "step": 3331 }, { "epoch": 0.8144707895380102, "grad_norm": 2.9402544498443604, "learning_rate": 1.751868078369372e-06, "loss": 0.5918, "step": 3332 }, { "epoch": 0.8147152285504766, "grad_norm": 2.7814300060272217, "learning_rate": 1.7473941688509511e-06, "loss": 0.4855, "step": 3333 }, { "epoch": 0.814959667562943, "grad_norm": 3.9253361225128174, "learning_rate": 1.7429254323923217e-06, "loss": 0.5666, "step": 3334 }, { "epoch": 0.8152041065754094, "grad_norm": 3.1928324699401855, "learning_rate": 1.7384618717946656e-06, "loss": 0.6218, "step": 3335 }, { "epoch": 0.8154485455878758, "grad_norm": 3.4124715328216553, "learning_rate": 1.7340034898559178e-06, "loss": 0.5449, "step": 3336 }, { "epoch": 0.8156929846003422, "grad_norm": 3.0815014839172363, "learning_rate": 1.7295502893707672e-06, "loss": 0.5887, "step": 3337 }, { "epoch": 0.8159374236128086, "grad_norm": 3.380671262741089, "learning_rate": 1.72510227313066e-06, "loss": 0.6312, "step": 3338 }, { "epoch": 0.816181862625275, "grad_norm": 4.207968711853027, "learning_rate": 1.7206594439237867e-06, "loss": 0.5467, "step": 3339 }, { "epoch": 0.8164263016377414, "grad_norm": 3.0521678924560547, "learning_rate": 1.7162218045350886e-06, "loss": 0.5587, "step": 3340 }, { "epoch": 0.8166707406502077, "grad_norm": 3.4466350078582764, "learning_rate": 1.7117893577462541e-06, "loss": 0.5124, "step": 3341 }, { "epoch": 0.8169151796626741, "grad_norm": 3.064985513687134, "learning_rate": 1.7073621063357149e-06, "loss": 0.5861, "step": 3342 }, { "epoch": 0.8171596186751405, "grad_norm": 3.235666275024414, "learning_rate": 1.7029400530786478e-06, "loss": 0.642, "step": 3343 }, { "epoch": 0.8174040576876069, "grad_norm": 3.1638059616088867, "learning_rate": 1.698523200746972e-06, "loss": 0.6843, "step": 3344 }, { "epoch": 0.8176484967000733, "grad_norm": 3.068049669265747, "learning_rate": 1.6941115521093433e-06, "loss": 0.5732, "step": 3345 }, { "epoch": 0.8178929357125397, "grad_norm": 4.2525506019592285, "learning_rate": 1.6897051099311578e-06, "loss": 0.5456, "step": 3346 }, { "epoch": 0.8181373747250061, "grad_norm": 2.8963608741760254, "learning_rate": 1.6853038769745466e-06, "loss": 0.5593, "step": 3347 }, { "epoch": 0.8183818137374725, "grad_norm": 2.282418727874756, "learning_rate": 1.6809078559983815e-06, "loss": 1.9292, "step": 3348 }, { "epoch": 0.8186262527499389, "grad_norm": 3.5449211597442627, "learning_rate": 1.6765170497582583e-06, "loss": 0.6164, "step": 3349 }, { "epoch": 0.8188706917624052, "grad_norm": 2.9903504848480225, "learning_rate": 1.6721314610065098e-06, "loss": 0.4726, "step": 3350 }, { "epoch": 0.8191151307748716, "grad_norm": 3.019068717956543, "learning_rate": 1.6677510924921958e-06, "loss": 0.6304, "step": 3351 }, { "epoch": 0.819359569787338, "grad_norm": 3.0083367824554443, "learning_rate": 1.6633759469611033e-06, "loss": 0.6567, "step": 3352 }, { "epoch": 0.8196040087998044, "grad_norm": 3.0527002811431885, "learning_rate": 1.6590060271557485e-06, "loss": 0.5369, "step": 3353 }, { "epoch": 0.8198484478122708, "grad_norm": 3.012636661529541, "learning_rate": 1.654641335815369e-06, "loss": 0.6116, "step": 3354 }, { "epoch": 0.8200928868247372, "grad_norm": 3.206631898880005, "learning_rate": 1.6502818756759275e-06, "loss": 0.5986, "step": 3355 }, { "epoch": 0.8203373258372036, "grad_norm": 3.120879888534546, "learning_rate": 1.6459276494701037e-06, "loss": 0.6617, "step": 3356 }, { "epoch": 0.82058176484967, "grad_norm": 4.448589324951172, "learning_rate": 1.6415786599272998e-06, "loss": 0.5226, "step": 3357 }, { "epoch": 0.8208262038621364, "grad_norm": 3.174437999725342, "learning_rate": 1.637234909773633e-06, "loss": 0.6756, "step": 3358 }, { "epoch": 0.8210706428746027, "grad_norm": 2.672804117202759, "learning_rate": 1.6328964017319415e-06, "loss": 1.9117, "step": 3359 }, { "epoch": 0.8213150818870691, "grad_norm": 3.058718204498291, "learning_rate": 1.6285631385217715e-06, "loss": 0.6112, "step": 3360 }, { "epoch": 0.8215595208995355, "grad_norm": 3.089479923248291, "learning_rate": 1.6242351228593833e-06, "loss": 0.567, "step": 3361 }, { "epoch": 0.821803959912002, "grad_norm": 3.1912524700164795, "learning_rate": 1.6199123574577501e-06, "loss": 0.5626, "step": 3362 }, { "epoch": 0.8220483989244683, "grad_norm": 2.8705239295959473, "learning_rate": 1.6155948450265501e-06, "loss": 0.6154, "step": 3363 }, { "epoch": 0.8222928379369348, "grad_norm": 2.5632574558258057, "learning_rate": 1.6112825882721727e-06, "loss": 0.4688, "step": 3364 }, { "epoch": 0.8225372769494012, "grad_norm": 3.148527145385742, "learning_rate": 1.6069755898977102e-06, "loss": 0.5945, "step": 3365 }, { "epoch": 0.8227817159618676, "grad_norm": 3.0986287593841553, "learning_rate": 1.6026738526029584e-06, "loss": 0.6132, "step": 3366 }, { "epoch": 0.823026154974334, "grad_norm": 3.508009672164917, "learning_rate": 1.5983773790844193e-06, "loss": 0.6409, "step": 3367 }, { "epoch": 0.8232705939868002, "grad_norm": 2.3688530921936035, "learning_rate": 1.5940861720352897e-06, "loss": 0.4476, "step": 3368 }, { "epoch": 0.8235150329992666, "grad_norm": 4.0446271896362305, "learning_rate": 1.5898002341454688e-06, "loss": 0.5516, "step": 3369 }, { "epoch": 0.823759472011733, "grad_norm": 3.177182197570801, "learning_rate": 1.5855195681015522e-06, "loss": 0.5728, "step": 3370 }, { "epoch": 0.8240039110241995, "grad_norm": 3.1529734134674072, "learning_rate": 1.5812441765868292e-06, "loss": 0.5391, "step": 3371 }, { "epoch": 0.8242483500366659, "grad_norm": 3.328282594680786, "learning_rate": 1.5769740622812878e-06, "loss": 0.5315, "step": 3372 }, { "epoch": 0.8244927890491323, "grad_norm": 3.349891185760498, "learning_rate": 1.5727092278616018e-06, "loss": 0.4858, "step": 3373 }, { "epoch": 0.8247372280615987, "grad_norm": 3.391483783721924, "learning_rate": 1.5684496760011402e-06, "loss": 0.6607, "step": 3374 }, { "epoch": 0.8249816670740651, "grad_norm": 3.472534656524658, "learning_rate": 1.5641954093699562e-06, "loss": 0.6045, "step": 3375 }, { "epoch": 0.8252261060865315, "grad_norm": 3.5358104705810547, "learning_rate": 1.5599464306347944e-06, "loss": 0.4378, "step": 3376 }, { "epoch": 0.8254705450989978, "grad_norm": 2.7267298698425293, "learning_rate": 1.5557027424590808e-06, "loss": 0.4642, "step": 3377 }, { "epoch": 0.8257149841114642, "grad_norm": 3.7782864570617676, "learning_rate": 1.551464347502929e-06, "loss": 0.5535, "step": 3378 }, { "epoch": 0.8259594231239306, "grad_norm": 2.8620481491088867, "learning_rate": 1.5472312484231322e-06, "loss": 0.6043, "step": 3379 }, { "epoch": 0.826203862136397, "grad_norm": 2.610053777694702, "learning_rate": 1.5430034478731626e-06, "loss": 0.5234, "step": 3380 }, { "epoch": 0.8264483011488634, "grad_norm": 3.294963836669922, "learning_rate": 1.5387809485031745e-06, "loss": 0.7222, "step": 3381 }, { "epoch": 0.8266927401613298, "grad_norm": 3.2242040634155273, "learning_rate": 1.5345637529599976e-06, "loss": 0.6174, "step": 3382 }, { "epoch": 0.8269371791737962, "grad_norm": 3.8701202869415283, "learning_rate": 1.5303518638871361e-06, "loss": 0.614, "step": 3383 }, { "epoch": 0.8271816181862626, "grad_norm": 2.8776087760925293, "learning_rate": 1.5261452839247693e-06, "loss": 0.5321, "step": 3384 }, { "epoch": 0.827426057198729, "grad_norm": 3.0642764568328857, "learning_rate": 1.5219440157097476e-06, "loss": 0.4576, "step": 3385 }, { "epoch": 0.8276704962111953, "grad_norm": 3.0332086086273193, "learning_rate": 1.517748061875589e-06, "loss": 0.5447, "step": 3386 }, { "epoch": 0.8279149352236617, "grad_norm": 3.202911138534546, "learning_rate": 1.5135574250524898e-06, "loss": 0.5537, "step": 3387 }, { "epoch": 0.8281593742361281, "grad_norm": 2.650573253631592, "learning_rate": 1.509372107867303e-06, "loss": 0.5489, "step": 3388 }, { "epoch": 0.8284038132485945, "grad_norm": 3.583961248397827, "learning_rate": 1.5051921129435521e-06, "loss": 0.581, "step": 3389 }, { "epoch": 0.8286482522610609, "grad_norm": 3.015089750289917, "learning_rate": 1.5010174429014223e-06, "loss": 0.4641, "step": 3390 }, { "epoch": 0.8288926912735273, "grad_norm": 3.078062057495117, "learning_rate": 1.4968481003577628e-06, "loss": 0.5389, "step": 3391 }, { "epoch": 0.8291371302859937, "grad_norm": 2.9358878135681152, "learning_rate": 1.4926840879260818e-06, "loss": 0.4643, "step": 3392 }, { "epoch": 0.8293815692984601, "grad_norm": 2.809329032897949, "learning_rate": 1.4885254082165479e-06, "loss": 0.5133, "step": 3393 }, { "epoch": 0.8296260083109265, "grad_norm": 3.1469004154205322, "learning_rate": 1.4843720638359848e-06, "loss": 0.6365, "step": 3394 }, { "epoch": 0.8298704473233928, "grad_norm": 3.38547420501709, "learning_rate": 1.4802240573878735e-06, "loss": 0.4907, "step": 3395 }, { "epoch": 0.8301148863358592, "grad_norm": 2.985200881958008, "learning_rate": 1.4760813914723493e-06, "loss": 0.5556, "step": 3396 }, { "epoch": 0.8303593253483256, "grad_norm": 3.3409643173217773, "learning_rate": 1.4719440686861975e-06, "loss": 0.7045, "step": 3397 }, { "epoch": 0.830603764360792, "grad_norm": 3.254333257675171, "learning_rate": 1.4678120916228577e-06, "loss": 0.6291, "step": 3398 }, { "epoch": 0.8308482033732584, "grad_norm": 3.3871660232543945, "learning_rate": 1.4636854628724152e-06, "loss": 0.5017, "step": 3399 }, { "epoch": 0.8310926423857248, "grad_norm": 2.778360366821289, "learning_rate": 1.4595641850216035e-06, "loss": 0.4855, "step": 3400 }, { "epoch": 0.8313370813981912, "grad_norm": 3.3966450691223145, "learning_rate": 1.4554482606538044e-06, "loss": 0.7041, "step": 3401 }, { "epoch": 0.8315815204106576, "grad_norm": 3.475071430206299, "learning_rate": 1.4513376923490397e-06, "loss": 0.7118, "step": 3402 }, { "epoch": 0.831825959423124, "grad_norm": 2.6346867084503174, "learning_rate": 1.447232482683979e-06, "loss": 2.0269, "step": 3403 }, { "epoch": 0.8320703984355903, "grad_norm": 3.1301729679107666, "learning_rate": 1.4431326342319297e-06, "loss": 0.5734, "step": 3404 }, { "epoch": 0.8323148374480567, "grad_norm": 3.976536512374878, "learning_rate": 1.439038149562838e-06, "loss": 0.5534, "step": 3405 }, { "epoch": 0.8325592764605231, "grad_norm": 3.441725730895996, "learning_rate": 1.4349490312432879e-06, "loss": 0.623, "step": 3406 }, { "epoch": 0.8328037154729895, "grad_norm": 2.3651645183563232, "learning_rate": 1.4308652818365031e-06, "loss": 1.8833, "step": 3407 }, { "epoch": 0.8330481544854559, "grad_norm": 2.689514636993408, "learning_rate": 1.4267869039023375e-06, "loss": 1.9719, "step": 3408 }, { "epoch": 0.8332925934979223, "grad_norm": 2.6300761699676514, "learning_rate": 1.4227138999972801e-06, "loss": 0.4334, "step": 3409 }, { "epoch": 0.8335370325103887, "grad_norm": 3.1888070106506348, "learning_rate": 1.41864627267445e-06, "loss": 0.5239, "step": 3410 }, { "epoch": 0.8337814715228551, "grad_norm": 3.563365936279297, "learning_rate": 1.4145840244835985e-06, "loss": 0.5463, "step": 3411 }, { "epoch": 0.8340259105353215, "grad_norm": 3.6456291675567627, "learning_rate": 1.4105271579711021e-06, "loss": 0.5268, "step": 3412 }, { "epoch": 0.8342703495477878, "grad_norm": 3.0830070972442627, "learning_rate": 1.4064756756799669e-06, "loss": 0.7429, "step": 3413 }, { "epoch": 0.8345147885602542, "grad_norm": 2.9476635456085205, "learning_rate": 1.40242958014982e-06, "loss": 0.4965, "step": 3414 }, { "epoch": 0.8347592275727206, "grad_norm": 3.2751858234405518, "learning_rate": 1.398388873916916e-06, "loss": 0.7193, "step": 3415 }, { "epoch": 0.835003666585187, "grad_norm": 3.2210662364959717, "learning_rate": 1.3943535595141288e-06, "loss": 0.6244, "step": 3416 }, { "epoch": 0.8352481055976534, "grad_norm": 3.0521740913391113, "learning_rate": 1.3903236394709507e-06, "loss": 0.5066, "step": 3417 }, { "epoch": 0.8354925446101198, "grad_norm": 2.9845948219299316, "learning_rate": 1.3862991163134997e-06, "loss": 0.4988, "step": 3418 }, { "epoch": 0.8357369836225862, "grad_norm": 2.6232802867889404, "learning_rate": 1.3822799925645036e-06, "loss": 0.435, "step": 3419 }, { "epoch": 0.8359814226350526, "grad_norm": 3.771132230758667, "learning_rate": 1.378266270743308e-06, "loss": 0.6209, "step": 3420 }, { "epoch": 0.836225861647519, "grad_norm": 2.725372314453125, "learning_rate": 1.3742579533658729e-06, "loss": 0.6027, "step": 3421 }, { "epoch": 0.8364703006599853, "grad_norm": 3.415480613708496, "learning_rate": 1.3702550429447704e-06, "loss": 0.6796, "step": 3422 }, { "epoch": 0.8367147396724517, "grad_norm": 3.163527488708496, "learning_rate": 1.3662575419891822e-06, "loss": 0.6231, "step": 3423 }, { "epoch": 0.8369591786849181, "grad_norm": 2.838290214538574, "learning_rate": 1.3622654530049018e-06, "loss": 0.4933, "step": 3424 }, { "epoch": 0.8372036176973845, "grad_norm": 3.074998378753662, "learning_rate": 1.3582787784943264e-06, "loss": 0.574, "step": 3425 }, { "epoch": 0.8374480567098509, "grad_norm": 2.949556589126587, "learning_rate": 1.3542975209564624e-06, "loss": 0.4989, "step": 3426 }, { "epoch": 0.8376924957223173, "grad_norm": 3.073032855987549, "learning_rate": 1.3503216828869192e-06, "loss": 0.5717, "step": 3427 }, { "epoch": 0.8379369347347837, "grad_norm": 3.236557722091675, "learning_rate": 1.34635126677791e-06, "loss": 0.6438, "step": 3428 }, { "epoch": 0.8381813737472501, "grad_norm": 3.076246500015259, "learning_rate": 1.342386275118247e-06, "loss": 0.5371, "step": 3429 }, { "epoch": 0.8384258127597165, "grad_norm": 3.4420979022979736, "learning_rate": 1.3384267103933467e-06, "loss": 0.5614, "step": 3430 }, { "epoch": 0.8386702517721828, "grad_norm": 3.550089120864868, "learning_rate": 1.3344725750852183e-06, "loss": 0.5129, "step": 3431 }, { "epoch": 0.8389146907846492, "grad_norm": 2.726255178451538, "learning_rate": 1.3305238716724712e-06, "loss": 0.5836, "step": 3432 }, { "epoch": 0.8391591297971156, "grad_norm": 3.8522744178771973, "learning_rate": 1.3265806026303063e-06, "loss": 0.5566, "step": 3433 }, { "epoch": 0.839403568809582, "grad_norm": 3.801321506500244, "learning_rate": 1.3226427704305257e-06, "loss": 0.653, "step": 3434 }, { "epoch": 0.8396480078220484, "grad_norm": 3.028841733932495, "learning_rate": 1.3187103775415156e-06, "loss": 0.5406, "step": 3435 }, { "epoch": 0.8398924468345148, "grad_norm": 2.624055862426758, "learning_rate": 1.3147834264282566e-06, "loss": 0.4619, "step": 3436 }, { "epoch": 0.8401368858469812, "grad_norm": 3.205291271209717, "learning_rate": 1.3108619195523153e-06, "loss": 0.6027, "step": 3437 }, { "epoch": 0.8403813248594476, "grad_norm": 2.5827434062957764, "learning_rate": 1.3069458593718487e-06, "loss": 0.5382, "step": 3438 }, { "epoch": 0.840625763871914, "grad_norm": 2.9252283573150635, "learning_rate": 1.3030352483415965e-06, "loss": 0.601, "step": 3439 }, { "epoch": 0.8408702028843803, "grad_norm": 3.566788911819458, "learning_rate": 1.2991300889128867e-06, "loss": 0.5668, "step": 3440 }, { "epoch": 0.8411146418968467, "grad_norm": 3.534867286682129, "learning_rate": 1.2952303835336256e-06, "loss": 0.6072, "step": 3441 }, { "epoch": 0.8413590809093131, "grad_norm": 2.8029935359954834, "learning_rate": 1.2913361346483044e-06, "loss": 0.5564, "step": 3442 }, { "epoch": 0.8416035199217795, "grad_norm": 3.5112040042877197, "learning_rate": 1.2874473446979917e-06, "loss": 0.5928, "step": 3443 }, { "epoch": 0.8418479589342459, "grad_norm": 3.430260181427002, "learning_rate": 1.2835640161203344e-06, "loss": 0.595, "step": 3444 }, { "epoch": 0.8420923979467123, "grad_norm": 3.0886754989624023, "learning_rate": 1.2796861513495572e-06, "loss": 0.7083, "step": 3445 }, { "epoch": 0.8423368369591787, "grad_norm": 2.705249547958374, "learning_rate": 1.2758137528164593e-06, "loss": 0.5929, "step": 3446 }, { "epoch": 0.8425812759716451, "grad_norm": 2.9894611835479736, "learning_rate": 1.2719468229484133e-06, "loss": 0.4958, "step": 3447 }, { "epoch": 0.8428257149841115, "grad_norm": 3.4993841648101807, "learning_rate": 1.2680853641693603e-06, "loss": 0.5782, "step": 3448 }, { "epoch": 0.8430701539965778, "grad_norm": 4.66158390045166, "learning_rate": 1.2642293788998206e-06, "loss": 0.5414, "step": 3449 }, { "epoch": 0.8433145930090442, "grad_norm": 2.817009925842285, "learning_rate": 1.2603788695568765e-06, "loss": 0.5355, "step": 3450 }, { "epoch": 0.8435590320215106, "grad_norm": 2.995457649230957, "learning_rate": 1.2565338385541792e-06, "loss": 0.5915, "step": 3451 }, { "epoch": 0.843803471033977, "grad_norm": 3.3391265869140625, "learning_rate": 1.2526942883019466e-06, "loss": 0.671, "step": 3452 }, { "epoch": 0.8440479100464434, "grad_norm": 2.5185141563415527, "learning_rate": 1.24886022120696e-06, "loss": 1.9917, "step": 3453 }, { "epoch": 0.8442923490589098, "grad_norm": 3.22603440284729, "learning_rate": 1.2450316396725647e-06, "loss": 0.6184, "step": 3454 }, { "epoch": 0.8445367880713762, "grad_norm": 3.209512948989868, "learning_rate": 1.2412085460986668e-06, "loss": 0.6196, "step": 3455 }, { "epoch": 0.8447812270838426, "grad_norm": 3.446316957473755, "learning_rate": 1.2373909428817333e-06, "loss": 0.4349, "step": 3456 }, { "epoch": 0.845025666096309, "grad_norm": 3.9681384563446045, "learning_rate": 1.233578832414788e-06, "loss": 0.5544, "step": 3457 }, { "epoch": 0.8452701051087753, "grad_norm": 2.933584213256836, "learning_rate": 1.2297722170874138e-06, "loss": 0.5319, "step": 3458 }, { "epoch": 0.8455145441212417, "grad_norm": 2.817800760269165, "learning_rate": 1.2259710992857464e-06, "loss": 0.5167, "step": 3459 }, { "epoch": 0.8457589831337081, "grad_norm": 3.023679256439209, "learning_rate": 1.222175481392478e-06, "loss": 0.5529, "step": 3460 }, { "epoch": 0.8460034221461745, "grad_norm": 3.3433263301849365, "learning_rate": 1.2183853657868504e-06, "loss": 0.5096, "step": 3461 }, { "epoch": 0.8462478611586409, "grad_norm": 3.537362813949585, "learning_rate": 1.2146007548446593e-06, "loss": 0.5955, "step": 3462 }, { "epoch": 0.8464923001711073, "grad_norm": 3.5287539958953857, "learning_rate": 1.2108216509382487e-06, "loss": 0.6435, "step": 3463 }, { "epoch": 0.8467367391835737, "grad_norm": 3.43084716796875, "learning_rate": 1.2070480564365084e-06, "loss": 0.6399, "step": 3464 }, { "epoch": 0.8469811781960401, "grad_norm": 3.1973459720611572, "learning_rate": 1.2032799737048795e-06, "loss": 0.5751, "step": 3465 }, { "epoch": 0.8472256172085065, "grad_norm": 3.478426218032837, "learning_rate": 1.1995174051053438e-06, "loss": 0.5172, "step": 3466 }, { "epoch": 0.8474700562209728, "grad_norm": 3.005103588104248, "learning_rate": 1.195760352996429e-06, "loss": 0.5434, "step": 3467 }, { "epoch": 0.8477144952334392, "grad_norm": 3.503016948699951, "learning_rate": 1.1920088197332037e-06, "loss": 0.6582, "step": 3468 }, { "epoch": 0.8479589342459056, "grad_norm": 3.18108868598938, "learning_rate": 1.1882628076672765e-06, "loss": 0.5774, "step": 3469 }, { "epoch": 0.848203373258372, "grad_norm": 3.7590348720550537, "learning_rate": 1.184522319146797e-06, "loss": 0.5038, "step": 3470 }, { "epoch": 0.8484478122708384, "grad_norm": 3.0808441638946533, "learning_rate": 1.1807873565164507e-06, "loss": 0.5621, "step": 3471 }, { "epoch": 0.8486922512833048, "grad_norm": 3.629685640335083, "learning_rate": 1.1770579221174606e-06, "loss": 0.5175, "step": 3472 }, { "epoch": 0.8489366902957712, "grad_norm": 2.5867950916290283, "learning_rate": 1.173334018287583e-06, "loss": 1.8647, "step": 3473 }, { "epoch": 0.8491811293082376, "grad_norm": 3.691901445388794, "learning_rate": 1.1696156473611076e-06, "loss": 0.477, "step": 3474 }, { "epoch": 0.849425568320704, "grad_norm": 3.2322871685028076, "learning_rate": 1.1659028116688575e-06, "loss": 0.6267, "step": 3475 }, { "epoch": 0.8496700073331703, "grad_norm": 7.0858259201049805, "learning_rate": 1.1621955135381846e-06, "loss": 0.676, "step": 3476 }, { "epoch": 0.8499144463456367, "grad_norm": 2.888760805130005, "learning_rate": 1.158493755292971e-06, "loss": 0.4867, "step": 3477 }, { "epoch": 0.8501588853581031, "grad_norm": 2.3710877895355225, "learning_rate": 1.1547975392536237e-06, "loss": 1.944, "step": 3478 }, { "epoch": 0.8504033243705695, "grad_norm": 2.343113422393799, "learning_rate": 1.151106867737075e-06, "loss": 1.8545, "step": 3479 }, { "epoch": 0.8506477633830359, "grad_norm": 3.9684882164001465, "learning_rate": 1.1474217430567891e-06, "loss": 0.6167, "step": 3480 }, { "epoch": 0.8508922023955023, "grad_norm": 2.7422499656677246, "learning_rate": 1.1437421675227457e-06, "loss": 0.51, "step": 3481 }, { "epoch": 0.8511366414079687, "grad_norm": 3.3167271614074707, "learning_rate": 1.1400681434414474e-06, "loss": 0.5601, "step": 3482 }, { "epoch": 0.8513810804204351, "grad_norm": 3.22688627243042, "learning_rate": 1.1363996731159188e-06, "loss": 0.5387, "step": 3483 }, { "epoch": 0.8516255194329015, "grad_norm": 2.901665449142456, "learning_rate": 1.1327367588457005e-06, "loss": 0.5228, "step": 3484 }, { "epoch": 0.8518699584453678, "grad_norm": 3.9015395641326904, "learning_rate": 1.1290794029268547e-06, "loss": 0.6894, "step": 3485 }, { "epoch": 0.8521143974578342, "grad_norm": 3.6681385040283203, "learning_rate": 1.1254276076519555e-06, "loss": 0.652, "step": 3486 }, { "epoch": 0.8523588364703006, "grad_norm": 2.9650065898895264, "learning_rate": 1.1217813753100914e-06, "loss": 0.5841, "step": 3487 }, { "epoch": 0.852603275482767, "grad_norm": 3.3798346519470215, "learning_rate": 1.1181407081868667e-06, "loss": 0.6314, "step": 3488 }, { "epoch": 0.8528477144952334, "grad_norm": 3.1485350131988525, "learning_rate": 1.1145056085643957e-06, "loss": 0.6248, "step": 3489 }, { "epoch": 0.8530921535076998, "grad_norm": 2.994342803955078, "learning_rate": 1.1108760787213014e-06, "loss": 1.967, "step": 3490 }, { "epoch": 0.8533365925201662, "grad_norm": 2.749772071838379, "learning_rate": 1.107252120932717e-06, "loss": 0.5881, "step": 3491 }, { "epoch": 0.8535810315326327, "grad_norm": 2.9013586044311523, "learning_rate": 1.1036337374702833e-06, "loss": 0.5159, "step": 3492 }, { "epoch": 0.853825470545099, "grad_norm": 3.1439692974090576, "learning_rate": 1.100020930602146e-06, "loss": 0.5081, "step": 3493 }, { "epoch": 0.8540699095575653, "grad_norm": 3.1041696071624756, "learning_rate": 1.0964137025929556e-06, "loss": 0.4956, "step": 3494 }, { "epoch": 0.8543143485700317, "grad_norm": 3.3415110111236572, "learning_rate": 1.0928120557038612e-06, "loss": 0.595, "step": 3495 }, { "epoch": 0.8545587875824981, "grad_norm": 5.222829341888428, "learning_rate": 1.089215992192525e-06, "loss": 0.7418, "step": 3496 }, { "epoch": 0.8548032265949645, "grad_norm": 2.5324299335479736, "learning_rate": 1.0856255143130967e-06, "loss": 2.0127, "step": 3497 }, { "epoch": 0.855047665607431, "grad_norm": 2.8876936435699463, "learning_rate": 1.0820406243162319e-06, "loss": 0.5061, "step": 3498 }, { "epoch": 0.8552921046198974, "grad_norm": 2.9262752532958984, "learning_rate": 1.0784613244490816e-06, "loss": 0.5307, "step": 3499 }, { "epoch": 0.8555365436323638, "grad_norm": 3.3010175228118896, "learning_rate": 1.0748876169552902e-06, "loss": 0.5321, "step": 3500 }, { "epoch": 0.8557809826448302, "grad_norm": 2.5630874633789062, "learning_rate": 1.0713195040750012e-06, "loss": 1.9169, "step": 3501 }, { "epoch": 0.8560254216572966, "grad_norm": 3.1026952266693115, "learning_rate": 1.0677569880448479e-06, "loss": 0.5615, "step": 3502 }, { "epoch": 0.8562698606697629, "grad_norm": 3.3464791774749756, "learning_rate": 1.0642000710979573e-06, "loss": 0.5555, "step": 3503 }, { "epoch": 0.8565142996822293, "grad_norm": 3.168520450592041, "learning_rate": 1.0606487554639445e-06, "loss": 0.568, "step": 3504 }, { "epoch": 0.8567587386946957, "grad_norm": 5.3720855712890625, "learning_rate": 1.0571030433689145e-06, "loss": 0.6874, "step": 3505 }, { "epoch": 0.8570031777071621, "grad_norm": 2.557122230529785, "learning_rate": 1.0535629370354594e-06, "loss": 2.0013, "step": 3506 }, { "epoch": 0.8572476167196285, "grad_norm": 4.0707173347473145, "learning_rate": 1.0500284386826598e-06, "loss": 0.6974, "step": 3507 }, { "epoch": 0.8574920557320949, "grad_norm": 2.953801393508911, "learning_rate": 1.046499550526079e-06, "loss": 0.5197, "step": 3508 }, { "epoch": 0.8577364947445613, "grad_norm": 2.8796932697296143, "learning_rate": 1.0429762747777627e-06, "loss": 0.4481, "step": 3509 }, { "epoch": 0.8579809337570277, "grad_norm": 2.8739702701568604, "learning_rate": 1.0394586136462392e-06, "loss": 0.4557, "step": 3510 }, { "epoch": 0.8582253727694941, "grad_norm": 2.9083657264709473, "learning_rate": 1.035946569336519e-06, "loss": 0.5681, "step": 3511 }, { "epoch": 0.8584698117819604, "grad_norm": 3.1416075229644775, "learning_rate": 1.0324401440500887e-06, "loss": 0.5766, "step": 3512 }, { "epoch": 0.8587142507944268, "grad_norm": 3.763436794281006, "learning_rate": 1.0289393399849156e-06, "loss": 0.4242, "step": 3513 }, { "epoch": 0.8589586898068932, "grad_norm": 3.609952211380005, "learning_rate": 1.0254441593354413e-06, "loss": 0.6058, "step": 3514 }, { "epoch": 0.8592031288193596, "grad_norm": 2.9349989891052246, "learning_rate": 1.0219546042925842e-06, "loss": 0.6789, "step": 3515 }, { "epoch": 0.859447567831826, "grad_norm": 2.800530195236206, "learning_rate": 1.0184706770437337e-06, "loss": 0.5381, "step": 3516 }, { "epoch": 0.8596920068442924, "grad_norm": 3.2204642295837402, "learning_rate": 1.0149923797727556e-06, "loss": 0.5504, "step": 3517 }, { "epoch": 0.8599364458567588, "grad_norm": 2.8866260051727295, "learning_rate": 1.0115197146599843e-06, "loss": 0.4532, "step": 3518 }, { "epoch": 0.8601808848692252, "grad_norm": 2.6637918949127197, "learning_rate": 1.0080526838822214e-06, "loss": 1.9445, "step": 3519 }, { "epoch": 0.8604253238816916, "grad_norm": 3.248420000076294, "learning_rate": 1.0045912896127408e-06, "loss": 0.5784, "step": 3520 }, { "epoch": 0.8606697628941579, "grad_norm": 3.388120174407959, "learning_rate": 1.0011355340212802e-06, "loss": 0.549, "step": 3521 }, { "epoch": 0.8609142019066243, "grad_norm": 2.701521635055542, "learning_rate": 9.976854192740437e-07, "loss": 0.4971, "step": 3522 }, { "epoch": 0.8611586409190907, "grad_norm": 2.9408528804779053, "learning_rate": 9.942409475337012e-07, "loss": 0.5046, "step": 3523 }, { "epoch": 0.8614030799315571, "grad_norm": 3.63046932220459, "learning_rate": 9.908021209593832e-07, "loss": 0.5875, "step": 3524 }, { "epoch": 0.8616475189440235, "grad_norm": 2.625814437866211, "learning_rate": 9.873689417066823e-07, "loss": 0.4405, "step": 3525 }, { "epoch": 0.8618919579564899, "grad_norm": 4.167357444763184, "learning_rate": 9.839414119276501e-07, "loss": 0.6265, "step": 3526 }, { "epoch": 0.8621363969689563, "grad_norm": 2.6651597023010254, "learning_rate": 9.805195337707962e-07, "loss": 1.9388, "step": 3527 }, { "epoch": 0.8623808359814227, "grad_norm": 4.415526866912842, "learning_rate": 9.771033093810945e-07, "loss": 0.6068, "step": 3528 }, { "epoch": 0.8626252749938891, "grad_norm": 3.076693058013916, "learning_rate": 9.736927408999653e-07, "loss": 0.4294, "step": 3529 }, { "epoch": 0.8628697140063554, "grad_norm": 2.883120059967041, "learning_rate": 9.702878304652908e-07, "loss": 0.6395, "step": 3530 }, { "epoch": 0.8631141530188218, "grad_norm": 3.9754087924957275, "learning_rate": 9.668885802114002e-07, "loss": 0.6143, "step": 3531 }, { "epoch": 0.8633585920312882, "grad_norm": 3.4229605197906494, "learning_rate": 9.634949922690806e-07, "loss": 0.5813, "step": 3532 }, { "epoch": 0.8636030310437546, "grad_norm": 3.067798137664795, "learning_rate": 9.601070687655667e-07, "loss": 0.5549, "step": 3533 }, { "epoch": 0.863847470056221, "grad_norm": 3.216827392578125, "learning_rate": 9.56724811824542e-07, "loss": 0.5268, "step": 3534 }, { "epoch": 0.8640919090686874, "grad_norm": 2.9125301837921143, "learning_rate": 9.533482235661395e-07, "loss": 1.9737, "step": 3535 }, { "epoch": 0.8643363480811538, "grad_norm": 2.989281177520752, "learning_rate": 9.499773061069384e-07, "loss": 0.5109, "step": 3536 }, { "epoch": 0.8645807870936202, "grad_norm": 2.971723794937134, "learning_rate": 9.466120615599628e-07, "loss": 0.5246, "step": 3537 }, { "epoch": 0.8648252261060866, "grad_norm": 2.9526774883270264, "learning_rate": 9.432524920346819e-07, "loss": 0.5929, "step": 3538 }, { "epoch": 0.8650696651185529, "grad_norm": 3.23332142829895, "learning_rate": 9.398985996370058e-07, "loss": 0.5422, "step": 3539 }, { "epoch": 0.8653141041310193, "grad_norm": 3.561800479888916, "learning_rate": 9.36550386469287e-07, "loss": 0.6373, "step": 3540 }, { "epoch": 0.8655585431434857, "grad_norm": 3.2027857303619385, "learning_rate": 9.33207854630317e-07, "loss": 0.5397, "step": 3541 }, { "epoch": 0.8658029821559521, "grad_norm": 3.2876229286193848, "learning_rate": 9.298710062153304e-07, "loss": 0.518, "step": 3542 }, { "epoch": 0.8660474211684185, "grad_norm": 4.465147972106934, "learning_rate": 9.265398433159933e-07, "loss": 0.5502, "step": 3543 }, { "epoch": 0.8662918601808849, "grad_norm": 4.417576313018799, "learning_rate": 9.232143680204109e-07, "loss": 0.5574, "step": 3544 }, { "epoch": 0.8665362991933513, "grad_norm": 2.8578336238861084, "learning_rate": 9.198945824131222e-07, "loss": 0.4954, "step": 3545 }, { "epoch": 0.8667807382058177, "grad_norm": 2.9006593227386475, "learning_rate": 9.165804885751006e-07, "loss": 0.5072, "step": 3546 }, { "epoch": 0.8670251772182841, "grad_norm": 2.952204465866089, "learning_rate": 9.132720885837509e-07, "loss": 0.5391, "step": 3547 }, { "epoch": 0.8672696162307504, "grad_norm": 2.539088726043701, "learning_rate": 9.099693845129087e-07, "loss": 1.8635, "step": 3548 }, { "epoch": 0.8675140552432168, "grad_norm": 2.843809127807617, "learning_rate": 9.066723784328412e-07, "loss": 0.6246, "step": 3549 }, { "epoch": 0.8677584942556832, "grad_norm": 3.1896400451660156, "learning_rate": 9.03381072410241e-07, "loss": 0.4663, "step": 3550 }, { "epoch": 0.8680029332681496, "grad_norm": 3.474761724472046, "learning_rate": 9.000954685082286e-07, "loss": 0.4892, "step": 3551 }, { "epoch": 0.868247372280616, "grad_norm": 3.1440491676330566, "learning_rate": 8.968155687863512e-07, "loss": 0.5951, "step": 3552 }, { "epoch": 0.8684918112930824, "grad_norm": 3.480292797088623, "learning_rate": 8.935413753005784e-07, "loss": 0.6139, "step": 3553 }, { "epoch": 0.8687362503055488, "grad_norm": 2.9116463661193848, "learning_rate": 8.902728901033065e-07, "loss": 0.6055, "step": 3554 }, { "epoch": 0.8689806893180152, "grad_norm": 3.0884416103363037, "learning_rate": 8.870101152433497e-07, "loss": 0.5688, "step": 3555 }, { "epoch": 0.8692251283304816, "grad_norm": 3.200479030609131, "learning_rate": 8.837530527659443e-07, "loss": 0.6305, "step": 3556 }, { "epoch": 0.8694695673429479, "grad_norm": 2.9874043464660645, "learning_rate": 8.805017047127451e-07, "loss": 0.5621, "step": 3557 }, { "epoch": 0.8697140063554143, "grad_norm": 3.9589836597442627, "learning_rate": 8.772560731218293e-07, "loss": 0.6652, "step": 3558 }, { "epoch": 0.8699584453678807, "grad_norm": 3.272484540939331, "learning_rate": 8.74016160027683e-07, "loss": 0.6541, "step": 3559 }, { "epoch": 0.8702028843803471, "grad_norm": 3.113635778427124, "learning_rate": 8.707819674612151e-07, "loss": 0.5202, "step": 3560 }, { "epoch": 0.8704473233928135, "grad_norm": 3.079364776611328, "learning_rate": 8.675534974497435e-07, "loss": 0.5399, "step": 3561 }, { "epoch": 0.8706917624052799, "grad_norm": 3.283508777618408, "learning_rate": 8.643307520170018e-07, "loss": 0.5759, "step": 3562 }, { "epoch": 0.8709362014177463, "grad_norm": 3.144463539123535, "learning_rate": 8.611137331831331e-07, "loss": 0.4659, "step": 3563 }, { "epoch": 0.8711806404302127, "grad_norm": 3.242224931716919, "learning_rate": 8.579024429646932e-07, "loss": 0.5753, "step": 3564 }, { "epoch": 0.8714250794426791, "grad_norm": 3.6955604553222656, "learning_rate": 8.54696883374646e-07, "loss": 0.5807, "step": 3565 }, { "epoch": 0.8716695184551454, "grad_norm": 3.042665719985962, "learning_rate": 8.514970564223623e-07, "loss": 0.5616, "step": 3566 }, { "epoch": 0.8719139574676118, "grad_norm": 4.710071563720703, "learning_rate": 8.483029641136209e-07, "loss": 0.6149, "step": 3567 }, { "epoch": 0.8721583964800782, "grad_norm": 2.4829370975494385, "learning_rate": 8.451146084506046e-07, "loss": 1.8073, "step": 3568 }, { "epoch": 0.8724028354925446, "grad_norm": 3.0676817893981934, "learning_rate": 8.419319914319002e-07, "loss": 0.54, "step": 3569 }, { "epoch": 0.872647274505011, "grad_norm": 3.504573345184326, "learning_rate": 8.387551150524998e-07, "loss": 0.4817, "step": 3570 }, { "epoch": 0.8728917135174774, "grad_norm": 3.500375270843506, "learning_rate": 8.355839813037936e-07, "loss": 0.5654, "step": 3571 }, { "epoch": 0.8731361525299438, "grad_norm": 2.4557383060455322, "learning_rate": 8.324185921735728e-07, "loss": 1.9117, "step": 3572 }, { "epoch": 0.8733805915424102, "grad_norm": 2.697639226913452, "learning_rate": 8.292589496460323e-07, "loss": 0.4705, "step": 3573 }, { "epoch": 0.8736250305548766, "grad_norm": 3.3228890895843506, "learning_rate": 8.26105055701758e-07, "loss": 0.5901, "step": 3574 }, { "epoch": 0.8738694695673429, "grad_norm": 3.011554479598999, "learning_rate": 8.229569123177361e-07, "loss": 0.6151, "step": 3575 }, { "epoch": 0.8741139085798093, "grad_norm": 2.707702875137329, "learning_rate": 8.198145214673481e-07, "loss": 0.5116, "step": 3576 }, { "epoch": 0.8743583475922757, "grad_norm": 3.353950023651123, "learning_rate": 8.166778851203672e-07, "loss": 0.646, "step": 3577 }, { "epoch": 0.8746027866047421, "grad_norm": 2.3703274726867676, "learning_rate": 8.135470052429628e-07, "loss": 1.8762, "step": 3578 }, { "epoch": 0.8748472256172085, "grad_norm": 3.6735572814941406, "learning_rate": 8.10421883797694e-07, "loss": 0.6678, "step": 3579 }, { "epoch": 0.8750916646296749, "grad_norm": 3.0874695777893066, "learning_rate": 8.073025227435093e-07, "loss": 0.6327, "step": 3580 }, { "epoch": 0.8753361036421413, "grad_norm": 2.686279296875, "learning_rate": 8.041889240357493e-07, "loss": 0.5964, "step": 3581 }, { "epoch": 0.8755805426546077, "grad_norm": 3.0673916339874268, "learning_rate": 8.0108108962614e-07, "loss": 0.6688, "step": 3582 }, { "epoch": 0.875824981667074, "grad_norm": 3.7942118644714355, "learning_rate": 7.979790214627937e-07, "loss": 0.5957, "step": 3583 }, { "epoch": 0.8760694206795404, "grad_norm": 4.573256015777588, "learning_rate": 7.948827214902122e-07, "loss": 0.5473, "step": 3584 }, { "epoch": 0.8763138596920068, "grad_norm": 3.371189832687378, "learning_rate": 7.917921916492765e-07, "loss": 0.5922, "step": 3585 }, { "epoch": 0.8765582987044732, "grad_norm": 3.7442123889923096, "learning_rate": 7.887074338772538e-07, "loss": 0.5704, "step": 3586 }, { "epoch": 0.8768027377169396, "grad_norm": 3.203345775604248, "learning_rate": 7.856284501077927e-07, "loss": 0.5671, "step": 3587 }, { "epoch": 0.877047176729406, "grad_norm": 2.4070796966552734, "learning_rate": 7.825552422709192e-07, "loss": 0.4202, "step": 3588 }, { "epoch": 0.8772916157418724, "grad_norm": 2.3015387058258057, "learning_rate": 7.79487812293046e-07, "loss": 1.8715, "step": 3589 }, { "epoch": 0.8775360547543388, "grad_norm": 3.0606741905212402, "learning_rate": 7.764261620969571e-07, "loss": 0.6387, "step": 3590 }, { "epoch": 0.8777804937668052, "grad_norm": 2.820016622543335, "learning_rate": 7.733702936018162e-07, "loss": 0.4976, "step": 3591 }, { "epoch": 0.8780249327792715, "grad_norm": 3.0214173793792725, "learning_rate": 7.703202087231632e-07, "loss": 0.5504, "step": 3592 }, { "epoch": 0.8782693717917379, "grad_norm": 3.4106690883636475, "learning_rate": 7.672759093729121e-07, "loss": 0.5901, "step": 3593 }, { "epoch": 0.8785138108042043, "grad_norm": 2.8477838039398193, "learning_rate": 7.642373974593487e-07, "loss": 0.4521, "step": 3594 }, { "epoch": 0.8787582498166707, "grad_norm": 3.21529221534729, "learning_rate": 7.612046748871327e-07, "loss": 0.6287, "step": 3595 }, { "epoch": 0.8790026888291371, "grad_norm": 2.869361639022827, "learning_rate": 7.581777435572956e-07, "loss": 0.5404, "step": 3596 }, { "epoch": 0.8792471278416035, "grad_norm": 2.8519232273101807, "learning_rate": 7.551566053672377e-07, "loss": 0.4786, "step": 3597 }, { "epoch": 0.8794915668540699, "grad_norm": 2.9038546085357666, "learning_rate": 7.521412622107271e-07, "loss": 0.5005, "step": 3598 }, { "epoch": 0.8797360058665363, "grad_norm": 11.24041748046875, "learning_rate": 7.49131715977901e-07, "loss": 0.5871, "step": 3599 }, { "epoch": 0.8799804448790027, "grad_norm": 2.5400288105010986, "learning_rate": 7.461279685552614e-07, "loss": 1.9806, "step": 3600 }, { "epoch": 0.880224883891469, "grad_norm": 2.685638666152954, "learning_rate": 7.431300218256754e-07, "loss": 0.5443, "step": 3601 }, { "epoch": 0.8804693229039354, "grad_norm": 3.251020669937134, "learning_rate": 7.401378776683754e-07, "loss": 0.614, "step": 3602 }, { "epoch": 0.8807137619164018, "grad_norm": 2.709386110305786, "learning_rate": 7.371515379589555e-07, "loss": 0.4861, "step": 3603 }, { "epoch": 0.8809582009288682, "grad_norm": 4.262940406799316, "learning_rate": 7.341710045693717e-07, "loss": 0.6105, "step": 3604 }, { "epoch": 0.8812026399413346, "grad_norm": 3.71225643157959, "learning_rate": 7.311962793679417e-07, "loss": 0.7028, "step": 3605 }, { "epoch": 0.881447078953801, "grad_norm": 2.9761314392089844, "learning_rate": 7.282273642193383e-07, "loss": 0.6167, "step": 3606 }, { "epoch": 0.8816915179662674, "grad_norm": 2.59669828414917, "learning_rate": 7.252642609845973e-07, "loss": 0.4977, "step": 3607 }, { "epoch": 0.8819359569787338, "grad_norm": 3.502819776535034, "learning_rate": 7.223069715211073e-07, "loss": 0.6519, "step": 3608 }, { "epoch": 0.8821803959912002, "grad_norm": 2.915161609649658, "learning_rate": 7.193554976826145e-07, "loss": 0.5265, "step": 3609 }, { "epoch": 0.8824248350036665, "grad_norm": 3.6396989822387695, "learning_rate": 7.16409841319221e-07, "loss": 0.6412, "step": 3610 }, { "epoch": 0.8826692740161329, "grad_norm": 3.3161749839782715, "learning_rate": 7.13470004277379e-07, "loss": 0.6229, "step": 3611 }, { "epoch": 0.8829137130285993, "grad_norm": 4.260815620422363, "learning_rate": 7.105359883998953e-07, "loss": 0.4812, "step": 3612 }, { "epoch": 0.8831581520410657, "grad_norm": 4.785272598266602, "learning_rate": 7.076077955259275e-07, "loss": 0.5726, "step": 3613 }, { "epoch": 0.8834025910535321, "grad_norm": 2.9758827686309814, "learning_rate": 7.046854274909809e-07, "loss": 0.4628, "step": 3614 }, { "epoch": 0.8836470300659985, "grad_norm": 2.9017090797424316, "learning_rate": 7.017688861269145e-07, "loss": 0.5333, "step": 3615 }, { "epoch": 0.8838914690784649, "grad_norm": 3.9723899364471436, "learning_rate": 6.988581732619293e-07, "loss": 0.5544, "step": 3616 }, { "epoch": 0.8841359080909313, "grad_norm": 3.3946757316589355, "learning_rate": 6.959532907205758e-07, "loss": 0.6218, "step": 3617 }, { "epoch": 0.8843803471033977, "grad_norm": 2.7967257499694824, "learning_rate": 6.930542403237506e-07, "loss": 0.5608, "step": 3618 }, { "epoch": 0.884624786115864, "grad_norm": 2.9319894313812256, "learning_rate": 6.901610238886891e-07, "loss": 0.5659, "step": 3619 }, { "epoch": 0.8848692251283304, "grad_norm": 3.273327350616455, "learning_rate": 6.872736432289794e-07, "loss": 0.6086, "step": 3620 }, { "epoch": 0.8851136641407968, "grad_norm": 3.4320311546325684, "learning_rate": 6.843921001545429e-07, "loss": 0.5059, "step": 3621 }, { "epoch": 0.8853581031532632, "grad_norm": 3.1227705478668213, "learning_rate": 6.815163964716453e-07, "loss": 0.4984, "step": 3622 }, { "epoch": 0.8856025421657296, "grad_norm": 3.0057742595672607, "learning_rate": 6.786465339828918e-07, "loss": 0.6151, "step": 3623 }, { "epoch": 0.885846981178196, "grad_norm": 3.298367500305176, "learning_rate": 6.757825144872255e-07, "loss": 0.6156, "step": 3624 }, { "epoch": 0.8860914201906624, "grad_norm": 3.862901210784912, "learning_rate": 6.729243397799268e-07, "loss": 0.6285, "step": 3625 }, { "epoch": 0.8863358592031289, "grad_norm": 3.6718273162841797, "learning_rate": 6.700720116526116e-07, "loss": 0.4351, "step": 3626 }, { "epoch": 0.8865802982155953, "grad_norm": 3.213244676589966, "learning_rate": 6.672255318932342e-07, "loss": 0.5081, "step": 3627 }, { "epoch": 0.8868247372280615, "grad_norm": 3.0694515705108643, "learning_rate": 6.643849022860794e-07, "loss": 0.6884, "step": 3628 }, { "epoch": 0.887069176240528, "grad_norm": 3.5863587856292725, "learning_rate": 6.615501246117673e-07, "loss": 0.6471, "step": 3629 }, { "epoch": 0.8873136152529943, "grad_norm": 3.412667989730835, "learning_rate": 6.587212006472465e-07, "loss": 0.5116, "step": 3630 }, { "epoch": 0.8875580542654607, "grad_norm": 3.131528615951538, "learning_rate": 6.558981321658009e-07, "loss": 0.529, "step": 3631 }, { "epoch": 0.8878024932779272, "grad_norm": 3.2789711952209473, "learning_rate": 6.530809209370403e-07, "loss": 0.5665, "step": 3632 }, { "epoch": 0.8880469322903936, "grad_norm": 3.286752700805664, "learning_rate": 6.502695687269034e-07, "loss": 0.661, "step": 3633 }, { "epoch": 0.88829137130286, "grad_norm": 3.495955467224121, "learning_rate": 6.474640772976571e-07, "loss": 0.6199, "step": 3634 }, { "epoch": 0.8885358103153264, "grad_norm": 2.993870973587036, "learning_rate": 6.446644484078968e-07, "loss": 0.5167, "step": 3635 }, { "epoch": 0.8887802493277928, "grad_norm": 2.757718324661255, "learning_rate": 6.418706838125388e-07, "loss": 0.5215, "step": 3636 }, { "epoch": 0.889024688340259, "grad_norm": 3.5278823375701904, "learning_rate": 6.390827852628256e-07, "loss": 0.5353, "step": 3637 }, { "epoch": 0.8892691273527255, "grad_norm": 3.1759135723114014, "learning_rate": 6.363007545063227e-07, "loss": 0.5631, "step": 3638 }, { "epoch": 0.8895135663651919, "grad_norm": 2.910161256790161, "learning_rate": 6.33524593286915e-07, "loss": 0.6147, "step": 3639 }, { "epoch": 0.8897580053776583, "grad_norm": 3.1023874282836914, "learning_rate": 6.307543033448127e-07, "loss": 0.5945, "step": 3640 }, { "epoch": 0.8900024443901247, "grad_norm": 40.76762390136719, "learning_rate": 6.279898864165423e-07, "loss": 0.5745, "step": 3641 }, { "epoch": 0.8902468834025911, "grad_norm": 4.0469231605529785, "learning_rate": 6.252313442349489e-07, "loss": 0.5816, "step": 3642 }, { "epoch": 0.8904913224150575, "grad_norm": 3.096689462661743, "learning_rate": 6.22478678529197e-07, "loss": 0.5289, "step": 3643 }, { "epoch": 0.8907357614275239, "grad_norm": 3.436739921569824, "learning_rate": 6.197318910247652e-07, "loss": 0.6384, "step": 3644 }, { "epoch": 0.8909802004399903, "grad_norm": 3.6540892124176025, "learning_rate": 6.169909834434484e-07, "loss": 0.6279, "step": 3645 }, { "epoch": 0.8912246394524566, "grad_norm": 3.2602572441101074, "learning_rate": 6.14255957503358e-07, "loss": 0.5458, "step": 3646 }, { "epoch": 0.891469078464923, "grad_norm": 3.537839651107788, "learning_rate": 6.115268149189147e-07, "loss": 0.6928, "step": 3647 }, { "epoch": 0.8917135174773894, "grad_norm": 6.434822082519531, "learning_rate": 6.088035574008532e-07, "loss": 0.57, "step": 3648 }, { "epoch": 0.8919579564898558, "grad_norm": 2.3490285873413086, "learning_rate": 6.060861866562196e-07, "loss": 1.9858, "step": 3649 }, { "epoch": 0.8922023955023222, "grad_norm": 3.2424468994140625, "learning_rate": 6.033747043883675e-07, "loss": 0.6247, "step": 3650 }, { "epoch": 0.8924468345147886, "grad_norm": 3.5415916442871094, "learning_rate": 6.006691122969644e-07, "loss": 0.5527, "step": 3651 }, { "epoch": 0.892691273527255, "grad_norm": 4.1693572998046875, "learning_rate": 5.979694120779811e-07, "loss": 0.6999, "step": 3652 }, { "epoch": 0.8929357125397214, "grad_norm": 3.3961987495422363, "learning_rate": 5.952756054236953e-07, "loss": 0.4981, "step": 3653 }, { "epoch": 0.8931801515521878, "grad_norm": 3.0379087924957275, "learning_rate": 5.925876940226926e-07, "loss": 0.7162, "step": 3654 }, { "epoch": 0.8934245905646541, "grad_norm": 3.3472681045532227, "learning_rate": 5.899056795598635e-07, "loss": 0.5355, "step": 3655 }, { "epoch": 0.8936690295771205, "grad_norm": 3.012376308441162, "learning_rate": 5.872295637163983e-07, "loss": 0.5354, "step": 3656 }, { "epoch": 0.8939134685895869, "grad_norm": 2.4523651599884033, "learning_rate": 5.845593481697931e-07, "loss": 1.8777, "step": 3657 }, { "epoch": 0.8941579076020533, "grad_norm": 3.6199769973754883, "learning_rate": 5.818950345938457e-07, "loss": 0.6562, "step": 3658 }, { "epoch": 0.8944023466145197, "grad_norm": 3.089425563812256, "learning_rate": 5.792366246586512e-07, "loss": 0.5348, "step": 3659 }, { "epoch": 0.8946467856269861, "grad_norm": 3.0774762630462646, "learning_rate": 5.765841200306088e-07, "loss": 0.7085, "step": 3660 }, { "epoch": 0.8948912246394525, "grad_norm": 3.439373731613159, "learning_rate": 5.739375223724108e-07, "loss": 0.6117, "step": 3661 }, { "epoch": 0.8951356636519189, "grad_norm": 2.7386152744293213, "learning_rate": 5.712968333430502e-07, "loss": 0.6133, "step": 3662 }, { "epoch": 0.8953801026643853, "grad_norm": 3.5213887691497803, "learning_rate": 5.686620545978161e-07, "loss": 0.5565, "step": 3663 }, { "epoch": 0.8956245416768516, "grad_norm": 3.4529879093170166, "learning_rate": 5.660331877882897e-07, "loss": 0.6187, "step": 3664 }, { "epoch": 0.895868980689318, "grad_norm": 2.5605435371398926, "learning_rate": 5.634102345623526e-07, "loss": 0.5098, "step": 3665 }, { "epoch": 0.8961134197017844, "grad_norm": 3.1192750930786133, "learning_rate": 5.607931965641733e-07, "loss": 0.4693, "step": 3666 }, { "epoch": 0.8963578587142508, "grad_norm": 2.7218759059906006, "learning_rate": 5.581820754342138e-07, "loss": 0.4582, "step": 3667 }, { "epoch": 0.8966022977267172, "grad_norm": 2.4536046981811523, "learning_rate": 5.555768728092304e-07, "loss": 1.8286, "step": 3668 }, { "epoch": 0.8968467367391836, "grad_norm": 3.203338384628296, "learning_rate": 5.529775903222656e-07, "loss": 0.5028, "step": 3669 }, { "epoch": 0.89709117575165, "grad_norm": 4.010106563568115, "learning_rate": 5.50384229602653e-07, "loss": 0.6932, "step": 3670 }, { "epoch": 0.8973356147641164, "grad_norm": 3.7111124992370605, "learning_rate": 5.477967922760141e-07, "loss": 0.6522, "step": 3671 }, { "epoch": 0.8975800537765828, "grad_norm": 3.3121304512023926, "learning_rate": 5.452152799642551e-07, "loss": 0.5106, "step": 3672 }, { "epoch": 0.8978244927890491, "grad_norm": 2.946337938308716, "learning_rate": 5.426396942855727e-07, "loss": 0.4833, "step": 3673 }, { "epoch": 0.8980689318015155, "grad_norm": 3.6235878467559814, "learning_rate": 5.400700368544442e-07, "loss": 0.6588, "step": 3674 }, { "epoch": 0.8983133708139819, "grad_norm": 3.8211817741394043, "learning_rate": 5.375063092816313e-07, "loss": 0.5454, "step": 3675 }, { "epoch": 0.8985578098264483, "grad_norm": 3.652818202972412, "learning_rate": 5.349485131741828e-07, "loss": 0.6498, "step": 3676 }, { "epoch": 0.8988022488389147, "grad_norm": 3.070875406265259, "learning_rate": 5.323966501354261e-07, "loss": 0.4917, "step": 3677 }, { "epoch": 0.8990466878513811, "grad_norm": 2.9537672996520996, "learning_rate": 5.298507217649684e-07, "loss": 0.525, "step": 3678 }, { "epoch": 0.8992911268638475, "grad_norm": 3.348336935043335, "learning_rate": 5.273107296586987e-07, "loss": 0.5026, "step": 3679 }, { "epoch": 0.8995355658763139, "grad_norm": 3.548139810562134, "learning_rate": 5.247766754087846e-07, "loss": 0.5901, "step": 3680 }, { "epoch": 0.8997800048887803, "grad_norm": 3.775935411453247, "learning_rate": 5.222485606036709e-07, "loss": 0.6468, "step": 3681 }, { "epoch": 0.9000244439012466, "grad_norm": 3.2515830993652344, "learning_rate": 5.197263868280821e-07, "loss": 0.66, "step": 3682 }, { "epoch": 0.900268882913713, "grad_norm": 2.97196888923645, "learning_rate": 5.172101556630149e-07, "loss": 0.5048, "step": 3683 }, { "epoch": 0.9005133219261794, "grad_norm": 3.2715892791748047, "learning_rate": 5.146998686857419e-07, "loss": 0.5409, "step": 3684 }, { "epoch": 0.9007577609386458, "grad_norm": 3.282442092895508, "learning_rate": 5.121955274698098e-07, "loss": 0.5975, "step": 3685 }, { "epoch": 0.9010021999511122, "grad_norm": 3.06097674369812, "learning_rate": 5.09697133585042e-07, "loss": 0.5095, "step": 3686 }, { "epoch": 0.9012466389635786, "grad_norm": 3.0278491973876953, "learning_rate": 5.072046885975291e-07, "loss": 0.5366, "step": 3687 }, { "epoch": 0.901491077976045, "grad_norm": 4.123024940490723, "learning_rate": 5.047181940696333e-07, "loss": 0.5028, "step": 3688 }, { "epoch": 0.9017355169885114, "grad_norm": 3.111257553100586, "learning_rate": 5.02237651559988e-07, "loss": 0.5759, "step": 3689 }, { "epoch": 0.9019799560009778, "grad_norm": 3.1422526836395264, "learning_rate": 4.997630626234962e-07, "loss": 0.4266, "step": 3690 }, { "epoch": 0.9022243950134441, "grad_norm": 3.3071560859680176, "learning_rate": 4.972944288113268e-07, "loss": 0.6064, "step": 3691 }, { "epoch": 0.9024688340259105, "grad_norm": 2.9000163078308105, "learning_rate": 4.948317516709177e-07, "loss": 1.9377, "step": 3692 }, { "epoch": 0.9027132730383769, "grad_norm": 2.9188854694366455, "learning_rate": 4.923750327459731e-07, "loss": 0.5908, "step": 3693 }, { "epoch": 0.9029577120508433, "grad_norm": 3.1524181365966797, "learning_rate": 4.899242735764609e-07, "loss": 0.563, "step": 3694 }, { "epoch": 0.9032021510633097, "grad_norm": 4.9470367431640625, "learning_rate": 4.874794756986145e-07, "loss": 0.5097, "step": 3695 }, { "epoch": 0.9034465900757761, "grad_norm": 3.1351184844970703, "learning_rate": 4.85040640644927e-07, "loss": 0.5333, "step": 3696 }, { "epoch": 0.9036910290882425, "grad_norm": 3.649569034576416, "learning_rate": 4.826077699441611e-07, "loss": 0.6437, "step": 3697 }, { "epoch": 0.9039354681007089, "grad_norm": 2.842440366744995, "learning_rate": 4.801808651213346e-07, "loss": 0.5591, "step": 3698 }, { "epoch": 0.9041799071131753, "grad_norm": 2.9804258346557617, "learning_rate": 4.777599276977262e-07, "loss": 0.6262, "step": 3699 }, { "epoch": 0.9044243461256416, "grad_norm": 4.149869918823242, "learning_rate": 4.7534495919087767e-07, "loss": 0.6289, "step": 3700 }, { "epoch": 0.904668785138108, "grad_norm": 2.6993746757507324, "learning_rate": 4.729359611145845e-07, "loss": 1.899, "step": 3701 }, { "epoch": 0.9049132241505744, "grad_norm": 2.8985257148742676, "learning_rate": 4.705329349789034e-07, "loss": 0.4559, "step": 3702 }, { "epoch": 0.9051576631630408, "grad_norm": 2.9051637649536133, "learning_rate": 4.6813588229014605e-07, "loss": 0.5752, "step": 3703 }, { "epoch": 0.9054021021755072, "grad_norm": 3.1530659198760986, "learning_rate": 4.6574480455087836e-07, "loss": 0.5387, "step": 3704 }, { "epoch": 0.9056465411879736, "grad_norm": 3.1871185302734375, "learning_rate": 4.6335970325992463e-07, "loss": 0.5843, "step": 3705 }, { "epoch": 0.90589098020044, "grad_norm": 2.9980342388153076, "learning_rate": 4.60980579912359e-07, "loss": 0.5831, "step": 3706 }, { "epoch": 0.9061354192129064, "grad_norm": 3.6054625511169434, "learning_rate": 4.5860743599951186e-07, "loss": 0.5821, "step": 3707 }, { "epoch": 0.9063798582253728, "grad_norm": 3.908937931060791, "learning_rate": 4.5624027300896214e-07, "loss": 0.5487, "step": 3708 }, { "epoch": 0.9066242972378391, "grad_norm": 3.200122833251953, "learning_rate": 4.538790924245418e-07, "loss": 0.4827, "step": 3709 }, { "epoch": 0.9068687362503055, "grad_norm": 3.2075066566467285, "learning_rate": 4.515238957263313e-07, "loss": 0.5491, "step": 3710 }, { "epoch": 0.9071131752627719, "grad_norm": 2.8623270988464355, "learning_rate": 4.49174684390663e-07, "loss": 0.6853, "step": 3711 }, { "epoch": 0.9073576142752383, "grad_norm": 3.1399548053741455, "learning_rate": 4.4683145989011356e-07, "loss": 0.6352, "step": 3712 }, { "epoch": 0.9076020532877047, "grad_norm": 3.4346516132354736, "learning_rate": 4.4449422369351013e-07, "loss": 0.7099, "step": 3713 }, { "epoch": 0.9078464923001711, "grad_norm": 2.989849805831909, "learning_rate": 4.421629772659242e-07, "loss": 0.4438, "step": 3714 }, { "epoch": 0.9080909313126375, "grad_norm": 3.322847604751587, "learning_rate": 4.3983772206867446e-07, "loss": 0.6186, "step": 3715 }, { "epoch": 0.9083353703251039, "grad_norm": 3.3168976306915283, "learning_rate": 4.375184595593207e-07, "loss": 0.4834, "step": 3716 }, { "epoch": 0.9085798093375703, "grad_norm": 3.5158891677856445, "learning_rate": 4.35205191191671e-07, "loss": 0.6321, "step": 3717 }, { "epoch": 0.9088242483500366, "grad_norm": 2.9080677032470703, "learning_rate": 4.3289791841577113e-07, "loss": 0.5509, "step": 3718 }, { "epoch": 0.909068687362503, "grad_norm": 3.4545187950134277, "learning_rate": 4.305966426779118e-07, "loss": 0.5999, "step": 3719 }, { "epoch": 0.9093131263749694, "grad_norm": 3.0747053623199463, "learning_rate": 4.283013654206236e-07, "loss": 0.5349, "step": 3720 }, { "epoch": 0.9095575653874358, "grad_norm": 2.9126248359680176, "learning_rate": 4.260120880826768e-07, "loss": 0.5253, "step": 3721 }, { "epoch": 0.9098020043999022, "grad_norm": 3.5154640674591064, "learning_rate": 4.2372881209908234e-07, "loss": 0.6083, "step": 3722 }, { "epoch": 0.9100464434123686, "grad_norm": 3.2760703563690186, "learning_rate": 4.214515389010865e-07, "loss": 0.4783, "step": 3723 }, { "epoch": 0.910290882424835, "grad_norm": 3.1350185871124268, "learning_rate": 4.1918026991617513e-07, "loss": 0.4299, "step": 3724 }, { "epoch": 0.9105353214373014, "grad_norm": 3.5196869373321533, "learning_rate": 4.1691500656806825e-07, "loss": 0.557, "step": 3725 }, { "epoch": 0.9107797604497678, "grad_norm": 3.4376039505004883, "learning_rate": 4.146557502767246e-07, "loss": 0.5684, "step": 3726 }, { "epoch": 0.9110241994622341, "grad_norm": 2.774477243423462, "learning_rate": 4.1240250245833226e-07, "loss": 0.4675, "step": 3727 }, { "epoch": 0.9112686384747005, "grad_norm": 2.8976266384124756, "learning_rate": 4.101552645253182e-07, "loss": 0.5628, "step": 3728 }, { "epoch": 0.9115130774871669, "grad_norm": 3.259958505630493, "learning_rate": 4.079140378863411e-07, "loss": 0.5347, "step": 3729 }, { "epoch": 0.9117575164996333, "grad_norm": 3.4463083744049072, "learning_rate": 4.056788239462894e-07, "loss": 0.6018, "step": 3730 }, { "epoch": 0.9120019555120997, "grad_norm": 3.383847951889038, "learning_rate": 4.034496241062824e-07, "loss": 0.473, "step": 3731 }, { "epoch": 0.9122463945245661, "grad_norm": 3.5528504848480225, "learning_rate": 4.0122643976367225e-07, "loss": 0.5487, "step": 3732 }, { "epoch": 0.9124908335370325, "grad_norm": 3.103224039077759, "learning_rate": 3.9900927231203866e-07, "loss": 0.7557, "step": 3733 }, { "epoch": 0.9127352725494989, "grad_norm": 3.223209857940674, "learning_rate": 3.967981231411888e-07, "loss": 0.5294, "step": 3734 }, { "epoch": 0.9129797115619653, "grad_norm": 2.878547191619873, "learning_rate": 3.945929936371595e-07, "loss": 0.5665, "step": 3735 }, { "epoch": 0.9132241505744316, "grad_norm": 3.334857225418091, "learning_rate": 3.9239388518221287e-07, "loss": 0.4665, "step": 3736 }, { "epoch": 0.913468589586898, "grad_norm": 3.6171045303344727, "learning_rate": 3.9020079915483733e-07, "loss": 0.5931, "step": 3737 }, { "epoch": 0.9137130285993644, "grad_norm": 3.136127471923828, "learning_rate": 3.8801373692974655e-07, "loss": 0.549, "step": 3738 }, { "epoch": 0.9139574676118308, "grad_norm": 4.839405536651611, "learning_rate": 3.8583269987787607e-07, "loss": 0.5256, "step": 3739 }, { "epoch": 0.9142019066242972, "grad_norm": 3.2706949710845947, "learning_rate": 3.836576893663868e-07, "loss": 0.5917, "step": 3740 }, { "epoch": 0.9144463456367636, "grad_norm": 2.8269474506378174, "learning_rate": 3.8148870675866145e-07, "loss": 0.5479, "step": 3741 }, { "epoch": 0.91469078464923, "grad_norm": 3.574561357498169, "learning_rate": 3.7932575341430467e-07, "loss": 0.6755, "step": 3742 }, { "epoch": 0.9149352236616964, "grad_norm": 4.047536849975586, "learning_rate": 3.7716883068913857e-07, "loss": 0.5917, "step": 3743 }, { "epoch": 0.9151796626741628, "grad_norm": 3.0007340908050537, "learning_rate": 3.7501793993521164e-07, "loss": 0.5201, "step": 3744 }, { "epoch": 0.9154241016866291, "grad_norm": 3.1705756187438965, "learning_rate": 3.7287308250078535e-07, "loss": 0.6814, "step": 3745 }, { "epoch": 0.9156685406990955, "grad_norm": 6.387424468994141, "learning_rate": 3.7073425973034095e-07, "loss": 0.6554, "step": 3746 }, { "epoch": 0.9159129797115619, "grad_norm": 3.1905219554901123, "learning_rate": 3.6860147296457814e-07, "loss": 0.5053, "step": 3747 }, { "epoch": 0.9161574187240283, "grad_norm": 5.621198654174805, "learning_rate": 3.664747235404109e-07, "loss": 0.5023, "step": 3748 }, { "epoch": 0.9164018577364947, "grad_norm": 3.318756341934204, "learning_rate": 3.643540127909695e-07, "loss": 0.5744, "step": 3749 }, { "epoch": 0.9166462967489611, "grad_norm": 2.960386276245117, "learning_rate": 3.6223934204560165e-07, "loss": 0.452, "step": 3750 }, { "epoch": 0.9168907357614275, "grad_norm": 3.3074445724487305, "learning_rate": 3.601307126298648e-07, "loss": 0.6977, "step": 3751 }, { "epoch": 0.917135174773894, "grad_norm": 3.0012786388397217, "learning_rate": 3.580281258655316e-07, "loss": 0.6826, "step": 3752 }, { "epoch": 0.9173796137863603, "grad_norm": 3.6927969455718994, "learning_rate": 3.559315830705856e-07, "loss": 0.5505, "step": 3753 }, { "epoch": 0.9176240527988266, "grad_norm": 3.082348585128784, "learning_rate": 3.5384108555922446e-07, "loss": 1.8339, "step": 3754 }, { "epoch": 0.917868491811293, "grad_norm": 2.996356725692749, "learning_rate": 3.517566346418544e-07, "loss": 0.5314, "step": 3755 }, { "epoch": 0.9181129308237594, "grad_norm": 2.9821434020996094, "learning_rate": 3.496782316250913e-07, "loss": 0.5003, "step": 3756 }, { "epoch": 0.9183573698362258, "grad_norm": 3.2172985076904297, "learning_rate": 3.476058778117608e-07, "loss": 0.5269, "step": 3757 }, { "epoch": 0.9186018088486922, "grad_norm": 4.517338752746582, "learning_rate": 3.4553957450089383e-07, "loss": 0.5819, "step": 3758 }, { "epoch": 0.9188462478611586, "grad_norm": 3.4000778198242188, "learning_rate": 3.434793229877353e-07, "loss": 0.5547, "step": 3759 }, { "epoch": 0.919090686873625, "grad_norm": 3.2452781200408936, "learning_rate": 3.41425124563729e-07, "loss": 0.6142, "step": 3760 }, { "epoch": 0.9193351258860915, "grad_norm": 3.8453309535980225, "learning_rate": 3.3937698051653034e-07, "loss": 0.5169, "step": 3761 }, { "epoch": 0.9195795648985579, "grad_norm": 3.3375754356384277, "learning_rate": 3.373348921299946e-07, "loss": 0.4744, "step": 3762 }, { "epoch": 0.9198240039110241, "grad_norm": 3.6361775398254395, "learning_rate": 3.352988606841845e-07, "loss": 0.59, "step": 3763 }, { "epoch": 0.9200684429234905, "grad_norm": 2.8407204151153564, "learning_rate": 3.3326888745536576e-07, "loss": 0.5369, "step": 3764 }, { "epoch": 0.920312881935957, "grad_norm": 4.514292240142822, "learning_rate": 3.312449737160051e-07, "loss": 0.4623, "step": 3765 }, { "epoch": 0.9205573209484234, "grad_norm": 2.8216071128845215, "learning_rate": 3.2922712073477194e-07, "loss": 0.4855, "step": 3766 }, { "epoch": 0.9208017599608898, "grad_norm": 3.497938394546509, "learning_rate": 3.272153297765368e-07, "loss": 0.582, "step": 3767 }, { "epoch": 0.9210461989733562, "grad_norm": 3.3942174911499023, "learning_rate": 3.252096021023687e-07, "loss": 0.5727, "step": 3768 }, { "epoch": 0.9212906379858226, "grad_norm": 2.20690655708313, "learning_rate": 3.2320993896953756e-07, "loss": 1.8639, "step": 3769 }, { "epoch": 0.921535076998289, "grad_norm": 2.8550937175750732, "learning_rate": 3.2121634163151285e-07, "loss": 0.5327, "step": 3770 }, { "epoch": 0.9217795160107554, "grad_norm": 3.311756134033203, "learning_rate": 3.1922881133795827e-07, "loss": 0.6002, "step": 3771 }, { "epoch": 0.9220239550232217, "grad_norm": 3.338653564453125, "learning_rate": 3.172473493347361e-07, "loss": 0.4517, "step": 3772 }, { "epoch": 0.922268394035688, "grad_norm": 3.409890651702881, "learning_rate": 3.152719568639073e-07, "loss": 0.5154, "step": 3773 }, { "epoch": 0.9225128330481545, "grad_norm": 3.480623960494995, "learning_rate": 3.133026351637236e-07, "loss": 0.5594, "step": 3774 }, { "epoch": 0.9227572720606209, "grad_norm": 5.149979591369629, "learning_rate": 3.113393854686353e-07, "loss": 0.6494, "step": 3775 }, { "epoch": 0.9230017110730873, "grad_norm": 3.3873095512390137, "learning_rate": 3.093822090092857e-07, "loss": 0.5442, "step": 3776 }, { "epoch": 0.9232461500855537, "grad_norm": 3.5246260166168213, "learning_rate": 3.0743110701250913e-07, "loss": 0.5315, "step": 3777 }, { "epoch": 0.9234905890980201, "grad_norm": 2.8095595836639404, "learning_rate": 3.054860807013327e-07, "loss": 1.9355, "step": 3778 }, { "epoch": 0.9237350281104865, "grad_norm": 3.0907297134399414, "learning_rate": 3.035471312949778e-07, "loss": 0.6488, "step": 3779 }, { "epoch": 0.9239794671229529, "grad_norm": 4.124264717102051, "learning_rate": 3.016142600088534e-07, "loss": 0.5738, "step": 3780 }, { "epoch": 0.9242239061354192, "grad_norm": 3.186234712600708, "learning_rate": 2.996874680545603e-07, "loss": 0.6444, "step": 3781 }, { "epoch": 0.9244683451478856, "grad_norm": 2.7483363151550293, "learning_rate": 2.9776675663988785e-07, "loss": 0.5399, "step": 3782 }, { "epoch": 0.924712784160352, "grad_norm": 3.520253896713257, "learning_rate": 2.9585212696881307e-07, "loss": 0.4852, "step": 3783 }, { "epoch": 0.9249572231728184, "grad_norm": 3.1906232833862305, "learning_rate": 2.939435802415025e-07, "loss": 0.5869, "step": 3784 }, { "epoch": 0.9252016621852848, "grad_norm": 3.0981388092041016, "learning_rate": 2.920411176543081e-07, "loss": 0.6187, "step": 3785 }, { "epoch": 0.9254461011977512, "grad_norm": 2.660426616668701, "learning_rate": 2.9014474039976923e-07, "loss": 0.4867, "step": 3786 }, { "epoch": 0.9256905402102176, "grad_norm": 2.9685277938842773, "learning_rate": 2.8825444966661064e-07, "loss": 0.4, "step": 3787 }, { "epoch": 0.925934979222684, "grad_norm": 9.820734024047852, "learning_rate": 2.8637024663974e-07, "loss": 0.5051, "step": 3788 }, { "epoch": 0.9261794182351504, "grad_norm": 2.707178831100464, "learning_rate": 2.844921325002492e-07, "loss": 2.026, "step": 3789 }, { "epoch": 0.9264238572476167, "grad_norm": 4.109099864959717, "learning_rate": 2.8262010842541874e-07, "loss": 0.6936, "step": 3790 }, { "epoch": 0.9266682962600831, "grad_norm": 3.1422524452209473, "learning_rate": 2.8075417558870333e-07, "loss": 0.5485, "step": 3791 }, { "epoch": 0.9269127352725495, "grad_norm": 4.1273698806762695, "learning_rate": 2.7889433515974616e-07, "loss": 0.5213, "step": 3792 }, { "epoch": 0.9271571742850159, "grad_norm": 2.814871072769165, "learning_rate": 2.7704058830436696e-07, "loss": 0.5673, "step": 3793 }, { "epoch": 0.9274016132974823, "grad_norm": 2.826104164123535, "learning_rate": 2.7519293618456845e-07, "loss": 0.4887, "step": 3794 }, { "epoch": 0.9276460523099487, "grad_norm": 4.0700764656066895, "learning_rate": 2.733513799585319e-07, "loss": 0.6081, "step": 3795 }, { "epoch": 0.9278904913224151, "grad_norm": 3.63450026512146, "learning_rate": 2.7151592078061726e-07, "loss": 0.4312, "step": 3796 }, { "epoch": 0.9281349303348815, "grad_norm": 3.3018198013305664, "learning_rate": 2.6968655980136425e-07, "loss": 0.5232, "step": 3797 }, { "epoch": 0.9283793693473479, "grad_norm": 3.0066380500793457, "learning_rate": 2.678632981674867e-07, "loss": 0.5133, "step": 3798 }, { "epoch": 0.9286238083598142, "grad_norm": 2.7969233989715576, "learning_rate": 2.660461370218792e-07, "loss": 0.6212, "step": 3799 }, { "epoch": 0.9288682473722806, "grad_norm": 2.9382822513580322, "learning_rate": 2.6423507750360844e-07, "loss": 0.5063, "step": 3800 }, { "epoch": 0.929112686384747, "grad_norm": 3.175105571746826, "learning_rate": 2.624301207479185e-07, "loss": 0.5232, "step": 3801 }, { "epoch": 0.9293571253972134, "grad_norm": 3.261991024017334, "learning_rate": 2.6063126788622883e-07, "loss": 0.7234, "step": 3802 }, { "epoch": 0.9296015644096798, "grad_norm": 3.413081407546997, "learning_rate": 2.588385200461307e-07, "loss": 0.6532, "step": 3803 }, { "epoch": 0.9298460034221462, "grad_norm": 3.2571053504943848, "learning_rate": 2.570518783513887e-07, "loss": 0.5096, "step": 3804 }, { "epoch": 0.9300904424346126, "grad_norm": 2.5494961738586426, "learning_rate": 2.552713439219412e-07, "loss": 1.8917, "step": 3805 }, { "epoch": 0.930334881447079, "grad_norm": 3.4841763973236084, "learning_rate": 2.5349691787389885e-07, "loss": 0.5292, "step": 3806 }, { "epoch": 0.9305793204595454, "grad_norm": 3.9227099418640137, "learning_rate": 2.517286013195397e-07, "loss": 0.6536, "step": 3807 }, { "epoch": 0.9308237594720117, "grad_norm": 3.9732210636138916, "learning_rate": 2.4996639536731706e-07, "loss": 0.7448, "step": 3808 }, { "epoch": 0.9310681984844781, "grad_norm": 3.2098023891448975, "learning_rate": 2.482103011218495e-07, "loss": 0.6618, "step": 3809 }, { "epoch": 0.9313126374969445, "grad_norm": 3.286810874938965, "learning_rate": 2.464603196839277e-07, "loss": 0.6486, "step": 3810 }, { "epoch": 0.9315570765094109, "grad_norm": 3.226602792739868, "learning_rate": 2.447164521505074e-07, "loss": 0.6143, "step": 3811 }, { "epoch": 0.9318015155218773, "grad_norm": 4.43597936630249, "learning_rate": 2.4297869961471544e-07, "loss": 0.5931, "step": 3812 }, { "epoch": 0.9320459545343437, "grad_norm": 4.943711757659912, "learning_rate": 2.4124706316584255e-07, "loss": 0.6558, "step": 3813 }, { "epoch": 0.9322903935468101, "grad_norm": 2.977363109588623, "learning_rate": 2.3952154388934833e-07, "loss": 0.5582, "step": 3814 }, { "epoch": 0.9325348325592765, "grad_norm": 3.788539409637451, "learning_rate": 2.3780214286685533e-07, "loss": 0.674, "step": 3815 }, { "epoch": 0.9327792715717429, "grad_norm": 2.4482150077819824, "learning_rate": 2.360888611761536e-07, "loss": 1.9792, "step": 3816 }, { "epoch": 0.9330237105842092, "grad_norm": 2.869302988052368, "learning_rate": 2.3438169989119408e-07, "loss": 0.4668, "step": 3817 }, { "epoch": 0.9332681495966756, "grad_norm": 3.960190773010254, "learning_rate": 2.32680660082093e-07, "loss": 0.5912, "step": 3818 }, { "epoch": 0.933512588609142, "grad_norm": 3.486650228500366, "learning_rate": 2.3098574281513185e-07, "loss": 0.3871, "step": 3819 }, { "epoch": 0.9337570276216084, "grad_norm": 3.2709710597991943, "learning_rate": 2.292969491527486e-07, "loss": 0.4614, "step": 3820 }, { "epoch": 0.9340014666340748, "grad_norm": 3.7016708850860596, "learning_rate": 2.276142801535486e-07, "loss": 0.488, "step": 3821 }, { "epoch": 0.9342459056465412, "grad_norm": 3.7128171920776367, "learning_rate": 2.259377368722937e-07, "loss": 0.5591, "step": 3822 }, { "epoch": 0.9344903446590076, "grad_norm": 2.55253267288208, "learning_rate": 2.2426732035990772e-07, "loss": 0.4151, "step": 3823 }, { "epoch": 0.934734783671474, "grad_norm": 4.539761543273926, "learning_rate": 2.226030316634753e-07, "loss": 0.5469, "step": 3824 }, { "epoch": 0.9349792226839404, "grad_norm": 4.752309322357178, "learning_rate": 2.2094487182623747e-07, "loss": 0.7591, "step": 3825 }, { "epoch": 0.9352236616964067, "grad_norm": 3.1529886722564697, "learning_rate": 2.192928418875928e-07, "loss": 0.5054, "step": 3826 }, { "epoch": 0.9354681007088731, "grad_norm": 3.589515209197998, "learning_rate": 2.1764694288310185e-07, "loss": 0.6118, "step": 3827 }, { "epoch": 0.9357125397213395, "grad_norm": 3.063481330871582, "learning_rate": 2.1600717584447706e-07, "loss": 0.5649, "step": 3828 }, { "epoch": 0.9359569787338059, "grad_norm": 3.080091953277588, "learning_rate": 2.143735417995907e-07, "loss": 0.5806, "step": 3829 }, { "epoch": 0.9362014177462723, "grad_norm": 3.8041274547576904, "learning_rate": 2.1274604177246804e-07, "loss": 0.5447, "step": 3830 }, { "epoch": 0.9364458567587387, "grad_norm": 3.509629726409912, "learning_rate": 2.1112467678329197e-07, "loss": 0.5552, "step": 3831 }, { "epoch": 0.9366902957712051, "grad_norm": 3.3342905044555664, "learning_rate": 2.095094478483961e-07, "loss": 0.5797, "step": 3832 }, { "epoch": 0.9369347347836715, "grad_norm": 2.862881898880005, "learning_rate": 2.079003559802706e-07, "loss": 0.4585, "step": 3833 }, { "epoch": 0.9371791737961379, "grad_norm": 2.8512535095214844, "learning_rate": 2.0629740218755857e-07, "loss": 0.4295, "step": 3834 }, { "epoch": 0.9374236128086042, "grad_norm": 2.828467607498169, "learning_rate": 2.0470058747505516e-07, "loss": 0.4893, "step": 3835 }, { "epoch": 0.9376680518210706, "grad_norm": 3.393681526184082, "learning_rate": 2.031099128437064e-07, "loss": 0.6233, "step": 3836 }, { "epoch": 0.937912490833537, "grad_norm": 2.908700942993164, "learning_rate": 2.0152537929061023e-07, "loss": 0.6536, "step": 3837 }, { "epoch": 0.9381569298460034, "grad_norm": 2.6761109828948975, "learning_rate": 1.9994698780901434e-07, "loss": 0.4504, "step": 3838 }, { "epoch": 0.9384013688584698, "grad_norm": 3.427279472351074, "learning_rate": 1.983747393883173e-07, "loss": 0.524, "step": 3839 }, { "epoch": 0.9386458078709362, "grad_norm": 3.2629621028900146, "learning_rate": 1.9680863501406744e-07, "loss": 0.5726, "step": 3840 }, { "epoch": 0.9388902468834026, "grad_norm": 2.9898974895477295, "learning_rate": 1.9524867566795945e-07, "loss": 0.4985, "step": 3841 }, { "epoch": 0.939134685895869, "grad_norm": 3.20027756690979, "learning_rate": 1.9369486232783896e-07, "loss": 0.5411, "step": 3842 }, { "epoch": 0.9393791249083354, "grad_norm": 5.192112922668457, "learning_rate": 1.921471959676957e-07, "loss": 0.5168, "step": 3843 }, { "epoch": 0.9396235639208017, "grad_norm": 3.0395219326019287, "learning_rate": 1.9060567755767035e-07, "loss": 0.546, "step": 3844 }, { "epoch": 0.9398680029332681, "grad_norm": 3.2350528240203857, "learning_rate": 1.890703080640466e-07, "loss": 0.5259, "step": 3845 }, { "epoch": 0.9401124419457345, "grad_norm": 4.550525188446045, "learning_rate": 1.8754108844925455e-07, "loss": 0.615, "step": 3846 }, { "epoch": 0.9403568809582009, "grad_norm": 2.9899308681488037, "learning_rate": 1.8601801967187083e-07, "loss": 0.4783, "step": 3847 }, { "epoch": 0.9406013199706673, "grad_norm": 3.2106664180755615, "learning_rate": 1.8450110268661392e-07, "loss": 0.5835, "step": 3848 }, { "epoch": 0.9408457589831337, "grad_norm": 3.7672276496887207, "learning_rate": 1.8299033844434878e-07, "loss": 0.575, "step": 3849 }, { "epoch": 0.9410901979956001, "grad_norm": 3.14738130569458, "learning_rate": 1.8148572789208007e-07, "loss": 0.6041, "step": 3850 }, { "epoch": 0.9413346370080665, "grad_norm": 3.2783117294311523, "learning_rate": 1.7998727197295785e-07, "loss": 0.475, "step": 3851 }, { "epoch": 0.9415790760205329, "grad_norm": 3.5579917430877686, "learning_rate": 1.7849497162627516e-07, "loss": 0.6771, "step": 3852 }, { "epoch": 0.9418235150329992, "grad_norm": 2.999138355255127, "learning_rate": 1.7700882778746374e-07, "loss": 0.556, "step": 3853 }, { "epoch": 0.9420679540454656, "grad_norm": 4.949743270874023, "learning_rate": 1.755288413880951e-07, "loss": 0.3769, "step": 3854 }, { "epoch": 0.942312393057932, "grad_norm": 2.9086415767669678, "learning_rate": 1.7405501335588825e-07, "loss": 0.4424, "step": 3855 }, { "epoch": 0.9425568320703984, "grad_norm": 3.3956940174102783, "learning_rate": 1.7258734461469307e-07, "loss": 0.6627, "step": 3856 }, { "epoch": 0.9428012710828648, "grad_norm": 3.2270140647888184, "learning_rate": 1.7112583608450361e-07, "loss": 0.6194, "step": 3857 }, { "epoch": 0.9430457100953312, "grad_norm": 3.090681552886963, "learning_rate": 1.6967048868145154e-07, "loss": 0.5783, "step": 3858 }, { "epoch": 0.9432901491077976, "grad_norm": 2.989459991455078, "learning_rate": 1.6822130331780485e-07, "loss": 0.549, "step": 3859 }, { "epoch": 0.943534588120264, "grad_norm": 3.7176594734191895, "learning_rate": 1.6677828090197245e-07, "loss": 0.672, "step": 3860 }, { "epoch": 0.9437790271327304, "grad_norm": 2.7825515270233154, "learning_rate": 1.6534142233849527e-07, "loss": 0.4756, "step": 3861 }, { "epoch": 0.9440234661451967, "grad_norm": 5.053201675415039, "learning_rate": 1.6391072852805613e-07, "loss": 0.6182, "step": 3862 }, { "epoch": 0.9442679051576631, "grad_norm": 3.03202486038208, "learning_rate": 1.6248620036746764e-07, "loss": 0.5124, "step": 3863 }, { "epoch": 0.9445123441701295, "grad_norm": 3.8698859214782715, "learning_rate": 1.6106783874968114e-07, "loss": 0.5856, "step": 3864 }, { "epoch": 0.9447567831825959, "grad_norm": 3.8033907413482666, "learning_rate": 1.5965564456378313e-07, "loss": 0.4567, "step": 3865 }, { "epoch": 0.9450012221950623, "grad_norm": 3.2855918407440186, "learning_rate": 1.582496186949911e-07, "loss": 0.5178, "step": 3866 }, { "epoch": 0.9452456612075287, "grad_norm": 3.8047099113464355, "learning_rate": 1.5684976202465786e-07, "loss": 0.5535, "step": 3867 }, { "epoch": 0.9454901002199951, "grad_norm": 4.872805118560791, "learning_rate": 1.554560754302703e-07, "loss": 0.6506, "step": 3868 }, { "epoch": 0.9457345392324615, "grad_norm": 3.565657138824463, "learning_rate": 1.5406855978544634e-07, "loss": 0.5024, "step": 3869 }, { "epoch": 0.9459789782449279, "grad_norm": 3.1939427852630615, "learning_rate": 1.5268721595993464e-07, "loss": 0.5648, "step": 3870 }, { "epoch": 0.9462234172573942, "grad_norm": 4.939906120300293, "learning_rate": 1.5131204481961592e-07, "loss": 0.4766, "step": 3871 }, { "epoch": 0.9464678562698606, "grad_norm": 3.5011534690856934, "learning_rate": 1.4994304722650176e-07, "loss": 0.556, "step": 3872 }, { "epoch": 0.946712295282327, "grad_norm": 3.5804851055145264, "learning_rate": 1.4858022403873573e-07, "loss": 0.5861, "step": 3873 }, { "epoch": 0.9469567342947934, "grad_norm": 3.398738384246826, "learning_rate": 1.472235761105878e-07, "loss": 0.5581, "step": 3874 }, { "epoch": 0.9472011733072598, "grad_norm": 3.1239571571350098, "learning_rate": 1.4587310429245883e-07, "loss": 0.6228, "step": 3875 }, { "epoch": 0.9474456123197262, "grad_norm": 3.6954193115234375, "learning_rate": 1.4452880943087722e-07, "loss": 0.468, "step": 3876 }, { "epoch": 0.9476900513321926, "grad_norm": 2.738016128540039, "learning_rate": 1.4319069236850224e-07, "loss": 0.542, "step": 3877 }, { "epoch": 0.947934490344659, "grad_norm": 3.486159086227417, "learning_rate": 1.418587539441174e-07, "loss": 0.6804, "step": 3878 }, { "epoch": 0.9481789293571254, "grad_norm": 3.5900747776031494, "learning_rate": 1.4053299499263374e-07, "loss": 0.7054, "step": 3879 }, { "epoch": 0.9484233683695917, "grad_norm": 3.0891590118408203, "learning_rate": 1.392134163450898e-07, "loss": 0.5256, "step": 3880 }, { "epoch": 0.9486678073820581, "grad_norm": 5.227392673492432, "learning_rate": 1.3790001882865056e-07, "loss": 0.5639, "step": 3881 }, { "epoch": 0.9489122463945245, "grad_norm": 4.298923969268799, "learning_rate": 1.365928032666042e-07, "loss": 0.5737, "step": 3882 }, { "epoch": 0.9491566854069909, "grad_norm": 3.931917667388916, "learning_rate": 1.3529177047836628e-07, "loss": 0.5685, "step": 3883 }, { "epoch": 0.9494011244194573, "grad_norm": 2.9893269538879395, "learning_rate": 1.3399692127947562e-07, "loss": 0.6057, "step": 3884 }, { "epoch": 0.9496455634319237, "grad_norm": 3.0450754165649414, "learning_rate": 1.3270825648159403e-07, "loss": 0.605, "step": 3885 }, { "epoch": 0.9498900024443901, "grad_norm": 3.5307817459106445, "learning_rate": 1.3142577689250756e-07, "loss": 0.5407, "step": 3886 }, { "epoch": 0.9501344414568565, "grad_norm": 3.128784656524658, "learning_rate": 1.3014948331612653e-07, "loss": 0.4743, "step": 3887 }, { "epoch": 0.950378880469323, "grad_norm": 4.463705062866211, "learning_rate": 1.2887937655248096e-07, "loss": 0.5312, "step": 3888 }, { "epoch": 0.9506233194817892, "grad_norm": 3.3671658039093018, "learning_rate": 1.2761545739772396e-07, "loss": 0.5913, "step": 3889 }, { "epoch": 0.9508677584942556, "grad_norm": 3.701791286468506, "learning_rate": 1.2635772664413072e-07, "loss": 0.6569, "step": 3890 }, { "epoch": 0.951112197506722, "grad_norm": 3.6596016883850098, "learning_rate": 1.251061850800961e-07, "loss": 0.5157, "step": 3891 }, { "epoch": 0.9513566365191884, "grad_norm": 3.160101890563965, "learning_rate": 1.23860833490137e-07, "loss": 0.6493, "step": 3892 }, { "epoch": 0.9516010755316548, "grad_norm": 2.915860652923584, "learning_rate": 1.226216726548879e-07, "loss": 0.5418, "step": 3893 }, { "epoch": 0.9518455145441213, "grad_norm": 3.316563844680786, "learning_rate": 1.213887033511041e-07, "loss": 0.5572, "step": 3894 }, { "epoch": 0.9520899535565877, "grad_norm": 3.3448314666748047, "learning_rate": 1.2016192635165958e-07, "loss": 0.5751, "step": 3895 }, { "epoch": 0.9523343925690541, "grad_norm": 3.304323196411133, "learning_rate": 1.1894134242554589e-07, "loss": 0.5206, "step": 3896 }, { "epoch": 0.9525788315815205, "grad_norm": 3.19252347946167, "learning_rate": 1.1772695233787545e-07, "loss": 0.6348, "step": 3897 }, { "epoch": 0.9528232705939867, "grad_norm": 3.2109947204589844, "learning_rate": 1.1651875684987268e-07, "loss": 0.6405, "step": 3898 }, { "epoch": 0.9530677096064532, "grad_norm": 3.1668834686279297, "learning_rate": 1.1531675671888621e-07, "loss": 0.5971, "step": 3899 }, { "epoch": 0.9533121486189196, "grad_norm": 2.928603172302246, "learning_rate": 1.1412095269837442e-07, "loss": 0.5108, "step": 3900 }, { "epoch": 0.953556587631386, "grad_norm": 3.901209831237793, "learning_rate": 1.1293134553791551e-07, "loss": 0.5936, "step": 3901 }, { "epoch": 0.9538010266438524, "grad_norm": 3.3114547729492188, "learning_rate": 1.1174793598320299e-07, "loss": 0.5659, "step": 3902 }, { "epoch": 0.9540454656563188, "grad_norm": 2.952723503112793, "learning_rate": 1.1057072477604347e-07, "loss": 0.5262, "step": 3903 }, { "epoch": 0.9542899046687852, "grad_norm": 3.1195309162139893, "learning_rate": 1.0939971265436e-07, "loss": 0.5468, "step": 3904 }, { "epoch": 0.9545343436812516, "grad_norm": 2.4515929222106934, "learning_rate": 1.0823490035218986e-07, "loss": 1.9989, "step": 3905 }, { "epoch": 0.954778782693718, "grad_norm": 2.760552406311035, "learning_rate": 1.0707628859968344e-07, "loss": 0.5545, "step": 3906 }, { "epoch": 0.9550232217061843, "grad_norm": 3.261154890060425, "learning_rate": 1.0592387812310312e-07, "loss": 0.424, "step": 3907 }, { "epoch": 0.9552676607186507, "grad_norm": 3.056074380874634, "learning_rate": 1.0477766964482772e-07, "loss": 0.5899, "step": 3908 }, { "epoch": 0.9555120997311171, "grad_norm": 4.080880165100098, "learning_rate": 1.0363766388334362e-07, "loss": 0.5603, "step": 3909 }, { "epoch": 0.9557565387435835, "grad_norm": 3.1623666286468506, "learning_rate": 1.0250386155325254e-07, "loss": 0.6329, "step": 3910 }, { "epoch": 0.9560009777560499, "grad_norm": 3.1618571281433105, "learning_rate": 1.0137626336526596e-07, "loss": 0.7661, "step": 3911 }, { "epoch": 0.9562454167685163, "grad_norm": 3.1398682594299316, "learning_rate": 1.0025487002620848e-07, "loss": 0.5826, "step": 3912 }, { "epoch": 0.9564898557809827, "grad_norm": 4.327209949493408, "learning_rate": 9.913968223901227e-08, "loss": 0.6099, "step": 3913 }, { "epoch": 0.9567342947934491, "grad_norm": 2.97302508354187, "learning_rate": 9.803070070272258e-08, "loss": 0.5439, "step": 3914 }, { "epoch": 0.9569787338059155, "grad_norm": 4.734954357147217, "learning_rate": 9.692792611249224e-08, "loss": 0.5333, "step": 3915 }, { "epoch": 0.9572231728183818, "grad_norm": 3.317031145095825, "learning_rate": 9.583135915958386e-08, "loss": 0.5139, "step": 3916 }, { "epoch": 0.9574676118308482, "grad_norm": 3.1889259815216064, "learning_rate": 9.47410005313687e-08, "loss": 0.651, "step": 3917 }, { "epoch": 0.9577120508433146, "grad_norm": 3.311048984527588, "learning_rate": 9.365685091132782e-08, "loss": 0.567, "step": 3918 }, { "epoch": 0.957956489855781, "grad_norm": 2.774733543395996, "learning_rate": 9.257891097904759e-08, "loss": 0.4583, "step": 3919 }, { "epoch": 0.9582009288682474, "grad_norm": 2.797528028488159, "learning_rate": 9.15071814102253e-08, "loss": 0.5282, "step": 3920 }, { "epoch": 0.9584453678807138, "grad_norm": 3.390241861343384, "learning_rate": 9.044166287666134e-08, "loss": 0.4714, "step": 3921 }, { "epoch": 0.9586898068931802, "grad_norm": 3.394399404525757, "learning_rate": 8.938235604626589e-08, "loss": 0.5817, "step": 3922 }, { "epoch": 0.9589342459056466, "grad_norm": 4.014432907104492, "learning_rate": 8.832926158305443e-08, "loss": 0.6375, "step": 3923 }, { "epoch": 0.959178684918113, "grad_norm": 2.896803379058838, "learning_rate": 8.728238014714785e-08, "loss": 0.5929, "step": 3924 }, { "epoch": 0.9594231239305793, "grad_norm": 3.5837626457214355, "learning_rate": 8.624171239477342e-08, "loss": 0.4898, "step": 3925 }, { "epoch": 0.9596675629430457, "grad_norm": 2.9800174236297607, "learning_rate": 8.520725897826265e-08, "loss": 0.5695, "step": 3926 }, { "epoch": 0.9599120019555121, "grad_norm": 3.6687331199645996, "learning_rate": 8.417902054605131e-08, "loss": 0.5069, "step": 3927 }, { "epoch": 0.9601564409679785, "grad_norm": 3.7042768001556396, "learning_rate": 8.315699774268048e-08, "loss": 0.5197, "step": 3928 }, { "epoch": 0.9604008799804449, "grad_norm": 3.6040287017822266, "learning_rate": 8.214119120879437e-08, "loss": 0.4721, "step": 3929 }, { "epoch": 0.9606453189929113, "grad_norm": 3.8857340812683105, "learning_rate": 8.11316015811403e-08, "loss": 0.5595, "step": 3930 }, { "epoch": 0.9608897580053777, "grad_norm": 2.8456661701202393, "learning_rate": 8.012822949256981e-08, "loss": 0.659, "step": 3931 }, { "epoch": 0.9611341970178441, "grad_norm": 3.1799733638763428, "learning_rate": 7.913107557203648e-08, "loss": 0.6267, "step": 3932 }, { "epoch": 0.9613786360303105, "grad_norm": 3.4712483882904053, "learning_rate": 7.814014044459362e-08, "loss": 0.622, "step": 3933 }, { "epoch": 0.9616230750427768, "grad_norm": 3.23189115524292, "learning_rate": 7.715542473140103e-08, "loss": 0.4777, "step": 3934 }, { "epoch": 0.9618675140552432, "grad_norm": 2.996572971343994, "learning_rate": 7.617692904971718e-08, "loss": 0.4598, "step": 3935 }, { "epoch": 0.9621119530677096, "grad_norm": 3.7120778560638428, "learning_rate": 7.520465401290033e-08, "loss": 0.5708, "step": 3936 }, { "epoch": 0.962356392080176, "grad_norm": 2.813230514526367, "learning_rate": 7.423860023041185e-08, "loss": 0.4853, "step": 3937 }, { "epoch": 0.9626008310926424, "grad_norm": 2.9314701557159424, "learning_rate": 7.327876830781288e-08, "loss": 0.5752, "step": 3938 }, { "epoch": 0.9628452701051088, "grad_norm": 3.0445821285247803, "learning_rate": 7.232515884676327e-08, "loss": 0.6062, "step": 3939 }, { "epoch": 0.9630897091175752, "grad_norm": 6.091731548309326, "learning_rate": 7.137777244502375e-08, "loss": 0.5332, "step": 3940 }, { "epoch": 0.9633341481300416, "grad_norm": 3.6530470848083496, "learning_rate": 7.043660969645261e-08, "loss": 0.5968, "step": 3941 }, { "epoch": 0.963578587142508, "grad_norm": 3.23164439201355, "learning_rate": 6.950167119100903e-08, "loss": 0.5435, "step": 3942 }, { "epoch": 0.9638230261549743, "grad_norm": 3.90278959274292, "learning_rate": 6.857295751474979e-08, "loss": 0.5857, "step": 3943 }, { "epoch": 0.9640674651674407, "grad_norm": 3.4893884658813477, "learning_rate": 6.765046924982699e-08, "loss": 0.5718, "step": 3944 }, { "epoch": 0.9643119041799071, "grad_norm": 2.482055425643921, "learning_rate": 6.673420697449473e-08, "loss": 1.9335, "step": 3945 }, { "epoch": 0.9645563431923735, "grad_norm": 3.010075092315674, "learning_rate": 6.582417126310359e-08, "loss": 0.5077, "step": 3946 }, { "epoch": 0.9648007822048399, "grad_norm": 2.269382953643799, "learning_rate": 6.492036268609725e-08, "loss": 1.8825, "step": 3947 }, { "epoch": 0.9650452212173063, "grad_norm": 4.631083011627197, "learning_rate": 6.402278181002031e-08, "loss": 0.5548, "step": 3948 }, { "epoch": 0.9652896602297727, "grad_norm": 3.2476084232330322, "learning_rate": 6.313142919751159e-08, "loss": 0.5657, "step": 3949 }, { "epoch": 0.9655340992422391, "grad_norm": 3.203571081161499, "learning_rate": 6.224630540730525e-08, "loss": 0.4498, "step": 3950 }, { "epoch": 0.9657785382547055, "grad_norm": 3.3352339267730713, "learning_rate": 6.136741099423416e-08, "loss": 0.4863, "step": 3951 }, { "epoch": 0.9660229772671718, "grad_norm": 3.1167092323303223, "learning_rate": 6.049474650922094e-08, "loss": 0.6072, "step": 3952 }, { "epoch": 0.9662674162796382, "grad_norm": 3.6150765419006348, "learning_rate": 5.962831249928913e-08, "loss": 0.5307, "step": 3953 }, { "epoch": 0.9665118552921046, "grad_norm": 3.6788432598114014, "learning_rate": 5.8768109507553185e-08, "loss": 0.4531, "step": 3954 }, { "epoch": 0.966756294304571, "grad_norm": 3.4593963623046875, "learning_rate": 5.791413807322066e-08, "loss": 0.4987, "step": 3955 }, { "epoch": 0.9670007333170374, "grad_norm": 3.258887767791748, "learning_rate": 5.7066398731596694e-08, "loss": 0.646, "step": 3956 }, { "epoch": 0.9672451723295038, "grad_norm": 3.145399570465088, "learning_rate": 5.6224892014077324e-08, "loss": 0.6184, "step": 3957 }, { "epoch": 0.9674896113419702, "grad_norm": 3.165806531906128, "learning_rate": 5.5389618448150606e-08, "loss": 0.5939, "step": 3958 }, { "epoch": 0.9677340503544366, "grad_norm": 3.396052360534668, "learning_rate": 5.456057855740215e-08, "loss": 0.4596, "step": 3959 }, { "epoch": 0.967978489366903, "grad_norm": 5.122045993804932, "learning_rate": 5.3737772861502945e-08, "loss": 0.6347, "step": 3960 }, { "epoch": 0.9682229283793693, "grad_norm": 3.333050489425659, "learning_rate": 5.2921201876223737e-08, "loss": 0.6124, "step": 3961 }, { "epoch": 0.9684673673918357, "grad_norm": 2.9732086658477783, "learning_rate": 5.2110866113421754e-08, "loss": 0.5867, "step": 3962 }, { "epoch": 0.9687118064043021, "grad_norm": 2.6987709999084473, "learning_rate": 5.1306766081048456e-08, "loss": 0.4852, "step": 3963 }, { "epoch": 0.9689562454167685, "grad_norm": 3.627744436264038, "learning_rate": 5.050890228314509e-08, "loss": 0.7174, "step": 3964 }, { "epoch": 0.9692006844292349, "grad_norm": 3.2260825634002686, "learning_rate": 4.971727521984382e-08, "loss": 0.4968, "step": 3965 }, { "epoch": 0.9694451234417013, "grad_norm": 3.5571961402893066, "learning_rate": 4.8931885387367706e-08, "loss": 0.5191, "step": 3966 }, { "epoch": 0.9696895624541677, "grad_norm": 3.4485998153686523, "learning_rate": 4.815273327803183e-08, "loss": 0.567, "step": 3967 }, { "epoch": 0.9699340014666341, "grad_norm": 4.239404201507568, "learning_rate": 4.7379819380236614e-08, "loss": 0.6098, "step": 3968 }, { "epoch": 0.9701784404791005, "grad_norm": 3.378331422805786, "learning_rate": 4.6613144178477844e-08, "loss": 0.4949, "step": 3969 }, { "epoch": 0.9704228794915668, "grad_norm": 3.5867037773132324, "learning_rate": 4.585270815333665e-08, "loss": 0.5589, "step": 3970 }, { "epoch": 0.9706673185040332, "grad_norm": 3.320458173751831, "learning_rate": 4.5098511781485056e-08, "loss": 0.5532, "step": 3971 }, { "epoch": 0.9709117575164996, "grad_norm": 3.188258171081543, "learning_rate": 4.435055553568268e-08, "loss": 0.6084, "step": 3972 }, { "epoch": 0.971156196528966, "grad_norm": 3.1021206378936768, "learning_rate": 4.360883988477782e-08, "loss": 0.6218, "step": 3973 }, { "epoch": 0.9714006355414324, "grad_norm": 3.52012300491333, "learning_rate": 4.2873365293706335e-08, "loss": 0.5511, "step": 3974 }, { "epoch": 0.9716450745538988, "grad_norm": 3.305572032928467, "learning_rate": 4.2144132223495e-08, "loss": 0.5442, "step": 3975 }, { "epoch": 0.9718895135663652, "grad_norm": 2.587251663208008, "learning_rate": 4.1421141131254835e-08, "loss": 1.852, "step": 3976 }, { "epoch": 0.9721339525788316, "grad_norm": 2.945148229598999, "learning_rate": 4.070439247018554e-08, "loss": 0.5045, "step": 3977 }, { "epoch": 0.972378391591298, "grad_norm": 2.8457350730895996, "learning_rate": 3.999388668957216e-08, "loss": 0.5637, "step": 3978 }, { "epoch": 0.9726228306037643, "grad_norm": 3.0436813831329346, "learning_rate": 3.928962423479066e-08, "loss": 0.4438, "step": 3979 }, { "epoch": 0.9728672696162307, "grad_norm": 3.1909372806549072, "learning_rate": 3.859160554729791e-08, "loss": 0.6261, "step": 3980 }, { "epoch": 0.9731117086286971, "grad_norm": 2.574711322784424, "learning_rate": 3.789983106464057e-08, "loss": 1.8583, "step": 3981 }, { "epoch": 0.9733561476411635, "grad_norm": 3.3103384971618652, "learning_rate": 3.7214301220450666e-08, "loss": 0.5687, "step": 3982 }, { "epoch": 0.9736005866536299, "grad_norm": 3.413031816482544, "learning_rate": 3.653501644444446e-08, "loss": 0.6336, "step": 3983 }, { "epoch": 0.9738450256660963, "grad_norm": 3.0209238529205322, "learning_rate": 3.586197716242579e-08, "loss": 0.5029, "step": 3984 }, { "epoch": 0.9740894646785627, "grad_norm": 4.465044021606445, "learning_rate": 3.519518379628162e-08, "loss": 0.7279, "step": 3985 }, { "epoch": 0.9743339036910291, "grad_norm": 3.354797124862671, "learning_rate": 3.453463676398427e-08, "loss": 0.5322, "step": 3986 }, { "epoch": 0.9745783427034955, "grad_norm": 9.969941139221191, "learning_rate": 3.3880336479590324e-08, "loss": 0.5312, "step": 3987 }, { "epoch": 0.9748227817159618, "grad_norm": 3.3734261989593506, "learning_rate": 3.32322833532428e-08, "loss": 0.5122, "step": 3988 }, { "epoch": 0.9750672207284282, "grad_norm": 3.315612554550171, "learning_rate": 3.259047779116453e-08, "loss": 0.511, "step": 3989 }, { "epoch": 0.9753116597408946, "grad_norm": 3.0684654712677, "learning_rate": 3.195492019566704e-08, "loss": 0.602, "step": 3990 }, { "epoch": 0.975556098753361, "grad_norm": 4.556325435638428, "learning_rate": 3.132561096514164e-08, "loss": 0.6097, "step": 3991 }, { "epoch": 0.9758005377658274, "grad_norm": 3.0438623428344727, "learning_rate": 3.07025504940639e-08, "loss": 0.4931, "step": 3992 }, { "epoch": 0.9760449767782938, "grad_norm": 3.167351245880127, "learning_rate": 3.0085739172993623e-08, "loss": 0.4976, "step": 3993 }, { "epoch": 0.9762894157907602, "grad_norm": 2.8869566917419434, "learning_rate": 2.9475177388573752e-08, "loss": 1.9206, "step": 3994 }, { "epoch": 0.9765338548032266, "grad_norm": 2.6476495265960693, "learning_rate": 2.8870865523525916e-08, "loss": 0.478, "step": 3995 }, { "epoch": 0.976778293815693, "grad_norm": 2.7549493312835693, "learning_rate": 2.8272803956658212e-08, "loss": 0.6107, "step": 3996 }, { "epoch": 0.9770227328281593, "grad_norm": 3.1494834423065186, "learning_rate": 2.7680993062858543e-08, "loss": 0.4316, "step": 3997 }, { "epoch": 0.9772671718406257, "grad_norm": 3.2198588848114014, "learning_rate": 2.7095433213097933e-08, "loss": 0.5957, "step": 3998 }, { "epoch": 0.9775116108530921, "grad_norm": 3.3557252883911133, "learning_rate": 2.651612477442833e-08, "loss": 0.6265, "step": 3999 }, { "epoch": 0.9777560498655585, "grad_norm": 4.323390007019043, "learning_rate": 2.5943068109983706e-08, "loss": 0.5282, "step": 4000 }, { "epoch": 0.9780004888780249, "grad_norm": 2.7845654487609863, "learning_rate": 2.5376263578977823e-08, "loss": 0.5089, "step": 4001 }, { "epoch": 0.9782449278904913, "grad_norm": 3.303093910217285, "learning_rate": 2.4815711536706478e-08, "loss": 0.5188, "step": 4002 }, { "epoch": 0.9784893669029577, "grad_norm": 3.255608558654785, "learning_rate": 2.4261412334546373e-08, "loss": 0.644, "step": 4003 }, { "epoch": 0.9787338059154241, "grad_norm": 3.067213773727417, "learning_rate": 2.3713366319954023e-08, "loss": 0.4166, "step": 4004 }, { "epoch": 0.9789782449278905, "grad_norm": 3.1581287384033203, "learning_rate": 2.3171573836464623e-08, "loss": 0.6115, "step": 4005 }, { "epoch": 0.9792226839403568, "grad_norm": 4.095109939575195, "learning_rate": 2.2636035223697616e-08, "loss": 0.5978, "step": 4006 }, { "epoch": 0.9794671229528232, "grad_norm": 2.942018747329712, "learning_rate": 2.210675081734892e-08, "loss": 0.5768, "step": 4007 }, { "epoch": 0.9797115619652896, "grad_norm": 3.5556507110595703, "learning_rate": 2.1583720949194253e-08, "loss": 0.5152, "step": 4008 }, { "epoch": 0.979956000977756, "grad_norm": 2.95241379737854, "learning_rate": 2.1066945947090244e-08, "loss": 0.5812, "step": 4009 }, { "epoch": 0.9802004399902224, "grad_norm": 3.049710988998413, "learning_rate": 2.055642613497111e-08, "loss": 0.5166, "step": 4010 }, { "epoch": 0.9804448790026888, "grad_norm": 3.40008807182312, "learning_rate": 2.0052161832850858e-08, "loss": 0.4201, "step": 4011 }, { "epoch": 0.9806893180151552, "grad_norm": 2.8084328174591064, "learning_rate": 1.9554153356823313e-08, "loss": 0.5247, "step": 4012 }, { "epoch": 0.9809337570276216, "grad_norm": 3.105757236480713, "learning_rate": 1.9062401019059874e-08, "loss": 0.467, "step": 4013 }, { "epoch": 0.981178196040088, "grad_norm": 2.792382001876831, "learning_rate": 1.857690512780952e-08, "loss": 0.6131, "step": 4014 }, { "epoch": 0.9814226350525543, "grad_norm": 3.0735909938812256, "learning_rate": 1.8097665987399925e-08, "loss": 0.5773, "step": 4015 }, { "epoch": 0.9816670740650207, "grad_norm": 3.092674493789673, "learning_rate": 1.762468389823857e-08, "loss": 0.6127, "step": 4016 }, { "epoch": 0.9819115130774871, "grad_norm": 2.8229806423187256, "learning_rate": 1.7157959156808292e-08, "loss": 0.6363, "step": 4017 }, { "epoch": 0.9821559520899535, "grad_norm": 3.3986260890960693, "learning_rate": 1.6697492055671725e-08, "loss": 0.5013, "step": 4018 }, { "epoch": 0.98240039110242, "grad_norm": 3.2098350524902344, "learning_rate": 1.6243282883465772e-08, "loss": 0.5575, "step": 4019 }, { "epoch": 0.9826448301148863, "grad_norm": 3.488957643508911, "learning_rate": 1.579533192490823e-08, "loss": 0.5848, "step": 4020 }, { "epoch": 0.9828892691273527, "grad_norm": 4.478259086608887, "learning_rate": 1.5353639460793378e-08, "loss": 0.6147, "step": 4021 }, { "epoch": 0.9831337081398192, "grad_norm": 4.556621074676514, "learning_rate": 1.4918205767989746e-08, "loss": 0.5966, "step": 4022 }, { "epoch": 0.9833781471522856, "grad_norm": 3.862208604812622, "learning_rate": 1.4489031119444553e-08, "loss": 0.4832, "step": 4023 }, { "epoch": 0.9836225861647518, "grad_norm": 3.3213610649108887, "learning_rate": 1.4066115784181488e-08, "loss": 0.5607, "step": 4024 }, { "epoch": 0.9838670251772182, "grad_norm": 2.868886947631836, "learning_rate": 1.364946002730183e-08, "loss": 0.5182, "step": 4025 }, { "epoch": 0.9841114641896846, "grad_norm": 3.2725284099578857, "learning_rate": 1.3239064109981103e-08, "loss": 0.6052, "step": 4026 }, { "epoch": 0.984355903202151, "grad_norm": 3.655763626098633, "learning_rate": 1.2834928289472415e-08, "loss": 0.5215, "step": 4027 }, { "epoch": 0.9846003422146175, "grad_norm": 3.8058059215545654, "learning_rate": 1.2437052819102013e-08, "loss": 0.5269, "step": 4028 }, { "epoch": 0.9848447812270839, "grad_norm": 3.2667770385742188, "learning_rate": 1.2045437948275952e-08, "loss": 0.5415, "step": 4029 }, { "epoch": 0.9850892202395503, "grad_norm": 2.810737133026123, "learning_rate": 1.166008392247342e-08, "loss": 1.9476, "step": 4030 }, { "epoch": 0.9853336592520167, "grad_norm": 3.1225063800811768, "learning_rate": 1.1280990983248975e-08, "loss": 0.5263, "step": 4031 }, { "epoch": 0.9855780982644831, "grad_norm": 2.9405412673950195, "learning_rate": 1.090815936823475e-08, "loss": 1.9106, "step": 4032 }, { "epoch": 0.9858225372769494, "grad_norm": 2.3317549228668213, "learning_rate": 1.0541589311133804e-08, "loss": 1.8014, "step": 4033 }, { "epoch": 0.9860669762894158, "grad_norm": 3.156327486038208, "learning_rate": 1.0181281041728996e-08, "loss": 0.5655, "step": 4034 }, { "epoch": 0.9863114153018822, "grad_norm": 2.7583189010620117, "learning_rate": 9.827234785874107e-09, "loss": 0.6005, "step": 4035 }, { "epoch": 0.9865558543143486, "grad_norm": 3.0450844764709473, "learning_rate": 9.479450765500497e-09, "loss": 0.4669, "step": 4036 }, { "epoch": 0.986800293326815, "grad_norm": 3.0492610931396484, "learning_rate": 9.137929198612672e-09, "loss": 0.5395, "step": 4037 }, { "epoch": 0.9870447323392814, "grad_norm": 3.032101631164551, "learning_rate": 8.802670299290494e-09, "loss": 0.5397, "step": 4038 }, { "epoch": 0.9872891713517478, "grad_norm": 3.1941475868225098, "learning_rate": 8.473674277685862e-09, "loss": 0.5214, "step": 4039 }, { "epoch": 0.9875336103642142, "grad_norm": 3.856670379638672, "learning_rate": 8.150941340027142e-09, "loss": 0.551, "step": 4040 }, { "epoch": 0.9877780493766806, "grad_norm": 3.1291728019714355, "learning_rate": 7.834471688616952e-09, "loss": 0.5206, "step": 4041 }, { "epoch": 0.9880224883891469, "grad_norm": 3.016200304031372, "learning_rate": 7.524265521831054e-09, "loss": 0.5845, "step": 4042 }, { "epoch": 0.9882669274016133, "grad_norm": 3.971785306930542, "learning_rate": 7.220323034117238e-09, "loss": 0.5915, "step": 4043 }, { "epoch": 0.9885113664140797, "grad_norm": 3.125352382659912, "learning_rate": 6.9226444159997684e-09, "loss": 0.4734, "step": 4044 }, { "epoch": 0.9887558054265461, "grad_norm": 2.851304531097412, "learning_rate": 6.631229854076049e-09, "loss": 0.5954, "step": 4045 }, { "epoch": 0.9890002444390125, "grad_norm": 2.7313387393951416, "learning_rate": 6.346079531015514e-09, "loss": 1.8465, "step": 4046 }, { "epoch": 0.9892446834514789, "grad_norm": 2.8360304832458496, "learning_rate": 6.067193625560741e-09, "loss": 0.488, "step": 4047 }, { "epoch": 0.9894891224639453, "grad_norm": 2.8465449810028076, "learning_rate": 5.794572312530777e-09, "loss": 0.4853, "step": 4048 }, { "epoch": 0.9897335614764117, "grad_norm": 3.0731120109558105, "learning_rate": 5.528215762813372e-09, "loss": 0.5777, "step": 4049 }, { "epoch": 0.9899780004888781, "grad_norm": 3.0907604694366455, "learning_rate": 5.268124143371633e-09, "loss": 0.6588, "step": 4050 }, { "epoch": 0.9902224395013444, "grad_norm": 3.1686196327209473, "learning_rate": 5.014297617242925e-09, "loss": 0.57, "step": 4051 }, { "epoch": 0.9904668785138108, "grad_norm": 3.1961419582366943, "learning_rate": 4.766736343533307e-09, "loss": 0.5639, "step": 4052 }, { "epoch": 0.9907113175262772, "grad_norm": 3.2424230575561523, "learning_rate": 4.525440477425314e-09, "loss": 0.605, "step": 4053 }, { "epoch": 0.9909557565387436, "grad_norm": 3.204475164413452, "learning_rate": 4.290410170173509e-09, "loss": 0.4764, "step": 4054 }, { "epoch": 0.99120019555121, "grad_norm": 10.434322357177734, "learning_rate": 4.061645569103379e-09, "loss": 0.6154, "step": 4055 }, { "epoch": 0.9914446345636764, "grad_norm": 3.9556498527526855, "learning_rate": 3.839146817612438e-09, "loss": 0.5348, "step": 4056 }, { "epoch": 0.9916890735761428, "grad_norm": 3.443718671798706, "learning_rate": 3.6229140551735655e-09, "loss": 0.5136, "step": 4057 }, { "epoch": 0.9919335125886092, "grad_norm": 2.643991708755493, "learning_rate": 3.4129474173294486e-09, "loss": 0.4989, "step": 4058 }, { "epoch": 0.9921779516010756, "grad_norm": 3.180394411087036, "learning_rate": 3.209247035694807e-09, "loss": 0.4559, "step": 4059 }, { "epoch": 0.9924223906135419, "grad_norm": 3.485826015472412, "learning_rate": 3.0118130379575005e-09, "loss": 0.5038, "step": 4060 }, { "epoch": 0.9926668296260083, "grad_norm": 3.4516539573669434, "learning_rate": 2.8206455478774206e-09, "loss": 0.5237, "step": 4061 }, { "epoch": 0.9929112686384747, "grad_norm": 3.217087507247925, "learning_rate": 2.6357446852853797e-09, "loss": 0.5465, "step": 4062 }, { "epoch": 0.9931557076509411, "grad_norm": 3.096763849258423, "learning_rate": 2.45711056608533e-09, "loss": 0.5782, "step": 4063 }, { "epoch": 0.9934001466634075, "grad_norm": 3.3539602756500244, "learning_rate": 2.284743302251036e-09, "loss": 0.5717, "step": 4064 }, { "epoch": 0.9936445856758739, "grad_norm": 3.5747156143188477, "learning_rate": 2.1186430018294015e-09, "loss": 0.4844, "step": 4065 }, { "epoch": 0.9938890246883403, "grad_norm": 3.2676870822906494, "learning_rate": 1.9588097689393627e-09, "loss": 0.6203, "step": 4066 }, { "epoch": 0.9941334637008067, "grad_norm": 3.1657159328460693, "learning_rate": 1.805243703770776e-09, "loss": 0.5333, "step": 4067 }, { "epoch": 0.9943779027132731, "grad_norm": 2.9021496772766113, "learning_rate": 1.6579449025844186e-09, "loss": 0.5604, "step": 4068 }, { "epoch": 0.9946223417257394, "grad_norm": 2.9915456771850586, "learning_rate": 1.5169134577119882e-09, "loss": 0.5709, "step": 4069 }, { "epoch": 0.9948667807382058, "grad_norm": 3.2231616973876953, "learning_rate": 1.3821494575594341e-09, "loss": 0.519, "step": 4070 }, { "epoch": 0.9951112197506722, "grad_norm": 4.008378505706787, "learning_rate": 1.2536529866014058e-09, "loss": 0.5854, "step": 4071 }, { "epoch": 0.9953556587631386, "grad_norm": 3.2395238876342773, "learning_rate": 1.1314241253834734e-09, "loss": 0.5486, "step": 4072 }, { "epoch": 0.995600097775605, "grad_norm": 3.368342161178589, "learning_rate": 1.0154629505254587e-09, "loss": 0.513, "step": 4073 }, { "epoch": 0.9958445367880714, "grad_norm": 3.0852789878845215, "learning_rate": 9.057695347158835e-10, "loss": 0.5585, "step": 4074 }, { "epoch": 0.9960889758005378, "grad_norm": 2.6465818881988525, "learning_rate": 8.023439467141902e-10, "loss": 1.8768, "step": 4075 }, { "epoch": 0.9963334148130042, "grad_norm": 5.087937355041504, "learning_rate": 7.051862513518526e-10, "loss": 0.6375, "step": 4076 }, { "epoch": 0.9965778538254706, "grad_norm": 3.305978536605835, "learning_rate": 6.142965095312648e-10, "loss": 0.6294, "step": 4077 }, { "epoch": 0.9968222928379369, "grad_norm": 3.551265001296997, "learning_rate": 5.296747782257416e-10, "loss": 0.5115, "step": 4078 }, { "epoch": 0.9970667318504033, "grad_norm": 3.1084768772125244, "learning_rate": 4.51321110479519e-10, "loss": 0.5921, "step": 4079 }, { "epoch": 0.9973111708628697, "grad_norm": 3.146742820739746, "learning_rate": 3.7923555540886336e-10, "loss": 0.4884, "step": 4080 }, { "epoch": 0.9975556098753361, "grad_norm": 3.0563406944274902, "learning_rate": 3.1341815819763146e-10, "loss": 0.5111, "step": 4081 }, { "epoch": 0.9978000488878025, "grad_norm": 3.5022075176239014, "learning_rate": 2.5386896010393126e-10, "loss": 0.6503, "step": 4082 }, { "epoch": 0.9980444879002689, "grad_norm": 3.6591250896453857, "learning_rate": 2.0058799845568134e-10, "loss": 0.6007, "step": 4083 }, { "epoch": 0.9982889269127353, "grad_norm": 3.736536979675293, "learning_rate": 1.5357530665172094e-10, "loss": 0.5575, "step": 4084 }, { "epoch": 0.9985333659252017, "grad_norm": 3.080501079559326, "learning_rate": 1.128309141606998e-10, "loss": 0.5473, "step": 4085 }, { "epoch": 0.9987778049376681, "grad_norm": 3.0575947761535645, "learning_rate": 7.835484652329861e-11, "loss": 0.4791, "step": 4086 }, { "epoch": 0.9990222439501344, "grad_norm": 3.32004976272583, "learning_rate": 5.014712535111876e-11, "loss": 0.5526, "step": 4087 }, { "epoch": 0.9992666829626008, "grad_norm": 3.2589986324310303, "learning_rate": 2.8207768324461928e-11, "loss": 0.5164, "step": 4088 }, { "epoch": 0.9995111219750672, "grad_norm": 3.316433906555176, "learning_rate": 1.2536789196770927e-11, "loss": 0.5081, "step": 4089 }, { "epoch": 0.9997555609875336, "grad_norm": 3.078674077987671, "learning_rate": 3.134197790188864e-12, "loss": 0.64, "step": 4090 }, { "epoch": 1.0, "grad_norm": 3.979902982711792, "learning_rate": 0.0, "loss": 0.7255, "step": 4091 }, { "epoch": 1.0, "step": 4091, "total_flos": 1.4976433291127685e+18, "train_loss": 0.902596783321838, "train_runtime": 27008.3323, "train_samples_per_second": 7.269, "train_steps_per_second": 0.151 } ], "logging_steps": 1.0, "max_steps": 4091, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 3500000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.4976433291127685e+18, "train_batch_size": 6, "trial_name": null, "trial_params": null }