{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 4656, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0006443298969072165, "grad_norm": 7.110601902008057, "learning_rate": 2.145922746781116e-08, "loss": 1.1276, "step": 1 }, { "epoch": 0.001288659793814433, "grad_norm": 7.2371721267700195, "learning_rate": 4.291845493562232e-08, "loss": 1.1072, "step": 2 }, { "epoch": 0.0019329896907216496, "grad_norm": 7.011204719543457, "learning_rate": 6.437768240343348e-08, "loss": 1.1194, "step": 3 }, { "epoch": 0.002577319587628866, "grad_norm": 7.473376750946045, "learning_rate": 8.583690987124465e-08, "loss": 1.1356, "step": 4 }, { "epoch": 0.0032216494845360823, "grad_norm": 7.445034503936768, "learning_rate": 1.072961373390558e-07, "loss": 1.1441, "step": 5 }, { "epoch": 0.003865979381443299, "grad_norm": 7.178295612335205, "learning_rate": 1.2875536480686695e-07, "loss": 1.1268, "step": 6 }, { "epoch": 0.004510309278350515, "grad_norm": 7.132780075073242, "learning_rate": 1.5021459227467812e-07, "loss": 1.1407, "step": 7 }, { "epoch": 0.005154639175257732, "grad_norm": 6.90869140625, "learning_rate": 1.716738197424893e-07, "loss": 1.119, "step": 8 }, { "epoch": 0.005798969072164949, "grad_norm": 7.242048740386963, "learning_rate": 1.9313304721030046e-07, "loss": 1.1212, "step": 9 }, { "epoch": 0.006443298969072165, "grad_norm": 6.913217544555664, "learning_rate": 2.145922746781116e-07, "loss": 1.1294, "step": 10 }, { "epoch": 0.007087628865979381, "grad_norm": 6.906850337982178, "learning_rate": 2.3605150214592278e-07, "loss": 1.1179, "step": 11 }, { "epoch": 0.007731958762886598, "grad_norm": 7.030967712402344, "learning_rate": 2.575107296137339e-07, "loss": 1.1117, "step": 12 }, { "epoch": 0.008376288659793814, "grad_norm": 7.052539825439453, "learning_rate": 2.789699570815451e-07, "loss": 1.1308, "step": 13 }, { "epoch": 0.00902061855670103, "grad_norm": 7.180584907531738, "learning_rate": 3.0042918454935624e-07, "loss": 1.1379, "step": 14 }, { "epoch": 0.009664948453608248, "grad_norm": 6.794429302215576, "learning_rate": 3.218884120171674e-07, "loss": 1.1254, "step": 15 }, { "epoch": 0.010309278350515464, "grad_norm": 6.6315460205078125, "learning_rate": 3.433476394849786e-07, "loss": 1.129, "step": 16 }, { "epoch": 0.01095360824742268, "grad_norm": 6.192741394042969, "learning_rate": 3.6480686695278973e-07, "loss": 1.0931, "step": 17 }, { "epoch": 0.011597938144329897, "grad_norm": 6.684213161468506, "learning_rate": 3.8626609442060093e-07, "loss": 1.097, "step": 18 }, { "epoch": 0.012242268041237113, "grad_norm": 6.633246898651123, "learning_rate": 4.07725321888412e-07, "loss": 1.1156, "step": 19 }, { "epoch": 0.01288659793814433, "grad_norm": 6.388881206512451, "learning_rate": 4.291845493562232e-07, "loss": 1.1355, "step": 20 }, { "epoch": 0.013530927835051547, "grad_norm": 5.435776710510254, "learning_rate": 4.5064377682403436e-07, "loss": 1.086, "step": 21 }, { "epoch": 0.014175257731958763, "grad_norm": 5.1827287673950195, "learning_rate": 4.7210300429184556e-07, "loss": 1.0772, "step": 22 }, { "epoch": 0.014819587628865979, "grad_norm": 5.062796115875244, "learning_rate": 4.935622317596567e-07, "loss": 1.0899, "step": 23 }, { "epoch": 0.015463917525773196, "grad_norm": 4.846042633056641, "learning_rate": 5.150214592274678e-07, "loss": 1.0445, "step": 24 }, { "epoch": 0.01610824742268041, "grad_norm": 4.710771560668945, "learning_rate": 5.36480686695279e-07, "loss": 1.0438, "step": 25 }, { "epoch": 0.01675257731958763, "grad_norm": 7.707144260406494, "learning_rate": 5.579399141630902e-07, "loss": 1.1404, "step": 26 }, { "epoch": 0.017396907216494846, "grad_norm": 4.591529369354248, "learning_rate": 5.793991416309013e-07, "loss": 1.0613, "step": 27 }, { "epoch": 0.01804123711340206, "grad_norm": 4.643542289733887, "learning_rate": 6.008583690987125e-07, "loss": 1.0745, "step": 28 }, { "epoch": 0.018685567010309278, "grad_norm": 2.991255521774292, "learning_rate": 6.223175965665236e-07, "loss": 1.0227, "step": 29 }, { "epoch": 0.019329896907216496, "grad_norm": 2.871891736984253, "learning_rate": 6.437768240343348e-07, "loss": 1.0161, "step": 30 }, { "epoch": 0.01997422680412371, "grad_norm": 2.757741689682007, "learning_rate": 6.652360515021459e-07, "loss": 1.0369, "step": 31 }, { "epoch": 0.020618556701030927, "grad_norm": 2.799226999282837, "learning_rate": 6.866952789699572e-07, "loss": 1.0143, "step": 32 }, { "epoch": 0.021262886597938145, "grad_norm": 2.6710588932037354, "learning_rate": 7.081545064377683e-07, "loss": 0.9664, "step": 33 }, { "epoch": 0.02190721649484536, "grad_norm": 2.6751835346221924, "learning_rate": 7.296137339055795e-07, "loss": 1.0091, "step": 34 }, { "epoch": 0.022551546391752577, "grad_norm": 2.653776168823242, "learning_rate": 7.510729613733907e-07, "loss": 1.0076, "step": 35 }, { "epoch": 0.023195876288659795, "grad_norm": 2.5108840465545654, "learning_rate": 7.725321888412019e-07, "loss": 0.9828, "step": 36 }, { "epoch": 0.02384020618556701, "grad_norm": 2.4935286045074463, "learning_rate": 7.939914163090129e-07, "loss": 0.9828, "step": 37 }, { "epoch": 0.024484536082474227, "grad_norm": 2.238776445388794, "learning_rate": 8.15450643776824e-07, "loss": 0.9822, "step": 38 }, { "epoch": 0.025128865979381444, "grad_norm": 1.7742258310317993, "learning_rate": 8.369098712446352e-07, "loss": 0.9661, "step": 39 }, { "epoch": 0.02577319587628866, "grad_norm": 1.9159789085388184, "learning_rate": 8.583690987124464e-07, "loss": 0.9514, "step": 40 }, { "epoch": 0.026417525773195876, "grad_norm": 1.9263840913772583, "learning_rate": 8.798283261802576e-07, "loss": 0.9566, "step": 41 }, { "epoch": 0.027061855670103094, "grad_norm": 1.9234426021575928, "learning_rate": 9.012875536480687e-07, "loss": 0.9241, "step": 42 }, { "epoch": 0.027706185567010308, "grad_norm": 1.7556010484695435, "learning_rate": 9.2274678111588e-07, "loss": 0.9226, "step": 43 }, { "epoch": 0.028350515463917526, "grad_norm": 1.7978771924972534, "learning_rate": 9.442060085836911e-07, "loss": 0.9332, "step": 44 }, { "epoch": 0.028994845360824743, "grad_norm": 1.6247144937515259, "learning_rate": 9.656652360515022e-07, "loss": 0.9279, "step": 45 }, { "epoch": 0.029639175257731958, "grad_norm": 1.5409694910049438, "learning_rate": 9.871244635193133e-07, "loss": 0.9143, "step": 46 }, { "epoch": 0.030283505154639175, "grad_norm": 1.3396236896514893, "learning_rate": 1.0085836909871245e-06, "loss": 0.9158, "step": 47 }, { "epoch": 0.030927835051546393, "grad_norm": 1.2827332019805908, "learning_rate": 1.0300429184549356e-06, "loss": 0.9117, "step": 48 }, { "epoch": 0.03157216494845361, "grad_norm": 1.2325736284255981, "learning_rate": 1.051502145922747e-06, "loss": 0.9237, "step": 49 }, { "epoch": 0.03221649484536082, "grad_norm": 1.2356582880020142, "learning_rate": 1.072961373390558e-06, "loss": 0.902, "step": 50 }, { "epoch": 0.03286082474226804, "grad_norm": 1.2174237966537476, "learning_rate": 1.0944206008583692e-06, "loss": 0.9067, "step": 51 }, { "epoch": 0.03350515463917526, "grad_norm": 1.205655813217163, "learning_rate": 1.1158798283261804e-06, "loss": 0.8716, "step": 52 }, { "epoch": 0.03414948453608247, "grad_norm": 1.3317782878875732, "learning_rate": 1.1373390557939915e-06, "loss": 0.8774, "step": 53 }, { "epoch": 0.03479381443298969, "grad_norm": 1.2049365043640137, "learning_rate": 1.1587982832618027e-06, "loss": 0.8715, "step": 54 }, { "epoch": 0.035438144329896906, "grad_norm": 1.2478984594345093, "learning_rate": 1.1802575107296138e-06, "loss": 0.8664, "step": 55 }, { "epoch": 0.03608247422680412, "grad_norm": 1.0256261825561523, "learning_rate": 1.201716738197425e-06, "loss": 0.8645, "step": 56 }, { "epoch": 0.03672680412371134, "grad_norm": 0.9772769212722778, "learning_rate": 1.2231759656652361e-06, "loss": 0.8501, "step": 57 }, { "epoch": 0.037371134020618556, "grad_norm": 1.046656847000122, "learning_rate": 1.2446351931330473e-06, "loss": 0.8214, "step": 58 }, { "epoch": 0.03801546391752577, "grad_norm": 1.0926765203475952, "learning_rate": 1.2660944206008586e-06, "loss": 0.8432, "step": 59 }, { "epoch": 0.03865979381443299, "grad_norm": 1.0493460893630981, "learning_rate": 1.2875536480686696e-06, "loss": 0.8339, "step": 60 }, { "epoch": 0.039304123711340205, "grad_norm": 1.1175580024719238, "learning_rate": 1.309012875536481e-06, "loss": 0.8355, "step": 61 }, { "epoch": 0.03994845360824742, "grad_norm": 0.9839034676551819, "learning_rate": 1.3304721030042918e-06, "loss": 0.8439, "step": 62 }, { "epoch": 0.04059278350515464, "grad_norm": 0.8569884300231934, "learning_rate": 1.3519313304721032e-06, "loss": 0.8457, "step": 63 }, { "epoch": 0.041237113402061855, "grad_norm": 0.7701215147972107, "learning_rate": 1.3733905579399143e-06, "loss": 0.8283, "step": 64 }, { "epoch": 0.04188144329896907, "grad_norm": 0.7985702157020569, "learning_rate": 1.3948497854077253e-06, "loss": 0.8317, "step": 65 }, { "epoch": 0.04252577319587629, "grad_norm": 0.7105412483215332, "learning_rate": 1.4163090128755366e-06, "loss": 0.8115, "step": 66 }, { "epoch": 0.043170103092783504, "grad_norm": 0.7274165153503418, "learning_rate": 1.4377682403433476e-06, "loss": 0.8118, "step": 67 }, { "epoch": 0.04381443298969072, "grad_norm": 0.746004045009613, "learning_rate": 1.459227467811159e-06, "loss": 0.7964, "step": 68 }, { "epoch": 0.04445876288659794, "grad_norm": 0.7390499711036682, "learning_rate": 1.48068669527897e-06, "loss": 0.8182, "step": 69 }, { "epoch": 0.045103092783505154, "grad_norm": 0.7096271514892578, "learning_rate": 1.5021459227467814e-06, "loss": 0.8045, "step": 70 }, { "epoch": 0.04574742268041237, "grad_norm": 0.7057626247406006, "learning_rate": 1.5236051502145924e-06, "loss": 0.8012, "step": 71 }, { "epoch": 0.04639175257731959, "grad_norm": 0.6706627607345581, "learning_rate": 1.5450643776824037e-06, "loss": 0.8104, "step": 72 }, { "epoch": 0.047036082474226804, "grad_norm": 0.6653320789337158, "learning_rate": 1.5665236051502147e-06, "loss": 0.8012, "step": 73 }, { "epoch": 0.04768041237113402, "grad_norm": 0.709685206413269, "learning_rate": 1.5879828326180258e-06, "loss": 0.8054, "step": 74 }, { "epoch": 0.04832474226804124, "grad_norm": 0.6996152400970459, "learning_rate": 1.6094420600858372e-06, "loss": 0.7904, "step": 75 }, { "epoch": 0.04896907216494845, "grad_norm": 0.6835272908210754, "learning_rate": 1.630901287553648e-06, "loss": 0.829, "step": 76 }, { "epoch": 0.04961340206185567, "grad_norm": 0.6945921778678894, "learning_rate": 1.6523605150214594e-06, "loss": 0.826, "step": 77 }, { "epoch": 0.05025773195876289, "grad_norm": 0.6877522468566895, "learning_rate": 1.6738197424892704e-06, "loss": 0.7916, "step": 78 }, { "epoch": 0.0509020618556701, "grad_norm": 0.7250533699989319, "learning_rate": 1.6952789699570817e-06, "loss": 0.7815, "step": 79 }, { "epoch": 0.05154639175257732, "grad_norm": 0.6545637845993042, "learning_rate": 1.7167381974248929e-06, "loss": 0.7994, "step": 80 }, { "epoch": 0.05219072164948454, "grad_norm": 0.6582580208778381, "learning_rate": 1.7381974248927038e-06, "loss": 0.8143, "step": 81 }, { "epoch": 0.05283505154639175, "grad_norm": 0.6158490180969238, "learning_rate": 1.7596566523605152e-06, "loss": 0.7834, "step": 82 }, { "epoch": 0.053479381443298966, "grad_norm": 0.7158565521240234, "learning_rate": 1.7811158798283263e-06, "loss": 0.7836, "step": 83 }, { "epoch": 0.05412371134020619, "grad_norm": 0.6770134568214417, "learning_rate": 1.8025751072961375e-06, "loss": 0.7926, "step": 84 }, { "epoch": 0.0547680412371134, "grad_norm": 0.623852550983429, "learning_rate": 1.8240343347639486e-06, "loss": 0.7943, "step": 85 }, { "epoch": 0.055412371134020616, "grad_norm": 0.6438775658607483, "learning_rate": 1.84549356223176e-06, "loss": 0.7795, "step": 86 }, { "epoch": 0.05605670103092784, "grad_norm": 0.6181722283363342, "learning_rate": 1.866952789699571e-06, "loss": 0.7691, "step": 87 }, { "epoch": 0.05670103092783505, "grad_norm": 0.6757891178131104, "learning_rate": 1.8884120171673823e-06, "loss": 0.769, "step": 88 }, { "epoch": 0.057345360824742266, "grad_norm": 0.5938519239425659, "learning_rate": 1.9098712446351934e-06, "loss": 0.7671, "step": 89 }, { "epoch": 0.05798969072164949, "grad_norm": 0.5742356777191162, "learning_rate": 1.9313304721030043e-06, "loss": 0.7487, "step": 90 }, { "epoch": 0.0586340206185567, "grad_norm": 0.620228111743927, "learning_rate": 1.9527896995708157e-06, "loss": 0.7508, "step": 91 }, { "epoch": 0.059278350515463915, "grad_norm": 0.6447807550430298, "learning_rate": 1.9742489270386266e-06, "loss": 0.7669, "step": 92 }, { "epoch": 0.059922680412371136, "grad_norm": 0.6638559103012085, "learning_rate": 1.995708154506438e-06, "loss": 0.7601, "step": 93 }, { "epoch": 0.06056701030927835, "grad_norm": 0.6439847350120544, "learning_rate": 2.017167381974249e-06, "loss": 0.7914, "step": 94 }, { "epoch": 0.061211340206185565, "grad_norm": 0.6288350820541382, "learning_rate": 2.0386266094420603e-06, "loss": 0.7622, "step": 95 }, { "epoch": 0.061855670103092786, "grad_norm": 0.6260641813278198, "learning_rate": 2.060085836909871e-06, "loss": 0.7411, "step": 96 }, { "epoch": 0.0625, "grad_norm": 0.5831416845321655, "learning_rate": 2.0815450643776826e-06, "loss": 0.7676, "step": 97 }, { "epoch": 0.06314432989690721, "grad_norm": 0.618773877620697, "learning_rate": 2.103004291845494e-06, "loss": 0.7715, "step": 98 }, { "epoch": 0.06378865979381443, "grad_norm": 0.6563497185707092, "learning_rate": 2.124463519313305e-06, "loss": 0.768, "step": 99 }, { "epoch": 0.06443298969072164, "grad_norm": 0.661193311214447, "learning_rate": 2.145922746781116e-06, "loss": 0.7594, "step": 100 }, { "epoch": 0.06507731958762887, "grad_norm": 0.5894092917442322, "learning_rate": 2.167381974248927e-06, "loss": 0.7461, "step": 101 }, { "epoch": 0.06572164948453608, "grad_norm": 0.5690293908119202, "learning_rate": 2.1888412017167385e-06, "loss": 0.7655, "step": 102 }, { "epoch": 0.0663659793814433, "grad_norm": 0.6250858902931213, "learning_rate": 2.2103004291845494e-06, "loss": 0.7479, "step": 103 }, { "epoch": 0.06701030927835051, "grad_norm": 0.6357895731925964, "learning_rate": 2.2317596566523608e-06, "loss": 0.76, "step": 104 }, { "epoch": 0.06765463917525773, "grad_norm": 0.572175920009613, "learning_rate": 2.2532188841201717e-06, "loss": 0.7364, "step": 105 }, { "epoch": 0.06829896907216494, "grad_norm": 0.5831284523010254, "learning_rate": 2.274678111587983e-06, "loss": 0.754, "step": 106 }, { "epoch": 0.06894329896907217, "grad_norm": 0.5630279183387756, "learning_rate": 2.2961373390557944e-06, "loss": 0.742, "step": 107 }, { "epoch": 0.06958762886597938, "grad_norm": 0.5738471150398254, "learning_rate": 2.3175965665236054e-06, "loss": 0.7516, "step": 108 }, { "epoch": 0.0702319587628866, "grad_norm": 0.5648314952850342, "learning_rate": 2.3390557939914167e-06, "loss": 0.7514, "step": 109 }, { "epoch": 0.07087628865979381, "grad_norm": 0.6147857904434204, "learning_rate": 2.3605150214592277e-06, "loss": 0.764, "step": 110 }, { "epoch": 0.07152061855670103, "grad_norm": 0.6403223872184753, "learning_rate": 2.381974248927039e-06, "loss": 0.7431, "step": 111 }, { "epoch": 0.07216494845360824, "grad_norm": 0.6109868288040161, "learning_rate": 2.40343347639485e-06, "loss": 0.717, "step": 112 }, { "epoch": 0.07280927835051547, "grad_norm": 0.68013995885849, "learning_rate": 2.424892703862661e-06, "loss": 0.743, "step": 113 }, { "epoch": 0.07345360824742268, "grad_norm": 0.5799630284309387, "learning_rate": 2.4463519313304722e-06, "loss": 0.7632, "step": 114 }, { "epoch": 0.0740979381443299, "grad_norm": 0.6456771492958069, "learning_rate": 2.467811158798283e-06, "loss": 0.759, "step": 115 }, { "epoch": 0.07474226804123711, "grad_norm": 0.6629730463027954, "learning_rate": 2.4892703862660945e-06, "loss": 0.7477, "step": 116 }, { "epoch": 0.07538659793814433, "grad_norm": 0.6278717517852783, "learning_rate": 2.510729613733906e-06, "loss": 0.7567, "step": 117 }, { "epoch": 0.07603092783505154, "grad_norm": 0.6200818419456482, "learning_rate": 2.5321888412017172e-06, "loss": 0.7372, "step": 118 }, { "epoch": 0.07667525773195877, "grad_norm": 0.6505746245384216, "learning_rate": 2.553648068669528e-06, "loss": 0.7407, "step": 119 }, { "epoch": 0.07731958762886598, "grad_norm": 0.6583534479141235, "learning_rate": 2.575107296137339e-06, "loss": 0.7632, "step": 120 }, { "epoch": 0.0779639175257732, "grad_norm": 0.5869385600090027, "learning_rate": 2.59656652360515e-06, "loss": 0.7242, "step": 121 }, { "epoch": 0.07860824742268041, "grad_norm": 0.5914641618728638, "learning_rate": 2.618025751072962e-06, "loss": 0.7621, "step": 122 }, { "epoch": 0.07925257731958762, "grad_norm": 0.5811408758163452, "learning_rate": 2.6394849785407728e-06, "loss": 0.7363, "step": 123 }, { "epoch": 0.07989690721649484, "grad_norm": 0.6350299715995789, "learning_rate": 2.6609442060085837e-06, "loss": 0.7262, "step": 124 }, { "epoch": 0.08054123711340207, "grad_norm": 0.6121471524238586, "learning_rate": 2.682403433476395e-06, "loss": 0.7344, "step": 125 }, { "epoch": 0.08118556701030928, "grad_norm": 0.6378530859947205, "learning_rate": 2.7038626609442064e-06, "loss": 0.7264, "step": 126 }, { "epoch": 0.0818298969072165, "grad_norm": 0.629152238368988, "learning_rate": 2.7253218884120173e-06, "loss": 0.7482, "step": 127 }, { "epoch": 0.08247422680412371, "grad_norm": 0.6727675199508667, "learning_rate": 2.7467811158798287e-06, "loss": 0.7544, "step": 128 }, { "epoch": 0.08311855670103092, "grad_norm": 0.5801218748092651, "learning_rate": 2.7682403433476396e-06, "loss": 0.7577, "step": 129 }, { "epoch": 0.08376288659793814, "grad_norm": 0.5930361151695251, "learning_rate": 2.7896995708154506e-06, "loss": 0.7296, "step": 130 }, { "epoch": 0.08440721649484537, "grad_norm": 0.6841540336608887, "learning_rate": 2.8111587982832623e-06, "loss": 0.7625, "step": 131 }, { "epoch": 0.08505154639175258, "grad_norm": 0.6143293976783752, "learning_rate": 2.8326180257510733e-06, "loss": 0.7226, "step": 132 }, { "epoch": 0.0856958762886598, "grad_norm": 0.6195946335792542, "learning_rate": 2.854077253218884e-06, "loss": 0.7342, "step": 133 }, { "epoch": 0.08634020618556701, "grad_norm": 0.6028604507446289, "learning_rate": 2.875536480686695e-06, "loss": 0.7451, "step": 134 }, { "epoch": 0.08698453608247422, "grad_norm": 0.6273115277290344, "learning_rate": 2.896995708154507e-06, "loss": 0.7342, "step": 135 }, { "epoch": 0.08762886597938144, "grad_norm": 0.5710777640342712, "learning_rate": 2.918454935622318e-06, "loss": 0.7175, "step": 136 }, { "epoch": 0.08827319587628867, "grad_norm": 0.566100001335144, "learning_rate": 2.939914163090129e-06, "loss": 0.6804, "step": 137 }, { "epoch": 0.08891752577319588, "grad_norm": 0.6627039313316345, "learning_rate": 2.96137339055794e-06, "loss": 0.7047, "step": 138 }, { "epoch": 0.0895618556701031, "grad_norm": 0.6538745164871216, "learning_rate": 2.982832618025751e-06, "loss": 0.7277, "step": 139 }, { "epoch": 0.09020618556701031, "grad_norm": 0.5521991848945618, "learning_rate": 3.004291845493563e-06, "loss": 0.7288, "step": 140 }, { "epoch": 0.09085051546391752, "grad_norm": 0.6788451671600342, "learning_rate": 3.025751072961374e-06, "loss": 0.7266, "step": 141 }, { "epoch": 0.09149484536082474, "grad_norm": 0.6423946022987366, "learning_rate": 3.0472103004291847e-06, "loss": 0.713, "step": 142 }, { "epoch": 0.09213917525773196, "grad_norm": 0.5802819132804871, "learning_rate": 3.0686695278969957e-06, "loss": 0.7403, "step": 143 }, { "epoch": 0.09278350515463918, "grad_norm": 0.6125937104225159, "learning_rate": 3.0901287553648074e-06, "loss": 0.7491, "step": 144 }, { "epoch": 0.09342783505154639, "grad_norm": 0.618902325630188, "learning_rate": 3.1115879828326184e-06, "loss": 0.7278, "step": 145 }, { "epoch": 0.09407216494845361, "grad_norm": 0.6590876579284668, "learning_rate": 3.1330472103004293e-06, "loss": 0.7208, "step": 146 }, { "epoch": 0.09471649484536082, "grad_norm": 0.6284641623497009, "learning_rate": 3.1545064377682407e-06, "loss": 0.7275, "step": 147 }, { "epoch": 0.09536082474226804, "grad_norm": 0.6145771145820618, "learning_rate": 3.1759656652360516e-06, "loss": 0.7187, "step": 148 }, { "epoch": 0.09600515463917526, "grad_norm": 0.6180451512336731, "learning_rate": 3.197424892703863e-06, "loss": 0.7185, "step": 149 }, { "epoch": 0.09664948453608248, "grad_norm": 0.5832404494285583, "learning_rate": 3.2188841201716743e-06, "loss": 0.7113, "step": 150 }, { "epoch": 0.09729381443298969, "grad_norm": 0.6325623989105225, "learning_rate": 3.2403433476394852e-06, "loss": 0.724, "step": 151 }, { "epoch": 0.0979381443298969, "grad_norm": 0.6722143292427063, "learning_rate": 3.261802575107296e-06, "loss": 0.7302, "step": 152 }, { "epoch": 0.09858247422680412, "grad_norm": 0.5832739472389221, "learning_rate": 3.283261802575107e-06, "loss": 0.7341, "step": 153 }, { "epoch": 0.09922680412371133, "grad_norm": 0.6487283110618591, "learning_rate": 3.304721030042919e-06, "loss": 0.7446, "step": 154 }, { "epoch": 0.09987113402061856, "grad_norm": 0.6175786852836609, "learning_rate": 3.32618025751073e-06, "loss": 0.722, "step": 155 }, { "epoch": 0.10051546391752578, "grad_norm": 0.6783549785614014, "learning_rate": 3.3476394849785408e-06, "loss": 0.7379, "step": 156 }, { "epoch": 0.10115979381443299, "grad_norm": 0.6923893690109253, "learning_rate": 3.369098712446352e-06, "loss": 0.7527, "step": 157 }, { "epoch": 0.1018041237113402, "grad_norm": 0.5557937026023865, "learning_rate": 3.3905579399141635e-06, "loss": 0.7232, "step": 158 }, { "epoch": 0.10244845360824742, "grad_norm": 0.5501450300216675, "learning_rate": 3.412017167381975e-06, "loss": 0.7257, "step": 159 }, { "epoch": 0.10309278350515463, "grad_norm": 0.602790117263794, "learning_rate": 3.4334763948497858e-06, "loss": 0.7153, "step": 160 }, { "epoch": 0.10373711340206186, "grad_norm": 0.6886022090911865, "learning_rate": 3.4549356223175967e-06, "loss": 0.721, "step": 161 }, { "epoch": 0.10438144329896908, "grad_norm": 0.620611846446991, "learning_rate": 3.4763948497854076e-06, "loss": 0.7315, "step": 162 }, { "epoch": 0.10502577319587629, "grad_norm": 0.644003689289093, "learning_rate": 3.4978540772532194e-06, "loss": 0.7033, "step": 163 }, { "epoch": 0.1056701030927835, "grad_norm": 0.609451413154602, "learning_rate": 3.5193133047210303e-06, "loss": 0.7258, "step": 164 }, { "epoch": 0.10631443298969072, "grad_norm": 0.5927817821502686, "learning_rate": 3.5407725321888413e-06, "loss": 0.7405, "step": 165 }, { "epoch": 0.10695876288659793, "grad_norm": 0.6208043098449707, "learning_rate": 3.5622317596566526e-06, "loss": 0.7049, "step": 166 }, { "epoch": 0.10760309278350516, "grad_norm": 0.6070868372917175, "learning_rate": 3.583690987124464e-06, "loss": 0.7189, "step": 167 }, { "epoch": 0.10824742268041238, "grad_norm": 0.607042670249939, "learning_rate": 3.605150214592275e-06, "loss": 0.7275, "step": 168 }, { "epoch": 0.10889175257731959, "grad_norm": 0.6275116801261902, "learning_rate": 3.6266094420600863e-06, "loss": 0.7133, "step": 169 }, { "epoch": 0.1095360824742268, "grad_norm": 0.6777527332305908, "learning_rate": 3.648068669527897e-06, "loss": 0.7187, "step": 170 }, { "epoch": 0.11018041237113402, "grad_norm": 0.7039538025856018, "learning_rate": 3.669527896995708e-06, "loss": 0.7238, "step": 171 }, { "epoch": 0.11082474226804123, "grad_norm": 0.6610651612281799, "learning_rate": 3.69098712446352e-06, "loss": 0.7367, "step": 172 }, { "epoch": 0.11146907216494846, "grad_norm": 0.6705362796783447, "learning_rate": 3.712446351931331e-06, "loss": 0.7079, "step": 173 }, { "epoch": 0.11211340206185567, "grad_norm": 0.785240888595581, "learning_rate": 3.733905579399142e-06, "loss": 0.7218, "step": 174 }, { "epoch": 0.11275773195876289, "grad_norm": 0.620299220085144, "learning_rate": 3.7553648068669527e-06, "loss": 0.6949, "step": 175 }, { "epoch": 0.1134020618556701, "grad_norm": 0.6333096027374268, "learning_rate": 3.7768240343347645e-06, "loss": 0.7151, "step": 176 }, { "epoch": 0.11404639175257732, "grad_norm": 0.6317524909973145, "learning_rate": 3.7982832618025754e-06, "loss": 0.7267, "step": 177 }, { "epoch": 0.11469072164948453, "grad_norm": 0.6243137121200562, "learning_rate": 3.819742489270387e-06, "loss": 0.734, "step": 178 }, { "epoch": 0.11533505154639176, "grad_norm": 0.5798303484916687, "learning_rate": 3.841201716738197e-06, "loss": 0.7033, "step": 179 }, { "epoch": 0.11597938144329897, "grad_norm": 0.5858868956565857, "learning_rate": 3.862660944206009e-06, "loss": 0.6914, "step": 180 }, { "epoch": 0.11662371134020619, "grad_norm": 0.5907038450241089, "learning_rate": 3.88412017167382e-06, "loss": 0.7168, "step": 181 }, { "epoch": 0.1172680412371134, "grad_norm": 0.6620161533355713, "learning_rate": 3.905579399141631e-06, "loss": 0.7002, "step": 182 }, { "epoch": 0.11791237113402062, "grad_norm": 0.6809520125389099, "learning_rate": 3.927038626609443e-06, "loss": 0.7222, "step": 183 }, { "epoch": 0.11855670103092783, "grad_norm": 0.6089593768119812, "learning_rate": 3.948497854077253e-06, "loss": 0.7097, "step": 184 }, { "epoch": 0.11920103092783506, "grad_norm": 0.6427510380744934, "learning_rate": 3.969957081545065e-06, "loss": 0.7258, "step": 185 }, { "epoch": 0.11984536082474227, "grad_norm": 0.5783995389938354, "learning_rate": 3.991416309012876e-06, "loss": 0.7139, "step": 186 }, { "epoch": 0.12048969072164949, "grad_norm": 0.6561513543128967, "learning_rate": 4.012875536480687e-06, "loss": 0.7071, "step": 187 }, { "epoch": 0.1211340206185567, "grad_norm": 0.5855488777160645, "learning_rate": 4.034334763948498e-06, "loss": 0.6925, "step": 188 }, { "epoch": 0.12177835051546392, "grad_norm": 0.6439988017082214, "learning_rate": 4.055793991416309e-06, "loss": 0.701, "step": 189 }, { "epoch": 0.12242268041237113, "grad_norm": 0.6012244820594788, "learning_rate": 4.0772532188841205e-06, "loss": 0.7194, "step": 190 }, { "epoch": 0.12306701030927836, "grad_norm": 0.6320216655731201, "learning_rate": 4.098712446351932e-06, "loss": 0.7269, "step": 191 }, { "epoch": 0.12371134020618557, "grad_norm": 0.6225422620773315, "learning_rate": 4.120171673819742e-06, "loss": 0.721, "step": 192 }, { "epoch": 0.12435567010309279, "grad_norm": 0.7879793643951416, "learning_rate": 4.141630901287554e-06, "loss": 0.7074, "step": 193 }, { "epoch": 0.125, "grad_norm": 0.6840083003044128, "learning_rate": 4.163090128755365e-06, "loss": 0.7099, "step": 194 }, { "epoch": 0.12564432989690721, "grad_norm": 0.6430121064186096, "learning_rate": 4.1845493562231765e-06, "loss": 0.7183, "step": 195 }, { "epoch": 0.12628865979381443, "grad_norm": 0.6003310084342957, "learning_rate": 4.206008583690988e-06, "loss": 0.7278, "step": 196 }, { "epoch": 0.12693298969072164, "grad_norm": 0.6260854601860046, "learning_rate": 4.227467811158798e-06, "loss": 0.6834, "step": 197 }, { "epoch": 0.12757731958762886, "grad_norm": 0.6256402730941772, "learning_rate": 4.24892703862661e-06, "loss": 0.7115, "step": 198 }, { "epoch": 0.12822164948453607, "grad_norm": 0.6426709294319153, "learning_rate": 4.270386266094421e-06, "loss": 0.6944, "step": 199 }, { "epoch": 0.12886597938144329, "grad_norm": 0.5736485123634338, "learning_rate": 4.291845493562232e-06, "loss": 0.7072, "step": 200 }, { "epoch": 0.12951030927835053, "grad_norm": 0.5523471832275391, "learning_rate": 4.313304721030043e-06, "loss": 0.7007, "step": 201 }, { "epoch": 0.13015463917525774, "grad_norm": 0.6589829921722412, "learning_rate": 4.334763948497854e-06, "loss": 0.7014, "step": 202 }, { "epoch": 0.13079896907216496, "grad_norm": 0.6589136719703674, "learning_rate": 4.356223175965666e-06, "loss": 0.7188, "step": 203 }, { "epoch": 0.13144329896907217, "grad_norm": 0.6531979441642761, "learning_rate": 4.377682403433477e-06, "loss": 0.7303, "step": 204 }, { "epoch": 0.13208762886597938, "grad_norm": 0.6530022621154785, "learning_rate": 4.399141630901288e-06, "loss": 0.674, "step": 205 }, { "epoch": 0.1327319587628866, "grad_norm": 0.6427558064460754, "learning_rate": 4.420600858369099e-06, "loss": 0.7181, "step": 206 }, { "epoch": 0.1333762886597938, "grad_norm": 0.6089393496513367, "learning_rate": 4.44206008583691e-06, "loss": 0.7155, "step": 207 }, { "epoch": 0.13402061855670103, "grad_norm": 0.6053621172904968, "learning_rate": 4.4635193133047216e-06, "loss": 0.7176, "step": 208 }, { "epoch": 0.13466494845360824, "grad_norm": 0.5928657650947571, "learning_rate": 4.484978540772533e-06, "loss": 0.713, "step": 209 }, { "epoch": 0.13530927835051546, "grad_norm": 0.5997255444526672, "learning_rate": 4.5064377682403434e-06, "loss": 0.711, "step": 210 }, { "epoch": 0.13595360824742267, "grad_norm": 0.6372370719909668, "learning_rate": 4.527896995708155e-06, "loss": 0.7036, "step": 211 }, { "epoch": 0.13659793814432988, "grad_norm": 0.627398669719696, "learning_rate": 4.549356223175966e-06, "loss": 0.7158, "step": 212 }, { "epoch": 0.13724226804123713, "grad_norm": 0.5548101663589478, "learning_rate": 4.5708154506437775e-06, "loss": 0.6941, "step": 213 }, { "epoch": 0.13788659793814434, "grad_norm": 0.6887850165367126, "learning_rate": 4.592274678111589e-06, "loss": 0.717, "step": 214 }, { "epoch": 0.13853092783505155, "grad_norm": 0.6508849859237671, "learning_rate": 4.613733905579399e-06, "loss": 0.7132, "step": 215 }, { "epoch": 0.13917525773195877, "grad_norm": 0.5906941294670105, "learning_rate": 4.635193133047211e-06, "loss": 0.7012, "step": 216 }, { "epoch": 0.13981958762886598, "grad_norm": 0.6980515718460083, "learning_rate": 4.656652360515021e-06, "loss": 0.6966, "step": 217 }, { "epoch": 0.1404639175257732, "grad_norm": 0.6323555111885071, "learning_rate": 4.6781115879828334e-06, "loss": 0.6824, "step": 218 }, { "epoch": 0.1411082474226804, "grad_norm": 0.6843428611755371, "learning_rate": 4.699570815450644e-06, "loss": 0.6778, "step": 219 }, { "epoch": 0.14175257731958762, "grad_norm": 0.6896873712539673, "learning_rate": 4.721030042918455e-06, "loss": 0.7004, "step": 220 }, { "epoch": 0.14239690721649484, "grad_norm": 0.6527490615844727, "learning_rate": 4.742489270386267e-06, "loss": 0.7063, "step": 221 }, { "epoch": 0.14304123711340205, "grad_norm": 0.688143253326416, "learning_rate": 4.763948497854078e-06, "loss": 0.7202, "step": 222 }, { "epoch": 0.14368556701030927, "grad_norm": 0.6378912925720215, "learning_rate": 4.7854077253218885e-06, "loss": 0.6782, "step": 223 }, { "epoch": 0.14432989690721648, "grad_norm": 0.6460157632827759, "learning_rate": 4.8068669527897e-06, "loss": 0.7044, "step": 224 }, { "epoch": 0.14497422680412372, "grad_norm": 0.6503344774246216, "learning_rate": 4.828326180257511e-06, "loss": 0.6967, "step": 225 }, { "epoch": 0.14561855670103094, "grad_norm": 0.6385593414306641, "learning_rate": 4.849785407725322e-06, "loss": 0.6974, "step": 226 }, { "epoch": 0.14626288659793815, "grad_norm": 0.643185555934906, "learning_rate": 4.871244635193134e-06, "loss": 0.7077, "step": 227 }, { "epoch": 0.14690721649484537, "grad_norm": 0.6844326853752136, "learning_rate": 4.8927038626609445e-06, "loss": 0.7014, "step": 228 }, { "epoch": 0.14755154639175258, "grad_norm": 0.6674286127090454, "learning_rate": 4.914163090128756e-06, "loss": 0.7065, "step": 229 }, { "epoch": 0.1481958762886598, "grad_norm": 0.5696401596069336, "learning_rate": 4.935622317596566e-06, "loss": 0.71, "step": 230 }, { "epoch": 0.148840206185567, "grad_norm": 0.7520720362663269, "learning_rate": 4.9570815450643785e-06, "loss": 0.7005, "step": 231 }, { "epoch": 0.14948453608247422, "grad_norm": 0.6747918128967285, "learning_rate": 4.978540772532189e-06, "loss": 0.709, "step": 232 }, { "epoch": 0.15012886597938144, "grad_norm": 0.7071784138679504, "learning_rate": 5e-06, "loss": 0.6943, "step": 233 }, { "epoch": 0.15077319587628865, "grad_norm": 0.6552026271820068, "learning_rate": 5.021459227467812e-06, "loss": 0.702, "step": 234 }, { "epoch": 0.15141752577319587, "grad_norm": 0.665221631526947, "learning_rate": 5.042918454935622e-06, "loss": 0.6931, "step": 235 }, { "epoch": 0.15206185567010308, "grad_norm": 0.6667265295982361, "learning_rate": 5.0643776824034345e-06, "loss": 0.6764, "step": 236 }, { "epoch": 0.15270618556701032, "grad_norm": 0.6807601451873779, "learning_rate": 5.085836909871244e-06, "loss": 0.6716, "step": 237 }, { "epoch": 0.15335051546391754, "grad_norm": 0.6183218359947205, "learning_rate": 5.107296137339056e-06, "loss": 0.6883, "step": 238 }, { "epoch": 0.15399484536082475, "grad_norm": 0.712308406829834, "learning_rate": 5.128755364806868e-06, "loss": 0.6953, "step": 239 }, { "epoch": 0.15463917525773196, "grad_norm": 0.651141881942749, "learning_rate": 5.150214592274678e-06, "loss": 0.7075, "step": 240 }, { "epoch": 0.15528350515463918, "grad_norm": 0.6598920822143555, "learning_rate": 5.1716738197424896e-06, "loss": 0.681, "step": 241 }, { "epoch": 0.1559278350515464, "grad_norm": 0.733690083026886, "learning_rate": 5.1931330472103e-06, "loss": 0.7033, "step": 242 }, { "epoch": 0.1565721649484536, "grad_norm": 0.6627342104911804, "learning_rate": 5.214592274678112e-06, "loss": 0.6939, "step": 243 }, { "epoch": 0.15721649484536082, "grad_norm": 0.6706998348236084, "learning_rate": 5.236051502145924e-06, "loss": 0.6902, "step": 244 }, { "epoch": 0.15786082474226804, "grad_norm": 0.676616370677948, "learning_rate": 5.257510729613734e-06, "loss": 0.6936, "step": 245 }, { "epoch": 0.15850515463917525, "grad_norm": 0.6022554039955139, "learning_rate": 5.2789699570815455e-06, "loss": 0.7145, "step": 246 }, { "epoch": 0.15914948453608246, "grad_norm": 0.6522243618965149, "learning_rate": 5.300429184549357e-06, "loss": 0.6925, "step": 247 }, { "epoch": 0.15979381443298968, "grad_norm": 0.658470869064331, "learning_rate": 5.321888412017167e-06, "loss": 0.6983, "step": 248 }, { "epoch": 0.16043814432989692, "grad_norm": 0.6702178716659546, "learning_rate": 5.34334763948498e-06, "loss": 0.6811, "step": 249 }, { "epoch": 0.16108247422680413, "grad_norm": 0.6123067140579224, "learning_rate": 5.36480686695279e-06, "loss": 0.6888, "step": 250 }, { "epoch": 0.16172680412371135, "grad_norm": 0.6775242686271667, "learning_rate": 5.3862660944206014e-06, "loss": 0.7184, "step": 251 }, { "epoch": 0.16237113402061856, "grad_norm": 0.684730052947998, "learning_rate": 5.407725321888413e-06, "loss": 0.6889, "step": 252 }, { "epoch": 0.16301546391752578, "grad_norm": 0.5833144187927246, "learning_rate": 5.429184549356223e-06, "loss": 0.7031, "step": 253 }, { "epoch": 0.163659793814433, "grad_norm": 0.710203230381012, "learning_rate": 5.450643776824035e-06, "loss": 0.6796, "step": 254 }, { "epoch": 0.1643041237113402, "grad_norm": 0.676637589931488, "learning_rate": 5.472103004291845e-06, "loss": 0.6898, "step": 255 }, { "epoch": 0.16494845360824742, "grad_norm": 0.7159347534179688, "learning_rate": 5.493562231759657e-06, "loss": 0.6792, "step": 256 }, { "epoch": 0.16559278350515463, "grad_norm": 0.7335567474365234, "learning_rate": 5.515021459227469e-06, "loss": 0.7004, "step": 257 }, { "epoch": 0.16623711340206185, "grad_norm": 0.685498058795929, "learning_rate": 5.536480686695279e-06, "loss": 0.6774, "step": 258 }, { "epoch": 0.16688144329896906, "grad_norm": 0.7149011492729187, "learning_rate": 5.557939914163091e-06, "loss": 0.6818, "step": 259 }, { "epoch": 0.16752577319587628, "grad_norm": 0.7026809453964233, "learning_rate": 5.579399141630901e-06, "loss": 0.6864, "step": 260 }, { "epoch": 0.16817010309278352, "grad_norm": 0.6316818594932556, "learning_rate": 5.6008583690987125e-06, "loss": 0.6917, "step": 261 }, { "epoch": 0.16881443298969073, "grad_norm": 0.6292291283607483, "learning_rate": 5.622317596566525e-06, "loss": 0.7124, "step": 262 }, { "epoch": 0.16945876288659795, "grad_norm": 0.6646528244018555, "learning_rate": 5.643776824034335e-06, "loss": 0.692, "step": 263 }, { "epoch": 0.17010309278350516, "grad_norm": 0.6756330132484436, "learning_rate": 5.6652360515021465e-06, "loss": 0.6798, "step": 264 }, { "epoch": 0.17074742268041238, "grad_norm": 0.7752998471260071, "learning_rate": 5.686695278969958e-06, "loss": 0.7157, "step": 265 }, { "epoch": 0.1713917525773196, "grad_norm": 0.6672255396842957, "learning_rate": 5.708154506437768e-06, "loss": 0.6921, "step": 266 }, { "epoch": 0.1720360824742268, "grad_norm": 0.6201971769332886, "learning_rate": 5.72961373390558e-06, "loss": 0.6698, "step": 267 }, { "epoch": 0.17268041237113402, "grad_norm": 0.8849310278892517, "learning_rate": 5.75107296137339e-06, "loss": 0.6774, "step": 268 }, { "epoch": 0.17332474226804123, "grad_norm": 0.6066883206367493, "learning_rate": 5.7725321888412025e-06, "loss": 0.7126, "step": 269 }, { "epoch": 0.17396907216494845, "grad_norm": 0.7163182497024536, "learning_rate": 5.793991416309014e-06, "loss": 0.7013, "step": 270 }, { "epoch": 0.17461340206185566, "grad_norm": 0.6985563635826111, "learning_rate": 5.815450643776824e-06, "loss": 0.6843, "step": 271 }, { "epoch": 0.17525773195876287, "grad_norm": 0.6979106664657593, "learning_rate": 5.836909871244636e-06, "loss": 0.6797, "step": 272 }, { "epoch": 0.17590206185567012, "grad_norm": 0.7043166160583496, "learning_rate": 5.858369098712446e-06, "loss": 0.6771, "step": 273 }, { "epoch": 0.17654639175257733, "grad_norm": 0.6490378975868225, "learning_rate": 5.879828326180258e-06, "loss": 0.6957, "step": 274 }, { "epoch": 0.17719072164948454, "grad_norm": 0.7085456252098083, "learning_rate": 5.90128755364807e-06, "loss": 0.7011, "step": 275 }, { "epoch": 0.17783505154639176, "grad_norm": 0.6790546178817749, "learning_rate": 5.92274678111588e-06, "loss": 0.6843, "step": 276 }, { "epoch": 0.17847938144329897, "grad_norm": 0.5952588319778442, "learning_rate": 5.944206008583692e-06, "loss": 0.6788, "step": 277 }, { "epoch": 0.1791237113402062, "grad_norm": 0.6639391779899597, "learning_rate": 5.965665236051502e-06, "loss": 0.6666, "step": 278 }, { "epoch": 0.1797680412371134, "grad_norm": 0.7042316794395447, "learning_rate": 5.9871244635193135e-06, "loss": 0.7046, "step": 279 }, { "epoch": 0.18041237113402062, "grad_norm": 0.6036393642425537, "learning_rate": 6.008583690987126e-06, "loss": 0.6841, "step": 280 }, { "epoch": 0.18105670103092783, "grad_norm": 0.6635739207267761, "learning_rate": 6.030042918454936e-06, "loss": 0.6977, "step": 281 }, { "epoch": 0.18170103092783504, "grad_norm": 0.5560936331748962, "learning_rate": 6.051502145922748e-06, "loss": 0.6751, "step": 282 }, { "epoch": 0.18234536082474226, "grad_norm": 0.5534313917160034, "learning_rate": 6.072961373390558e-06, "loss": 0.6905, "step": 283 }, { "epoch": 0.18298969072164947, "grad_norm": 0.5958704352378845, "learning_rate": 6.0944206008583694e-06, "loss": 0.6702, "step": 284 }, { "epoch": 0.18363402061855671, "grad_norm": 0.6193600296974182, "learning_rate": 6.115879828326181e-06, "loss": 0.6995, "step": 285 }, { "epoch": 0.18427835051546393, "grad_norm": 0.6361005306243896, "learning_rate": 6.137339055793991e-06, "loss": 0.6716, "step": 286 }, { "epoch": 0.18492268041237114, "grad_norm": 0.6057776212692261, "learning_rate": 6.1587982832618035e-06, "loss": 0.6743, "step": 287 }, { "epoch": 0.18556701030927836, "grad_norm": 0.6024848222732544, "learning_rate": 6.180257510729615e-06, "loss": 0.6951, "step": 288 }, { "epoch": 0.18621134020618557, "grad_norm": 0.7520748376846313, "learning_rate": 6.201716738197425e-06, "loss": 0.6665, "step": 289 }, { "epoch": 0.18685567010309279, "grad_norm": 0.6366672515869141, "learning_rate": 6.223175965665237e-06, "loss": 0.6756, "step": 290 }, { "epoch": 0.1875, "grad_norm": 0.6061584949493408, "learning_rate": 6.244635193133047e-06, "loss": 0.6872, "step": 291 }, { "epoch": 0.18814432989690721, "grad_norm": 0.659074068069458, "learning_rate": 6.266094420600859e-06, "loss": 0.7011, "step": 292 }, { "epoch": 0.18878865979381443, "grad_norm": 0.5834708213806152, "learning_rate": 6.287553648068671e-06, "loss": 0.68, "step": 293 }, { "epoch": 0.18943298969072164, "grad_norm": 0.6897289752960205, "learning_rate": 6.309012875536481e-06, "loss": 0.6886, "step": 294 }, { "epoch": 0.19007731958762886, "grad_norm": 0.6058546304702759, "learning_rate": 6.330472103004293e-06, "loss": 0.69, "step": 295 }, { "epoch": 0.19072164948453607, "grad_norm": 0.6750144362449646, "learning_rate": 6.351931330472103e-06, "loss": 0.6903, "step": 296 }, { "epoch": 0.19136597938144329, "grad_norm": 0.6852880120277405, "learning_rate": 6.3733905579399145e-06, "loss": 0.6969, "step": 297 }, { "epoch": 0.19201030927835053, "grad_norm": 0.6959925889968872, "learning_rate": 6.394849785407726e-06, "loss": 0.6846, "step": 298 }, { "epoch": 0.19265463917525774, "grad_norm": 0.7114775776863098, "learning_rate": 6.416309012875536e-06, "loss": 0.6817, "step": 299 }, { "epoch": 0.19329896907216496, "grad_norm": 0.7066349387168884, "learning_rate": 6.437768240343349e-06, "loss": 0.7085, "step": 300 }, { "epoch": 0.19394329896907217, "grad_norm": 0.7317931652069092, "learning_rate": 6.459227467811159e-06, "loss": 0.7101, "step": 301 }, { "epoch": 0.19458762886597938, "grad_norm": 0.6549123525619507, "learning_rate": 6.4806866952789705e-06, "loss": 0.6752, "step": 302 }, { "epoch": 0.1952319587628866, "grad_norm": 0.6251623630523682, "learning_rate": 6.502145922746782e-06, "loss": 0.7056, "step": 303 }, { "epoch": 0.1958762886597938, "grad_norm": 0.620888352394104, "learning_rate": 6.523605150214592e-06, "loss": 0.6682, "step": 304 }, { "epoch": 0.19652061855670103, "grad_norm": 0.6473459601402283, "learning_rate": 6.545064377682404e-06, "loss": 0.6778, "step": 305 }, { "epoch": 0.19716494845360824, "grad_norm": 0.6732136607170105, "learning_rate": 6.566523605150214e-06, "loss": 0.6837, "step": 306 }, { "epoch": 0.19780927835051546, "grad_norm": 0.709600567817688, "learning_rate": 6.587982832618026e-06, "loss": 0.6929, "step": 307 }, { "epoch": 0.19845360824742267, "grad_norm": 0.6508067846298218, "learning_rate": 6.609442060085838e-06, "loss": 0.6773, "step": 308 }, { "epoch": 0.19909793814432988, "grad_norm": 0.7103069424629211, "learning_rate": 6.630901287553648e-06, "loss": 0.6934, "step": 309 }, { "epoch": 0.19974226804123713, "grad_norm": 0.683337390422821, "learning_rate": 6.65236051502146e-06, "loss": 0.6774, "step": 310 }, { "epoch": 0.20038659793814434, "grad_norm": 0.595637321472168, "learning_rate": 6.673819742489272e-06, "loss": 0.6832, "step": 311 }, { "epoch": 0.20103092783505155, "grad_norm": 0.6746698617935181, "learning_rate": 6.6952789699570815e-06, "loss": 0.6695, "step": 312 }, { "epoch": 0.20167525773195877, "grad_norm": 0.6899095177650452, "learning_rate": 6.716738197424894e-06, "loss": 0.7003, "step": 313 }, { "epoch": 0.20231958762886598, "grad_norm": 0.6622353196144104, "learning_rate": 6.738197424892704e-06, "loss": 0.6715, "step": 314 }, { "epoch": 0.2029639175257732, "grad_norm": 0.7099422812461853, "learning_rate": 6.759656652360516e-06, "loss": 0.6911, "step": 315 }, { "epoch": 0.2036082474226804, "grad_norm": 0.6878989934921265, "learning_rate": 6.781115879828327e-06, "loss": 0.6652, "step": 316 }, { "epoch": 0.20425257731958762, "grad_norm": 0.6950581073760986, "learning_rate": 6.8025751072961374e-06, "loss": 0.6642, "step": 317 }, { "epoch": 0.20489690721649484, "grad_norm": 0.7173905968666077, "learning_rate": 6.82403433476395e-06, "loss": 0.6785, "step": 318 }, { "epoch": 0.20554123711340205, "grad_norm": 0.6895129680633545, "learning_rate": 6.84549356223176e-06, "loss": 0.6704, "step": 319 }, { "epoch": 0.20618556701030927, "grad_norm": 0.7093017101287842, "learning_rate": 6.8669527896995715e-06, "loss": 0.6573, "step": 320 }, { "epoch": 0.20682989690721648, "grad_norm": 0.6341881155967712, "learning_rate": 6.888412017167383e-06, "loss": 0.6948, "step": 321 }, { "epoch": 0.20747422680412372, "grad_norm": 0.7234093546867371, "learning_rate": 6.909871244635193e-06, "loss": 0.6894, "step": 322 }, { "epoch": 0.20811855670103094, "grad_norm": 0.6822957992553711, "learning_rate": 6.931330472103005e-06, "loss": 0.6732, "step": 323 }, { "epoch": 0.20876288659793815, "grad_norm": 0.6355515718460083, "learning_rate": 6.952789699570815e-06, "loss": 0.6859, "step": 324 }, { "epoch": 0.20940721649484537, "grad_norm": 0.7011468410491943, "learning_rate": 6.9742489270386275e-06, "loss": 0.6767, "step": 325 }, { "epoch": 0.21005154639175258, "grad_norm": 0.727989137172699, "learning_rate": 6.995708154506439e-06, "loss": 0.6614, "step": 326 }, { "epoch": 0.2106958762886598, "grad_norm": 0.7674729228019714, "learning_rate": 7.017167381974249e-06, "loss": 0.6798, "step": 327 }, { "epoch": 0.211340206185567, "grad_norm": 0.6890509128570557, "learning_rate": 7.038626609442061e-06, "loss": 0.691, "step": 328 }, { "epoch": 0.21198453608247422, "grad_norm": 0.6729913949966431, "learning_rate": 7.060085836909872e-06, "loss": 0.6944, "step": 329 }, { "epoch": 0.21262886597938144, "grad_norm": 0.7068386077880859, "learning_rate": 7.0815450643776825e-06, "loss": 0.6657, "step": 330 }, { "epoch": 0.21327319587628865, "grad_norm": 0.6367886662483215, "learning_rate": 7.103004291845495e-06, "loss": 0.6653, "step": 331 }, { "epoch": 0.21391752577319587, "grad_norm": 0.6715766787528992, "learning_rate": 7.124463519313305e-06, "loss": 0.6736, "step": 332 }, { "epoch": 0.21456185567010308, "grad_norm": 0.7752699851989746, "learning_rate": 7.145922746781117e-06, "loss": 0.6938, "step": 333 }, { "epoch": 0.21520618556701032, "grad_norm": 0.7189640402793884, "learning_rate": 7.167381974248928e-06, "loss": 0.6959, "step": 334 }, { "epoch": 0.21585051546391754, "grad_norm": 0.638003945350647, "learning_rate": 7.1888412017167385e-06, "loss": 0.687, "step": 335 }, { "epoch": 0.21649484536082475, "grad_norm": 0.7306050658226013, "learning_rate": 7.21030042918455e-06, "loss": 0.6922, "step": 336 }, { "epoch": 0.21713917525773196, "grad_norm": 0.672145664691925, "learning_rate": 7.23175965665236e-06, "loss": 0.6615, "step": 337 }, { "epoch": 0.21778350515463918, "grad_norm": 0.7220679521560669, "learning_rate": 7.2532188841201726e-06, "loss": 0.6946, "step": 338 }, { "epoch": 0.2184278350515464, "grad_norm": 0.7008548378944397, "learning_rate": 7.274678111587984e-06, "loss": 0.6527, "step": 339 }, { "epoch": 0.2190721649484536, "grad_norm": 0.681744396686554, "learning_rate": 7.296137339055794e-06, "loss": 0.6695, "step": 340 }, { "epoch": 0.21971649484536082, "grad_norm": 0.7134683728218079, "learning_rate": 7.317596566523606e-06, "loss": 0.6881, "step": 341 }, { "epoch": 0.22036082474226804, "grad_norm": 0.6850399971008301, "learning_rate": 7.339055793991416e-06, "loss": 0.6825, "step": 342 }, { "epoch": 0.22100515463917525, "grad_norm": 0.670241117477417, "learning_rate": 7.360515021459228e-06, "loss": 0.6861, "step": 343 }, { "epoch": 0.22164948453608246, "grad_norm": 0.6723126173019409, "learning_rate": 7.38197424892704e-06, "loss": 0.6709, "step": 344 }, { "epoch": 0.22229381443298968, "grad_norm": 0.6781944632530212, "learning_rate": 7.40343347639485e-06, "loss": 0.6696, "step": 345 }, { "epoch": 0.22293814432989692, "grad_norm": 0.6173579096794128, "learning_rate": 7.424892703862662e-06, "loss": 0.674, "step": 346 }, { "epoch": 0.22358247422680413, "grad_norm": 0.7341597676277161, "learning_rate": 7.446351931330472e-06, "loss": 0.6892, "step": 347 }, { "epoch": 0.22422680412371135, "grad_norm": 0.6928067207336426, "learning_rate": 7.467811158798284e-06, "loss": 0.6894, "step": 348 }, { "epoch": 0.22487113402061856, "grad_norm": 0.6428167223930359, "learning_rate": 7.489270386266096e-06, "loss": 0.6896, "step": 349 }, { "epoch": 0.22551546391752578, "grad_norm": 0.6473916172981262, "learning_rate": 7.5107296137339054e-06, "loss": 0.6679, "step": 350 }, { "epoch": 0.226159793814433, "grad_norm": 0.6726614832878113, "learning_rate": 7.532188841201718e-06, "loss": 0.6742, "step": 351 }, { "epoch": 0.2268041237113402, "grad_norm": 0.6203340888023376, "learning_rate": 7.553648068669529e-06, "loss": 0.6799, "step": 352 }, { "epoch": 0.22744845360824742, "grad_norm": 0.6794568300247192, "learning_rate": 7.5751072961373395e-06, "loss": 0.687, "step": 353 }, { "epoch": 0.22809278350515463, "grad_norm": 0.6228857040405273, "learning_rate": 7.596566523605151e-06, "loss": 0.6798, "step": 354 }, { "epoch": 0.22873711340206185, "grad_norm": 0.7089255452156067, "learning_rate": 7.618025751072961e-06, "loss": 0.6877, "step": 355 }, { "epoch": 0.22938144329896906, "grad_norm": 0.7471839785575867, "learning_rate": 7.639484978540774e-06, "loss": 0.7091, "step": 356 }, { "epoch": 0.23002577319587628, "grad_norm": 0.7836564779281616, "learning_rate": 7.660944206008584e-06, "loss": 0.6785, "step": 357 }, { "epoch": 0.23067010309278352, "grad_norm": 0.7077754735946655, "learning_rate": 7.682403433476395e-06, "loss": 0.6573, "step": 358 }, { "epoch": 0.23131443298969073, "grad_norm": 0.8452844023704529, "learning_rate": 7.703862660944207e-06, "loss": 0.672, "step": 359 }, { "epoch": 0.23195876288659795, "grad_norm": 0.6950657963752747, "learning_rate": 7.725321888412017e-06, "loss": 0.6685, "step": 360 }, { "epoch": 0.23260309278350516, "grad_norm": 0.895047128200531, "learning_rate": 7.74678111587983e-06, "loss": 0.6894, "step": 361 }, { "epoch": 0.23324742268041238, "grad_norm": 0.7426673769950867, "learning_rate": 7.76824034334764e-06, "loss": 0.6776, "step": 362 }, { "epoch": 0.2338917525773196, "grad_norm": 0.7076351642608643, "learning_rate": 7.78969957081545e-06, "loss": 0.6829, "step": 363 }, { "epoch": 0.2345360824742268, "grad_norm": 0.9385673999786377, "learning_rate": 7.811158798283263e-06, "loss": 0.6762, "step": 364 }, { "epoch": 0.23518041237113402, "grad_norm": 0.7567678689956665, "learning_rate": 7.832618025751073e-06, "loss": 0.6634, "step": 365 }, { "epoch": 0.23582474226804123, "grad_norm": 0.9127271771430969, "learning_rate": 7.854077253218885e-06, "loss": 0.6957, "step": 366 }, { "epoch": 0.23646907216494845, "grad_norm": 0.8499244451522827, "learning_rate": 7.875536480686696e-06, "loss": 0.6799, "step": 367 }, { "epoch": 0.23711340206185566, "grad_norm": 0.6736699938774109, "learning_rate": 7.896995708154506e-06, "loss": 0.692, "step": 368 }, { "epoch": 0.23775773195876287, "grad_norm": 0.78948575258255, "learning_rate": 7.918454935622319e-06, "loss": 0.6785, "step": 369 }, { "epoch": 0.23840206185567012, "grad_norm": 0.7162641286849976, "learning_rate": 7.93991416309013e-06, "loss": 0.6813, "step": 370 }, { "epoch": 0.23904639175257733, "grad_norm": 0.8003594875335693, "learning_rate": 7.96137339055794e-06, "loss": 0.6869, "step": 371 }, { "epoch": 0.23969072164948454, "grad_norm": 0.7845168113708496, "learning_rate": 7.982832618025752e-06, "loss": 0.6961, "step": 372 }, { "epoch": 0.24033505154639176, "grad_norm": 0.6650747656822205, "learning_rate": 8.004291845493562e-06, "loss": 0.6799, "step": 373 }, { "epoch": 0.24097938144329897, "grad_norm": 0.795129656791687, "learning_rate": 8.025751072961375e-06, "loss": 0.6625, "step": 374 }, { "epoch": 0.2416237113402062, "grad_norm": 0.7700155973434448, "learning_rate": 8.047210300429185e-06, "loss": 0.6789, "step": 375 }, { "epoch": 0.2422680412371134, "grad_norm": 0.8150885701179504, "learning_rate": 8.068669527896996e-06, "loss": 0.6937, "step": 376 }, { "epoch": 0.24291237113402062, "grad_norm": 0.6507502794265747, "learning_rate": 8.090128755364808e-06, "loss": 0.6753, "step": 377 }, { "epoch": 0.24355670103092783, "grad_norm": 0.7538992166519165, "learning_rate": 8.111587982832618e-06, "loss": 0.6768, "step": 378 }, { "epoch": 0.24420103092783504, "grad_norm": 0.7078290581703186, "learning_rate": 8.13304721030043e-06, "loss": 0.6808, "step": 379 }, { "epoch": 0.24484536082474226, "grad_norm": 0.7651002407073975, "learning_rate": 8.154506437768241e-06, "loss": 0.6756, "step": 380 }, { "epoch": 0.24548969072164947, "grad_norm": 0.6297709345817566, "learning_rate": 8.175965665236052e-06, "loss": 0.6689, "step": 381 }, { "epoch": 0.24613402061855671, "grad_norm": 0.7845105528831482, "learning_rate": 8.197424892703864e-06, "loss": 0.673, "step": 382 }, { "epoch": 0.24677835051546393, "grad_norm": 0.7264772653579712, "learning_rate": 8.218884120171674e-06, "loss": 0.6883, "step": 383 }, { "epoch": 0.24742268041237114, "grad_norm": 0.6383763551712036, "learning_rate": 8.240343347639485e-06, "loss": 0.6603, "step": 384 }, { "epoch": 0.24806701030927836, "grad_norm": 0.6653419733047485, "learning_rate": 8.261802575107297e-06, "loss": 0.6792, "step": 385 }, { "epoch": 0.24871134020618557, "grad_norm": 0.637873649597168, "learning_rate": 8.283261802575108e-06, "loss": 0.6768, "step": 386 }, { "epoch": 0.24935567010309279, "grad_norm": 0.6562079787254333, "learning_rate": 8.30472103004292e-06, "loss": 0.6733, "step": 387 }, { "epoch": 0.25, "grad_norm": 0.6641515493392944, "learning_rate": 8.32618025751073e-06, "loss": 0.6847, "step": 388 }, { "epoch": 0.2506443298969072, "grad_norm": 0.6497541666030884, "learning_rate": 8.34763948497854e-06, "loss": 0.6474, "step": 389 }, { "epoch": 0.25128865979381443, "grad_norm": 0.7080177664756775, "learning_rate": 8.369098712446353e-06, "loss": 0.6674, "step": 390 }, { "epoch": 0.25193298969072164, "grad_norm": 0.619655966758728, "learning_rate": 8.390557939914163e-06, "loss": 0.6612, "step": 391 }, { "epoch": 0.25257731958762886, "grad_norm": 0.6577807068824768, "learning_rate": 8.412017167381976e-06, "loss": 0.6807, "step": 392 }, { "epoch": 0.25322164948453607, "grad_norm": 0.6744721531867981, "learning_rate": 8.433476394849786e-06, "loss": 0.6841, "step": 393 }, { "epoch": 0.2538659793814433, "grad_norm": 0.6748406291007996, "learning_rate": 8.454935622317597e-06, "loss": 0.6565, "step": 394 }, { "epoch": 0.2545103092783505, "grad_norm": 0.6534087061882019, "learning_rate": 8.476394849785409e-06, "loss": 0.6773, "step": 395 }, { "epoch": 0.2551546391752577, "grad_norm": 0.6861433386802673, "learning_rate": 8.49785407725322e-06, "loss": 0.6631, "step": 396 }, { "epoch": 0.25579896907216493, "grad_norm": 0.6713009476661682, "learning_rate": 8.519313304721032e-06, "loss": 0.6768, "step": 397 }, { "epoch": 0.25644329896907214, "grad_norm": 0.7301905155181885, "learning_rate": 8.540772532188842e-06, "loss": 0.6725, "step": 398 }, { "epoch": 0.25708762886597936, "grad_norm": 0.6636376976966858, "learning_rate": 8.562231759656653e-06, "loss": 0.6464, "step": 399 }, { "epoch": 0.25773195876288657, "grad_norm": 0.7043092250823975, "learning_rate": 8.583690987124465e-06, "loss": 0.6576, "step": 400 }, { "epoch": 0.25837628865979384, "grad_norm": 0.7424471378326416, "learning_rate": 8.605150214592275e-06, "loss": 0.6713, "step": 401 }, { "epoch": 0.25902061855670105, "grad_norm": 0.7654013633728027, "learning_rate": 8.626609442060086e-06, "loss": 0.691, "step": 402 }, { "epoch": 0.25966494845360827, "grad_norm": 0.75190269947052, "learning_rate": 8.648068669527898e-06, "loss": 0.6771, "step": 403 }, { "epoch": 0.2603092783505155, "grad_norm": 0.6555822491645813, "learning_rate": 8.669527896995709e-06, "loss": 0.6693, "step": 404 }, { "epoch": 0.2609536082474227, "grad_norm": 0.7377997636795044, "learning_rate": 8.69098712446352e-06, "loss": 0.6864, "step": 405 }, { "epoch": 0.2615979381443299, "grad_norm": 0.6632713675498962, "learning_rate": 8.712446351931331e-06, "loss": 0.6478, "step": 406 }, { "epoch": 0.2622422680412371, "grad_norm": 0.8292452692985535, "learning_rate": 8.733905579399142e-06, "loss": 0.7002, "step": 407 }, { "epoch": 0.26288659793814434, "grad_norm": 0.7562007308006287, "learning_rate": 8.755364806866954e-06, "loss": 0.6917, "step": 408 }, { "epoch": 0.26353092783505155, "grad_norm": 0.7227399349212646, "learning_rate": 8.776824034334764e-06, "loss": 0.6755, "step": 409 }, { "epoch": 0.26417525773195877, "grad_norm": 0.7944807410240173, "learning_rate": 8.798283261802577e-06, "loss": 0.6754, "step": 410 }, { "epoch": 0.264819587628866, "grad_norm": 0.6523758172988892, "learning_rate": 8.819742489270387e-06, "loss": 0.6779, "step": 411 }, { "epoch": 0.2654639175257732, "grad_norm": 0.8107074499130249, "learning_rate": 8.841201716738198e-06, "loss": 0.6738, "step": 412 }, { "epoch": 0.2661082474226804, "grad_norm": 0.672264575958252, "learning_rate": 8.86266094420601e-06, "loss": 0.6687, "step": 413 }, { "epoch": 0.2667525773195876, "grad_norm": 0.6971896290779114, "learning_rate": 8.88412017167382e-06, "loss": 0.6464, "step": 414 }, { "epoch": 0.26739690721649484, "grad_norm": 0.7722048163414001, "learning_rate": 8.905579399141631e-06, "loss": 0.6573, "step": 415 }, { "epoch": 0.26804123711340205, "grad_norm": 0.7833477258682251, "learning_rate": 8.927038626609443e-06, "loss": 0.6759, "step": 416 }, { "epoch": 0.26868556701030927, "grad_norm": 0.6647096276283264, "learning_rate": 8.948497854077254e-06, "loss": 0.6664, "step": 417 }, { "epoch": 0.2693298969072165, "grad_norm": 0.7119582891464233, "learning_rate": 8.969957081545066e-06, "loss": 0.6829, "step": 418 }, { "epoch": 0.2699742268041237, "grad_norm": 0.7793509364128113, "learning_rate": 8.991416309012876e-06, "loss": 0.6703, "step": 419 }, { "epoch": 0.2706185567010309, "grad_norm": 0.6854057312011719, "learning_rate": 9.012875536480687e-06, "loss": 0.6897, "step": 420 }, { "epoch": 0.2712628865979381, "grad_norm": 0.7024256587028503, "learning_rate": 9.034334763948499e-06, "loss": 0.6524, "step": 421 }, { "epoch": 0.27190721649484534, "grad_norm": 0.7091918587684631, "learning_rate": 9.05579399141631e-06, "loss": 0.7057, "step": 422 }, { "epoch": 0.27255154639175255, "grad_norm": 0.7316577434539795, "learning_rate": 9.077253218884122e-06, "loss": 0.6752, "step": 423 }, { "epoch": 0.27319587628865977, "grad_norm": 0.72195965051651, "learning_rate": 9.098712446351932e-06, "loss": 0.6678, "step": 424 }, { "epoch": 0.27384020618556704, "grad_norm": 0.798934280872345, "learning_rate": 9.120171673819743e-06, "loss": 0.6474, "step": 425 }, { "epoch": 0.27448453608247425, "grad_norm": 0.6303582787513733, "learning_rate": 9.141630901287555e-06, "loss": 0.6749, "step": 426 }, { "epoch": 0.27512886597938147, "grad_norm": 0.7523912191390991, "learning_rate": 9.163090128755366e-06, "loss": 0.6729, "step": 427 }, { "epoch": 0.2757731958762887, "grad_norm": 0.70609450340271, "learning_rate": 9.184549356223178e-06, "loss": 0.664, "step": 428 }, { "epoch": 0.2764175257731959, "grad_norm": 0.653485119342804, "learning_rate": 9.206008583690987e-06, "loss": 0.6697, "step": 429 }, { "epoch": 0.2770618556701031, "grad_norm": 0.697785496711731, "learning_rate": 9.227467811158799e-06, "loss": 0.6801, "step": 430 }, { "epoch": 0.2777061855670103, "grad_norm": 0.6799833178520203, "learning_rate": 9.248927038626611e-06, "loss": 0.6823, "step": 431 }, { "epoch": 0.27835051546391754, "grad_norm": 0.773899495601654, "learning_rate": 9.270386266094421e-06, "loss": 0.6767, "step": 432 }, { "epoch": 0.27899484536082475, "grad_norm": 0.7388724088668823, "learning_rate": 9.291845493562232e-06, "loss": 0.659, "step": 433 }, { "epoch": 0.27963917525773196, "grad_norm": 0.7416597008705139, "learning_rate": 9.313304721030042e-06, "loss": 0.6617, "step": 434 }, { "epoch": 0.2802835051546392, "grad_norm": 0.8551017045974731, "learning_rate": 9.334763948497855e-06, "loss": 0.6491, "step": 435 }, { "epoch": 0.2809278350515464, "grad_norm": 0.7309742569923401, "learning_rate": 9.356223175965667e-06, "loss": 0.6694, "step": 436 }, { "epoch": 0.2815721649484536, "grad_norm": 0.7063896059989929, "learning_rate": 9.377682403433477e-06, "loss": 0.6642, "step": 437 }, { "epoch": 0.2822164948453608, "grad_norm": 0.8130571246147156, "learning_rate": 9.399141630901288e-06, "loss": 0.6771, "step": 438 }, { "epoch": 0.28286082474226804, "grad_norm": 0.7543779611587524, "learning_rate": 9.4206008583691e-06, "loss": 0.6695, "step": 439 }, { "epoch": 0.28350515463917525, "grad_norm": 0.6949403285980225, "learning_rate": 9.44206008583691e-06, "loss": 0.6628, "step": 440 }, { "epoch": 0.28414948453608246, "grad_norm": 0.717494547367096, "learning_rate": 9.463519313304723e-06, "loss": 0.6647, "step": 441 }, { "epoch": 0.2847938144329897, "grad_norm": 0.7556158900260925, "learning_rate": 9.484978540772533e-06, "loss": 0.6822, "step": 442 }, { "epoch": 0.2854381443298969, "grad_norm": 0.6916296482086182, "learning_rate": 9.506437768240344e-06, "loss": 0.6843, "step": 443 }, { "epoch": 0.2860824742268041, "grad_norm": 0.7675018906593323, "learning_rate": 9.527896995708156e-06, "loss": 0.648, "step": 444 }, { "epoch": 0.2867268041237113, "grad_norm": 0.7227826714515686, "learning_rate": 9.549356223175967e-06, "loss": 0.6673, "step": 445 }, { "epoch": 0.28737113402061853, "grad_norm": 0.8025615215301514, "learning_rate": 9.570815450643777e-06, "loss": 0.6632, "step": 446 }, { "epoch": 0.28801546391752575, "grad_norm": 0.736897349357605, "learning_rate": 9.592274678111588e-06, "loss": 0.6784, "step": 447 }, { "epoch": 0.28865979381443296, "grad_norm": 0.7570528388023376, "learning_rate": 9.6137339055794e-06, "loss": 0.6593, "step": 448 }, { "epoch": 0.28930412371134023, "grad_norm": 0.8205536007881165, "learning_rate": 9.635193133047212e-06, "loss": 0.6764, "step": 449 }, { "epoch": 0.28994845360824745, "grad_norm": 0.6664119958877563, "learning_rate": 9.656652360515022e-06, "loss": 0.6579, "step": 450 }, { "epoch": 0.29059278350515466, "grad_norm": 0.8049569725990295, "learning_rate": 9.678111587982833e-06, "loss": 0.6642, "step": 451 }, { "epoch": 0.2912371134020619, "grad_norm": 0.7655144333839417, "learning_rate": 9.699570815450644e-06, "loss": 0.6678, "step": 452 }, { "epoch": 0.2918814432989691, "grad_norm": 0.6981635093688965, "learning_rate": 9.721030042918456e-06, "loss": 0.6667, "step": 453 }, { "epoch": 0.2925257731958763, "grad_norm": 0.7977787852287292, "learning_rate": 9.742489270386268e-06, "loss": 0.6759, "step": 454 }, { "epoch": 0.2931701030927835, "grad_norm": 0.7327467799186707, "learning_rate": 9.763948497854078e-06, "loss": 0.6717, "step": 455 }, { "epoch": 0.29381443298969073, "grad_norm": 0.6356126070022583, "learning_rate": 9.785407725321889e-06, "loss": 0.6716, "step": 456 }, { "epoch": 0.29445876288659795, "grad_norm": 0.6932129859924316, "learning_rate": 9.806866952789701e-06, "loss": 0.6742, "step": 457 }, { "epoch": 0.29510309278350516, "grad_norm": 0.6636216640472412, "learning_rate": 9.828326180257512e-06, "loss": 0.66, "step": 458 }, { "epoch": 0.2957474226804124, "grad_norm": 0.7990880608558655, "learning_rate": 9.849785407725322e-06, "loss": 0.6798, "step": 459 }, { "epoch": 0.2963917525773196, "grad_norm": 0.7023597359657288, "learning_rate": 9.871244635193133e-06, "loss": 0.6664, "step": 460 }, { "epoch": 0.2970360824742268, "grad_norm": 0.7127171158790588, "learning_rate": 9.892703862660945e-06, "loss": 0.6676, "step": 461 }, { "epoch": 0.297680412371134, "grad_norm": 0.6860979199409485, "learning_rate": 9.914163090128757e-06, "loss": 0.6647, "step": 462 }, { "epoch": 0.29832474226804123, "grad_norm": 0.7535109519958496, "learning_rate": 9.935622317596568e-06, "loss": 0.6952, "step": 463 }, { "epoch": 0.29896907216494845, "grad_norm": 0.7200997471809387, "learning_rate": 9.957081545064378e-06, "loss": 0.6623, "step": 464 }, { "epoch": 0.29961340206185566, "grad_norm": 0.7762708067893982, "learning_rate": 9.978540772532189e-06, "loss": 0.6595, "step": 465 }, { "epoch": 0.3002577319587629, "grad_norm": 0.8472055196762085, "learning_rate": 1e-05, "loss": 0.66, "step": 466 }, { "epoch": 0.3009020618556701, "grad_norm": 0.7369788885116577, "learning_rate": 9.999998594562013e-06, "loss": 0.6597, "step": 467 }, { "epoch": 0.3015463917525773, "grad_norm": 0.901429295539856, "learning_rate": 9.999994378248844e-06, "loss": 0.6728, "step": 468 }, { "epoch": 0.3021907216494845, "grad_norm": 0.7305338382720947, "learning_rate": 9.999987351062856e-06, "loss": 0.6401, "step": 469 }, { "epoch": 0.30283505154639173, "grad_norm": 0.6805121898651123, "learning_rate": 9.999977513008008e-06, "loss": 0.6837, "step": 470 }, { "epoch": 0.30347938144329895, "grad_norm": 0.8592498898506165, "learning_rate": 9.999964864089827e-06, "loss": 0.6665, "step": 471 }, { "epoch": 0.30412371134020616, "grad_norm": 0.7479709386825562, "learning_rate": 9.999949404315424e-06, "loss": 0.684, "step": 472 }, { "epoch": 0.30476804123711343, "grad_norm": 0.6966965198516846, "learning_rate": 9.999931133693489e-06, "loss": 0.689, "step": 473 }, { "epoch": 0.30541237113402064, "grad_norm": 0.7703639268875122, "learning_rate": 9.999910052234295e-06, "loss": 0.6761, "step": 474 }, { "epoch": 0.30605670103092786, "grad_norm": 0.7725124955177307, "learning_rate": 9.999886159949694e-06, "loss": 0.6471, "step": 475 }, { "epoch": 0.30670103092783507, "grad_norm": 0.781389057636261, "learning_rate": 9.999859456853116e-06, "loss": 0.6984, "step": 476 }, { "epoch": 0.3073453608247423, "grad_norm": 0.8902122974395752, "learning_rate": 9.999829942959574e-06, "loss": 0.6612, "step": 477 }, { "epoch": 0.3079896907216495, "grad_norm": 0.6777233481407166, "learning_rate": 9.999797618285658e-06, "loss": 0.6656, "step": 478 }, { "epoch": 0.3086340206185567, "grad_norm": 0.8558103442192078, "learning_rate": 9.999762482849544e-06, "loss": 0.6706, "step": 479 }, { "epoch": 0.30927835051546393, "grad_norm": 0.7024961113929749, "learning_rate": 9.99972453667098e-06, "loss": 0.6784, "step": 480 }, { "epoch": 0.30992268041237114, "grad_norm": 0.6966982483863831, "learning_rate": 9.999683779771301e-06, "loss": 0.6661, "step": 481 }, { "epoch": 0.31056701030927836, "grad_norm": 0.804500162601471, "learning_rate": 9.999640212173418e-06, "loss": 0.6769, "step": 482 }, { "epoch": 0.31121134020618557, "grad_norm": 0.6727991700172424, "learning_rate": 9.999593833901827e-06, "loss": 0.6576, "step": 483 }, { "epoch": 0.3118556701030928, "grad_norm": 0.7652475833892822, "learning_rate": 9.999544644982596e-06, "loss": 0.6745, "step": 484 }, { "epoch": 0.3125, "grad_norm": 0.7107188701629639, "learning_rate": 9.999492645443378e-06, "loss": 0.6697, "step": 485 }, { "epoch": 0.3131443298969072, "grad_norm": 0.7786276340484619, "learning_rate": 9.99943783531341e-06, "loss": 0.654, "step": 486 }, { "epoch": 0.31378865979381443, "grad_norm": 0.8017816543579102, "learning_rate": 9.999380214623503e-06, "loss": 0.663, "step": 487 }, { "epoch": 0.31443298969072164, "grad_norm": 0.7324289679527283, "learning_rate": 9.99931978340605e-06, "loss": 0.6715, "step": 488 }, { "epoch": 0.31507731958762886, "grad_norm": 0.8244476318359375, "learning_rate": 9.999256541695023e-06, "loss": 0.6659, "step": 489 }, { "epoch": 0.31572164948453607, "grad_norm": 0.6325238347053528, "learning_rate": 9.999190489525973e-06, "loss": 0.6835, "step": 490 }, { "epoch": 0.3163659793814433, "grad_norm": 0.6994988918304443, "learning_rate": 9.999121626936038e-06, "loss": 0.6627, "step": 491 }, { "epoch": 0.3170103092783505, "grad_norm": 0.7523902654647827, "learning_rate": 9.999049953963929e-06, "loss": 0.658, "step": 492 }, { "epoch": 0.3176546391752577, "grad_norm": 0.6999568343162537, "learning_rate": 9.998975470649936e-06, "loss": 0.6664, "step": 493 }, { "epoch": 0.31829896907216493, "grad_norm": 0.6783645153045654, "learning_rate": 9.998898177035935e-06, "loss": 0.67, "step": 494 }, { "epoch": 0.31894329896907214, "grad_norm": 0.7489144206047058, "learning_rate": 9.998818073165377e-06, "loss": 0.6923, "step": 495 }, { "epoch": 0.31958762886597936, "grad_norm": 0.6934759616851807, "learning_rate": 9.998735159083295e-06, "loss": 0.6952, "step": 496 }, { "epoch": 0.32023195876288657, "grad_norm": 0.7267290353775024, "learning_rate": 9.998649434836298e-06, "loss": 0.6539, "step": 497 }, { "epoch": 0.32087628865979384, "grad_norm": 0.7578147053718567, "learning_rate": 9.998560900472583e-06, "loss": 0.658, "step": 498 }, { "epoch": 0.32152061855670105, "grad_norm": 0.7360075116157532, "learning_rate": 9.99846955604192e-06, "loss": 0.6693, "step": 499 }, { "epoch": 0.32216494845360827, "grad_norm": 0.8120567202568054, "learning_rate": 9.99837540159566e-06, "loss": 0.6593, "step": 500 }, { "epoch": 0.3228092783505155, "grad_norm": 0.7034911513328552, "learning_rate": 9.998278437186734e-06, "loss": 0.6521, "step": 501 }, { "epoch": 0.3234536082474227, "grad_norm": 0.7283563613891602, "learning_rate": 9.998178662869653e-06, "loss": 0.6692, "step": 502 }, { "epoch": 0.3240979381443299, "grad_norm": 0.7492255568504333, "learning_rate": 9.998076078700509e-06, "loss": 0.6751, "step": 503 }, { "epoch": 0.3247422680412371, "grad_norm": 0.7625614404678345, "learning_rate": 9.997970684736972e-06, "loss": 0.6671, "step": 504 }, { "epoch": 0.32538659793814434, "grad_norm": 0.6705223917961121, "learning_rate": 9.99786248103829e-06, "loss": 0.6514, "step": 505 }, { "epoch": 0.32603092783505155, "grad_norm": 0.733304500579834, "learning_rate": 9.997751467665295e-06, "loss": 0.6483, "step": 506 }, { "epoch": 0.32667525773195877, "grad_norm": 0.6643393039703369, "learning_rate": 9.997637644680395e-06, "loss": 0.6461, "step": 507 }, { "epoch": 0.327319587628866, "grad_norm": 0.8055205345153809, "learning_rate": 9.997521012147577e-06, "loss": 0.6719, "step": 508 }, { "epoch": 0.3279639175257732, "grad_norm": 0.6707114577293396, "learning_rate": 9.99740157013241e-06, "loss": 0.6511, "step": 509 }, { "epoch": 0.3286082474226804, "grad_norm": 0.6611571907997131, "learning_rate": 9.997279318702043e-06, "loss": 0.6651, "step": 510 }, { "epoch": 0.3292525773195876, "grad_norm": 0.7622340321540833, "learning_rate": 9.997154257925199e-06, "loss": 0.6689, "step": 511 }, { "epoch": 0.32989690721649484, "grad_norm": 0.6912282705307007, "learning_rate": 9.99702638787219e-06, "loss": 0.659, "step": 512 }, { "epoch": 0.33054123711340205, "grad_norm": 0.7328752875328064, "learning_rate": 9.996895708614893e-06, "loss": 0.6558, "step": 513 }, { "epoch": 0.33118556701030927, "grad_norm": 0.6681116819381714, "learning_rate": 9.996762220226779e-06, "loss": 0.6509, "step": 514 }, { "epoch": 0.3318298969072165, "grad_norm": 0.6662226915359497, "learning_rate": 9.99662592278289e-06, "loss": 0.658, "step": 515 }, { "epoch": 0.3324742268041237, "grad_norm": 0.7532008290290833, "learning_rate": 9.996486816359851e-06, "loss": 0.6656, "step": 516 }, { "epoch": 0.3331185567010309, "grad_norm": 0.6898446679115295, "learning_rate": 9.99634490103586e-06, "loss": 0.6722, "step": 517 }, { "epoch": 0.3337628865979381, "grad_norm": 0.7055858373641968, "learning_rate": 9.996200176890702e-06, "loss": 0.6443, "step": 518 }, { "epoch": 0.33440721649484534, "grad_norm": 0.698577344417572, "learning_rate": 9.996052644005736e-06, "loss": 0.6532, "step": 519 }, { "epoch": 0.33505154639175255, "grad_norm": 0.6636202335357666, "learning_rate": 9.995902302463901e-06, "loss": 0.6741, "step": 520 }, { "epoch": 0.33569587628865977, "grad_norm": 0.6271678805351257, "learning_rate": 9.995749152349715e-06, "loss": 0.66, "step": 521 }, { "epoch": 0.33634020618556704, "grad_norm": 0.5747771859169006, "learning_rate": 9.995593193749275e-06, "loss": 0.6497, "step": 522 }, { "epoch": 0.33698453608247425, "grad_norm": 0.7136269807815552, "learning_rate": 9.99543442675026e-06, "loss": 0.6854, "step": 523 }, { "epoch": 0.33762886597938147, "grad_norm": 0.7354745268821716, "learning_rate": 9.99527285144192e-06, "loss": 0.6877, "step": 524 }, { "epoch": 0.3382731958762887, "grad_norm": 0.7314122915267944, "learning_rate": 9.995108467915093e-06, "loss": 0.6709, "step": 525 }, { "epoch": 0.3389175257731959, "grad_norm": 0.6720836162567139, "learning_rate": 9.994941276262188e-06, "loss": 0.6783, "step": 526 }, { "epoch": 0.3395618556701031, "grad_norm": 0.6846986413002014, "learning_rate": 9.9947712765772e-06, "loss": 0.6847, "step": 527 }, { "epoch": 0.3402061855670103, "grad_norm": 0.7138612270355225, "learning_rate": 9.994598468955697e-06, "loss": 0.6695, "step": 528 }, { "epoch": 0.34085051546391754, "grad_norm": 0.7742570042610168, "learning_rate": 9.994422853494823e-06, "loss": 0.6689, "step": 529 }, { "epoch": 0.34149484536082475, "grad_norm": 0.7206528782844543, "learning_rate": 9.99424443029331e-06, "loss": 0.6574, "step": 530 }, { "epoch": 0.34213917525773196, "grad_norm": 0.6567821502685547, "learning_rate": 9.99406319945146e-06, "loss": 0.6723, "step": 531 }, { "epoch": 0.3427835051546392, "grad_norm": 0.6330087184906006, "learning_rate": 9.993879161071159e-06, "loss": 0.6837, "step": 532 }, { "epoch": 0.3434278350515464, "grad_norm": 0.6835585236549377, "learning_rate": 9.993692315255867e-06, "loss": 0.673, "step": 533 }, { "epoch": 0.3440721649484536, "grad_norm": 0.6567046046257019, "learning_rate": 9.993502662110626e-06, "loss": 0.659, "step": 534 }, { "epoch": 0.3447164948453608, "grad_norm": 0.7065005302429199, "learning_rate": 9.99331020174205e-06, "loss": 0.6803, "step": 535 }, { "epoch": 0.34536082474226804, "grad_norm": 0.6424482464790344, "learning_rate": 9.99311493425834e-06, "loss": 0.6657, "step": 536 }, { "epoch": 0.34600515463917525, "grad_norm": 0.6590161919593811, "learning_rate": 9.992916859769268e-06, "loss": 0.6738, "step": 537 }, { "epoch": 0.34664948453608246, "grad_norm": 0.7740004658699036, "learning_rate": 9.992715978386188e-06, "loss": 0.6429, "step": 538 }, { "epoch": 0.3472938144329897, "grad_norm": 0.7301279306411743, "learning_rate": 9.99251229022203e-06, "loss": 0.6545, "step": 539 }, { "epoch": 0.3479381443298969, "grad_norm": 0.6926845908164978, "learning_rate": 9.992305795391301e-06, "loss": 0.6518, "step": 540 }, { "epoch": 0.3485824742268041, "grad_norm": 0.6792284250259399, "learning_rate": 9.99209649401009e-06, "loss": 0.6567, "step": 541 }, { "epoch": 0.3492268041237113, "grad_norm": 0.7534132599830627, "learning_rate": 9.991884386196061e-06, "loss": 0.6695, "step": 542 }, { "epoch": 0.34987113402061853, "grad_norm": 0.7620640993118286, "learning_rate": 9.991669472068451e-06, "loss": 0.6481, "step": 543 }, { "epoch": 0.35051546391752575, "grad_norm": 0.6658647656440735, "learning_rate": 9.991451751748085e-06, "loss": 0.6768, "step": 544 }, { "epoch": 0.35115979381443296, "grad_norm": 0.7681078314781189, "learning_rate": 9.991231225357356e-06, "loss": 0.657, "step": 545 }, { "epoch": 0.35180412371134023, "grad_norm": 0.6953827142715454, "learning_rate": 9.991007893020242e-06, "loss": 0.6543, "step": 546 }, { "epoch": 0.35244845360824745, "grad_norm": 0.7611663341522217, "learning_rate": 9.990781754862294e-06, "loss": 0.682, "step": 547 }, { "epoch": 0.35309278350515466, "grad_norm": 0.6446352005004883, "learning_rate": 9.990552811010638e-06, "loss": 0.6624, "step": 548 }, { "epoch": 0.3537371134020619, "grad_norm": 0.6192432045936584, "learning_rate": 9.990321061593983e-06, "loss": 0.6442, "step": 549 }, { "epoch": 0.3543814432989691, "grad_norm": 0.7998104691505432, "learning_rate": 9.990086506742614e-06, "loss": 0.6494, "step": 550 }, { "epoch": 0.3550257731958763, "grad_norm": 0.7128908634185791, "learning_rate": 9.98984914658839e-06, "loss": 0.6557, "step": 551 }, { "epoch": 0.3556701030927835, "grad_norm": 0.6812767386436462, "learning_rate": 9.98960898126475e-06, "loss": 0.6864, "step": 552 }, { "epoch": 0.35631443298969073, "grad_norm": 0.6752882599830627, "learning_rate": 9.98936601090671e-06, "loss": 0.6924, "step": 553 }, { "epoch": 0.35695876288659795, "grad_norm": 0.6289911866188049, "learning_rate": 9.989120235650858e-06, "loss": 0.6725, "step": 554 }, { "epoch": 0.35760309278350516, "grad_norm": 0.6408771872520447, "learning_rate": 9.988871655635366e-06, "loss": 0.664, "step": 555 }, { "epoch": 0.3582474226804124, "grad_norm": 0.805759847164154, "learning_rate": 9.98862027099998e-06, "loss": 0.6632, "step": 556 }, { "epoch": 0.3588917525773196, "grad_norm": 0.8032969236373901, "learning_rate": 9.988366081886019e-06, "loss": 0.6759, "step": 557 }, { "epoch": 0.3595360824742268, "grad_norm": 0.6474846601486206, "learning_rate": 9.988109088436386e-06, "loss": 0.6715, "step": 558 }, { "epoch": 0.360180412371134, "grad_norm": 0.743305504322052, "learning_rate": 9.987849290795552e-06, "loss": 0.6465, "step": 559 }, { "epoch": 0.36082474226804123, "grad_norm": 0.6293637752532959, "learning_rate": 9.987586689109572e-06, "loss": 0.635, "step": 560 }, { "epoch": 0.36146907216494845, "grad_norm": 0.6612563133239746, "learning_rate": 9.987321283526072e-06, "loss": 0.6754, "step": 561 }, { "epoch": 0.36211340206185566, "grad_norm": 0.7013714909553528, "learning_rate": 9.98705307419426e-06, "loss": 0.6607, "step": 562 }, { "epoch": 0.3627577319587629, "grad_norm": 0.6544200778007507, "learning_rate": 9.986782061264912e-06, "loss": 0.6614, "step": 563 }, { "epoch": 0.3634020618556701, "grad_norm": 0.6963590979576111, "learning_rate": 9.986508244890388e-06, "loss": 0.6589, "step": 564 }, { "epoch": 0.3640463917525773, "grad_norm": 0.7100768685340881, "learning_rate": 9.986231625224621e-06, "loss": 0.6708, "step": 565 }, { "epoch": 0.3646907216494845, "grad_norm": 0.7091698050498962, "learning_rate": 9.985952202423116e-06, "loss": 0.6568, "step": 566 }, { "epoch": 0.36533505154639173, "grad_norm": 0.7451876997947693, "learning_rate": 9.98566997664296e-06, "loss": 0.6498, "step": 567 }, { "epoch": 0.36597938144329895, "grad_norm": 0.8014804720878601, "learning_rate": 9.985384948042814e-06, "loss": 0.6672, "step": 568 }, { "epoch": 0.36662371134020616, "grad_norm": 0.6944939494132996, "learning_rate": 9.985097116782913e-06, "loss": 0.6501, "step": 569 }, { "epoch": 0.36726804123711343, "grad_norm": 0.7749470472335815, "learning_rate": 9.984806483025071e-06, "loss": 0.6306, "step": 570 }, { "epoch": 0.36791237113402064, "grad_norm": 0.7899306416511536, "learning_rate": 9.98451304693267e-06, "loss": 0.6536, "step": 571 }, { "epoch": 0.36855670103092786, "grad_norm": 0.7376818656921387, "learning_rate": 9.984216808670676e-06, "loss": 0.6738, "step": 572 }, { "epoch": 0.36920103092783507, "grad_norm": 0.8312161564826965, "learning_rate": 9.98391776840563e-06, "loss": 0.6437, "step": 573 }, { "epoch": 0.3698453608247423, "grad_norm": 0.6843551993370056, "learning_rate": 9.983615926305636e-06, "loss": 0.6535, "step": 574 }, { "epoch": 0.3704896907216495, "grad_norm": 0.7812162041664124, "learning_rate": 9.98331128254039e-06, "loss": 0.6502, "step": 575 }, { "epoch": 0.3711340206185567, "grad_norm": 0.7541844248771667, "learning_rate": 9.983003837281152e-06, "loss": 0.657, "step": 576 }, { "epoch": 0.37177835051546393, "grad_norm": 0.7040230631828308, "learning_rate": 9.982693590700763e-06, "loss": 0.6713, "step": 577 }, { "epoch": 0.37242268041237114, "grad_norm": 0.7184350490570068, "learning_rate": 9.982380542973632e-06, "loss": 0.6476, "step": 578 }, { "epoch": 0.37306701030927836, "grad_norm": 0.6380530595779419, "learning_rate": 9.982064694275748e-06, "loss": 0.6609, "step": 579 }, { "epoch": 0.37371134020618557, "grad_norm": 0.8075182437896729, "learning_rate": 9.981746044784674e-06, "loss": 0.6521, "step": 580 }, { "epoch": 0.3743556701030928, "grad_norm": 0.6488863825798035, "learning_rate": 9.981424594679545e-06, "loss": 0.6591, "step": 581 }, { "epoch": 0.375, "grad_norm": 0.6529377102851868, "learning_rate": 9.981100344141075e-06, "loss": 0.6568, "step": 582 }, { "epoch": 0.3756443298969072, "grad_norm": 0.7539911866188049, "learning_rate": 9.980773293351549e-06, "loss": 0.6621, "step": 583 }, { "epoch": 0.37628865979381443, "grad_norm": 0.7068541646003723, "learning_rate": 9.980443442494826e-06, "loss": 0.6549, "step": 584 }, { "epoch": 0.37693298969072164, "grad_norm": 0.6766762733459473, "learning_rate": 9.98011079175634e-06, "loss": 0.6936, "step": 585 }, { "epoch": 0.37757731958762886, "grad_norm": 0.7405463457107544, "learning_rate": 9.979775341323097e-06, "loss": 0.665, "step": 586 }, { "epoch": 0.37822164948453607, "grad_norm": 0.7993252277374268, "learning_rate": 9.979437091383684e-06, "loss": 0.6495, "step": 587 }, { "epoch": 0.3788659793814433, "grad_norm": 0.7793001532554626, "learning_rate": 9.979096042128254e-06, "loss": 0.6475, "step": 588 }, { "epoch": 0.3795103092783505, "grad_norm": 0.7632677555084229, "learning_rate": 9.978752193748533e-06, "loss": 0.6527, "step": 589 }, { "epoch": 0.3801546391752577, "grad_norm": 0.7380077242851257, "learning_rate": 9.97840554643783e-06, "loss": 0.6442, "step": 590 }, { "epoch": 0.38079896907216493, "grad_norm": 0.7229974865913391, "learning_rate": 9.978056100391017e-06, "loss": 0.6417, "step": 591 }, { "epoch": 0.38144329896907214, "grad_norm": 0.7002800107002258, "learning_rate": 9.977703855804549e-06, "loss": 0.6726, "step": 592 }, { "epoch": 0.38208762886597936, "grad_norm": 0.6747826337814331, "learning_rate": 9.977348812876442e-06, "loss": 0.6638, "step": 593 }, { "epoch": 0.38273195876288657, "grad_norm": 0.700654149055481, "learning_rate": 9.976990971806298e-06, "loss": 0.6438, "step": 594 }, { "epoch": 0.38337628865979384, "grad_norm": 0.6943671703338623, "learning_rate": 9.976630332795283e-06, "loss": 0.6728, "step": 595 }, { "epoch": 0.38402061855670105, "grad_norm": 0.7686260938644409, "learning_rate": 9.976266896046143e-06, "loss": 0.6547, "step": 596 }, { "epoch": 0.38466494845360827, "grad_norm": 0.8210086226463318, "learning_rate": 9.97590066176319e-06, "loss": 0.6453, "step": 597 }, { "epoch": 0.3853092783505155, "grad_norm": 0.6884353160858154, "learning_rate": 9.975531630152314e-06, "loss": 0.6436, "step": 598 }, { "epoch": 0.3859536082474227, "grad_norm": 1.0077779293060303, "learning_rate": 9.975159801420972e-06, "loss": 0.6586, "step": 599 }, { "epoch": 0.3865979381443299, "grad_norm": 0.72463059425354, "learning_rate": 9.9747851757782e-06, "loss": 0.6496, "step": 600 }, { "epoch": 0.3872422680412371, "grad_norm": 0.8664722442626953, "learning_rate": 9.974407753434604e-06, "loss": 0.6658, "step": 601 }, { "epoch": 0.38788659793814434, "grad_norm": 0.6920607089996338, "learning_rate": 9.974027534602358e-06, "loss": 0.6634, "step": 602 }, { "epoch": 0.38853092783505155, "grad_norm": 0.8034618496894836, "learning_rate": 9.973644519495213e-06, "loss": 0.6931, "step": 603 }, { "epoch": 0.38917525773195877, "grad_norm": 0.7935554385185242, "learning_rate": 9.973258708328492e-06, "loss": 0.6813, "step": 604 }, { "epoch": 0.389819587628866, "grad_norm": 0.8173457980155945, "learning_rate": 9.972870101319086e-06, "loss": 0.669, "step": 605 }, { "epoch": 0.3904639175257732, "grad_norm": 0.8653724789619446, "learning_rate": 9.972478698685463e-06, "loss": 0.6934, "step": 606 }, { "epoch": 0.3911082474226804, "grad_norm": 0.666444718837738, "learning_rate": 9.972084500647658e-06, "loss": 0.6398, "step": 607 }, { "epoch": 0.3917525773195876, "grad_norm": 0.6567226648330688, "learning_rate": 9.971687507427282e-06, "loss": 0.6606, "step": 608 }, { "epoch": 0.39239690721649484, "grad_norm": 0.9212610125541687, "learning_rate": 9.97128771924751e-06, "loss": 0.6425, "step": 609 }, { "epoch": 0.39304123711340205, "grad_norm": 0.8650708794593811, "learning_rate": 9.970885136333097e-06, "loss": 0.6831, "step": 610 }, { "epoch": 0.39368556701030927, "grad_norm": 0.702287495136261, "learning_rate": 9.970479758910365e-06, "loss": 0.6586, "step": 611 }, { "epoch": 0.3943298969072165, "grad_norm": 1.0810761451721191, "learning_rate": 9.970071587207204e-06, "loss": 0.682, "step": 612 }, { "epoch": 0.3949742268041237, "grad_norm": 0.8088216781616211, "learning_rate": 9.96966062145308e-06, "loss": 0.6625, "step": 613 }, { "epoch": 0.3956185567010309, "grad_norm": 0.8049780130386353, "learning_rate": 9.969246861879028e-06, "loss": 0.6419, "step": 614 }, { "epoch": 0.3962628865979381, "grad_norm": 0.7979827523231506, "learning_rate": 9.968830308717653e-06, "loss": 0.6632, "step": 615 }, { "epoch": 0.39690721649484534, "grad_norm": 0.6848436594009399, "learning_rate": 9.968410962203131e-06, "loss": 0.6727, "step": 616 }, { "epoch": 0.39755154639175255, "grad_norm": 0.8701410293579102, "learning_rate": 9.967988822571209e-06, "loss": 0.6573, "step": 617 }, { "epoch": 0.39819587628865977, "grad_norm": 0.7648741006851196, "learning_rate": 9.9675638900592e-06, "loss": 0.6642, "step": 618 }, { "epoch": 0.39884020618556704, "grad_norm": 0.8561636209487915, "learning_rate": 9.967136164905995e-06, "loss": 0.6607, "step": 619 }, { "epoch": 0.39948453608247425, "grad_norm": 0.7378113865852356, "learning_rate": 9.966705647352048e-06, "loss": 0.6644, "step": 620 }, { "epoch": 0.40012886597938147, "grad_norm": 0.7523707151412964, "learning_rate": 9.966272337639386e-06, "loss": 0.6535, "step": 621 }, { "epoch": 0.4007731958762887, "grad_norm": 0.7643794417381287, "learning_rate": 9.965836236011606e-06, "loss": 0.6567, "step": 622 }, { "epoch": 0.4014175257731959, "grad_norm": 0.7148403525352478, "learning_rate": 9.965397342713872e-06, "loss": 0.6552, "step": 623 }, { "epoch": 0.4020618556701031, "grad_norm": 0.7030014395713806, "learning_rate": 9.964955657992919e-06, "loss": 0.6682, "step": 624 }, { "epoch": 0.4027061855670103, "grad_norm": 0.7459947466850281, "learning_rate": 9.96451118209705e-06, "loss": 0.6664, "step": 625 }, { "epoch": 0.40335051546391754, "grad_norm": 0.6205683350563049, "learning_rate": 9.964063915276141e-06, "loss": 0.6552, "step": 626 }, { "epoch": 0.40399484536082475, "grad_norm": 0.7183192372322083, "learning_rate": 9.963613857781632e-06, "loss": 0.6539, "step": 627 }, { "epoch": 0.40463917525773196, "grad_norm": 0.6289457082748413, "learning_rate": 9.963161009866536e-06, "loss": 0.6437, "step": 628 }, { "epoch": 0.4052835051546392, "grad_norm": 0.6621333360671997, "learning_rate": 9.962705371785431e-06, "loss": 0.6626, "step": 629 }, { "epoch": 0.4059278350515464, "grad_norm": 0.6919174790382385, "learning_rate": 9.96224694379447e-06, "loss": 0.6636, "step": 630 }, { "epoch": 0.4065721649484536, "grad_norm": 0.6646978259086609, "learning_rate": 9.961785726151363e-06, "loss": 0.658, "step": 631 }, { "epoch": 0.4072164948453608, "grad_norm": 0.6607853770256042, "learning_rate": 9.961321719115398e-06, "loss": 0.6649, "step": 632 }, { "epoch": 0.40786082474226804, "grad_norm": 0.6004449725151062, "learning_rate": 9.96085492294743e-06, "loss": 0.6538, "step": 633 }, { "epoch": 0.40850515463917525, "grad_norm": 0.7338732481002808, "learning_rate": 9.960385337909879e-06, "loss": 0.658, "step": 634 }, { "epoch": 0.40914948453608246, "grad_norm": 0.6983126997947693, "learning_rate": 9.959912964266734e-06, "loss": 0.6488, "step": 635 }, { "epoch": 0.4097938144329897, "grad_norm": 0.6953639984130859, "learning_rate": 9.959437802283552e-06, "loss": 0.648, "step": 636 }, { "epoch": 0.4104381443298969, "grad_norm": 0.6399908661842346, "learning_rate": 9.958959852227458e-06, "loss": 0.673, "step": 637 }, { "epoch": 0.4110824742268041, "grad_norm": 0.6631487011909485, "learning_rate": 9.958479114367141e-06, "loss": 0.6564, "step": 638 }, { "epoch": 0.4117268041237113, "grad_norm": 0.6557109355926514, "learning_rate": 9.957995588972862e-06, "loss": 0.6487, "step": 639 }, { "epoch": 0.41237113402061853, "grad_norm": 0.6144114136695862, "learning_rate": 9.957509276316447e-06, "loss": 0.6561, "step": 640 }, { "epoch": 0.41301546391752575, "grad_norm": 0.6313749551773071, "learning_rate": 9.957020176671289e-06, "loss": 0.6666, "step": 641 }, { "epoch": 0.41365979381443296, "grad_norm": 0.5763276815414429, "learning_rate": 9.956528290312345e-06, "loss": 0.663, "step": 642 }, { "epoch": 0.41430412371134023, "grad_norm": 0.7578215599060059, "learning_rate": 9.956033617516146e-06, "loss": 0.6401, "step": 643 }, { "epoch": 0.41494845360824745, "grad_norm": 0.6944020986557007, "learning_rate": 9.955536158560781e-06, "loss": 0.6628, "step": 644 }, { "epoch": 0.41559278350515466, "grad_norm": 0.6362676620483398, "learning_rate": 9.95503591372591e-06, "loss": 0.6516, "step": 645 }, { "epoch": 0.4162371134020619, "grad_norm": 0.6670147180557251, "learning_rate": 9.954532883292761e-06, "loss": 0.6647, "step": 646 }, { "epoch": 0.4168814432989691, "grad_norm": 0.6974989771842957, "learning_rate": 9.95402706754412e-06, "loss": 0.6533, "step": 647 }, { "epoch": 0.4175257731958763, "grad_norm": 0.6202733516693115, "learning_rate": 9.953518466764348e-06, "loss": 0.654, "step": 648 }, { "epoch": 0.4181701030927835, "grad_norm": 0.7575302124023438, "learning_rate": 9.953007081239366e-06, "loss": 0.6644, "step": 649 }, { "epoch": 0.41881443298969073, "grad_norm": 0.5805395841598511, "learning_rate": 9.952492911256664e-06, "loss": 0.6509, "step": 650 }, { "epoch": 0.41945876288659795, "grad_norm": 0.6375502347946167, "learning_rate": 9.951975957105293e-06, "loss": 0.6515, "step": 651 }, { "epoch": 0.42010309278350516, "grad_norm": 0.6652111411094666, "learning_rate": 9.951456219075874e-06, "loss": 0.6534, "step": 652 }, { "epoch": 0.4207474226804124, "grad_norm": 0.6118087768554688, "learning_rate": 9.950933697460592e-06, "loss": 0.6669, "step": 653 }, { "epoch": 0.4213917525773196, "grad_norm": 0.6529296636581421, "learning_rate": 9.950408392553192e-06, "loss": 0.6393, "step": 654 }, { "epoch": 0.4220360824742268, "grad_norm": 0.6510635614395142, "learning_rate": 9.949880304648989e-06, "loss": 0.6662, "step": 655 }, { "epoch": 0.422680412371134, "grad_norm": 0.617573082447052, "learning_rate": 9.949349434044862e-06, "loss": 0.6565, "step": 656 }, { "epoch": 0.42332474226804123, "grad_norm": 0.6556816101074219, "learning_rate": 9.948815781039251e-06, "loss": 0.6266, "step": 657 }, { "epoch": 0.42396907216494845, "grad_norm": 0.6485823392868042, "learning_rate": 9.948279345932164e-06, "loss": 0.6718, "step": 658 }, { "epoch": 0.42461340206185566, "grad_norm": 0.7297622561454773, "learning_rate": 9.947740129025174e-06, "loss": 0.6521, "step": 659 }, { "epoch": 0.4252577319587629, "grad_norm": 0.6451907157897949, "learning_rate": 9.94719813062141e-06, "loss": 0.6706, "step": 660 }, { "epoch": 0.4259020618556701, "grad_norm": 0.7467945218086243, "learning_rate": 9.946653351025575e-06, "loss": 0.6503, "step": 661 }, { "epoch": 0.4265463917525773, "grad_norm": 0.734566330909729, "learning_rate": 9.946105790543927e-06, "loss": 0.6473, "step": 662 }, { "epoch": 0.4271907216494845, "grad_norm": 0.6143317222595215, "learning_rate": 9.945555449484292e-06, "loss": 0.6715, "step": 663 }, { "epoch": 0.42783505154639173, "grad_norm": 0.7910267114639282, "learning_rate": 9.945002328156059e-06, "loss": 0.6561, "step": 664 }, { "epoch": 0.42847938144329895, "grad_norm": 0.7222806215286255, "learning_rate": 9.944446426870179e-06, "loss": 0.6416, "step": 665 }, { "epoch": 0.42912371134020616, "grad_norm": 0.6078263521194458, "learning_rate": 9.943887745939164e-06, "loss": 0.6807, "step": 666 }, { "epoch": 0.42976804123711343, "grad_norm": 0.7931578755378723, "learning_rate": 9.943326285677092e-06, "loss": 0.6429, "step": 667 }, { "epoch": 0.43041237113402064, "grad_norm": 0.716678261756897, "learning_rate": 9.942762046399606e-06, "loss": 0.6588, "step": 668 }, { "epoch": 0.43105670103092786, "grad_norm": 0.7097393274307251, "learning_rate": 9.942195028423899e-06, "loss": 0.6495, "step": 669 }, { "epoch": 0.43170103092783507, "grad_norm": 0.7564566135406494, "learning_rate": 9.94162523206874e-06, "loss": 0.6583, "step": 670 }, { "epoch": 0.4323453608247423, "grad_norm": 0.7557837963104248, "learning_rate": 9.941052657654453e-06, "loss": 0.6578, "step": 671 }, { "epoch": 0.4329896907216495, "grad_norm": 0.8230401873588562, "learning_rate": 9.940477305502927e-06, "loss": 0.6611, "step": 672 }, { "epoch": 0.4336340206185567, "grad_norm": 0.849672794342041, "learning_rate": 9.939899175937607e-06, "loss": 0.6207, "step": 673 }, { "epoch": 0.43427835051546393, "grad_norm": 0.6921935081481934, "learning_rate": 9.939318269283506e-06, "loss": 0.6584, "step": 674 }, { "epoch": 0.43492268041237114, "grad_norm": 0.8069269061088562, "learning_rate": 9.938734585867194e-06, "loss": 0.6563, "step": 675 }, { "epoch": 0.43556701030927836, "grad_norm": 0.7240656018257141, "learning_rate": 9.938148126016805e-06, "loss": 0.6688, "step": 676 }, { "epoch": 0.43621134020618557, "grad_norm": 0.6675538420677185, "learning_rate": 9.93755889006203e-06, "loss": 0.6546, "step": 677 }, { "epoch": 0.4368556701030928, "grad_norm": 0.7601011395454407, "learning_rate": 9.936966878334122e-06, "loss": 0.6476, "step": 678 }, { "epoch": 0.4375, "grad_norm": 0.6560327410697937, "learning_rate": 9.9363720911659e-06, "loss": 0.6357, "step": 679 }, { "epoch": 0.4381443298969072, "grad_norm": 0.6408637762069702, "learning_rate": 9.935774528891732e-06, "loss": 0.635, "step": 680 }, { "epoch": 0.43878865979381443, "grad_norm": 0.7452948093414307, "learning_rate": 9.93517419184756e-06, "loss": 0.6566, "step": 681 }, { "epoch": 0.43943298969072164, "grad_norm": 0.6937955021858215, "learning_rate": 9.934571080370874e-06, "loss": 0.6513, "step": 682 }, { "epoch": 0.44007731958762886, "grad_norm": 0.7588962912559509, "learning_rate": 9.93396519480073e-06, "loss": 0.6768, "step": 683 }, { "epoch": 0.44072164948453607, "grad_norm": 0.7010397911071777, "learning_rate": 9.93335653547774e-06, "loss": 0.6663, "step": 684 }, { "epoch": 0.4413659793814433, "grad_norm": 0.6582136750221252, "learning_rate": 9.932745102744079e-06, "loss": 0.6319, "step": 685 }, { "epoch": 0.4420103092783505, "grad_norm": 0.6827657222747803, "learning_rate": 9.932130896943477e-06, "loss": 0.655, "step": 686 }, { "epoch": 0.4426546391752577, "grad_norm": 0.7098948359489441, "learning_rate": 9.931513918421227e-06, "loss": 0.6654, "step": 687 }, { "epoch": 0.44329896907216493, "grad_norm": 0.6884098649024963, "learning_rate": 9.930894167524181e-06, "loss": 0.6536, "step": 688 }, { "epoch": 0.44394329896907214, "grad_norm": 0.8623906970024109, "learning_rate": 9.930271644600744e-06, "loss": 0.6573, "step": 689 }, { "epoch": 0.44458762886597936, "grad_norm": 0.7174708247184753, "learning_rate": 9.929646350000885e-06, "loss": 0.6737, "step": 690 }, { "epoch": 0.44523195876288657, "grad_norm": 0.6763615012168884, "learning_rate": 9.929018284076128e-06, "loss": 0.6369, "step": 691 }, { "epoch": 0.44587628865979384, "grad_norm": 0.7395650148391724, "learning_rate": 9.928387447179558e-06, "loss": 0.6779, "step": 692 }, { "epoch": 0.44652061855670105, "grad_norm": 0.6918244361877441, "learning_rate": 9.927753839665812e-06, "loss": 0.6331, "step": 693 }, { "epoch": 0.44716494845360827, "grad_norm": 0.6395847797393799, "learning_rate": 9.927117461891093e-06, "loss": 0.6619, "step": 694 }, { "epoch": 0.4478092783505155, "grad_norm": 0.6821173429489136, "learning_rate": 9.926478314213152e-06, "loss": 0.6664, "step": 695 }, { "epoch": 0.4484536082474227, "grad_norm": 0.714694619178772, "learning_rate": 9.925836396991309e-06, "loss": 0.6437, "step": 696 }, { "epoch": 0.4490979381443299, "grad_norm": 0.660469114780426, "learning_rate": 9.925191710586426e-06, "loss": 0.6432, "step": 697 }, { "epoch": 0.4497422680412371, "grad_norm": 0.6885974407196045, "learning_rate": 9.924544255360935e-06, "loss": 0.6525, "step": 698 }, { "epoch": 0.45038659793814434, "grad_norm": 0.7293381690979004, "learning_rate": 9.923894031678817e-06, "loss": 0.6522, "step": 699 }, { "epoch": 0.45103092783505155, "grad_norm": 0.6417496204376221, "learning_rate": 9.923241039905612e-06, "loss": 0.6387, "step": 700 }, { "epoch": 0.45167525773195877, "grad_norm": 0.644286572933197, "learning_rate": 9.922585280408417e-06, "loss": 0.6661, "step": 701 }, { "epoch": 0.452319587628866, "grad_norm": 0.8172661662101746, "learning_rate": 9.921926753555882e-06, "loss": 0.6416, "step": 702 }, { "epoch": 0.4529639175257732, "grad_norm": 0.7171735763549805, "learning_rate": 9.921265459718215e-06, "loss": 0.6516, "step": 703 }, { "epoch": 0.4536082474226804, "grad_norm": 0.6718586683273315, "learning_rate": 9.92060139926718e-06, "loss": 0.6708, "step": 704 }, { "epoch": 0.4542525773195876, "grad_norm": 0.9136710166931152, "learning_rate": 9.919934572576093e-06, "loss": 0.6662, "step": 705 }, { "epoch": 0.45489690721649484, "grad_norm": 0.6810513734817505, "learning_rate": 9.919264980019829e-06, "loss": 0.6406, "step": 706 }, { "epoch": 0.45554123711340205, "grad_norm": 0.8355632424354553, "learning_rate": 9.918592621974818e-06, "loss": 0.6315, "step": 707 }, { "epoch": 0.45618556701030927, "grad_norm": 0.7449923753738403, "learning_rate": 9.917917498819038e-06, "loss": 0.6569, "step": 708 }, { "epoch": 0.4568298969072165, "grad_norm": 0.6943632364273071, "learning_rate": 9.917239610932033e-06, "loss": 0.6517, "step": 709 }, { "epoch": 0.4574742268041237, "grad_norm": 0.8251372575759888, "learning_rate": 9.916558958694891e-06, "loss": 0.6781, "step": 710 }, { "epoch": 0.4581185567010309, "grad_norm": 0.6846953630447388, "learning_rate": 9.915875542490257e-06, "loss": 0.6485, "step": 711 }, { "epoch": 0.4587628865979381, "grad_norm": 0.6775670647621155, "learning_rate": 9.915189362702333e-06, "loss": 0.6592, "step": 712 }, { "epoch": 0.45940721649484534, "grad_norm": 0.686984121799469, "learning_rate": 9.91450041971687e-06, "loss": 0.6628, "step": 713 }, { "epoch": 0.46005154639175255, "grad_norm": 0.7133525609970093, "learning_rate": 9.913808713921178e-06, "loss": 0.6421, "step": 714 }, { "epoch": 0.46069587628865977, "grad_norm": 0.6325277090072632, "learning_rate": 9.913114245704112e-06, "loss": 0.6472, "step": 715 }, { "epoch": 0.46134020618556704, "grad_norm": 0.6093975305557251, "learning_rate": 9.912417015456088e-06, "loss": 0.6347, "step": 716 }, { "epoch": 0.46198453608247425, "grad_norm": 0.6919620633125305, "learning_rate": 9.911717023569073e-06, "loss": 0.6329, "step": 717 }, { "epoch": 0.46262886597938147, "grad_norm": 0.6013222336769104, "learning_rate": 9.911014270436582e-06, "loss": 0.6449, "step": 718 }, { "epoch": 0.4632731958762887, "grad_norm": 0.6290733218193054, "learning_rate": 9.910308756453686e-06, "loss": 0.6474, "step": 719 }, { "epoch": 0.4639175257731959, "grad_norm": 0.6419352293014526, "learning_rate": 9.909600482017008e-06, "loss": 0.6559, "step": 720 }, { "epoch": 0.4645618556701031, "grad_norm": 0.6451571583747864, "learning_rate": 9.908889447524724e-06, "loss": 0.6435, "step": 721 }, { "epoch": 0.4652061855670103, "grad_norm": 0.6828376054763794, "learning_rate": 9.908175653376556e-06, "loss": 0.6585, "step": 722 }, { "epoch": 0.46585051546391754, "grad_norm": 0.6695808172225952, "learning_rate": 9.907459099973783e-06, "loss": 0.6451, "step": 723 }, { "epoch": 0.46649484536082475, "grad_norm": 0.5593898296356201, "learning_rate": 9.906739787719238e-06, "loss": 0.664, "step": 724 }, { "epoch": 0.46713917525773196, "grad_norm": 0.6320433020591736, "learning_rate": 9.906017717017292e-06, "loss": 0.6652, "step": 725 }, { "epoch": 0.4677835051546392, "grad_norm": 0.633308470249176, "learning_rate": 9.905292888273883e-06, "loss": 0.6384, "step": 726 }, { "epoch": 0.4684278350515464, "grad_norm": 0.7001979947090149, "learning_rate": 9.904565301896487e-06, "loss": 0.6667, "step": 727 }, { "epoch": 0.4690721649484536, "grad_norm": 0.7869543433189392, "learning_rate": 9.903834958294138e-06, "loss": 0.6392, "step": 728 }, { "epoch": 0.4697164948453608, "grad_norm": 0.6712290048599243, "learning_rate": 9.903101857877415e-06, "loss": 0.6515, "step": 729 }, { "epoch": 0.47036082474226804, "grad_norm": 0.7499169707298279, "learning_rate": 9.90236600105845e-06, "loss": 0.6302, "step": 730 }, { "epoch": 0.47100515463917525, "grad_norm": 0.7289696335792542, "learning_rate": 9.901627388250923e-06, "loss": 0.658, "step": 731 }, { "epoch": 0.47164948453608246, "grad_norm": 0.7225127816200256, "learning_rate": 9.900886019870063e-06, "loss": 0.6514, "step": 732 }, { "epoch": 0.4722938144329897, "grad_norm": 0.6776766777038574, "learning_rate": 9.90014189633265e-06, "loss": 0.6482, "step": 733 }, { "epoch": 0.4729381443298969, "grad_norm": 0.7037966847419739, "learning_rate": 9.899395018057011e-06, "loss": 0.6696, "step": 734 }, { "epoch": 0.4735824742268041, "grad_norm": 0.6831173300743103, "learning_rate": 9.898645385463024e-06, "loss": 0.6787, "step": 735 }, { "epoch": 0.4742268041237113, "grad_norm": 0.6221543550491333, "learning_rate": 9.897892998972113e-06, "loss": 0.6539, "step": 736 }, { "epoch": 0.47487113402061853, "grad_norm": 0.6944236755371094, "learning_rate": 9.89713785900725e-06, "loss": 0.6355, "step": 737 }, { "epoch": 0.47551546391752575, "grad_norm": 0.6441633105278015, "learning_rate": 9.896379965992958e-06, "loss": 0.6507, "step": 738 }, { "epoch": 0.47615979381443296, "grad_norm": 0.7550521492958069, "learning_rate": 9.895619320355304e-06, "loss": 0.6532, "step": 739 }, { "epoch": 0.47680412371134023, "grad_norm": 0.5728148221969604, "learning_rate": 9.894855922521903e-06, "loss": 0.6256, "step": 740 }, { "epoch": 0.47744845360824745, "grad_norm": 0.6324602365493774, "learning_rate": 9.89408977292192e-06, "loss": 0.6452, "step": 741 }, { "epoch": 0.47809278350515466, "grad_norm": 0.6948319673538208, "learning_rate": 9.893320871986067e-06, "loss": 0.6836, "step": 742 }, { "epoch": 0.4787371134020619, "grad_norm": 0.6458345055580139, "learning_rate": 9.892549220146599e-06, "loss": 0.661, "step": 743 }, { "epoch": 0.4793814432989691, "grad_norm": 0.6058452129364014, "learning_rate": 9.891774817837319e-06, "loss": 0.6612, "step": 744 }, { "epoch": 0.4800257731958763, "grad_norm": 0.6588382720947266, "learning_rate": 9.890997665493577e-06, "loss": 0.6584, "step": 745 }, { "epoch": 0.4806701030927835, "grad_norm": 0.7011035680770874, "learning_rate": 9.89021776355227e-06, "loss": 0.6506, "step": 746 }, { "epoch": 0.48131443298969073, "grad_norm": 0.7796095013618469, "learning_rate": 9.889435112451838e-06, "loss": 0.6382, "step": 747 }, { "epoch": 0.48195876288659795, "grad_norm": 0.6670638918876648, "learning_rate": 9.88864971263227e-06, "loss": 0.6455, "step": 748 }, { "epoch": 0.48260309278350516, "grad_norm": 0.8352872729301453, "learning_rate": 9.887861564535097e-06, "loss": 0.6556, "step": 749 }, { "epoch": 0.4832474226804124, "grad_norm": 0.7859240770339966, "learning_rate": 9.887070668603395e-06, "loss": 0.6616, "step": 750 }, { "epoch": 0.4838917525773196, "grad_norm": 0.6562676429748535, "learning_rate": 9.88627702528179e-06, "loss": 0.6292, "step": 751 }, { "epoch": 0.4845360824742268, "grad_norm": 0.7065544128417969, "learning_rate": 9.885480635016444e-06, "loss": 0.6504, "step": 752 }, { "epoch": 0.485180412371134, "grad_norm": 0.8521860241889954, "learning_rate": 9.884681498255071e-06, "loss": 0.621, "step": 753 }, { "epoch": 0.48582474226804123, "grad_norm": 0.732755184173584, "learning_rate": 9.883879615446924e-06, "loss": 0.6494, "step": 754 }, { "epoch": 0.48646907216494845, "grad_norm": 0.8244949579238892, "learning_rate": 9.883074987042803e-06, "loss": 0.6464, "step": 755 }, { "epoch": 0.48711340206185566, "grad_norm": 0.7181119322776794, "learning_rate": 9.882267613495049e-06, "loss": 0.6625, "step": 756 }, { "epoch": 0.4877577319587629, "grad_norm": 0.7305814623832703, "learning_rate": 9.88145749525755e-06, "loss": 0.6479, "step": 757 }, { "epoch": 0.4884020618556701, "grad_norm": 0.8221374750137329, "learning_rate": 9.88064463278573e-06, "loss": 0.6594, "step": 758 }, { "epoch": 0.4890463917525773, "grad_norm": 0.6673597693443298, "learning_rate": 9.879829026536562e-06, "loss": 0.6536, "step": 759 }, { "epoch": 0.4896907216494845, "grad_norm": 0.7194707989692688, "learning_rate": 9.879010676968563e-06, "loss": 0.6153, "step": 760 }, { "epoch": 0.49033505154639173, "grad_norm": 0.75184166431427, "learning_rate": 9.878189584541783e-06, "loss": 0.6663, "step": 761 }, { "epoch": 0.49097938144329895, "grad_norm": 0.8279308080673218, "learning_rate": 9.877365749717824e-06, "loss": 0.6377, "step": 762 }, { "epoch": 0.49162371134020616, "grad_norm": 0.6935127377510071, "learning_rate": 9.876539172959824e-06, "loss": 0.6404, "step": 763 }, { "epoch": 0.49226804123711343, "grad_norm": 0.7935934662818909, "learning_rate": 9.875709854732466e-06, "loss": 0.6486, "step": 764 }, { "epoch": 0.49291237113402064, "grad_norm": 0.7463295459747314, "learning_rate": 9.874877795501966e-06, "loss": 0.6475, "step": 765 }, { "epoch": 0.49355670103092786, "grad_norm": 0.6833794713020325, "learning_rate": 9.874042995736095e-06, "loss": 0.664, "step": 766 }, { "epoch": 0.49420103092783507, "grad_norm": 0.7334388494491577, "learning_rate": 9.873205455904152e-06, "loss": 0.6428, "step": 767 }, { "epoch": 0.4948453608247423, "grad_norm": 0.7040857672691345, "learning_rate": 9.872365176476982e-06, "loss": 0.6495, "step": 768 }, { "epoch": 0.4954896907216495, "grad_norm": 0.6679001450538635, "learning_rate": 9.87152215792697e-06, "loss": 0.6453, "step": 769 }, { "epoch": 0.4961340206185567, "grad_norm": 0.7275574207305908, "learning_rate": 9.87067640072804e-06, "loss": 0.6582, "step": 770 }, { "epoch": 0.49677835051546393, "grad_norm": 0.657684862613678, "learning_rate": 9.869827905355654e-06, "loss": 0.6492, "step": 771 }, { "epoch": 0.49742268041237114, "grad_norm": 0.7498141527175903, "learning_rate": 9.868976672286815e-06, "loss": 0.6509, "step": 772 }, { "epoch": 0.49806701030927836, "grad_norm": 0.7310913801193237, "learning_rate": 9.868122702000068e-06, "loss": 0.6644, "step": 773 }, { "epoch": 0.49871134020618557, "grad_norm": 0.6822633743286133, "learning_rate": 9.867265994975492e-06, "loss": 0.64, "step": 774 }, { "epoch": 0.4993556701030928, "grad_norm": 0.8016406893730164, "learning_rate": 9.866406551694707e-06, "loss": 0.6528, "step": 775 }, { "epoch": 0.5, "grad_norm": 0.8265242576599121, "learning_rate": 9.865544372640872e-06, "loss": 0.6497, "step": 776 }, { "epoch": 0.5006443298969072, "grad_norm": 0.8079116344451904, "learning_rate": 9.864679458298679e-06, "loss": 0.6599, "step": 777 }, { "epoch": 0.5012886597938144, "grad_norm": 0.6394626498222351, "learning_rate": 9.863811809154364e-06, "loss": 0.6265, "step": 778 }, { "epoch": 0.5019329896907216, "grad_norm": 0.7904717922210693, "learning_rate": 9.862941425695699e-06, "loss": 0.6513, "step": 779 }, { "epoch": 0.5025773195876289, "grad_norm": 0.7867199182510376, "learning_rate": 9.862068308411986e-06, "loss": 0.6471, "step": 780 }, { "epoch": 0.5032216494845361, "grad_norm": 0.5558522939682007, "learning_rate": 9.861192457794077e-06, "loss": 0.6364, "step": 781 }, { "epoch": 0.5038659793814433, "grad_norm": 0.7100971341133118, "learning_rate": 9.860313874334352e-06, "loss": 0.6511, "step": 782 }, { "epoch": 0.5045103092783505, "grad_norm": 0.6625097990036011, "learning_rate": 9.859432558526727e-06, "loss": 0.6533, "step": 783 }, { "epoch": 0.5051546391752577, "grad_norm": 0.6785762310028076, "learning_rate": 9.858548510866655e-06, "loss": 0.6688, "step": 784 }, { "epoch": 0.5057989690721649, "grad_norm": 0.8249129056930542, "learning_rate": 9.857661731851129e-06, "loss": 0.6499, "step": 785 }, { "epoch": 0.5064432989690721, "grad_norm": 0.5959396362304688, "learning_rate": 9.85677222197867e-06, "loss": 0.6294, "step": 786 }, { "epoch": 0.5070876288659794, "grad_norm": 0.6872497797012329, "learning_rate": 9.855879981749344e-06, "loss": 0.6261, "step": 787 }, { "epoch": 0.5077319587628866, "grad_norm": 0.6959662437438965, "learning_rate": 9.85498501166474e-06, "loss": 0.6536, "step": 788 }, { "epoch": 0.5083762886597938, "grad_norm": 0.6125643849372864, "learning_rate": 9.854087312227992e-06, "loss": 0.6662, "step": 789 }, { "epoch": 0.509020618556701, "grad_norm": 0.6781906485557556, "learning_rate": 9.853186883943762e-06, "loss": 0.6269, "step": 790 }, { "epoch": 0.5096649484536082, "grad_norm": 0.7217610478401184, "learning_rate": 9.852283727318251e-06, "loss": 0.6434, "step": 791 }, { "epoch": 0.5103092783505154, "grad_norm": 0.6464298963546753, "learning_rate": 9.851377842859191e-06, "loss": 0.6415, "step": 792 }, { "epoch": 0.5109536082474226, "grad_norm": 0.6849439740180969, "learning_rate": 9.850469231075845e-06, "loss": 0.6342, "step": 793 }, { "epoch": 0.5115979381443299, "grad_norm": 0.6877132058143616, "learning_rate": 9.849557892479015e-06, "loss": 0.6612, "step": 794 }, { "epoch": 0.5122422680412371, "grad_norm": 0.6660096049308777, "learning_rate": 9.84864382758103e-06, "loss": 0.66, "step": 795 }, { "epoch": 0.5128865979381443, "grad_norm": 0.7804885506629944, "learning_rate": 9.847727036895759e-06, "loss": 0.664, "step": 796 }, { "epoch": 0.5135309278350515, "grad_norm": 0.6140522360801697, "learning_rate": 9.846807520938594e-06, "loss": 0.6374, "step": 797 }, { "epoch": 0.5141752577319587, "grad_norm": 0.6900352239608765, "learning_rate": 9.845885280226469e-06, "loss": 0.6492, "step": 798 }, { "epoch": 0.5148195876288659, "grad_norm": 0.6793680191040039, "learning_rate": 9.844960315277839e-06, "loss": 0.6808, "step": 799 }, { "epoch": 0.5154639175257731, "grad_norm": 0.621385395526886, "learning_rate": 9.844032626612701e-06, "loss": 0.6519, "step": 800 }, { "epoch": 0.5161082474226805, "grad_norm": 0.6685365438461304, "learning_rate": 9.843102214752577e-06, "loss": 0.6352, "step": 801 }, { "epoch": 0.5167525773195877, "grad_norm": 0.6775370836257935, "learning_rate": 9.842169080220523e-06, "loss": 0.6521, "step": 802 }, { "epoch": 0.5173969072164949, "grad_norm": 0.6927893161773682, "learning_rate": 9.841233223541121e-06, "loss": 0.6494, "step": 803 }, { "epoch": 0.5180412371134021, "grad_norm": 0.7237555384635925, "learning_rate": 9.840294645240488e-06, "loss": 0.6767, "step": 804 }, { "epoch": 0.5186855670103093, "grad_norm": 0.6732221245765686, "learning_rate": 9.839353345846272e-06, "loss": 0.6473, "step": 805 }, { "epoch": 0.5193298969072165, "grad_norm": 0.6946284174919128, "learning_rate": 9.838409325887643e-06, "loss": 0.6439, "step": 806 }, { "epoch": 0.5199742268041238, "grad_norm": 0.7950090765953064, "learning_rate": 9.83746258589531e-06, "loss": 0.6528, "step": 807 }, { "epoch": 0.520618556701031, "grad_norm": 0.6466878652572632, "learning_rate": 9.836513126401504e-06, "loss": 0.6245, "step": 808 }, { "epoch": 0.5212628865979382, "grad_norm": 0.7175922393798828, "learning_rate": 9.83556094793999e-06, "loss": 0.635, "step": 809 }, { "epoch": 0.5219072164948454, "grad_norm": 0.8202354311943054, "learning_rate": 9.834606051046056e-06, "loss": 0.6607, "step": 810 }, { "epoch": 0.5225515463917526, "grad_norm": 0.6628302931785583, "learning_rate": 9.833648436256525e-06, "loss": 0.6272, "step": 811 }, { "epoch": 0.5231958762886598, "grad_norm": 0.9413242340087891, "learning_rate": 9.832688104109742e-06, "loss": 0.6626, "step": 812 }, { "epoch": 0.523840206185567, "grad_norm": 0.7113590240478516, "learning_rate": 9.831725055145581e-06, "loss": 0.6473, "step": 813 }, { "epoch": 0.5244845360824743, "grad_norm": 0.7008445262908936, "learning_rate": 9.830759289905448e-06, "loss": 0.6449, "step": 814 }, { "epoch": 0.5251288659793815, "grad_norm": 0.8174973726272583, "learning_rate": 9.829790808932266e-06, "loss": 0.6365, "step": 815 }, { "epoch": 0.5257731958762887, "grad_norm": 0.6829904317855835, "learning_rate": 9.828819612770497e-06, "loss": 0.6664, "step": 816 }, { "epoch": 0.5264175257731959, "grad_norm": 0.860411524772644, "learning_rate": 9.827845701966121e-06, "loss": 0.6318, "step": 817 }, { "epoch": 0.5270618556701031, "grad_norm": 0.7295666933059692, "learning_rate": 9.826869077066648e-06, "loss": 0.6585, "step": 818 }, { "epoch": 0.5277061855670103, "grad_norm": 0.8204461336135864, "learning_rate": 9.825889738621109e-06, "loss": 0.6497, "step": 819 }, { "epoch": 0.5283505154639175, "grad_norm": 0.9523586630821228, "learning_rate": 9.824907687180067e-06, "loss": 0.6589, "step": 820 }, { "epoch": 0.5289948453608248, "grad_norm": 0.7110240459442139, "learning_rate": 9.823922923295606e-06, "loss": 0.6452, "step": 821 }, { "epoch": 0.529639175257732, "grad_norm": 0.8268753290176392, "learning_rate": 9.822935447521333e-06, "loss": 0.6453, "step": 822 }, { "epoch": 0.5302835051546392, "grad_norm": 0.9293363094329834, "learning_rate": 9.821945260412388e-06, "loss": 0.6632, "step": 823 }, { "epoch": 0.5309278350515464, "grad_norm": 0.8253251910209656, "learning_rate": 9.820952362525424e-06, "loss": 0.6474, "step": 824 }, { "epoch": 0.5315721649484536, "grad_norm": 0.8333533406257629, "learning_rate": 9.819956754418628e-06, "loss": 0.6341, "step": 825 }, { "epoch": 0.5322164948453608, "grad_norm": 0.933760941028595, "learning_rate": 9.818958436651704e-06, "loss": 0.6386, "step": 826 }, { "epoch": 0.532860824742268, "grad_norm": 0.8404245972633362, "learning_rate": 9.81795740978588e-06, "loss": 0.6684, "step": 827 }, { "epoch": 0.5335051546391752, "grad_norm": 0.865915834903717, "learning_rate": 9.816953674383913e-06, "loss": 0.6483, "step": 828 }, { "epoch": 0.5341494845360825, "grad_norm": 0.9834263920783997, "learning_rate": 9.815947231010075e-06, "loss": 0.6379, "step": 829 }, { "epoch": 0.5347938144329897, "grad_norm": 0.7735155820846558, "learning_rate": 9.814938080230163e-06, "loss": 0.6367, "step": 830 }, { "epoch": 0.5354381443298969, "grad_norm": 1.0027036666870117, "learning_rate": 9.813926222611496e-06, "loss": 0.6482, "step": 831 }, { "epoch": 0.5360824742268041, "grad_norm": 0.8820791840553284, "learning_rate": 9.81291165872292e-06, "loss": 0.6379, "step": 832 }, { "epoch": 0.5367268041237113, "grad_norm": 0.6414960026741028, "learning_rate": 9.811894389134792e-06, "loss": 0.6323, "step": 833 }, { "epoch": 0.5373711340206185, "grad_norm": 0.752196192741394, "learning_rate": 9.810874414418998e-06, "loss": 0.6407, "step": 834 }, { "epoch": 0.5380154639175257, "grad_norm": 0.6465397477149963, "learning_rate": 9.809851735148943e-06, "loss": 0.6481, "step": 835 }, { "epoch": 0.538659793814433, "grad_norm": 0.7583353519439697, "learning_rate": 9.808826351899551e-06, "loss": 0.6479, "step": 836 }, { "epoch": 0.5393041237113402, "grad_norm": 0.6689878702163696, "learning_rate": 9.807798265247269e-06, "loss": 0.6414, "step": 837 }, { "epoch": 0.5399484536082474, "grad_norm": 0.8775922060012817, "learning_rate": 9.80676747577006e-06, "loss": 0.6425, "step": 838 }, { "epoch": 0.5405927835051546, "grad_norm": 0.6851148009300232, "learning_rate": 9.805733984047408e-06, "loss": 0.6412, "step": 839 }, { "epoch": 0.5412371134020618, "grad_norm": 0.6769123673439026, "learning_rate": 9.804697790660317e-06, "loss": 0.6484, "step": 840 }, { "epoch": 0.541881443298969, "grad_norm": 0.8902427554130554, "learning_rate": 9.80365889619131e-06, "loss": 0.6486, "step": 841 }, { "epoch": 0.5425257731958762, "grad_norm": 0.6505954265594482, "learning_rate": 9.802617301224425e-06, "loss": 0.6355, "step": 842 }, { "epoch": 0.5431701030927835, "grad_norm": 0.8050699234008789, "learning_rate": 9.801573006345224e-06, "loss": 0.6261, "step": 843 }, { "epoch": 0.5438144329896907, "grad_norm": 0.8468318581581116, "learning_rate": 9.800526012140782e-06, "loss": 0.641, "step": 844 }, { "epoch": 0.5444587628865979, "grad_norm": 0.7054587006568909, "learning_rate": 9.799476319199695e-06, "loss": 0.6546, "step": 845 }, { "epoch": 0.5451030927835051, "grad_norm": 0.7490087747573853, "learning_rate": 9.798423928112071e-06, "loss": 0.6488, "step": 846 }, { "epoch": 0.5457474226804123, "grad_norm": 0.7574702501296997, "learning_rate": 9.797368839469542e-06, "loss": 0.6519, "step": 847 }, { "epoch": 0.5463917525773195, "grad_norm": 0.7323676943778992, "learning_rate": 9.79631105386525e-06, "loss": 0.6826, "step": 848 }, { "epoch": 0.5470360824742269, "grad_norm": 0.6424057483673096, "learning_rate": 9.795250571893857e-06, "loss": 0.644, "step": 849 }, { "epoch": 0.5476804123711341, "grad_norm": 0.696093738079071, "learning_rate": 9.794187394151538e-06, "loss": 0.6407, "step": 850 }, { "epoch": 0.5483247422680413, "grad_norm": 0.6338746547698975, "learning_rate": 9.793121521235988e-06, "loss": 0.6591, "step": 851 }, { "epoch": 0.5489690721649485, "grad_norm": 0.6039814352989197, "learning_rate": 9.792052953746411e-06, "loss": 0.6579, "step": 852 }, { "epoch": 0.5496134020618557, "grad_norm": 0.7008209228515625, "learning_rate": 9.790981692283532e-06, "loss": 0.6224, "step": 853 }, { "epoch": 0.5502577319587629, "grad_norm": 0.6647630929946899, "learning_rate": 9.789907737449588e-06, "loss": 0.6503, "step": 854 }, { "epoch": 0.5509020618556701, "grad_norm": 0.5996343493461609, "learning_rate": 9.788831089848327e-06, "loss": 0.6512, "step": 855 }, { "epoch": 0.5515463917525774, "grad_norm": 0.691163182258606, "learning_rate": 9.787751750085014e-06, "loss": 0.6642, "step": 856 }, { "epoch": 0.5521907216494846, "grad_norm": 0.6191086769104004, "learning_rate": 9.786669718766429e-06, "loss": 0.6218, "step": 857 }, { "epoch": 0.5528350515463918, "grad_norm": 0.6323687434196472, "learning_rate": 9.785584996500862e-06, "loss": 0.6316, "step": 858 }, { "epoch": 0.553479381443299, "grad_norm": 0.6885550618171692, "learning_rate": 9.784497583898116e-06, "loss": 0.6439, "step": 859 }, { "epoch": 0.5541237113402062, "grad_norm": 0.6412866711616516, "learning_rate": 9.783407481569509e-06, "loss": 0.624, "step": 860 }, { "epoch": 0.5547680412371134, "grad_norm": 0.6438667178153992, "learning_rate": 9.782314690127867e-06, "loss": 0.6555, "step": 861 }, { "epoch": 0.5554123711340206, "grad_norm": 0.7188820838928223, "learning_rate": 9.781219210187535e-06, "loss": 0.6407, "step": 862 }, { "epoch": 0.5560567010309279, "grad_norm": 0.7178824543952942, "learning_rate": 9.780121042364362e-06, "loss": 0.6368, "step": 863 }, { "epoch": 0.5567010309278351, "grad_norm": 0.7491236925125122, "learning_rate": 9.779020187275709e-06, "loss": 0.6448, "step": 864 }, { "epoch": 0.5573453608247423, "grad_norm": 0.8216255903244019, "learning_rate": 9.77791664554045e-06, "loss": 0.6436, "step": 865 }, { "epoch": 0.5579896907216495, "grad_norm": 0.6939603686332703, "learning_rate": 9.77681041777897e-06, "loss": 0.6331, "step": 866 }, { "epoch": 0.5586340206185567, "grad_norm": 0.7152724266052246, "learning_rate": 9.775701504613164e-06, "loss": 0.6568, "step": 867 }, { "epoch": 0.5592783505154639, "grad_norm": 0.7545168399810791, "learning_rate": 9.774589906666435e-06, "loss": 0.637, "step": 868 }, { "epoch": 0.5599226804123711, "grad_norm": 0.7088530659675598, "learning_rate": 9.773475624563691e-06, "loss": 0.6458, "step": 869 }, { "epoch": 0.5605670103092784, "grad_norm": 0.594097375869751, "learning_rate": 9.77235865893136e-06, "loss": 0.6475, "step": 870 }, { "epoch": 0.5612113402061856, "grad_norm": 0.675575852394104, "learning_rate": 9.77123901039737e-06, "loss": 0.6518, "step": 871 }, { "epoch": 0.5618556701030928, "grad_norm": 0.728264331817627, "learning_rate": 9.77011667959116e-06, "loss": 0.6497, "step": 872 }, { "epoch": 0.5625, "grad_norm": 0.858462393283844, "learning_rate": 9.768991667143674e-06, "loss": 0.6533, "step": 873 }, { "epoch": 0.5631443298969072, "grad_norm": 0.6419100165367126, "learning_rate": 9.76786397368737e-06, "loss": 0.6385, "step": 874 }, { "epoch": 0.5637886597938144, "grad_norm": 0.821353554725647, "learning_rate": 9.766733599856207e-06, "loss": 0.6296, "step": 875 }, { "epoch": 0.5644329896907216, "grad_norm": 0.7434787154197693, "learning_rate": 9.765600546285654e-06, "loss": 0.6599, "step": 876 }, { "epoch": 0.5650773195876289, "grad_norm": 0.6288697123527527, "learning_rate": 9.764464813612684e-06, "loss": 0.6476, "step": 877 }, { "epoch": 0.5657216494845361, "grad_norm": 0.7098308205604553, "learning_rate": 9.763326402475778e-06, "loss": 0.6551, "step": 878 }, { "epoch": 0.5663659793814433, "grad_norm": 0.710366427898407, "learning_rate": 9.762185313514927e-06, "loss": 0.6343, "step": 879 }, { "epoch": 0.5670103092783505, "grad_norm": 0.5693610906600952, "learning_rate": 9.761041547371617e-06, "loss": 0.6204, "step": 880 }, { "epoch": 0.5676546391752577, "grad_norm": 0.6205378770828247, "learning_rate": 9.759895104688848e-06, "loss": 0.6394, "step": 881 }, { "epoch": 0.5682989690721649, "grad_norm": 0.6765633821487427, "learning_rate": 9.75874598611112e-06, "loss": 0.6275, "step": 882 }, { "epoch": 0.5689432989690721, "grad_norm": 0.6129748225212097, "learning_rate": 9.757594192284441e-06, "loss": 0.6395, "step": 883 }, { "epoch": 0.5695876288659794, "grad_norm": 0.6062330007553101, "learning_rate": 9.756439723856319e-06, "loss": 0.6313, "step": 884 }, { "epoch": 0.5702319587628866, "grad_norm": 0.6747280955314636, "learning_rate": 9.755282581475769e-06, "loss": 0.632, "step": 885 }, { "epoch": 0.5708762886597938, "grad_norm": 0.7212796807289124, "learning_rate": 9.754122765793306e-06, "loss": 0.6373, "step": 886 }, { "epoch": 0.571520618556701, "grad_norm": 0.6104435324668884, "learning_rate": 9.752960277460951e-06, "loss": 0.6509, "step": 887 }, { "epoch": 0.5721649484536082, "grad_norm": 0.6594733595848083, "learning_rate": 9.751795117132227e-06, "loss": 0.6271, "step": 888 }, { "epoch": 0.5728092783505154, "grad_norm": 0.6633055806159973, "learning_rate": 9.750627285462156e-06, "loss": 0.6273, "step": 889 }, { "epoch": 0.5734536082474226, "grad_norm": 0.6113850474357605, "learning_rate": 9.749456783107266e-06, "loss": 0.6217, "step": 890 }, { "epoch": 0.5740979381443299, "grad_norm": 0.6833528280258179, "learning_rate": 9.748283610725581e-06, "loss": 0.6551, "step": 891 }, { "epoch": 0.5747422680412371, "grad_norm": 0.7211122512817383, "learning_rate": 9.747107768976635e-06, "loss": 0.6271, "step": 892 }, { "epoch": 0.5753865979381443, "grad_norm": 0.6235044002532959, "learning_rate": 9.745929258521452e-06, "loss": 0.6511, "step": 893 }, { "epoch": 0.5760309278350515, "grad_norm": 0.6488106846809387, "learning_rate": 9.744748080022563e-06, "loss": 0.6475, "step": 894 }, { "epoch": 0.5766752577319587, "grad_norm": 0.61222904920578, "learning_rate": 9.743564234143999e-06, "loss": 0.6615, "step": 895 }, { "epoch": 0.5773195876288659, "grad_norm": 0.7023225426673889, "learning_rate": 9.742377721551286e-06, "loss": 0.6586, "step": 896 }, { "epoch": 0.5779639175257731, "grad_norm": 0.6103045344352722, "learning_rate": 9.741188542911454e-06, "loss": 0.6417, "step": 897 }, { "epoch": 0.5786082474226805, "grad_norm": 0.6861590147018433, "learning_rate": 9.739996698893028e-06, "loss": 0.6504, "step": 898 }, { "epoch": 0.5792525773195877, "grad_norm": 0.7182576656341553, "learning_rate": 9.738802190166035e-06, "loss": 0.6632, "step": 899 }, { "epoch": 0.5798969072164949, "grad_norm": 0.631467342376709, "learning_rate": 9.737605017401997e-06, "loss": 0.63, "step": 900 }, { "epoch": 0.5805412371134021, "grad_norm": 0.6721380352973938, "learning_rate": 9.736405181273935e-06, "loss": 0.6375, "step": 901 }, { "epoch": 0.5811855670103093, "grad_norm": 0.6552727818489075, "learning_rate": 9.735202682456368e-06, "loss": 0.6407, "step": 902 }, { "epoch": 0.5818298969072165, "grad_norm": 0.6513555645942688, "learning_rate": 9.733997521625309e-06, "loss": 0.6262, "step": 903 }, { "epoch": 0.5824742268041238, "grad_norm": 0.7494980692863464, "learning_rate": 9.732789699458272e-06, "loss": 0.6462, "step": 904 }, { "epoch": 0.583118556701031, "grad_norm": 0.6122732162475586, "learning_rate": 9.731579216634261e-06, "loss": 0.6174, "step": 905 }, { "epoch": 0.5837628865979382, "grad_norm": 0.6170387864112854, "learning_rate": 9.730366073833785e-06, "loss": 0.6198, "step": 906 }, { "epoch": 0.5844072164948454, "grad_norm": 0.631680428981781, "learning_rate": 9.729150271738839e-06, "loss": 0.6318, "step": 907 }, { "epoch": 0.5850515463917526, "grad_norm": 0.6586320996284485, "learning_rate": 9.727931811032916e-06, "loss": 0.6479, "step": 908 }, { "epoch": 0.5856958762886598, "grad_norm": 0.6931595802307129, "learning_rate": 9.726710692401006e-06, "loss": 0.6683, "step": 909 }, { "epoch": 0.586340206185567, "grad_norm": 0.6524629592895508, "learning_rate": 9.725486916529592e-06, "loss": 0.6342, "step": 910 }, { "epoch": 0.5869845360824743, "grad_norm": 0.6370339393615723, "learning_rate": 9.72426048410665e-06, "loss": 0.6164, "step": 911 }, { "epoch": 0.5876288659793815, "grad_norm": 0.6214847564697266, "learning_rate": 9.723031395821649e-06, "loss": 0.6246, "step": 912 }, { "epoch": 0.5882731958762887, "grad_norm": 0.6768206357955933, "learning_rate": 9.721799652365553e-06, "loss": 0.6424, "step": 913 }, { "epoch": 0.5889175257731959, "grad_norm": 0.6190541982650757, "learning_rate": 9.720565254430815e-06, "loss": 0.6376, "step": 914 }, { "epoch": 0.5895618556701031, "grad_norm": 0.6091976165771484, "learning_rate": 9.719328202711388e-06, "loss": 0.6463, "step": 915 }, { "epoch": 0.5902061855670103, "grad_norm": 0.6777482628822327, "learning_rate": 9.718088497902709e-06, "loss": 0.6453, "step": 916 }, { "epoch": 0.5908505154639175, "grad_norm": 0.5904690027236938, "learning_rate": 9.716846140701707e-06, "loss": 0.6268, "step": 917 }, { "epoch": 0.5914948453608248, "grad_norm": 0.7139037251472473, "learning_rate": 9.71560113180681e-06, "loss": 0.6309, "step": 918 }, { "epoch": 0.592139175257732, "grad_norm": 0.672453761100769, "learning_rate": 9.714353471917926e-06, "loss": 0.6467, "step": 919 }, { "epoch": 0.5927835051546392, "grad_norm": 0.676138162612915, "learning_rate": 9.71310316173646e-06, "loss": 0.6518, "step": 920 }, { "epoch": 0.5934278350515464, "grad_norm": 0.6425537467002869, "learning_rate": 9.711850201965305e-06, "loss": 0.6307, "step": 921 }, { "epoch": 0.5940721649484536, "grad_norm": 0.736703097820282, "learning_rate": 9.710594593308846e-06, "loss": 0.6413, "step": 922 }, { "epoch": 0.5947164948453608, "grad_norm": 0.6477802991867065, "learning_rate": 9.709336336472953e-06, "loss": 0.6346, "step": 923 }, { "epoch": 0.595360824742268, "grad_norm": 0.6371743083000183, "learning_rate": 9.708075432164988e-06, "loss": 0.645, "step": 924 }, { "epoch": 0.5960051546391752, "grad_norm": 0.6381727457046509, "learning_rate": 9.706811881093799e-06, "loss": 0.6259, "step": 925 }, { "epoch": 0.5966494845360825, "grad_norm": 0.6847191452980042, "learning_rate": 9.705545683969722e-06, "loss": 0.6508, "step": 926 }, { "epoch": 0.5972938144329897, "grad_norm": 0.6423772573471069, "learning_rate": 9.704276841504586e-06, "loss": 0.6475, "step": 927 }, { "epoch": 0.5979381443298969, "grad_norm": 0.6457873582839966, "learning_rate": 9.703005354411696e-06, "loss": 0.6597, "step": 928 }, { "epoch": 0.5985824742268041, "grad_norm": 0.5851149559020996, "learning_rate": 9.701731223405857e-06, "loss": 0.6424, "step": 929 }, { "epoch": 0.5992268041237113, "grad_norm": 0.6814531683921814, "learning_rate": 9.700454449203351e-06, "loss": 0.6417, "step": 930 }, { "epoch": 0.5998711340206185, "grad_norm": 0.6776267886161804, "learning_rate": 9.699175032521949e-06, "loss": 0.6149, "step": 931 }, { "epoch": 0.6005154639175257, "grad_norm": 0.8106547594070435, "learning_rate": 9.697892974080907e-06, "loss": 0.6403, "step": 932 }, { "epoch": 0.601159793814433, "grad_norm": 0.7483506202697754, "learning_rate": 9.696608274600966e-06, "loss": 0.6394, "step": 933 }, { "epoch": 0.6018041237113402, "grad_norm": 0.6360734105110168, "learning_rate": 9.695320934804356e-06, "loss": 0.6502, "step": 934 }, { "epoch": 0.6024484536082474, "grad_norm": 0.7431177496910095, "learning_rate": 9.694030955414781e-06, "loss": 0.6377, "step": 935 }, { "epoch": 0.6030927835051546, "grad_norm": 0.6655299663543701, "learning_rate": 9.692738337157441e-06, "loss": 0.6361, "step": 936 }, { "epoch": 0.6037371134020618, "grad_norm": 0.716605007648468, "learning_rate": 9.69144308075901e-06, "loss": 0.65, "step": 937 }, { "epoch": 0.604381443298969, "grad_norm": 0.8165088295936584, "learning_rate": 9.690145186947653e-06, "loss": 0.622, "step": 938 }, { "epoch": 0.6050257731958762, "grad_norm": 0.673112690448761, "learning_rate": 9.68884465645301e-06, "loss": 0.63, "step": 939 }, { "epoch": 0.6056701030927835, "grad_norm": 0.6738924980163574, "learning_rate": 9.687541490006209e-06, "loss": 0.6368, "step": 940 }, { "epoch": 0.6063144329896907, "grad_norm": 0.7261308431625366, "learning_rate": 9.686235688339857e-06, "loss": 0.632, "step": 941 }, { "epoch": 0.6069587628865979, "grad_norm": 0.6608893871307373, "learning_rate": 9.684927252188046e-06, "loss": 0.6333, "step": 942 }, { "epoch": 0.6076030927835051, "grad_norm": 0.6324573755264282, "learning_rate": 9.683616182286342e-06, "loss": 0.6675, "step": 943 }, { "epoch": 0.6082474226804123, "grad_norm": 0.8667783141136169, "learning_rate": 9.682302479371797e-06, "loss": 0.6453, "step": 944 }, { "epoch": 0.6088917525773195, "grad_norm": 0.6096985340118408, "learning_rate": 9.680986144182944e-06, "loss": 0.6386, "step": 945 }, { "epoch": 0.6095360824742269, "grad_norm": 0.8256608843803406, "learning_rate": 9.679667177459794e-06, "loss": 0.6523, "step": 946 }, { "epoch": 0.6101804123711341, "grad_norm": 0.7617649435997009, "learning_rate": 9.678345579943837e-06, "loss": 0.6437, "step": 947 }, { "epoch": 0.6108247422680413, "grad_norm": 0.6732947826385498, "learning_rate": 9.67702135237804e-06, "loss": 0.6349, "step": 948 }, { "epoch": 0.6114690721649485, "grad_norm": 0.700886607170105, "learning_rate": 9.675694495506854e-06, "loss": 0.6578, "step": 949 }, { "epoch": 0.6121134020618557, "grad_norm": 0.6399227380752563, "learning_rate": 9.674365010076203e-06, "loss": 0.6396, "step": 950 }, { "epoch": 0.6127577319587629, "grad_norm": 0.7727089524269104, "learning_rate": 9.673032896833493e-06, "loss": 0.6433, "step": 951 }, { "epoch": 0.6134020618556701, "grad_norm": 0.6507975459098816, "learning_rate": 9.671698156527603e-06, "loss": 0.6343, "step": 952 }, { "epoch": 0.6140463917525774, "grad_norm": 0.6477310061454773, "learning_rate": 9.67036078990889e-06, "loss": 0.6442, "step": 953 }, { "epoch": 0.6146907216494846, "grad_norm": 0.9328968524932861, "learning_rate": 9.669020797729192e-06, "loss": 0.6555, "step": 954 }, { "epoch": 0.6153350515463918, "grad_norm": 0.6176644563674927, "learning_rate": 9.667678180741817e-06, "loss": 0.6529, "step": 955 }, { "epoch": 0.615979381443299, "grad_norm": 0.8628554940223694, "learning_rate": 9.66633293970155e-06, "loss": 0.6515, "step": 956 }, { "epoch": 0.6166237113402062, "grad_norm": 0.747087836265564, "learning_rate": 9.664985075364654e-06, "loss": 0.6237, "step": 957 }, { "epoch": 0.6172680412371134, "grad_norm": 0.7322733998298645, "learning_rate": 9.663634588488864e-06, "loss": 0.6433, "step": 958 }, { "epoch": 0.6179123711340206, "grad_norm": 0.6610790491104126, "learning_rate": 9.662281479833389e-06, "loss": 0.6369, "step": 959 }, { "epoch": 0.6185567010309279, "grad_norm": 0.7392152547836304, "learning_rate": 9.660925750158917e-06, "loss": 0.6595, "step": 960 }, { "epoch": 0.6192010309278351, "grad_norm": 0.6075034737586975, "learning_rate": 9.6595674002276e-06, "loss": 0.6347, "step": 961 }, { "epoch": 0.6198453608247423, "grad_norm": 0.6183353066444397, "learning_rate": 9.658206430803074e-06, "loss": 0.6503, "step": 962 }, { "epoch": 0.6204896907216495, "grad_norm": 0.8094824552536011, "learning_rate": 9.65684284265044e-06, "loss": 0.648, "step": 963 }, { "epoch": 0.6211340206185567, "grad_norm": 0.6207033395767212, "learning_rate": 9.655476636536273e-06, "loss": 0.6442, "step": 964 }, { "epoch": 0.6217783505154639, "grad_norm": 0.6459785103797913, "learning_rate": 9.654107813228622e-06, "loss": 0.6633, "step": 965 }, { "epoch": 0.6224226804123711, "grad_norm": 0.7386115193367004, "learning_rate": 9.652736373497001e-06, "loss": 0.6483, "step": 966 }, { "epoch": 0.6230670103092784, "grad_norm": 0.5501590967178345, "learning_rate": 9.651362318112405e-06, "loss": 0.6336, "step": 967 }, { "epoch": 0.6237113402061856, "grad_norm": 0.7040377259254456, "learning_rate": 9.649985647847291e-06, "loss": 0.6335, "step": 968 }, { "epoch": 0.6243556701030928, "grad_norm": 0.6006501317024231, "learning_rate": 9.648606363475588e-06, "loss": 0.6496, "step": 969 }, { "epoch": 0.625, "grad_norm": 0.6655983924865723, "learning_rate": 9.647224465772698e-06, "loss": 0.6327, "step": 970 }, { "epoch": 0.6256443298969072, "grad_norm": 0.6774782538414001, "learning_rate": 9.645839955515488e-06, "loss": 0.6437, "step": 971 }, { "epoch": 0.6262886597938144, "grad_norm": 0.5998676419258118, "learning_rate": 9.644452833482296e-06, "loss": 0.6301, "step": 972 }, { "epoch": 0.6269329896907216, "grad_norm": 0.8207651376724243, "learning_rate": 9.643063100452927e-06, "loss": 0.6639, "step": 973 }, { "epoch": 0.6275773195876289, "grad_norm": 0.8253977298736572, "learning_rate": 9.641670757208655e-06, "loss": 0.6267, "step": 974 }, { "epoch": 0.6282216494845361, "grad_norm": 0.6204148530960083, "learning_rate": 9.64027580453222e-06, "loss": 0.6456, "step": 975 }, { "epoch": 0.6288659793814433, "grad_norm": 0.6561158299446106, "learning_rate": 9.63887824320783e-06, "loss": 0.6337, "step": 976 }, { "epoch": 0.6295103092783505, "grad_norm": 0.7014433741569519, "learning_rate": 9.637478074021161e-06, "loss": 0.6325, "step": 977 }, { "epoch": 0.6301546391752577, "grad_norm": 0.6084746718406677, "learning_rate": 9.636075297759351e-06, "loss": 0.6575, "step": 978 }, { "epoch": 0.6307989690721649, "grad_norm": 0.6832376718521118, "learning_rate": 9.634669915211005e-06, "loss": 0.6523, "step": 979 }, { "epoch": 0.6314432989690721, "grad_norm": 0.7221859097480774, "learning_rate": 9.633261927166199e-06, "loss": 0.643, "step": 980 }, { "epoch": 0.6320876288659794, "grad_norm": 0.672120213508606, "learning_rate": 9.631851334416467e-06, "loss": 0.6066, "step": 981 }, { "epoch": 0.6327319587628866, "grad_norm": 0.7328267693519592, "learning_rate": 9.630438137754807e-06, "loss": 0.6576, "step": 982 }, { "epoch": 0.6333762886597938, "grad_norm": 0.7860979437828064, "learning_rate": 9.629022337975684e-06, "loss": 0.648, "step": 983 }, { "epoch": 0.634020618556701, "grad_norm": 0.7111921310424805, "learning_rate": 9.627603935875027e-06, "loss": 0.628, "step": 984 }, { "epoch": 0.6346649484536082, "grad_norm": 0.7360318899154663, "learning_rate": 9.626182932250225e-06, "loss": 0.6376, "step": 985 }, { "epoch": 0.6353092783505154, "grad_norm": 0.7733191251754761, "learning_rate": 9.624759327900131e-06, "loss": 0.6452, "step": 986 }, { "epoch": 0.6359536082474226, "grad_norm": 0.6809870004653931, "learning_rate": 9.623333123625062e-06, "loss": 0.6235, "step": 987 }, { "epoch": 0.6365979381443299, "grad_norm": 0.6524026393890381, "learning_rate": 9.621904320226792e-06, "loss": 0.6408, "step": 988 }, { "epoch": 0.6372422680412371, "grad_norm": 0.6878443956375122, "learning_rate": 9.620472918508561e-06, "loss": 0.6351, "step": 989 }, { "epoch": 0.6378865979381443, "grad_norm": 0.7105034589767456, "learning_rate": 9.619038919275067e-06, "loss": 0.6515, "step": 990 }, { "epoch": 0.6385309278350515, "grad_norm": 0.648108184337616, "learning_rate": 9.617602323332467e-06, "loss": 0.6433, "step": 991 }, { "epoch": 0.6391752577319587, "grad_norm": 0.7021229267120361, "learning_rate": 9.616163131488384e-06, "loss": 0.649, "step": 992 }, { "epoch": 0.6398195876288659, "grad_norm": 0.5918980836868286, "learning_rate": 9.61472134455189e-06, "loss": 0.6112, "step": 993 }, { "epoch": 0.6404639175257731, "grad_norm": 0.6715604066848755, "learning_rate": 9.613276963333525e-06, "loss": 0.6258, "step": 994 }, { "epoch": 0.6411082474226805, "grad_norm": 0.6819112300872803, "learning_rate": 9.611829988645285e-06, "loss": 0.6342, "step": 995 }, { "epoch": 0.6417525773195877, "grad_norm": 0.6377294063568115, "learning_rate": 9.610380421300623e-06, "loss": 0.6228, "step": 996 }, { "epoch": 0.6423969072164949, "grad_norm": 0.7220823168754578, "learning_rate": 9.608928262114448e-06, "loss": 0.6404, "step": 997 }, { "epoch": 0.6430412371134021, "grad_norm": 0.664067804813385, "learning_rate": 9.607473511903129e-06, "loss": 0.634, "step": 998 }, { "epoch": 0.6436855670103093, "grad_norm": 0.6406931281089783, "learning_rate": 9.606016171484491e-06, "loss": 0.6415, "step": 999 }, { "epoch": 0.6443298969072165, "grad_norm": 0.6940628290176392, "learning_rate": 9.604556241677813e-06, "loss": 0.6629, "step": 1000 }, { "epoch": 0.6449742268041238, "grad_norm": 0.6435523629188538, "learning_rate": 9.603093723303835e-06, "loss": 0.6329, "step": 1001 }, { "epoch": 0.645618556701031, "grad_norm": 0.6660202741622925, "learning_rate": 9.601628617184744e-06, "loss": 0.6321, "step": 1002 }, { "epoch": 0.6462628865979382, "grad_norm": 0.6589183807373047, "learning_rate": 9.600160924144191e-06, "loss": 0.6357, "step": 1003 }, { "epoch": 0.6469072164948454, "grad_norm": 0.7592248916625977, "learning_rate": 9.598690645007271e-06, "loss": 0.6268, "step": 1004 }, { "epoch": 0.6475515463917526, "grad_norm": 0.6070667505264282, "learning_rate": 9.597217780600543e-06, "loss": 0.631, "step": 1005 }, { "epoch": 0.6481958762886598, "grad_norm": 0.8970237374305725, "learning_rate": 9.595742331752014e-06, "loss": 0.6629, "step": 1006 }, { "epoch": 0.648840206185567, "grad_norm": 0.6678323149681091, "learning_rate": 9.594264299291142e-06, "loss": 0.6302, "step": 1007 }, { "epoch": 0.6494845360824743, "grad_norm": 0.6294822692871094, "learning_rate": 9.592783684048844e-06, "loss": 0.6414, "step": 1008 }, { "epoch": 0.6501288659793815, "grad_norm": 0.6644755601882935, "learning_rate": 9.591300486857481e-06, "loss": 0.6378, "step": 1009 }, { "epoch": 0.6507731958762887, "grad_norm": 0.6900251507759094, "learning_rate": 9.589814708550876e-06, "loss": 0.6397, "step": 1010 }, { "epoch": 0.6514175257731959, "grad_norm": 0.626064658164978, "learning_rate": 9.58832634996429e-06, "loss": 0.6448, "step": 1011 }, { "epoch": 0.6520618556701031, "grad_norm": 0.64618980884552, "learning_rate": 9.586835411934445e-06, "loss": 0.6586, "step": 1012 }, { "epoch": 0.6527061855670103, "grad_norm": 0.6432942748069763, "learning_rate": 9.585341895299507e-06, "loss": 0.6462, "step": 1013 }, { "epoch": 0.6533505154639175, "grad_norm": 0.6582573652267456, "learning_rate": 9.583845800899097e-06, "loss": 0.6413, "step": 1014 }, { "epoch": 0.6539948453608248, "grad_norm": 0.6557276844978333, "learning_rate": 9.58234712957428e-06, "loss": 0.6599, "step": 1015 }, { "epoch": 0.654639175257732, "grad_norm": 0.7248553037643433, "learning_rate": 9.580845882167574e-06, "loss": 0.6592, "step": 1016 }, { "epoch": 0.6552835051546392, "grad_norm": 0.6509864330291748, "learning_rate": 9.57934205952294e-06, "loss": 0.6569, "step": 1017 }, { "epoch": 0.6559278350515464, "grad_norm": 0.6445841193199158, "learning_rate": 9.57783566248579e-06, "loss": 0.6178, "step": 1018 }, { "epoch": 0.6565721649484536, "grad_norm": 0.6260314583778381, "learning_rate": 9.576326691902986e-06, "loss": 0.6256, "step": 1019 }, { "epoch": 0.6572164948453608, "grad_norm": 0.6364720463752747, "learning_rate": 9.574815148622833e-06, "loss": 0.6541, "step": 1020 }, { "epoch": 0.657860824742268, "grad_norm": 0.6762931942939758, "learning_rate": 9.57330103349508e-06, "loss": 0.6492, "step": 1021 }, { "epoch": 0.6585051546391752, "grad_norm": 0.638401985168457, "learning_rate": 9.571784347370929e-06, "loss": 0.6542, "step": 1022 }, { "epoch": 0.6591494845360825, "grad_norm": 0.666255533695221, "learning_rate": 9.570265091103021e-06, "loss": 0.6471, "step": 1023 }, { "epoch": 0.6597938144329897, "grad_norm": 0.6140169501304626, "learning_rate": 9.568743265545445e-06, "loss": 0.6402, "step": 1024 }, { "epoch": 0.6604381443298969, "grad_norm": 0.6449477672576904, "learning_rate": 9.567218871553734e-06, "loss": 0.6329, "step": 1025 }, { "epoch": 0.6610824742268041, "grad_norm": 0.6372910737991333, "learning_rate": 9.565691909984864e-06, "loss": 0.6657, "step": 1026 }, { "epoch": 0.6617268041237113, "grad_norm": 0.7547515630722046, "learning_rate": 9.564162381697253e-06, "loss": 0.6494, "step": 1027 }, { "epoch": 0.6623711340206185, "grad_norm": 0.6681149005889893, "learning_rate": 9.562630287550766e-06, "loss": 0.6594, "step": 1028 }, { "epoch": 0.6630154639175257, "grad_norm": 0.8291139006614685, "learning_rate": 9.561095628406708e-06, "loss": 0.6312, "step": 1029 }, { "epoch": 0.663659793814433, "grad_norm": 0.690674901008606, "learning_rate": 9.55955840512783e-06, "loss": 0.6202, "step": 1030 }, { "epoch": 0.6643041237113402, "grad_norm": 0.7419590950012207, "learning_rate": 9.558018618578314e-06, "loss": 0.6224, "step": 1031 }, { "epoch": 0.6649484536082474, "grad_norm": 0.8142262697219849, "learning_rate": 9.55647626962379e-06, "loss": 0.6583, "step": 1032 }, { "epoch": 0.6655927835051546, "grad_norm": 0.7585706114768982, "learning_rate": 9.554931359131335e-06, "loss": 0.6393, "step": 1033 }, { "epoch": 0.6662371134020618, "grad_norm": 0.6978089809417725, "learning_rate": 9.553383887969454e-06, "loss": 0.6592, "step": 1034 }, { "epoch": 0.666881443298969, "grad_norm": 0.7058124542236328, "learning_rate": 9.5518338570081e-06, "loss": 0.6485, "step": 1035 }, { "epoch": 0.6675257731958762, "grad_norm": 0.6158116459846497, "learning_rate": 9.550281267118659e-06, "loss": 0.6393, "step": 1036 }, { "epoch": 0.6681701030927835, "grad_norm": 0.731252133846283, "learning_rate": 9.54872611917396e-06, "loss": 0.6373, "step": 1037 }, { "epoch": 0.6688144329896907, "grad_norm": 0.7668895721435547, "learning_rate": 9.54716841404827e-06, "loss": 0.6615, "step": 1038 }, { "epoch": 0.6694587628865979, "grad_norm": 0.7548523545265198, "learning_rate": 9.545608152617288e-06, "loss": 0.6307, "step": 1039 }, { "epoch": 0.6701030927835051, "grad_norm": 0.7310301661491394, "learning_rate": 9.544045335758158e-06, "loss": 0.6433, "step": 1040 }, { "epoch": 0.6707474226804123, "grad_norm": 0.6051482558250427, "learning_rate": 9.542479964349456e-06, "loss": 0.6266, "step": 1041 }, { "epoch": 0.6713917525773195, "grad_norm": 0.7835216522216797, "learning_rate": 9.540912039271197e-06, "loss": 0.649, "step": 1042 }, { "epoch": 0.6720360824742269, "grad_norm": 0.5967505574226379, "learning_rate": 9.539341561404824e-06, "loss": 0.6418, "step": 1043 }, { "epoch": 0.6726804123711341, "grad_norm": 0.6458664536476135, "learning_rate": 9.537768531633227e-06, "loss": 0.6403, "step": 1044 }, { "epoch": 0.6733247422680413, "grad_norm": 0.6303918361663818, "learning_rate": 9.53619295084072e-06, "loss": 0.6236, "step": 1045 }, { "epoch": 0.6739690721649485, "grad_norm": 0.5829802751541138, "learning_rate": 9.534614819913056e-06, "loss": 0.622, "step": 1046 }, { "epoch": 0.6746134020618557, "grad_norm": 0.673724889755249, "learning_rate": 9.533034139737424e-06, "loss": 0.641, "step": 1047 }, { "epoch": 0.6752577319587629, "grad_norm": 0.6364939212799072, "learning_rate": 9.53145091120244e-06, "loss": 0.63, "step": 1048 }, { "epoch": 0.6759020618556701, "grad_norm": 0.6499471664428711, "learning_rate": 9.529865135198157e-06, "loss": 0.6186, "step": 1049 }, { "epoch": 0.6765463917525774, "grad_norm": 0.6180210113525391, "learning_rate": 9.528276812616058e-06, "loss": 0.6331, "step": 1050 }, { "epoch": 0.6771907216494846, "grad_norm": 0.6137977242469788, "learning_rate": 9.526685944349061e-06, "loss": 0.6497, "step": 1051 }, { "epoch": 0.6778350515463918, "grad_norm": 0.6026824116706848, "learning_rate": 9.52509253129151e-06, "loss": 0.6369, "step": 1052 }, { "epoch": 0.678479381443299, "grad_norm": 0.6312832832336426, "learning_rate": 9.523496574339184e-06, "loss": 0.6371, "step": 1053 }, { "epoch": 0.6791237113402062, "grad_norm": 0.6155292391777039, "learning_rate": 9.52189807438929e-06, "loss": 0.6378, "step": 1054 }, { "epoch": 0.6797680412371134, "grad_norm": 0.6416810154914856, "learning_rate": 9.520297032340464e-06, "loss": 0.6437, "step": 1055 }, { "epoch": 0.6804123711340206, "grad_norm": 0.6995428204536438, "learning_rate": 9.518693449092772e-06, "loss": 0.6319, "step": 1056 }, { "epoch": 0.6810567010309279, "grad_norm": 0.6115797162055969, "learning_rate": 9.517087325547713e-06, "loss": 0.6519, "step": 1057 }, { "epoch": 0.6817010309278351, "grad_norm": 0.857643723487854, "learning_rate": 9.515478662608205e-06, "loss": 0.6387, "step": 1058 }, { "epoch": 0.6823453608247423, "grad_norm": 0.8577443957328796, "learning_rate": 9.513867461178598e-06, "loss": 0.6355, "step": 1059 }, { "epoch": 0.6829896907216495, "grad_norm": 0.5894935727119446, "learning_rate": 9.512253722164673e-06, "loss": 0.6491, "step": 1060 }, { "epoch": 0.6836340206185567, "grad_norm": 0.9004279375076294, "learning_rate": 9.510637446473633e-06, "loss": 0.6327, "step": 1061 }, { "epoch": 0.6842783505154639, "grad_norm": 0.8127424716949463, "learning_rate": 9.509018635014106e-06, "loss": 0.6613, "step": 1062 }, { "epoch": 0.6849226804123711, "grad_norm": 0.6970890164375305, "learning_rate": 9.50739728869615e-06, "loss": 0.6554, "step": 1063 }, { "epoch": 0.6855670103092784, "grad_norm": 0.6836268305778503, "learning_rate": 9.505773408431246e-06, "loss": 0.6425, "step": 1064 }, { "epoch": 0.6862113402061856, "grad_norm": 0.6830376982688904, "learning_rate": 9.504146995132298e-06, "loss": 0.644, "step": 1065 }, { "epoch": 0.6868556701030928, "grad_norm": 0.7666815519332886, "learning_rate": 9.502518049713633e-06, "loss": 0.6323, "step": 1066 }, { "epoch": 0.6875, "grad_norm": 0.676910400390625, "learning_rate": 9.500886573091007e-06, "loss": 0.67, "step": 1067 }, { "epoch": 0.6881443298969072, "grad_norm": 0.673403263092041, "learning_rate": 9.499252566181595e-06, "loss": 0.629, "step": 1068 }, { "epoch": 0.6887886597938144, "grad_norm": 0.6549422144889832, "learning_rate": 9.497616029903994e-06, "loss": 0.6291, "step": 1069 }, { "epoch": 0.6894329896907216, "grad_norm": 0.6966941356658936, "learning_rate": 9.495976965178227e-06, "loss": 0.6317, "step": 1070 }, { "epoch": 0.6900773195876289, "grad_norm": 0.6863613724708557, "learning_rate": 9.49433537292573e-06, "loss": 0.6287, "step": 1071 }, { "epoch": 0.6907216494845361, "grad_norm": 0.5814676284790039, "learning_rate": 9.492691254069372e-06, "loss": 0.6316, "step": 1072 }, { "epoch": 0.6913659793814433, "grad_norm": 0.6514321565628052, "learning_rate": 9.491044609533431e-06, "loss": 0.6508, "step": 1073 }, { "epoch": 0.6920103092783505, "grad_norm": 0.6035879850387573, "learning_rate": 9.489395440243613e-06, "loss": 0.6217, "step": 1074 }, { "epoch": 0.6926546391752577, "grad_norm": 0.68366938829422, "learning_rate": 9.487743747127039e-06, "loss": 0.6043, "step": 1075 }, { "epoch": 0.6932989690721649, "grad_norm": 0.679527223110199, "learning_rate": 9.486089531112247e-06, "loss": 0.6302, "step": 1076 }, { "epoch": 0.6939432989690721, "grad_norm": 0.6752008199691772, "learning_rate": 9.4844327931292e-06, "loss": 0.6521, "step": 1077 }, { "epoch": 0.6945876288659794, "grad_norm": 0.6166187524795532, "learning_rate": 9.482773534109272e-06, "loss": 0.6581, "step": 1078 }, { "epoch": 0.6952319587628866, "grad_norm": 0.7203593254089355, "learning_rate": 9.481111754985262e-06, "loss": 0.6418, "step": 1079 }, { "epoch": 0.6958762886597938, "grad_norm": 0.656432032585144, "learning_rate": 9.479447456691376e-06, "loss": 0.6356, "step": 1080 }, { "epoch": 0.696520618556701, "grad_norm": 0.6752628684043884, "learning_rate": 9.477780640163242e-06, "loss": 0.6146, "step": 1081 }, { "epoch": 0.6971649484536082, "grad_norm": 0.6441132426261902, "learning_rate": 9.476111306337906e-06, "loss": 0.6296, "step": 1082 }, { "epoch": 0.6978092783505154, "grad_norm": 0.6800857186317444, "learning_rate": 9.474439456153824e-06, "loss": 0.6392, "step": 1083 }, { "epoch": 0.6984536082474226, "grad_norm": 0.6030648946762085, "learning_rate": 9.472765090550866e-06, "loss": 0.6447, "step": 1084 }, { "epoch": 0.6990979381443299, "grad_norm": 0.6066173315048218, "learning_rate": 9.471088210470325e-06, "loss": 0.6318, "step": 1085 }, { "epoch": 0.6997422680412371, "grad_norm": 0.6738420128822327, "learning_rate": 9.469408816854898e-06, "loss": 0.621, "step": 1086 }, { "epoch": 0.7003865979381443, "grad_norm": 0.701508641242981, "learning_rate": 9.467726910648696e-06, "loss": 0.6305, "step": 1087 }, { "epoch": 0.7010309278350515, "grad_norm": 0.679986298084259, "learning_rate": 9.466042492797247e-06, "loss": 0.6425, "step": 1088 }, { "epoch": 0.7016752577319587, "grad_norm": 0.5870832800865173, "learning_rate": 9.464355564247492e-06, "loss": 0.6264, "step": 1089 }, { "epoch": 0.7023195876288659, "grad_norm": 0.6965056657791138, "learning_rate": 9.462666125947776e-06, "loss": 0.6455, "step": 1090 }, { "epoch": 0.7029639175257731, "grad_norm": 0.681724488735199, "learning_rate": 9.460974178847861e-06, "loss": 0.6361, "step": 1091 }, { "epoch": 0.7036082474226805, "grad_norm": 0.7101638913154602, "learning_rate": 9.459279723898918e-06, "loss": 0.638, "step": 1092 }, { "epoch": 0.7042525773195877, "grad_norm": 0.5915218591690063, "learning_rate": 9.457582762053528e-06, "loss": 0.6281, "step": 1093 }, { "epoch": 0.7048969072164949, "grad_norm": 0.7152234315872192, "learning_rate": 9.45588329426568e-06, "loss": 0.6495, "step": 1094 }, { "epoch": 0.7055412371134021, "grad_norm": 0.6532676219940186, "learning_rate": 9.454181321490772e-06, "loss": 0.6238, "step": 1095 }, { "epoch": 0.7061855670103093, "grad_norm": 0.6351122856140137, "learning_rate": 9.452476844685611e-06, "loss": 0.6413, "step": 1096 }, { "epoch": 0.7068298969072165, "grad_norm": 0.6054962277412415, "learning_rate": 9.450769864808413e-06, "loss": 0.6336, "step": 1097 }, { "epoch": 0.7074742268041238, "grad_norm": 0.6233380436897278, "learning_rate": 9.449060382818798e-06, "loss": 0.6247, "step": 1098 }, { "epoch": 0.708118556701031, "grad_norm": 0.6613344550132751, "learning_rate": 9.447348399677797e-06, "loss": 0.6274, "step": 1099 }, { "epoch": 0.7087628865979382, "grad_norm": 0.646092414855957, "learning_rate": 9.445633916347842e-06, "loss": 0.6267, "step": 1100 }, { "epoch": 0.7094072164948454, "grad_norm": 0.6846522688865662, "learning_rate": 9.443916933792774e-06, "loss": 0.6284, "step": 1101 }, { "epoch": 0.7100515463917526, "grad_norm": 0.6655775308609009, "learning_rate": 9.442197452977836e-06, "loss": 0.6267, "step": 1102 }, { "epoch": 0.7106958762886598, "grad_norm": 0.6958496570587158, "learning_rate": 9.440475474869681e-06, "loss": 0.6208, "step": 1103 }, { "epoch": 0.711340206185567, "grad_norm": 0.8020801544189453, "learning_rate": 9.438751000436359e-06, "loss": 0.6353, "step": 1104 }, { "epoch": 0.7119845360824743, "grad_norm": 0.7182430028915405, "learning_rate": 9.437024030647331e-06, "loss": 0.6392, "step": 1105 }, { "epoch": 0.7126288659793815, "grad_norm": 0.7050268650054932, "learning_rate": 9.435294566473453e-06, "loss": 0.6379, "step": 1106 }, { "epoch": 0.7132731958762887, "grad_norm": 0.6995869278907776, "learning_rate": 9.433562608886986e-06, "loss": 0.6262, "step": 1107 }, { "epoch": 0.7139175257731959, "grad_norm": 0.628134548664093, "learning_rate": 9.431828158861595e-06, "loss": 0.6334, "step": 1108 }, { "epoch": 0.7145618556701031, "grad_norm": 0.8187612295150757, "learning_rate": 9.430091217372348e-06, "loss": 0.6395, "step": 1109 }, { "epoch": 0.7152061855670103, "grad_norm": 0.6395763754844666, "learning_rate": 9.428351785395707e-06, "loss": 0.6452, "step": 1110 }, { "epoch": 0.7158505154639175, "grad_norm": 0.7134526371955872, "learning_rate": 9.426609863909537e-06, "loss": 0.6327, "step": 1111 }, { "epoch": 0.7164948453608248, "grad_norm": 0.6984047889709473, "learning_rate": 9.424865453893104e-06, "loss": 0.6413, "step": 1112 }, { "epoch": 0.717139175257732, "grad_norm": 0.6183768510818481, "learning_rate": 9.423118556327073e-06, "loss": 0.6308, "step": 1113 }, { "epoch": 0.7177835051546392, "grad_norm": 0.6566948294639587, "learning_rate": 9.421369172193504e-06, "loss": 0.6384, "step": 1114 }, { "epoch": 0.7184278350515464, "grad_norm": 0.703401505947113, "learning_rate": 9.41961730247586e-06, "loss": 0.6438, "step": 1115 }, { "epoch": 0.7190721649484536, "grad_norm": 0.6829988956451416, "learning_rate": 9.417862948158997e-06, "loss": 0.6453, "step": 1116 }, { "epoch": 0.7197164948453608, "grad_norm": 0.7416354417800903, "learning_rate": 9.416106110229172e-06, "loss": 0.649, "step": 1117 }, { "epoch": 0.720360824742268, "grad_norm": 0.6525890231132507, "learning_rate": 9.414346789674034e-06, "loss": 0.6216, "step": 1118 }, { "epoch": 0.7210051546391752, "grad_norm": 0.6824215054512024, "learning_rate": 9.412584987482628e-06, "loss": 0.6275, "step": 1119 }, { "epoch": 0.7216494845360825, "grad_norm": 0.7183943390846252, "learning_rate": 9.410820704645397e-06, "loss": 0.6471, "step": 1120 }, { "epoch": 0.7222938144329897, "grad_norm": 0.6618180274963379, "learning_rate": 9.409053942154177e-06, "loss": 0.6239, "step": 1121 }, { "epoch": 0.7229381443298969, "grad_norm": 0.7154481410980225, "learning_rate": 9.407284701002197e-06, "loss": 0.6342, "step": 1122 }, { "epoch": 0.7235824742268041, "grad_norm": 0.7052229642868042, "learning_rate": 9.405512982184082e-06, "loss": 0.6609, "step": 1123 }, { "epoch": 0.7242268041237113, "grad_norm": 0.6403135061264038, "learning_rate": 9.403738786695848e-06, "loss": 0.6258, "step": 1124 }, { "epoch": 0.7248711340206185, "grad_norm": 0.8580516576766968, "learning_rate": 9.401962115534905e-06, "loss": 0.6311, "step": 1125 }, { "epoch": 0.7255154639175257, "grad_norm": 0.6727852821350098, "learning_rate": 9.40018296970005e-06, "loss": 0.6232, "step": 1126 }, { "epoch": 0.726159793814433, "grad_norm": 0.8249883055686951, "learning_rate": 9.398401350191475e-06, "loss": 0.6071, "step": 1127 }, { "epoch": 0.7268041237113402, "grad_norm": 0.9529392123222351, "learning_rate": 9.396617258010765e-06, "loss": 0.6331, "step": 1128 }, { "epoch": 0.7274484536082474, "grad_norm": 0.6339847445487976, "learning_rate": 9.394830694160892e-06, "loss": 0.6342, "step": 1129 }, { "epoch": 0.7280927835051546, "grad_norm": 0.9731364250183105, "learning_rate": 9.393041659646219e-06, "loss": 0.6465, "step": 1130 }, { "epoch": 0.7287371134020618, "grad_norm": 0.8255374431610107, "learning_rate": 9.391250155472492e-06, "loss": 0.6507, "step": 1131 }, { "epoch": 0.729381443298969, "grad_norm": 0.6622611284255981, "learning_rate": 9.389456182646853e-06, "loss": 0.6613, "step": 1132 }, { "epoch": 0.7300257731958762, "grad_norm": 0.853882908821106, "learning_rate": 9.38765974217783e-06, "loss": 0.6625, "step": 1133 }, { "epoch": 0.7306701030927835, "grad_norm": 0.7556077241897583, "learning_rate": 9.385860835075336e-06, "loss": 0.6344, "step": 1134 }, { "epoch": 0.7313144329896907, "grad_norm": 0.663593053817749, "learning_rate": 9.384059462350673e-06, "loss": 0.6381, "step": 1135 }, { "epoch": 0.7319587628865979, "grad_norm": 0.7425535917282104, "learning_rate": 9.382255625016527e-06, "loss": 0.6322, "step": 1136 }, { "epoch": 0.7326030927835051, "grad_norm": 0.6761476993560791, "learning_rate": 9.38044932408697e-06, "loss": 0.6553, "step": 1137 }, { "epoch": 0.7332474226804123, "grad_norm": 0.8109030723571777, "learning_rate": 9.378640560577461e-06, "loss": 0.6211, "step": 1138 }, { "epoch": 0.7338917525773195, "grad_norm": 0.8072793483734131, "learning_rate": 9.376829335504843e-06, "loss": 0.6103, "step": 1139 }, { "epoch": 0.7345360824742269, "grad_norm": 0.6076831817626953, "learning_rate": 9.37501564988734e-06, "loss": 0.6308, "step": 1140 }, { "epoch": 0.7351804123711341, "grad_norm": 0.9256844520568848, "learning_rate": 9.37319950474456e-06, "loss": 0.6475, "step": 1141 }, { "epoch": 0.7358247422680413, "grad_norm": 0.8054786324501038, "learning_rate": 9.371380901097498e-06, "loss": 0.6506, "step": 1142 }, { "epoch": 0.7364690721649485, "grad_norm": 0.7207463383674622, "learning_rate": 9.369559839968525e-06, "loss": 0.6335, "step": 1143 }, { "epoch": 0.7371134020618557, "grad_norm": 0.7570593953132629, "learning_rate": 9.367736322381399e-06, "loss": 0.6472, "step": 1144 }, { "epoch": 0.7377577319587629, "grad_norm": 0.5968641042709351, "learning_rate": 9.365910349361253e-06, "loss": 0.6464, "step": 1145 }, { "epoch": 0.7384020618556701, "grad_norm": 0.6915954947471619, "learning_rate": 9.364081921934607e-06, "loss": 0.6281, "step": 1146 }, { "epoch": 0.7390463917525774, "grad_norm": 0.6595852375030518, "learning_rate": 9.362251041129355e-06, "loss": 0.6275, "step": 1147 }, { "epoch": 0.7396907216494846, "grad_norm": 0.578289270401001, "learning_rate": 9.360417707974773e-06, "loss": 0.6499, "step": 1148 }, { "epoch": 0.7403350515463918, "grad_norm": 0.6894807815551758, "learning_rate": 9.358581923501517e-06, "loss": 0.649, "step": 1149 }, { "epoch": 0.740979381443299, "grad_norm": 0.6760059595108032, "learning_rate": 9.356743688741621e-06, "loss": 0.6374, "step": 1150 }, { "epoch": 0.7416237113402062, "grad_norm": 0.6723427176475525, "learning_rate": 9.354903004728491e-06, "loss": 0.6319, "step": 1151 }, { "epoch": 0.7422680412371134, "grad_norm": 0.648849368095398, "learning_rate": 9.353059872496914e-06, "loss": 0.6368, "step": 1152 }, { "epoch": 0.7429123711340206, "grad_norm": 0.6194808483123779, "learning_rate": 9.351214293083057e-06, "loss": 0.6388, "step": 1153 }, { "epoch": 0.7435567010309279, "grad_norm": 0.725946307182312, "learning_rate": 9.349366267524456e-06, "loss": 0.6421, "step": 1154 }, { "epoch": 0.7442010309278351, "grad_norm": 0.6427181959152222, "learning_rate": 9.347515796860027e-06, "loss": 0.663, "step": 1155 }, { "epoch": 0.7448453608247423, "grad_norm": 0.6697866916656494, "learning_rate": 9.345662882130056e-06, "loss": 0.6362, "step": 1156 }, { "epoch": 0.7454896907216495, "grad_norm": 0.6654358506202698, "learning_rate": 9.34380752437621e-06, "loss": 0.6335, "step": 1157 }, { "epoch": 0.7461340206185567, "grad_norm": 0.6664418578147888, "learning_rate": 9.34194972464152e-06, "loss": 0.6316, "step": 1158 }, { "epoch": 0.7467783505154639, "grad_norm": 0.763221025466919, "learning_rate": 9.340089483970398e-06, "loss": 0.6268, "step": 1159 }, { "epoch": 0.7474226804123711, "grad_norm": 0.6662298440933228, "learning_rate": 9.338226803408625e-06, "loss": 0.6224, "step": 1160 }, { "epoch": 0.7480670103092784, "grad_norm": 0.6292816996574402, "learning_rate": 9.336361684003353e-06, "loss": 0.6374, "step": 1161 }, { "epoch": 0.7487113402061856, "grad_norm": 0.5897899866104126, "learning_rate": 9.334494126803106e-06, "loss": 0.6353, "step": 1162 }, { "epoch": 0.7493556701030928, "grad_norm": 0.6881434321403503, "learning_rate": 9.33262413285778e-06, "loss": 0.6408, "step": 1163 }, { "epoch": 0.75, "grad_norm": 0.6135979294776917, "learning_rate": 9.330751703218636e-06, "loss": 0.629, "step": 1164 }, { "epoch": 0.7506443298969072, "grad_norm": 0.6195429563522339, "learning_rate": 9.328876838938312e-06, "loss": 0.6434, "step": 1165 }, { "epoch": 0.7512886597938144, "grad_norm": 0.5902673006057739, "learning_rate": 9.326999541070804e-06, "loss": 0.6372, "step": 1166 }, { "epoch": 0.7519329896907216, "grad_norm": 0.5894961953163147, "learning_rate": 9.325119810671488e-06, "loss": 0.6297, "step": 1167 }, { "epoch": 0.7525773195876289, "grad_norm": 0.6033127903938293, "learning_rate": 9.323237648797099e-06, "loss": 0.635, "step": 1168 }, { "epoch": 0.7532216494845361, "grad_norm": 0.6696355938911438, "learning_rate": 9.321353056505742e-06, "loss": 0.6345, "step": 1169 }, { "epoch": 0.7538659793814433, "grad_norm": 0.5878171324729919, "learning_rate": 9.31946603485689e-06, "loss": 0.6349, "step": 1170 }, { "epoch": 0.7545103092783505, "grad_norm": 0.641944169998169, "learning_rate": 9.317576584911376e-06, "loss": 0.6427, "step": 1171 }, { "epoch": 0.7551546391752577, "grad_norm": 0.6330858469009399, "learning_rate": 9.315684707731405e-06, "loss": 0.6425, "step": 1172 }, { "epoch": 0.7557989690721649, "grad_norm": 0.7213586568832397, "learning_rate": 9.313790404380542e-06, "loss": 0.6681, "step": 1173 }, { "epoch": 0.7564432989690721, "grad_norm": 0.6121848225593567, "learning_rate": 9.311893675923717e-06, "loss": 0.6331, "step": 1174 }, { "epoch": 0.7570876288659794, "grad_norm": 0.6929645538330078, "learning_rate": 9.309994523427226e-06, "loss": 0.6249, "step": 1175 }, { "epoch": 0.7577319587628866, "grad_norm": 0.5917505025863647, "learning_rate": 9.308092947958725e-06, "loss": 0.6424, "step": 1176 }, { "epoch": 0.7583762886597938, "grad_norm": 0.6805191040039062, "learning_rate": 9.30618895058723e-06, "loss": 0.6302, "step": 1177 }, { "epoch": 0.759020618556701, "grad_norm": 0.7083525061607361, "learning_rate": 9.30428253238312e-06, "loss": 0.6299, "step": 1178 }, { "epoch": 0.7596649484536082, "grad_norm": 0.6037490367889404, "learning_rate": 9.302373694418143e-06, "loss": 0.6378, "step": 1179 }, { "epoch": 0.7603092783505154, "grad_norm": 0.7738140821456909, "learning_rate": 9.300462437765394e-06, "loss": 0.6497, "step": 1180 }, { "epoch": 0.7609536082474226, "grad_norm": 0.6869704127311707, "learning_rate": 9.298548763499336e-06, "loss": 0.6434, "step": 1181 }, { "epoch": 0.7615979381443299, "grad_norm": 0.7071198225021362, "learning_rate": 9.296632672695788e-06, "loss": 0.6368, "step": 1182 }, { "epoch": 0.7622422680412371, "grad_norm": 0.6063685417175293, "learning_rate": 9.294714166431931e-06, "loss": 0.6359, "step": 1183 }, { "epoch": 0.7628865979381443, "grad_norm": 0.777434766292572, "learning_rate": 9.2927932457863e-06, "loss": 0.6461, "step": 1184 }, { "epoch": 0.7635309278350515, "grad_norm": 0.7584032416343689, "learning_rate": 9.29086991183879e-06, "loss": 0.6413, "step": 1185 }, { "epoch": 0.7641752577319587, "grad_norm": 0.7103497385978699, "learning_rate": 9.288944165670651e-06, "loss": 0.6424, "step": 1186 }, { "epoch": 0.7648195876288659, "grad_norm": 0.7027865052223206, "learning_rate": 9.28701600836449e-06, "loss": 0.6358, "step": 1187 }, { "epoch": 0.7654639175257731, "grad_norm": 0.6008939743041992, "learning_rate": 9.28508544100427e-06, "loss": 0.6142, "step": 1188 }, { "epoch": 0.7661082474226805, "grad_norm": 0.6687167882919312, "learning_rate": 9.283152464675305e-06, "loss": 0.6291, "step": 1189 }, { "epoch": 0.7667525773195877, "grad_norm": 0.6580998301506042, "learning_rate": 9.281217080464271e-06, "loss": 0.6251, "step": 1190 }, { "epoch": 0.7673969072164949, "grad_norm": 0.6297574639320374, "learning_rate": 9.279279289459188e-06, "loss": 0.6375, "step": 1191 }, { "epoch": 0.7680412371134021, "grad_norm": 0.7317301630973816, "learning_rate": 9.277339092749437e-06, "loss": 0.6342, "step": 1192 }, { "epoch": 0.7686855670103093, "grad_norm": 0.6630852818489075, "learning_rate": 9.27539649142575e-06, "loss": 0.6263, "step": 1193 }, { "epoch": 0.7693298969072165, "grad_norm": 0.6121829748153687, "learning_rate": 9.273451486580206e-06, "loss": 0.6356, "step": 1194 }, { "epoch": 0.7699742268041238, "grad_norm": 0.647666335105896, "learning_rate": 9.27150407930624e-06, "loss": 0.6287, "step": 1195 }, { "epoch": 0.770618556701031, "grad_norm": 0.6869192123413086, "learning_rate": 9.269554270698636e-06, "loss": 0.6209, "step": 1196 }, { "epoch": 0.7712628865979382, "grad_norm": 0.5346981287002563, "learning_rate": 9.267602061853527e-06, "loss": 0.6346, "step": 1197 }, { "epoch": 0.7719072164948454, "grad_norm": 0.6869325637817383, "learning_rate": 9.265647453868398e-06, "loss": 0.6367, "step": 1198 }, { "epoch": 0.7725515463917526, "grad_norm": 0.6395169496536255, "learning_rate": 9.26369044784208e-06, "loss": 0.6239, "step": 1199 }, { "epoch": 0.7731958762886598, "grad_norm": 0.5956645011901855, "learning_rate": 9.261731044874753e-06, "loss": 0.6241, "step": 1200 }, { "epoch": 0.773840206185567, "grad_norm": 0.6869899034500122, "learning_rate": 9.259769246067946e-06, "loss": 0.6229, "step": 1201 }, { "epoch": 0.7744845360824743, "grad_norm": 0.6170216202735901, "learning_rate": 9.257805052524533e-06, "loss": 0.6298, "step": 1202 }, { "epoch": 0.7751288659793815, "grad_norm": 0.5844255685806274, "learning_rate": 9.255838465348735e-06, "loss": 0.6149, "step": 1203 }, { "epoch": 0.7757731958762887, "grad_norm": 0.6804046630859375, "learning_rate": 9.253869485646118e-06, "loss": 0.6311, "step": 1204 }, { "epoch": 0.7764175257731959, "grad_norm": 0.6878448724746704, "learning_rate": 9.251898114523594e-06, "loss": 0.6422, "step": 1205 }, { "epoch": 0.7770618556701031, "grad_norm": 0.5971963405609131, "learning_rate": 9.24992435308942e-06, "loss": 0.6153, "step": 1206 }, { "epoch": 0.7777061855670103, "grad_norm": 0.6268125176429749, "learning_rate": 9.247948202453193e-06, "loss": 0.629, "step": 1207 }, { "epoch": 0.7783505154639175, "grad_norm": 0.6019789576530457, "learning_rate": 9.245969663725857e-06, "loss": 0.6188, "step": 1208 }, { "epoch": 0.7789948453608248, "grad_norm": 0.6480079293251038, "learning_rate": 9.2439887380197e-06, "loss": 0.6274, "step": 1209 }, { "epoch": 0.779639175257732, "grad_norm": 0.5775524973869324, "learning_rate": 9.242005426448345e-06, "loss": 0.6177, "step": 1210 }, { "epoch": 0.7802835051546392, "grad_norm": 0.5800431370735168, "learning_rate": 9.240019730126764e-06, "loss": 0.6401, "step": 1211 }, { "epoch": 0.7809278350515464, "grad_norm": 0.6220250725746155, "learning_rate": 9.238031650171266e-06, "loss": 0.6169, "step": 1212 }, { "epoch": 0.7815721649484536, "grad_norm": 0.6265754103660583, "learning_rate": 9.236041187699499e-06, "loss": 0.6588, "step": 1213 }, { "epoch": 0.7822164948453608, "grad_norm": 0.5856478810310364, "learning_rate": 9.23404834383045e-06, "loss": 0.6425, "step": 1214 }, { "epoch": 0.782860824742268, "grad_norm": 0.5323241949081421, "learning_rate": 9.232053119684448e-06, "loss": 0.6222, "step": 1215 }, { "epoch": 0.7835051546391752, "grad_norm": 0.5987434983253479, "learning_rate": 9.23005551638316e-06, "loss": 0.6232, "step": 1216 }, { "epoch": 0.7841494845360825, "grad_norm": 0.6116014122962952, "learning_rate": 9.228055535049587e-06, "loss": 0.6249, "step": 1217 }, { "epoch": 0.7847938144329897, "grad_norm": 0.5936328768730164, "learning_rate": 9.226053176808071e-06, "loss": 0.6264, "step": 1218 }, { "epoch": 0.7854381443298969, "grad_norm": 0.6615980267524719, "learning_rate": 9.224048442784286e-06, "loss": 0.6295, "step": 1219 }, { "epoch": 0.7860824742268041, "grad_norm": 0.6283928751945496, "learning_rate": 9.222041334105244e-06, "loss": 0.6559, "step": 1220 }, { "epoch": 0.7867268041237113, "grad_norm": 0.5877431631088257, "learning_rate": 9.220031851899293e-06, "loss": 0.6006, "step": 1221 }, { "epoch": 0.7873711340206185, "grad_norm": 0.6209008097648621, "learning_rate": 9.218019997296112e-06, "loss": 0.6353, "step": 1222 }, { "epoch": 0.7880154639175257, "grad_norm": 0.6559786796569824, "learning_rate": 9.216005771426718e-06, "loss": 0.643, "step": 1223 }, { "epoch": 0.788659793814433, "grad_norm": 0.5746191143989563, "learning_rate": 9.21398917542346e-06, "loss": 0.6579, "step": 1224 }, { "epoch": 0.7893041237113402, "grad_norm": 0.7109358906745911, "learning_rate": 9.211970210420013e-06, "loss": 0.6348, "step": 1225 }, { "epoch": 0.7899484536082474, "grad_norm": 0.6633077263832092, "learning_rate": 9.209948877551393e-06, "loss": 0.6159, "step": 1226 }, { "epoch": 0.7905927835051546, "grad_norm": 0.6718607544898987, "learning_rate": 9.207925177953942e-06, "loss": 0.6168, "step": 1227 }, { "epoch": 0.7912371134020618, "grad_norm": 0.6582311391830444, "learning_rate": 9.205899112765335e-06, "loss": 0.6524, "step": 1228 }, { "epoch": 0.791881443298969, "grad_norm": 0.6267915964126587, "learning_rate": 9.203870683124574e-06, "loss": 0.6236, "step": 1229 }, { "epoch": 0.7925257731958762, "grad_norm": 0.6537051200866699, "learning_rate": 9.201839890171994e-06, "loss": 0.6426, "step": 1230 }, { "epoch": 0.7931701030927835, "grad_norm": 0.6531317830085754, "learning_rate": 9.199806735049253e-06, "loss": 0.6506, "step": 1231 }, { "epoch": 0.7938144329896907, "grad_norm": 0.5801550149917603, "learning_rate": 9.197771218899342e-06, "loss": 0.6252, "step": 1232 }, { "epoch": 0.7944587628865979, "grad_norm": 0.6708964109420776, "learning_rate": 9.19573334286658e-06, "loss": 0.649, "step": 1233 }, { "epoch": 0.7951030927835051, "grad_norm": 0.7125882506370544, "learning_rate": 9.193693108096606e-06, "loss": 0.6232, "step": 1234 }, { "epoch": 0.7957474226804123, "grad_norm": 0.6477491855621338, "learning_rate": 9.191650515736395e-06, "loss": 0.6178, "step": 1235 }, { "epoch": 0.7963917525773195, "grad_norm": 0.7643678188323975, "learning_rate": 9.189605566934235e-06, "loss": 0.6304, "step": 1236 }, { "epoch": 0.7970360824742269, "grad_norm": 0.626975417137146, "learning_rate": 9.187558262839751e-06, "loss": 0.641, "step": 1237 }, { "epoch": 0.7976804123711341, "grad_norm": 0.6939786672592163, "learning_rate": 9.185508604603884e-06, "loss": 0.6326, "step": 1238 }, { "epoch": 0.7983247422680413, "grad_norm": 0.6059709191322327, "learning_rate": 9.183456593378901e-06, "loss": 0.6103, "step": 1239 }, { "epoch": 0.7989690721649485, "grad_norm": 0.7225053906440735, "learning_rate": 9.181402230318393e-06, "loss": 0.6244, "step": 1240 }, { "epoch": 0.7996134020618557, "grad_norm": 0.7372803688049316, "learning_rate": 9.179345516577271e-06, "loss": 0.6461, "step": 1241 }, { "epoch": 0.8002577319587629, "grad_norm": 0.5972424149513245, "learning_rate": 9.177286453311768e-06, "loss": 0.6009, "step": 1242 }, { "epoch": 0.8009020618556701, "grad_norm": 0.6514745354652405, "learning_rate": 9.17522504167944e-06, "loss": 0.6283, "step": 1243 }, { "epoch": 0.8015463917525774, "grad_norm": 0.7760336995124817, "learning_rate": 9.173161282839162e-06, "loss": 0.625, "step": 1244 }, { "epoch": 0.8021907216494846, "grad_norm": 0.7245351076126099, "learning_rate": 9.171095177951125e-06, "loss": 0.6279, "step": 1245 }, { "epoch": 0.8028350515463918, "grad_norm": 0.668097734451294, "learning_rate": 9.169026728176845e-06, "loss": 0.6598, "step": 1246 }, { "epoch": 0.803479381443299, "grad_norm": 0.7276373505592346, "learning_rate": 9.166955934679149e-06, "loss": 0.6277, "step": 1247 }, { "epoch": 0.8041237113402062, "grad_norm": 0.6909542679786682, "learning_rate": 9.16488279862219e-06, "loss": 0.6581, "step": 1248 }, { "epoch": 0.8047680412371134, "grad_norm": 0.697545051574707, "learning_rate": 9.162807321171431e-06, "loss": 0.6383, "step": 1249 }, { "epoch": 0.8054123711340206, "grad_norm": 0.7005974054336548, "learning_rate": 9.160729503493656e-06, "loss": 0.654, "step": 1250 }, { "epoch": 0.8060567010309279, "grad_norm": 0.7004820704460144, "learning_rate": 9.158649346756962e-06, "loss": 0.6482, "step": 1251 }, { "epoch": 0.8067010309278351, "grad_norm": 0.7356707453727722, "learning_rate": 9.15656685213076e-06, "loss": 0.6398, "step": 1252 }, { "epoch": 0.8073453608247423, "grad_norm": 0.6297895908355713, "learning_rate": 9.154482020785778e-06, "loss": 0.6414, "step": 1253 }, { "epoch": 0.8079896907216495, "grad_norm": 0.629740297794342, "learning_rate": 9.152394853894057e-06, "loss": 0.6414, "step": 1254 }, { "epoch": 0.8086340206185567, "grad_norm": 0.6084477305412292, "learning_rate": 9.15030535262895e-06, "loss": 0.653, "step": 1255 }, { "epoch": 0.8092783505154639, "grad_norm": 0.5940077900886536, "learning_rate": 9.148213518165121e-06, "loss": 0.6304, "step": 1256 }, { "epoch": 0.8099226804123711, "grad_norm": 0.7205477356910706, "learning_rate": 9.146119351678551e-06, "loss": 0.6454, "step": 1257 }, { "epoch": 0.8105670103092784, "grad_norm": 0.6184560656547546, "learning_rate": 9.144022854346526e-06, "loss": 0.6176, "step": 1258 }, { "epoch": 0.8112113402061856, "grad_norm": 0.7526706457138062, "learning_rate": 9.141924027347643e-06, "loss": 0.6302, "step": 1259 }, { "epoch": 0.8118556701030928, "grad_norm": 0.6701117753982544, "learning_rate": 9.139822871861816e-06, "loss": 0.6083, "step": 1260 }, { "epoch": 0.8125, "grad_norm": 0.6877208352088928, "learning_rate": 9.137719389070259e-06, "loss": 0.6534, "step": 1261 }, { "epoch": 0.8131443298969072, "grad_norm": 0.7126523852348328, "learning_rate": 9.135613580155495e-06, "loss": 0.6511, "step": 1262 }, { "epoch": 0.8137886597938144, "grad_norm": 0.5888392925262451, "learning_rate": 9.133505446301364e-06, "loss": 0.6275, "step": 1263 }, { "epoch": 0.8144329896907216, "grad_norm": 0.7268720269203186, "learning_rate": 9.131394988693001e-06, "loss": 0.6344, "step": 1264 }, { "epoch": 0.8150773195876289, "grad_norm": 0.7011691927909851, "learning_rate": 9.129282208516857e-06, "loss": 0.6224, "step": 1265 }, { "epoch": 0.8157216494845361, "grad_norm": 0.5642725229263306, "learning_rate": 9.127167106960682e-06, "loss": 0.6238, "step": 1266 }, { "epoch": 0.8163659793814433, "grad_norm": 0.7066255807876587, "learning_rate": 9.125049685213534e-06, "loss": 0.6396, "step": 1267 }, { "epoch": 0.8170103092783505, "grad_norm": 0.6752985715866089, "learning_rate": 9.122929944465775e-06, "loss": 0.6293, "step": 1268 }, { "epoch": 0.8176546391752577, "grad_norm": 0.5502704977989197, "learning_rate": 9.120807885909073e-06, "loss": 0.6361, "step": 1269 }, { "epoch": 0.8182989690721649, "grad_norm": 0.549164354801178, "learning_rate": 9.118683510736393e-06, "loss": 0.6323, "step": 1270 }, { "epoch": 0.8189432989690721, "grad_norm": 0.6615608930587769, "learning_rate": 9.116556820142007e-06, "loss": 0.643, "step": 1271 }, { "epoch": 0.8195876288659794, "grad_norm": 0.760270893573761, "learning_rate": 9.11442781532149e-06, "loss": 0.6185, "step": 1272 }, { "epoch": 0.8202319587628866, "grad_norm": 0.6817582845687866, "learning_rate": 9.112296497471713e-06, "loss": 0.6305, "step": 1273 }, { "epoch": 0.8208762886597938, "grad_norm": 0.6941832900047302, "learning_rate": 9.110162867790852e-06, "loss": 0.6372, "step": 1274 }, { "epoch": 0.821520618556701, "grad_norm": 0.6526524424552917, "learning_rate": 9.10802692747838e-06, "loss": 0.6264, "step": 1275 }, { "epoch": 0.8221649484536082, "grad_norm": 0.573545515537262, "learning_rate": 9.105888677735069e-06, "loss": 0.6006, "step": 1276 }, { "epoch": 0.8228092783505154, "grad_norm": 0.5860809087753296, "learning_rate": 9.10374811976299e-06, "loss": 0.6193, "step": 1277 }, { "epoch": 0.8234536082474226, "grad_norm": 0.7927619218826294, "learning_rate": 9.101605254765513e-06, "loss": 0.626, "step": 1278 }, { "epoch": 0.8240979381443299, "grad_norm": 0.6322084069252014, "learning_rate": 9.099460083947303e-06, "loss": 0.6412, "step": 1279 }, { "epoch": 0.8247422680412371, "grad_norm": 0.6806204319000244, "learning_rate": 9.097312608514319e-06, "loss": 0.6372, "step": 1280 }, { "epoch": 0.8253865979381443, "grad_norm": 0.6651644110679626, "learning_rate": 9.095162829673824e-06, "loss": 0.6522, "step": 1281 }, { "epoch": 0.8260309278350515, "grad_norm": 0.5957874059677124, "learning_rate": 9.093010748634366e-06, "loss": 0.6338, "step": 1282 }, { "epoch": 0.8266752577319587, "grad_norm": 0.6260101795196533, "learning_rate": 9.090856366605791e-06, "loss": 0.6377, "step": 1283 }, { "epoch": 0.8273195876288659, "grad_norm": 0.6074230074882507, "learning_rate": 9.088699684799243e-06, "loss": 0.6439, "step": 1284 }, { "epoch": 0.8279639175257731, "grad_norm": 0.6371315717697144, "learning_rate": 9.086540704427152e-06, "loss": 0.6377, "step": 1285 }, { "epoch": 0.8286082474226805, "grad_norm": 0.6877110600471497, "learning_rate": 9.084379426703245e-06, "loss": 0.6476, "step": 1286 }, { "epoch": 0.8292525773195877, "grad_norm": 0.5881201028823853, "learning_rate": 9.082215852842538e-06, "loss": 0.6294, "step": 1287 }, { "epoch": 0.8298969072164949, "grad_norm": 0.6913458108901978, "learning_rate": 9.080049984061337e-06, "loss": 0.6463, "step": 1288 }, { "epoch": 0.8305412371134021, "grad_norm": 0.6276844143867493, "learning_rate": 9.077881821577242e-06, "loss": 0.6394, "step": 1289 }, { "epoch": 0.8311855670103093, "grad_norm": 0.6117535829544067, "learning_rate": 9.07571136660914e-06, "loss": 0.6296, "step": 1290 }, { "epoch": 0.8318298969072165, "grad_norm": 0.6249837279319763, "learning_rate": 9.073538620377203e-06, "loss": 0.6336, "step": 1291 }, { "epoch": 0.8324742268041238, "grad_norm": 0.6574905514717102, "learning_rate": 9.071363584102902e-06, "loss": 0.6208, "step": 1292 }, { "epoch": 0.833118556701031, "grad_norm": 0.5752551555633545, "learning_rate": 9.069186259008983e-06, "loss": 0.6266, "step": 1293 }, { "epoch": 0.8337628865979382, "grad_norm": 0.6885551810264587, "learning_rate": 9.067006646319484e-06, "loss": 0.6457, "step": 1294 }, { "epoch": 0.8344072164948454, "grad_norm": 0.6401488184928894, "learning_rate": 9.064824747259734e-06, "loss": 0.6468, "step": 1295 }, { "epoch": 0.8350515463917526, "grad_norm": 0.6018393039703369, "learning_rate": 9.062640563056339e-06, "loss": 0.6399, "step": 1296 }, { "epoch": 0.8356958762886598, "grad_norm": 0.5771751999855042, "learning_rate": 9.060454094937194e-06, "loss": 0.6236, "step": 1297 }, { "epoch": 0.836340206185567, "grad_norm": 0.6653033494949341, "learning_rate": 9.058265344131475e-06, "loss": 0.635, "step": 1298 }, { "epoch": 0.8369845360824743, "grad_norm": 0.6165934801101685, "learning_rate": 9.056074311869649e-06, "loss": 0.6584, "step": 1299 }, { "epoch": 0.8376288659793815, "grad_norm": 0.5951228737831116, "learning_rate": 9.053880999383453e-06, "loss": 0.6283, "step": 1300 }, { "epoch": 0.8382731958762887, "grad_norm": 0.6772807836532593, "learning_rate": 9.051685407905916e-06, "loss": 0.6497, "step": 1301 }, { "epoch": 0.8389175257731959, "grad_norm": 0.5966616868972778, "learning_rate": 9.049487538671348e-06, "loss": 0.6315, "step": 1302 }, { "epoch": 0.8395618556701031, "grad_norm": 0.6960387825965881, "learning_rate": 9.047287392915333e-06, "loss": 0.6561, "step": 1303 }, { "epoch": 0.8402061855670103, "grad_norm": 0.7062965631484985, "learning_rate": 9.045084971874738e-06, "loss": 0.6294, "step": 1304 }, { "epoch": 0.8408505154639175, "grad_norm": 0.6865085363388062, "learning_rate": 9.04288027678771e-06, "loss": 0.627, "step": 1305 }, { "epoch": 0.8414948453608248, "grad_norm": 0.7023534178733826, "learning_rate": 9.040673308893677e-06, "loss": 0.6293, "step": 1306 }, { "epoch": 0.842139175257732, "grad_norm": 0.6442386507987976, "learning_rate": 9.038464069433337e-06, "loss": 0.6203, "step": 1307 }, { "epoch": 0.8427835051546392, "grad_norm": 0.6743494272232056, "learning_rate": 9.036252559648672e-06, "loss": 0.6162, "step": 1308 }, { "epoch": 0.8434278350515464, "grad_norm": 0.6483396887779236, "learning_rate": 9.034038780782939e-06, "loss": 0.6151, "step": 1309 }, { "epoch": 0.8440721649484536, "grad_norm": 0.6018944382667542, "learning_rate": 9.031822734080669e-06, "loss": 0.6369, "step": 1310 }, { "epoch": 0.8447164948453608, "grad_norm": 0.7228425145149231, "learning_rate": 9.029604420787666e-06, "loss": 0.6247, "step": 1311 }, { "epoch": 0.845360824742268, "grad_norm": 0.6157804727554321, "learning_rate": 9.027383842151011e-06, "loss": 0.633, "step": 1312 }, { "epoch": 0.8460051546391752, "grad_norm": 0.5487989783287048, "learning_rate": 9.025160999419062e-06, "loss": 0.6181, "step": 1313 }, { "epoch": 0.8466494845360825, "grad_norm": 0.5861343741416931, "learning_rate": 9.022935893841445e-06, "loss": 0.6585, "step": 1314 }, { "epoch": 0.8472938144329897, "grad_norm": 0.6552554965019226, "learning_rate": 9.020708526669056e-06, "loss": 0.6243, "step": 1315 }, { "epoch": 0.8479381443298969, "grad_norm": 0.6097061634063721, "learning_rate": 9.018478899154068e-06, "loss": 0.6464, "step": 1316 }, { "epoch": 0.8485824742268041, "grad_norm": 0.7119239568710327, "learning_rate": 9.016247012549922e-06, "loss": 0.6535, "step": 1317 }, { "epoch": 0.8492268041237113, "grad_norm": 0.641550600528717, "learning_rate": 9.014012868111326e-06, "loss": 0.6426, "step": 1318 }, { "epoch": 0.8498711340206185, "grad_norm": 0.5911536812782288, "learning_rate": 9.011776467094267e-06, "loss": 0.6405, "step": 1319 }, { "epoch": 0.8505154639175257, "grad_norm": 0.616237223148346, "learning_rate": 9.00953781075599e-06, "loss": 0.6209, "step": 1320 }, { "epoch": 0.851159793814433, "grad_norm": 0.7206349968910217, "learning_rate": 9.007296900355013e-06, "loss": 0.6392, "step": 1321 }, { "epoch": 0.8518041237113402, "grad_norm": 0.644565761089325, "learning_rate": 9.00505373715112e-06, "loss": 0.6465, "step": 1322 }, { "epoch": 0.8524484536082474, "grad_norm": 0.5719360113143921, "learning_rate": 9.002808322405363e-06, "loss": 0.6376, "step": 1323 }, { "epoch": 0.8530927835051546, "grad_norm": 0.6766362190246582, "learning_rate": 9.000560657380055e-06, "loss": 0.6398, "step": 1324 }, { "epoch": 0.8537371134020618, "grad_norm": 0.6746644377708435, "learning_rate": 8.998310743338781e-06, "loss": 0.6257, "step": 1325 }, { "epoch": 0.854381443298969, "grad_norm": 0.6285490989685059, "learning_rate": 8.996058581546386e-06, "loss": 0.6502, "step": 1326 }, { "epoch": 0.8550257731958762, "grad_norm": 0.6991211771965027, "learning_rate": 8.993804173268978e-06, "loss": 0.6274, "step": 1327 }, { "epoch": 0.8556701030927835, "grad_norm": 0.8534426093101501, "learning_rate": 8.991547519773933e-06, "loss": 0.6326, "step": 1328 }, { "epoch": 0.8563144329896907, "grad_norm": 0.5743288397789001, "learning_rate": 8.989288622329881e-06, "loss": 0.6348, "step": 1329 }, { "epoch": 0.8569587628865979, "grad_norm": 0.7626894116401672, "learning_rate": 8.987027482206722e-06, "loss": 0.6148, "step": 1330 }, { "epoch": 0.8576030927835051, "grad_norm": 0.755601167678833, "learning_rate": 8.98476410067561e-06, "loss": 0.627, "step": 1331 }, { "epoch": 0.8582474226804123, "grad_norm": 0.5494433045387268, "learning_rate": 8.982498479008963e-06, "loss": 0.6242, "step": 1332 }, { "epoch": 0.8588917525773195, "grad_norm": 0.6186935901641846, "learning_rate": 8.980230618480459e-06, "loss": 0.602, "step": 1333 }, { "epoch": 0.8595360824742269, "grad_norm": 0.7979156970977783, "learning_rate": 8.97796052036503e-06, "loss": 0.6219, "step": 1334 }, { "epoch": 0.8601804123711341, "grad_norm": 0.5860368013381958, "learning_rate": 8.975688185938872e-06, "loss": 0.6109, "step": 1335 }, { "epoch": 0.8608247422680413, "grad_norm": 0.7756735682487488, "learning_rate": 8.973413616479429e-06, "loss": 0.6373, "step": 1336 }, { "epoch": 0.8614690721649485, "grad_norm": 0.6802050471305847, "learning_rate": 8.971136813265416e-06, "loss": 0.6444, "step": 1337 }, { "epoch": 0.8621134020618557, "grad_norm": 0.6538388729095459, "learning_rate": 8.96885777757679e-06, "loss": 0.6173, "step": 1338 }, { "epoch": 0.8627577319587629, "grad_norm": 0.6921644806861877, "learning_rate": 8.966576510694769e-06, "loss": 0.6482, "step": 1339 }, { "epoch": 0.8634020618556701, "grad_norm": 0.8006290793418884, "learning_rate": 8.964293013901826e-06, "loss": 0.6487, "step": 1340 }, { "epoch": 0.8640463917525774, "grad_norm": 0.7534888982772827, "learning_rate": 8.962007288481684e-06, "loss": 0.6261, "step": 1341 }, { "epoch": 0.8646907216494846, "grad_norm": 0.7748348712921143, "learning_rate": 8.959719335719324e-06, "loss": 0.6553, "step": 1342 }, { "epoch": 0.8653350515463918, "grad_norm": 0.7349157333374023, "learning_rate": 8.957429156900974e-06, "loss": 0.632, "step": 1343 }, { "epoch": 0.865979381443299, "grad_norm": 0.6257424354553223, "learning_rate": 8.955136753314117e-06, "loss": 0.627, "step": 1344 }, { "epoch": 0.8666237113402062, "grad_norm": 0.6956115961074829, "learning_rate": 8.952842126247486e-06, "loss": 0.634, "step": 1345 }, { "epoch": 0.8672680412371134, "grad_norm": 0.6520851850509644, "learning_rate": 8.95054527699106e-06, "loss": 0.6214, "step": 1346 }, { "epoch": 0.8679123711340206, "grad_norm": 0.5806514620780945, "learning_rate": 8.948246206836073e-06, "loss": 0.6212, "step": 1347 }, { "epoch": 0.8685567010309279, "grad_norm": 0.6330692768096924, "learning_rate": 8.945944917075007e-06, "loss": 0.6314, "step": 1348 }, { "epoch": 0.8692010309278351, "grad_norm": 0.693496584892273, "learning_rate": 8.943641409001587e-06, "loss": 0.6427, "step": 1349 }, { "epoch": 0.8698453608247423, "grad_norm": 0.6225863695144653, "learning_rate": 8.94133568391079e-06, "loss": 0.6237, "step": 1350 }, { "epoch": 0.8704896907216495, "grad_norm": 0.6819723844528198, "learning_rate": 8.939027743098838e-06, "loss": 0.6257, "step": 1351 }, { "epoch": 0.8711340206185567, "grad_norm": 0.6456234455108643, "learning_rate": 8.936717587863195e-06, "loss": 0.6344, "step": 1352 }, { "epoch": 0.8717783505154639, "grad_norm": 0.6102128028869629, "learning_rate": 8.934405219502575e-06, "loss": 0.6427, "step": 1353 }, { "epoch": 0.8724226804123711, "grad_norm": 0.6515916585922241, "learning_rate": 8.932090639316936e-06, "loss": 0.6363, "step": 1354 }, { "epoch": 0.8730670103092784, "grad_norm": 0.7615947723388672, "learning_rate": 8.929773848607475e-06, "loss": 0.6094, "step": 1355 }, { "epoch": 0.8737113402061856, "grad_norm": 0.6025731563568115, "learning_rate": 8.927454848676633e-06, "loss": 0.6274, "step": 1356 }, { "epoch": 0.8743556701030928, "grad_norm": 0.5699487924575806, "learning_rate": 8.925133640828098e-06, "loss": 0.6494, "step": 1357 }, { "epoch": 0.875, "grad_norm": 0.7110111713409424, "learning_rate": 8.922810226366791e-06, "loss": 0.6185, "step": 1358 }, { "epoch": 0.8756443298969072, "grad_norm": 0.690891444683075, "learning_rate": 8.920484606598883e-06, "loss": 0.6168, "step": 1359 }, { "epoch": 0.8762886597938144, "grad_norm": 0.5997257232666016, "learning_rate": 8.918156782831777e-06, "loss": 0.6467, "step": 1360 }, { "epoch": 0.8769329896907216, "grad_norm": 0.6461016535758972, "learning_rate": 8.915826756374118e-06, "loss": 0.6184, "step": 1361 }, { "epoch": 0.8775773195876289, "grad_norm": 0.6262242794036865, "learning_rate": 8.913494528535788e-06, "loss": 0.6153, "step": 1362 }, { "epoch": 0.8782216494845361, "grad_norm": 0.6815193891525269, "learning_rate": 8.91116010062791e-06, "loss": 0.6379, "step": 1363 }, { "epoch": 0.8788659793814433, "grad_norm": 0.6638871431350708, "learning_rate": 8.90882347396284e-06, "loss": 0.6396, "step": 1364 }, { "epoch": 0.8795103092783505, "grad_norm": 0.711566150188446, "learning_rate": 8.906484649854172e-06, "loss": 0.6569, "step": 1365 }, { "epoch": 0.8801546391752577, "grad_norm": 0.6691820025444031, "learning_rate": 8.904143629616735e-06, "loss": 0.6411, "step": 1366 }, { "epoch": 0.8807989690721649, "grad_norm": 0.7313368320465088, "learning_rate": 8.90180041456659e-06, "loss": 0.6503, "step": 1367 }, { "epoch": 0.8814432989690721, "grad_norm": 0.5888932943344116, "learning_rate": 8.899455006021037e-06, "loss": 0.6229, "step": 1368 }, { "epoch": 0.8820876288659794, "grad_norm": 0.6746011972427368, "learning_rate": 8.897107405298608e-06, "loss": 0.6094, "step": 1369 }, { "epoch": 0.8827319587628866, "grad_norm": 0.7311916947364807, "learning_rate": 8.894757613719064e-06, "loss": 0.6212, "step": 1370 }, { "epoch": 0.8833762886597938, "grad_norm": 0.5725377798080444, "learning_rate": 8.892405632603397e-06, "loss": 0.6519, "step": 1371 }, { "epoch": 0.884020618556701, "grad_norm": 0.6139603853225708, "learning_rate": 8.89005146327384e-06, "loss": 0.628, "step": 1372 }, { "epoch": 0.8846649484536082, "grad_norm": 0.6665766835212708, "learning_rate": 8.887695107053837e-06, "loss": 0.5992, "step": 1373 }, { "epoch": 0.8853092783505154, "grad_norm": 0.5948442816734314, "learning_rate": 8.885336565268084e-06, "loss": 0.6379, "step": 1374 }, { "epoch": 0.8859536082474226, "grad_norm": 0.639834463596344, "learning_rate": 8.88297583924249e-06, "loss": 0.6392, "step": 1375 }, { "epoch": 0.8865979381443299, "grad_norm": 0.6827385425567627, "learning_rate": 8.880612930304196e-06, "loss": 0.6415, "step": 1376 }, { "epoch": 0.8872422680412371, "grad_norm": 0.737837016582489, "learning_rate": 8.87824783978157e-06, "loss": 0.6192, "step": 1377 }, { "epoch": 0.8878865979381443, "grad_norm": 0.688591718673706, "learning_rate": 8.87588056900421e-06, "loss": 0.6354, "step": 1378 }, { "epoch": 0.8885309278350515, "grad_norm": 0.659796416759491, "learning_rate": 8.873511119302938e-06, "loss": 0.6267, "step": 1379 }, { "epoch": 0.8891752577319587, "grad_norm": 0.666473925113678, "learning_rate": 8.871139492009796e-06, "loss": 0.6557, "step": 1380 }, { "epoch": 0.8898195876288659, "grad_norm": 0.5870363116264343, "learning_rate": 8.868765688458056e-06, "loss": 0.6339, "step": 1381 }, { "epoch": 0.8904639175257731, "grad_norm": 0.6369422674179077, "learning_rate": 8.86638970998221e-06, "loss": 0.6471, "step": 1382 }, { "epoch": 0.8911082474226805, "grad_norm": 0.6492503881454468, "learning_rate": 8.864011557917977e-06, "loss": 0.6203, "step": 1383 }, { "epoch": 0.8917525773195877, "grad_norm": 0.6294646859169006, "learning_rate": 8.861631233602293e-06, "loss": 0.6374, "step": 1384 }, { "epoch": 0.8923969072164949, "grad_norm": 0.672601044178009, "learning_rate": 8.859248738373316e-06, "loss": 0.6445, "step": 1385 }, { "epoch": 0.8930412371134021, "grad_norm": 0.767219066619873, "learning_rate": 8.856864073570429e-06, "loss": 0.626, "step": 1386 }, { "epoch": 0.8936855670103093, "grad_norm": 0.6164502501487732, "learning_rate": 8.85447724053423e-06, "loss": 0.6321, "step": 1387 }, { "epoch": 0.8943298969072165, "grad_norm": 0.7088932991027832, "learning_rate": 8.852088240606537e-06, "loss": 0.6552, "step": 1388 }, { "epoch": 0.8949742268041238, "grad_norm": 0.6432206630706787, "learning_rate": 8.849697075130387e-06, "loss": 0.6297, "step": 1389 }, { "epoch": 0.895618556701031, "grad_norm": 0.6774529814720154, "learning_rate": 8.847303745450033e-06, "loss": 0.6108, "step": 1390 }, { "epoch": 0.8962628865979382, "grad_norm": 0.6912225484848022, "learning_rate": 8.844908252910947e-06, "loss": 0.6283, "step": 1391 }, { "epoch": 0.8969072164948454, "grad_norm": 0.6528540849685669, "learning_rate": 8.842510598859813e-06, "loss": 0.6292, "step": 1392 }, { "epoch": 0.8975515463917526, "grad_norm": 0.747994065284729, "learning_rate": 8.840110784644535e-06, "loss": 0.6213, "step": 1393 }, { "epoch": 0.8981958762886598, "grad_norm": 0.7140428423881531, "learning_rate": 8.837708811614228e-06, "loss": 0.6351, "step": 1394 }, { "epoch": 0.898840206185567, "grad_norm": 0.6564962863922119, "learning_rate": 8.835304681119223e-06, "loss": 0.6379, "step": 1395 }, { "epoch": 0.8994845360824743, "grad_norm": 0.720706045627594, "learning_rate": 8.83289839451106e-06, "loss": 0.647, "step": 1396 }, { "epoch": 0.9001288659793815, "grad_norm": 0.7378793954849243, "learning_rate": 8.830489953142497e-06, "loss": 0.6353, "step": 1397 }, { "epoch": 0.9007731958762887, "grad_norm": 0.5812186002731323, "learning_rate": 8.828079358367497e-06, "loss": 0.6101, "step": 1398 }, { "epoch": 0.9014175257731959, "grad_norm": 0.6637722253799438, "learning_rate": 8.825666611541238e-06, "loss": 0.6291, "step": 1399 }, { "epoch": 0.9020618556701031, "grad_norm": 0.7379080653190613, "learning_rate": 8.823251714020106e-06, "loss": 0.6398, "step": 1400 }, { "epoch": 0.9027061855670103, "grad_norm": 0.6158630847930908, "learning_rate": 8.820834667161698e-06, "loss": 0.6299, "step": 1401 }, { "epoch": 0.9033505154639175, "grad_norm": 0.7135244011878967, "learning_rate": 8.818415472324814e-06, "loss": 0.6368, "step": 1402 }, { "epoch": 0.9039948453608248, "grad_norm": 0.6369975805282593, "learning_rate": 8.815994130869469e-06, "loss": 0.6172, "step": 1403 }, { "epoch": 0.904639175257732, "grad_norm": 0.7419471144676208, "learning_rate": 8.813570644156879e-06, "loss": 0.6369, "step": 1404 }, { "epoch": 0.9052835051546392, "grad_norm": 0.633112907409668, "learning_rate": 8.811145013549469e-06, "loss": 0.6364, "step": 1405 }, { "epoch": 0.9059278350515464, "grad_norm": 0.8166525959968567, "learning_rate": 8.80871724041087e-06, "loss": 0.6422, "step": 1406 }, { "epoch": 0.9065721649484536, "grad_norm": 0.6880943179130554, "learning_rate": 8.806287326105913e-06, "loss": 0.6343, "step": 1407 }, { "epoch": 0.9072164948453608, "grad_norm": 0.7599040269851685, "learning_rate": 8.803855272000637e-06, "loss": 0.6239, "step": 1408 }, { "epoch": 0.907860824742268, "grad_norm": 0.6690489053726196, "learning_rate": 8.801421079462283e-06, "loss": 0.6393, "step": 1409 }, { "epoch": 0.9085051546391752, "grad_norm": 0.6866702437400818, "learning_rate": 8.79898474985929e-06, "loss": 0.6352, "step": 1410 }, { "epoch": 0.9091494845360825, "grad_norm": 0.6156057119369507, "learning_rate": 8.796546284561307e-06, "loss": 0.6456, "step": 1411 }, { "epoch": 0.9097938144329897, "grad_norm": 0.8203875422477722, "learning_rate": 8.794105684939178e-06, "loss": 0.6341, "step": 1412 }, { "epoch": 0.9104381443298969, "grad_norm": 0.6302875280380249, "learning_rate": 8.791662952364944e-06, "loss": 0.6006, "step": 1413 }, { "epoch": 0.9110824742268041, "grad_norm": 0.7221728563308716, "learning_rate": 8.78921808821185e-06, "loss": 0.6294, "step": 1414 }, { "epoch": 0.9117268041237113, "grad_norm": 0.7059857845306396, "learning_rate": 8.786771093854342e-06, "loss": 0.616, "step": 1415 }, { "epoch": 0.9123711340206185, "grad_norm": 0.725558876991272, "learning_rate": 8.784321970668054e-06, "loss": 0.6291, "step": 1416 }, { "epoch": 0.9130154639175257, "grad_norm": 0.6105697154998779, "learning_rate": 8.781870720029826e-06, "loss": 0.6225, "step": 1417 }, { "epoch": 0.913659793814433, "grad_norm": 0.6322675943374634, "learning_rate": 8.779417343317687e-06, "loss": 0.6341, "step": 1418 }, { "epoch": 0.9143041237113402, "grad_norm": 0.6295060515403748, "learning_rate": 8.776961841910869e-06, "loss": 0.6468, "step": 1419 }, { "epoch": 0.9149484536082474, "grad_norm": 0.5917021036148071, "learning_rate": 8.77450421718979e-06, "loss": 0.6199, "step": 1420 }, { "epoch": 0.9155927835051546, "grad_norm": 0.6251276135444641, "learning_rate": 8.772044470536067e-06, "loss": 0.6103, "step": 1421 }, { "epoch": 0.9162371134020618, "grad_norm": 0.6316661834716797, "learning_rate": 8.769582603332509e-06, "loss": 0.6347, "step": 1422 }, { "epoch": 0.916881443298969, "grad_norm": 0.7327718138694763, "learning_rate": 8.767118616963117e-06, "loss": 0.6155, "step": 1423 }, { "epoch": 0.9175257731958762, "grad_norm": 0.6367208361625671, "learning_rate": 8.764652512813082e-06, "loss": 0.6198, "step": 1424 }, { "epoch": 0.9181701030927835, "grad_norm": 0.6724476218223572, "learning_rate": 8.762184292268788e-06, "loss": 0.6227, "step": 1425 }, { "epoch": 0.9188144329896907, "grad_norm": 0.7400143146514893, "learning_rate": 8.759713956717804e-06, "loss": 0.6144, "step": 1426 }, { "epoch": 0.9194587628865979, "grad_norm": 0.5900826454162598, "learning_rate": 8.757241507548895e-06, "loss": 0.6369, "step": 1427 }, { "epoch": 0.9201030927835051, "grad_norm": 0.7563096284866333, "learning_rate": 8.754766946152009e-06, "loss": 0.6326, "step": 1428 }, { "epoch": 0.9207474226804123, "grad_norm": 0.6724210381507874, "learning_rate": 8.752290273918285e-06, "loss": 0.6047, "step": 1429 }, { "epoch": 0.9213917525773195, "grad_norm": 0.6104978919029236, "learning_rate": 8.749811492240043e-06, "loss": 0.6429, "step": 1430 }, { "epoch": 0.9220360824742269, "grad_norm": 0.6784375309944153, "learning_rate": 8.747330602510796e-06, "loss": 0.6258, "step": 1431 }, { "epoch": 0.9226804123711341, "grad_norm": 0.6390827298164368, "learning_rate": 8.744847606125235e-06, "loss": 0.632, "step": 1432 }, { "epoch": 0.9233247422680413, "grad_norm": 0.6358007192611694, "learning_rate": 8.742362504479244e-06, "loss": 0.6291, "step": 1433 }, { "epoch": 0.9239690721649485, "grad_norm": 0.7355823516845703, "learning_rate": 8.739875298969881e-06, "loss": 0.6197, "step": 1434 }, { "epoch": 0.9246134020618557, "grad_norm": 0.7125774621963501, "learning_rate": 8.737385990995395e-06, "loss": 0.6319, "step": 1435 }, { "epoch": 0.9252577319587629, "grad_norm": 0.8087342977523804, "learning_rate": 8.734894581955208e-06, "loss": 0.6329, "step": 1436 }, { "epoch": 0.9259020618556701, "grad_norm": 0.6417868137359619, "learning_rate": 8.732401073249934e-06, "loss": 0.6274, "step": 1437 }, { "epoch": 0.9265463917525774, "grad_norm": 0.6716688871383667, "learning_rate": 8.729905466281359e-06, "loss": 0.6422, "step": 1438 }, { "epoch": 0.9271907216494846, "grad_norm": 0.7770051956176758, "learning_rate": 8.727407762452452e-06, "loss": 0.6321, "step": 1439 }, { "epoch": 0.9278350515463918, "grad_norm": 0.6159130930900574, "learning_rate": 8.724907963167357e-06, "loss": 0.6165, "step": 1440 }, { "epoch": 0.928479381443299, "grad_norm": 0.5580605864524841, "learning_rate": 8.722406069831403e-06, "loss": 0.6381, "step": 1441 }, { "epoch": 0.9291237113402062, "grad_norm": 0.596988320350647, "learning_rate": 8.719902083851095e-06, "loss": 0.633, "step": 1442 }, { "epoch": 0.9297680412371134, "grad_norm": 0.5962273478507996, "learning_rate": 8.717396006634104e-06, "loss": 0.6505, "step": 1443 }, { "epoch": 0.9304123711340206, "grad_norm": 0.6372034549713135, "learning_rate": 8.71488783958929e-06, "loss": 0.6009, "step": 1444 }, { "epoch": 0.9310567010309279, "grad_norm": 0.6673142313957214, "learning_rate": 8.71237758412668e-06, "loss": 0.6395, "step": 1445 }, { "epoch": 0.9317010309278351, "grad_norm": 0.6965785026550293, "learning_rate": 8.70986524165748e-06, "loss": 0.6149, "step": 1446 }, { "epoch": 0.9323453608247423, "grad_norm": 0.754651665687561, "learning_rate": 8.707350813594064e-06, "loss": 0.6206, "step": 1447 }, { "epoch": 0.9329896907216495, "grad_norm": 0.7132344245910645, "learning_rate": 8.704834301349982e-06, "loss": 0.6387, "step": 1448 }, { "epoch": 0.9336340206185567, "grad_norm": 0.7555209994316101, "learning_rate": 8.702315706339954e-06, "loss": 0.6078, "step": 1449 }, { "epoch": 0.9342783505154639, "grad_norm": 0.7580965161323547, "learning_rate": 8.699795029979871e-06, "loss": 0.6135, "step": 1450 }, { "epoch": 0.9349226804123711, "grad_norm": 0.7330949902534485, "learning_rate": 8.697272273686798e-06, "loss": 0.6433, "step": 1451 }, { "epoch": 0.9355670103092784, "grad_norm": 0.6449919939041138, "learning_rate": 8.694747438878965e-06, "loss": 0.6342, "step": 1452 }, { "epoch": 0.9362113402061856, "grad_norm": 0.7653663158416748, "learning_rate": 8.69222052697577e-06, "loss": 0.6112, "step": 1453 }, { "epoch": 0.9368556701030928, "grad_norm": 0.7962027192115784, "learning_rate": 8.689691539397778e-06, "loss": 0.6206, "step": 1454 }, { "epoch": 0.9375, "grad_norm": 0.6870254278182983, "learning_rate": 8.68716047756673e-06, "loss": 0.6431, "step": 1455 }, { "epoch": 0.9381443298969072, "grad_norm": 0.7381996512413025, "learning_rate": 8.684627342905519e-06, "loss": 0.6359, "step": 1456 }, { "epoch": 0.9387886597938144, "grad_norm": 0.6400518417358398, "learning_rate": 8.682092136838214e-06, "loss": 0.5937, "step": 1457 }, { "epoch": 0.9394329896907216, "grad_norm": 0.7030803561210632, "learning_rate": 8.679554860790045e-06, "loss": 0.6087, "step": 1458 }, { "epoch": 0.9400773195876289, "grad_norm": 0.8021583557128906, "learning_rate": 8.677015516187404e-06, "loss": 0.6243, "step": 1459 }, { "epoch": 0.9407216494845361, "grad_norm": 0.6371223330497742, "learning_rate": 8.67447410445785e-06, "loss": 0.5904, "step": 1460 }, { "epoch": 0.9413659793814433, "grad_norm": 0.7774003148078918, "learning_rate": 8.6719306270301e-06, "loss": 0.6285, "step": 1461 }, { "epoch": 0.9420103092783505, "grad_norm": 0.6610166430473328, "learning_rate": 8.669385085334034e-06, "loss": 0.6019, "step": 1462 }, { "epoch": 0.9426546391752577, "grad_norm": 0.6110290884971619, "learning_rate": 8.666837480800691e-06, "loss": 0.6287, "step": 1463 }, { "epoch": 0.9432989690721649, "grad_norm": 0.7102707624435425, "learning_rate": 8.664287814862276e-06, "loss": 0.6168, "step": 1464 }, { "epoch": 0.9439432989690721, "grad_norm": 0.6994916200637817, "learning_rate": 8.661736088952141e-06, "loss": 0.6197, "step": 1465 }, { "epoch": 0.9445876288659794, "grad_norm": 0.6474416255950928, "learning_rate": 8.65918230450481e-06, "loss": 0.6344, "step": 1466 }, { "epoch": 0.9452319587628866, "grad_norm": 0.6552828550338745, "learning_rate": 8.656626462955951e-06, "loss": 0.6289, "step": 1467 }, { "epoch": 0.9458762886597938, "grad_norm": 0.5672217607498169, "learning_rate": 8.654068565742399e-06, "loss": 0.6033, "step": 1468 }, { "epoch": 0.946520618556701, "grad_norm": 0.6580991744995117, "learning_rate": 8.65150861430214e-06, "loss": 0.6229, "step": 1469 }, { "epoch": 0.9471649484536082, "grad_norm": 0.6503843069076538, "learning_rate": 8.648946610074312e-06, "loss": 0.6344, "step": 1470 }, { "epoch": 0.9478092783505154, "grad_norm": 0.681161105632782, "learning_rate": 8.646382554499212e-06, "loss": 0.6401, "step": 1471 }, { "epoch": 0.9484536082474226, "grad_norm": 0.6704784631729126, "learning_rate": 8.643816449018292e-06, "loss": 0.6008, "step": 1472 }, { "epoch": 0.9490979381443299, "grad_norm": 0.6512477397918701, "learning_rate": 8.641248295074147e-06, "loss": 0.6161, "step": 1473 }, { "epoch": 0.9497422680412371, "grad_norm": 0.5624778270721436, "learning_rate": 8.638678094110533e-06, "loss": 0.6165, "step": 1474 }, { "epoch": 0.9503865979381443, "grad_norm": 0.8310864567756653, "learning_rate": 8.636105847572354e-06, "loss": 0.6315, "step": 1475 }, { "epoch": 0.9510309278350515, "grad_norm": 0.6216065287590027, "learning_rate": 8.63353155690566e-06, "loss": 0.6206, "step": 1476 }, { "epoch": 0.9516752577319587, "grad_norm": 0.6350827217102051, "learning_rate": 8.630955223557653e-06, "loss": 0.6217, "step": 1477 }, { "epoch": 0.9523195876288659, "grad_norm": 0.6727639436721802, "learning_rate": 8.628376848976688e-06, "loss": 0.6135, "step": 1478 }, { "epoch": 0.9529639175257731, "grad_norm": 0.6734524965286255, "learning_rate": 8.625796434612262e-06, "loss": 0.6462, "step": 1479 }, { "epoch": 0.9536082474226805, "grad_norm": 0.6236312389373779, "learning_rate": 8.623213981915015e-06, "loss": 0.6213, "step": 1480 }, { "epoch": 0.9542525773195877, "grad_norm": 0.7418127059936523, "learning_rate": 8.620629492336745e-06, "loss": 0.6445, "step": 1481 }, { "epoch": 0.9548969072164949, "grad_norm": 0.647305965423584, "learning_rate": 8.618042967330384e-06, "loss": 0.6541, "step": 1482 }, { "epoch": 0.9555412371134021, "grad_norm": 0.7332753539085388, "learning_rate": 8.615454408350012e-06, "loss": 0.6442, "step": 1483 }, { "epoch": 0.9561855670103093, "grad_norm": 0.6507708430290222, "learning_rate": 8.612863816850853e-06, "loss": 0.6176, "step": 1484 }, { "epoch": 0.9568298969072165, "grad_norm": 0.6407703161239624, "learning_rate": 8.610271194289275e-06, "loss": 0.6215, "step": 1485 }, { "epoch": 0.9574742268041238, "grad_norm": 0.7398317456245422, "learning_rate": 8.607676542122782e-06, "loss": 0.6082, "step": 1486 }, { "epoch": 0.958118556701031, "grad_norm": 0.6126739978790283, "learning_rate": 8.605079861810029e-06, "loss": 0.6121, "step": 1487 }, { "epoch": 0.9587628865979382, "grad_norm": 0.6356568932533264, "learning_rate": 8.602481154810803e-06, "loss": 0.6186, "step": 1488 }, { "epoch": 0.9594072164948454, "grad_norm": 0.6487914323806763, "learning_rate": 8.599880422586029e-06, "loss": 0.6038, "step": 1489 }, { "epoch": 0.9600515463917526, "grad_norm": 0.6306661367416382, "learning_rate": 8.597277666597776e-06, "loss": 0.6395, "step": 1490 }, { "epoch": 0.9606958762886598, "grad_norm": 0.6028286218643188, "learning_rate": 8.594672888309252e-06, "loss": 0.6098, "step": 1491 }, { "epoch": 0.961340206185567, "grad_norm": 0.6481478810310364, "learning_rate": 8.592066089184796e-06, "loss": 0.6372, "step": 1492 }, { "epoch": 0.9619845360824743, "grad_norm": 0.5717189908027649, "learning_rate": 8.589457270689887e-06, "loss": 0.6247, "step": 1493 }, { "epoch": 0.9626288659793815, "grad_norm": 0.5669311881065369, "learning_rate": 8.586846434291136e-06, "loss": 0.6287, "step": 1494 }, { "epoch": 0.9632731958762887, "grad_norm": 0.5625038743019104, "learning_rate": 8.584233581456294e-06, "loss": 0.6203, "step": 1495 }, { "epoch": 0.9639175257731959, "grad_norm": 0.5797958374023438, "learning_rate": 8.581618713654239e-06, "loss": 0.6249, "step": 1496 }, { "epoch": 0.9645618556701031, "grad_norm": 0.5655548572540283, "learning_rate": 8.579001832354986e-06, "loss": 0.6182, "step": 1497 }, { "epoch": 0.9652061855670103, "grad_norm": 0.59474778175354, "learning_rate": 8.576382939029677e-06, "loss": 0.6271, "step": 1498 }, { "epoch": 0.9658505154639175, "grad_norm": 0.5697249174118042, "learning_rate": 8.573762035150595e-06, "loss": 0.6247, "step": 1499 }, { "epoch": 0.9664948453608248, "grad_norm": 0.7169575691223145, "learning_rate": 8.571139122191145e-06, "loss": 0.6273, "step": 1500 }, { "epoch": 0.967139175257732, "grad_norm": 0.612878680229187, "learning_rate": 8.568514201625862e-06, "loss": 0.6757, "step": 1501 }, { "epoch": 0.9677835051546392, "grad_norm": 0.5841500163078308, "learning_rate": 8.565887274930413e-06, "loss": 0.6153, "step": 1502 }, { "epoch": 0.9684278350515464, "grad_norm": 0.6811551451683044, "learning_rate": 8.563258343581591e-06, "loss": 0.6119, "step": 1503 }, { "epoch": 0.9690721649484536, "grad_norm": 0.6028275489807129, "learning_rate": 8.560627409057314e-06, "loss": 0.6164, "step": 1504 }, { "epoch": 0.9697164948453608, "grad_norm": 0.5996479392051697, "learning_rate": 8.557994472836631e-06, "loss": 0.6287, "step": 1505 }, { "epoch": 0.970360824742268, "grad_norm": 0.6174697279930115, "learning_rate": 8.55535953639971e-06, "loss": 0.6209, "step": 1506 }, { "epoch": 0.9710051546391752, "grad_norm": 0.7079206109046936, "learning_rate": 8.552722601227849e-06, "loss": 0.6319, "step": 1507 }, { "epoch": 0.9716494845360825, "grad_norm": 0.6217359900474548, "learning_rate": 8.550083668803468e-06, "loss": 0.6401, "step": 1508 }, { "epoch": 0.9722938144329897, "grad_norm": 0.7058160305023193, "learning_rate": 8.547442740610109e-06, "loss": 0.6139, "step": 1509 }, { "epoch": 0.9729381443298969, "grad_norm": 0.7916936278343201, "learning_rate": 8.544799818132434e-06, "loss": 0.6265, "step": 1510 }, { "epoch": 0.9735824742268041, "grad_norm": 0.6902171969413757, "learning_rate": 8.542154902856232e-06, "loss": 0.6367, "step": 1511 }, { "epoch": 0.9742268041237113, "grad_norm": 0.6656110882759094, "learning_rate": 8.539507996268409e-06, "loss": 0.6227, "step": 1512 }, { "epoch": 0.9748711340206185, "grad_norm": 0.710136353969574, "learning_rate": 8.536859099856984e-06, "loss": 0.6261, "step": 1513 }, { "epoch": 0.9755154639175257, "grad_norm": 0.6649792194366455, "learning_rate": 8.534208215111109e-06, "loss": 0.6343, "step": 1514 }, { "epoch": 0.976159793814433, "grad_norm": 0.622371256351471, "learning_rate": 8.53155534352104e-06, "loss": 0.6077, "step": 1515 }, { "epoch": 0.9768041237113402, "grad_norm": 0.7487042546272278, "learning_rate": 8.528900486578158e-06, "loss": 0.6446, "step": 1516 }, { "epoch": 0.9774484536082474, "grad_norm": 0.6053404211997986, "learning_rate": 8.526243645774959e-06, "loss": 0.6169, "step": 1517 }, { "epoch": 0.9780927835051546, "grad_norm": 0.6313574314117432, "learning_rate": 8.52358482260505e-06, "loss": 0.6121, "step": 1518 }, { "epoch": 0.9787371134020618, "grad_norm": 0.7067204713821411, "learning_rate": 8.520924018563156e-06, "loss": 0.6052, "step": 1519 }, { "epoch": 0.979381443298969, "grad_norm": 0.6128863096237183, "learning_rate": 8.518261235145117e-06, "loss": 0.6088, "step": 1520 }, { "epoch": 0.9800257731958762, "grad_norm": 0.6884070038795471, "learning_rate": 8.515596473847882e-06, "loss": 0.6302, "step": 1521 }, { "epoch": 0.9806701030927835, "grad_norm": 0.6993927955627441, "learning_rate": 8.512929736169513e-06, "loss": 0.6297, "step": 1522 }, { "epoch": 0.9813144329896907, "grad_norm": 0.6248343586921692, "learning_rate": 8.510261023609187e-06, "loss": 0.6149, "step": 1523 }, { "epoch": 0.9819587628865979, "grad_norm": 0.6966699361801147, "learning_rate": 8.507590337667185e-06, "loss": 0.6433, "step": 1524 }, { "epoch": 0.9826030927835051, "grad_norm": 0.6148841381072998, "learning_rate": 8.5049176798449e-06, "loss": 0.6198, "step": 1525 }, { "epoch": 0.9832474226804123, "grad_norm": 0.7284646034240723, "learning_rate": 8.502243051644838e-06, "loss": 0.6268, "step": 1526 }, { "epoch": 0.9838917525773195, "grad_norm": 0.6387619972229004, "learning_rate": 8.499566454570605e-06, "loss": 0.6216, "step": 1527 }, { "epoch": 0.9845360824742269, "grad_norm": 0.5855134129524231, "learning_rate": 8.496887890126917e-06, "loss": 0.6136, "step": 1528 }, { "epoch": 0.9851804123711341, "grad_norm": 0.8115233182907104, "learning_rate": 8.4942073598196e-06, "loss": 0.6442, "step": 1529 }, { "epoch": 0.9858247422680413, "grad_norm": 0.6333781480789185, "learning_rate": 8.491524865155579e-06, "loss": 0.6417, "step": 1530 }, { "epoch": 0.9864690721649485, "grad_norm": 0.646173357963562, "learning_rate": 8.488840407642885e-06, "loss": 0.6384, "step": 1531 }, { "epoch": 0.9871134020618557, "grad_norm": 0.6546114683151245, "learning_rate": 8.486153988790657e-06, "loss": 0.6133, "step": 1532 }, { "epoch": 0.9877577319587629, "grad_norm": 0.5937528610229492, "learning_rate": 8.48346561010913e-06, "loss": 0.6127, "step": 1533 }, { "epoch": 0.9884020618556701, "grad_norm": 0.6373299360275269, "learning_rate": 8.480775273109643e-06, "loss": 0.6332, "step": 1534 }, { "epoch": 0.9890463917525774, "grad_norm": 0.5954495668411255, "learning_rate": 8.478082979304642e-06, "loss": 0.6218, "step": 1535 }, { "epoch": 0.9896907216494846, "grad_norm": 0.686431348323822, "learning_rate": 8.475388730207662e-06, "loss": 0.6005, "step": 1536 }, { "epoch": 0.9903350515463918, "grad_norm": 0.6426213979721069, "learning_rate": 8.472692527333345e-06, "loss": 0.6409, "step": 1537 }, { "epoch": 0.990979381443299, "grad_norm": 0.5912914276123047, "learning_rate": 8.469994372197431e-06, "loss": 0.6075, "step": 1538 }, { "epoch": 0.9916237113402062, "grad_norm": 0.6084652543067932, "learning_rate": 8.467294266316755e-06, "loss": 0.6295, "step": 1539 }, { "epoch": 0.9922680412371134, "grad_norm": 0.6249634623527527, "learning_rate": 8.464592211209248e-06, "loss": 0.6224, "step": 1540 }, { "epoch": 0.9929123711340206, "grad_norm": 0.6904609203338623, "learning_rate": 8.46188820839394e-06, "loss": 0.6264, "step": 1541 }, { "epoch": 0.9935567010309279, "grad_norm": 0.5945738554000854, "learning_rate": 8.459182259390954e-06, "loss": 0.6149, "step": 1542 }, { "epoch": 0.9942010309278351, "grad_norm": 0.5944358110427856, "learning_rate": 8.456474365721508e-06, "loss": 0.619, "step": 1543 }, { "epoch": 0.9948453608247423, "grad_norm": 0.6204288005828857, "learning_rate": 8.453764528907911e-06, "loss": 0.6381, "step": 1544 }, { "epoch": 0.9954896907216495, "grad_norm": 0.5949960947036743, "learning_rate": 8.451052750473568e-06, "loss": 0.6455, "step": 1545 }, { "epoch": 0.9961340206185567, "grad_norm": 0.6642283797264099, "learning_rate": 8.44833903194297e-06, "loss": 0.6525, "step": 1546 }, { "epoch": 0.9967783505154639, "grad_norm": 0.6218704581260681, "learning_rate": 8.445623374841705e-06, "loss": 0.61, "step": 1547 }, { "epoch": 0.9974226804123711, "grad_norm": 0.6486876606941223, "learning_rate": 8.44290578069645e-06, "loss": 0.6146, "step": 1548 }, { "epoch": 0.9980670103092784, "grad_norm": 0.7156948447227478, "learning_rate": 8.440186251034966e-06, "loss": 0.6489, "step": 1549 }, { "epoch": 0.9987113402061856, "grad_norm": 0.5796813368797302, "learning_rate": 8.437464787386105e-06, "loss": 0.6046, "step": 1550 }, { "epoch": 0.9993556701030928, "grad_norm": 0.6534943580627441, "learning_rate": 8.434741391279809e-06, "loss": 0.6185, "step": 1551 }, { "epoch": 1.0, "grad_norm": 0.6907224059104919, "learning_rate": 8.432016064247099e-06, "loss": 0.6397, "step": 1552 }, { "epoch": 1.0006443298969072, "grad_norm": 0.6651949286460876, "learning_rate": 8.429288807820092e-06, "loss": 0.5789, "step": 1553 }, { "epoch": 1.0012886597938144, "grad_norm": 0.6989370584487915, "learning_rate": 8.42655962353198e-06, "loss": 0.5851, "step": 1554 }, { "epoch": 1.0019329896907216, "grad_norm": 0.6211574077606201, "learning_rate": 8.423828512917044e-06, "loss": 0.5816, "step": 1555 }, { "epoch": 1.0025773195876289, "grad_norm": 0.563105583190918, "learning_rate": 8.421095477510648e-06, "loss": 0.5651, "step": 1556 }, { "epoch": 1.003221649484536, "grad_norm": 0.7259016633033752, "learning_rate": 8.418360518849233e-06, "loss": 0.6017, "step": 1557 }, { "epoch": 1.0038659793814433, "grad_norm": 0.5652526021003723, "learning_rate": 8.415623638470328e-06, "loss": 0.5699, "step": 1558 }, { "epoch": 1.0045103092783505, "grad_norm": 0.5854575634002686, "learning_rate": 8.412884837912539e-06, "loss": 0.5747, "step": 1559 }, { "epoch": 1.0051546391752577, "grad_norm": 0.6120525598526001, "learning_rate": 8.41014411871555e-06, "loss": 0.581, "step": 1560 }, { "epoch": 1.005798969072165, "grad_norm": 0.6588024497032166, "learning_rate": 8.407401482420129e-06, "loss": 0.5757, "step": 1561 }, { "epoch": 1.0064432989690721, "grad_norm": 0.5976163148880005, "learning_rate": 8.404656930568113e-06, "loss": 0.5894, "step": 1562 }, { "epoch": 1.0070876288659794, "grad_norm": 0.604404091835022, "learning_rate": 8.401910464702425e-06, "loss": 0.5975, "step": 1563 }, { "epoch": 1.0077319587628866, "grad_norm": 0.6331729292869568, "learning_rate": 8.399162086367056e-06, "loss": 0.5594, "step": 1564 }, { "epoch": 1.0083762886597938, "grad_norm": 0.5600042939186096, "learning_rate": 8.396411797107082e-06, "loss": 0.5853, "step": 1565 }, { "epoch": 1.009020618556701, "grad_norm": 0.6132659912109375, "learning_rate": 8.393659598468644e-06, "loss": 0.5838, "step": 1566 }, { "epoch": 1.0096649484536082, "grad_norm": 0.58404541015625, "learning_rate": 8.390905491998957e-06, "loss": 0.5745, "step": 1567 }, { "epoch": 1.0103092783505154, "grad_norm": 0.6773462295532227, "learning_rate": 8.388149479246316e-06, "loss": 0.5819, "step": 1568 }, { "epoch": 1.0109536082474226, "grad_norm": 0.5442595481872559, "learning_rate": 8.38539156176008e-06, "loss": 0.5649, "step": 1569 }, { "epoch": 1.0115979381443299, "grad_norm": 0.576801061630249, "learning_rate": 8.382631741090686e-06, "loss": 0.5631, "step": 1570 }, { "epoch": 1.012242268041237, "grad_norm": 0.5792935490608215, "learning_rate": 8.379870018789632e-06, "loss": 0.5742, "step": 1571 }, { "epoch": 1.0128865979381443, "grad_norm": 0.562649667263031, "learning_rate": 8.377106396409491e-06, "loss": 0.5751, "step": 1572 }, { "epoch": 1.0135309278350515, "grad_norm": 0.5557625889778137, "learning_rate": 8.374340875503905e-06, "loss": 0.5732, "step": 1573 }, { "epoch": 1.0141752577319587, "grad_norm": 0.5549947619438171, "learning_rate": 8.371573457627577e-06, "loss": 0.5816, "step": 1574 }, { "epoch": 1.014819587628866, "grad_norm": 0.556542158126831, "learning_rate": 8.368804144336287e-06, "loss": 0.581, "step": 1575 }, { "epoch": 1.0154639175257731, "grad_norm": 0.5656431913375854, "learning_rate": 8.366032937186869e-06, "loss": 0.571, "step": 1576 }, { "epoch": 1.0161082474226804, "grad_norm": 0.6035471558570862, "learning_rate": 8.363259837737228e-06, "loss": 0.5435, "step": 1577 }, { "epoch": 1.0167525773195876, "grad_norm": 0.6120109558105469, "learning_rate": 8.360484847546332e-06, "loss": 0.5843, "step": 1578 }, { "epoch": 1.0173969072164948, "grad_norm": 0.5975964069366455, "learning_rate": 8.357707968174214e-06, "loss": 0.5621, "step": 1579 }, { "epoch": 1.018041237113402, "grad_norm": 0.5442563891410828, "learning_rate": 8.354929201181963e-06, "loss": 0.5621, "step": 1580 }, { "epoch": 1.0186855670103092, "grad_norm": 0.6110741496086121, "learning_rate": 8.352148548131736e-06, "loss": 0.5623, "step": 1581 }, { "epoch": 1.0193298969072164, "grad_norm": 0.5796793699264526, "learning_rate": 8.349366010586743e-06, "loss": 0.5715, "step": 1582 }, { "epoch": 1.0199742268041236, "grad_norm": 0.6133847832679749, "learning_rate": 8.346581590111263e-06, "loss": 0.5823, "step": 1583 }, { "epoch": 1.0206185567010309, "grad_norm": 0.6006589531898499, "learning_rate": 8.343795288270624e-06, "loss": 0.5773, "step": 1584 }, { "epoch": 1.021262886597938, "grad_norm": 0.5368292331695557, "learning_rate": 8.341007106631217e-06, "loss": 0.5601, "step": 1585 }, { "epoch": 1.0219072164948453, "grad_norm": 0.5745546221733093, "learning_rate": 8.33821704676049e-06, "loss": 0.5632, "step": 1586 }, { "epoch": 1.0225515463917525, "grad_norm": 0.6488891839981079, "learning_rate": 8.335425110226944e-06, "loss": 0.5623, "step": 1587 }, { "epoch": 1.0231958762886597, "grad_norm": 0.5996876358985901, "learning_rate": 8.332631298600136e-06, "loss": 0.573, "step": 1588 }, { "epoch": 1.023840206185567, "grad_norm": 0.7342520952224731, "learning_rate": 8.329835613450678e-06, "loss": 0.5818, "step": 1589 }, { "epoch": 1.0244845360824741, "grad_norm": 0.5877978801727295, "learning_rate": 8.327038056350236e-06, "loss": 0.5942, "step": 1590 }, { "epoch": 1.0251288659793814, "grad_norm": 0.7290641069412231, "learning_rate": 8.324238628871526e-06, "loss": 0.564, "step": 1591 }, { "epoch": 1.0257731958762886, "grad_norm": 0.6779270172119141, "learning_rate": 8.321437332588316e-06, "loss": 0.5605, "step": 1592 }, { "epoch": 1.0264175257731958, "grad_norm": 0.6248373985290527, "learning_rate": 8.318634169075428e-06, "loss": 0.581, "step": 1593 }, { "epoch": 1.027061855670103, "grad_norm": 0.7351896166801453, "learning_rate": 8.315829139908728e-06, "loss": 0.5727, "step": 1594 }, { "epoch": 1.0277061855670102, "grad_norm": 0.6194167137145996, "learning_rate": 8.313022246665135e-06, "loss": 0.569, "step": 1595 }, { "epoch": 1.0283505154639174, "grad_norm": 0.5790055990219116, "learning_rate": 8.310213490922616e-06, "loss": 0.5615, "step": 1596 }, { "epoch": 1.0289948453608246, "grad_norm": 0.6195235252380371, "learning_rate": 8.307402874260183e-06, "loss": 0.5734, "step": 1597 }, { "epoch": 1.0296391752577319, "grad_norm": 0.56621915102005, "learning_rate": 8.304590398257893e-06, "loss": 0.5783, "step": 1598 }, { "epoch": 1.030283505154639, "grad_norm": 0.6291447877883911, "learning_rate": 8.301776064496853e-06, "loss": 0.5635, "step": 1599 }, { "epoch": 1.0309278350515463, "grad_norm": 0.5548785924911499, "learning_rate": 8.29895987455921e-06, "loss": 0.563, "step": 1600 }, { "epoch": 1.0315721649484537, "grad_norm": 0.5734769105911255, "learning_rate": 8.296141830028157e-06, "loss": 0.5554, "step": 1601 }, { "epoch": 1.0322164948453607, "grad_norm": 0.6178357005119324, "learning_rate": 8.293321932487928e-06, "loss": 0.5756, "step": 1602 }, { "epoch": 1.0328608247422681, "grad_norm": 0.5657501816749573, "learning_rate": 8.2905001835238e-06, "loss": 0.5656, "step": 1603 }, { "epoch": 1.0335051546391754, "grad_norm": 0.6490956544876099, "learning_rate": 8.28767658472209e-06, "loss": 0.5823, "step": 1604 }, { "epoch": 1.0341494845360826, "grad_norm": 0.5873987674713135, "learning_rate": 8.284851137670156e-06, "loss": 0.5519, "step": 1605 }, { "epoch": 1.0347938144329898, "grad_norm": 0.5865747332572937, "learning_rate": 8.282023843956392e-06, "loss": 0.5547, "step": 1606 }, { "epoch": 1.035438144329897, "grad_norm": 0.6641358137130737, "learning_rate": 8.279194705170235e-06, "loss": 0.5554, "step": 1607 }, { "epoch": 1.0360824742268042, "grad_norm": 0.572713315486908, "learning_rate": 8.276363722902156e-06, "loss": 0.581, "step": 1608 }, { "epoch": 1.0367268041237114, "grad_norm": 0.6787334680557251, "learning_rate": 8.273530898743662e-06, "loss": 0.5687, "step": 1609 }, { "epoch": 1.0373711340206186, "grad_norm": 0.6074557304382324, "learning_rate": 8.270696234287297e-06, "loss": 0.5656, "step": 1610 }, { "epoch": 1.0380154639175259, "grad_norm": 0.6209086775779724, "learning_rate": 8.26785973112664e-06, "loss": 0.5542, "step": 1611 }, { "epoch": 1.038659793814433, "grad_norm": 0.5608944892883301, "learning_rate": 8.2650213908563e-06, "loss": 0.569, "step": 1612 }, { "epoch": 1.0393041237113403, "grad_norm": 0.6669208407402039, "learning_rate": 8.262181215071925e-06, "loss": 0.5719, "step": 1613 }, { "epoch": 1.0399484536082475, "grad_norm": 0.6176864504814148, "learning_rate": 8.25933920537019e-06, "loss": 0.5754, "step": 1614 }, { "epoch": 1.0405927835051547, "grad_norm": 0.6220034956932068, "learning_rate": 8.2564953633488e-06, "loss": 0.5544, "step": 1615 }, { "epoch": 1.041237113402062, "grad_norm": 0.6698017120361328, "learning_rate": 8.253649690606495e-06, "loss": 0.5785, "step": 1616 }, { "epoch": 1.0418814432989691, "grad_norm": 0.5491378903388977, "learning_rate": 8.250802188743042e-06, "loss": 0.5693, "step": 1617 }, { "epoch": 1.0425257731958764, "grad_norm": 0.6057620048522949, "learning_rate": 8.247952859359235e-06, "loss": 0.5788, "step": 1618 }, { "epoch": 1.0431701030927836, "grad_norm": 0.6617067456245422, "learning_rate": 8.245101704056896e-06, "loss": 0.6078, "step": 1619 }, { "epoch": 1.0438144329896908, "grad_norm": 0.6160987019538879, "learning_rate": 8.242248724438875e-06, "loss": 0.574, "step": 1620 }, { "epoch": 1.044458762886598, "grad_norm": 0.62845778465271, "learning_rate": 8.239393922109045e-06, "loss": 0.5633, "step": 1621 }, { "epoch": 1.0451030927835052, "grad_norm": 0.698736846446991, "learning_rate": 8.236537298672304e-06, "loss": 0.5645, "step": 1622 }, { "epoch": 1.0457474226804124, "grad_norm": 0.5302225947380066, "learning_rate": 8.23367885573458e-06, "loss": 0.5696, "step": 1623 }, { "epoch": 1.0463917525773196, "grad_norm": 0.7500773072242737, "learning_rate": 8.230818594902811e-06, "loss": 0.5731, "step": 1624 }, { "epoch": 1.0470360824742269, "grad_norm": 0.5700222849845886, "learning_rate": 8.227956517784972e-06, "loss": 0.5645, "step": 1625 }, { "epoch": 1.047680412371134, "grad_norm": 0.5717722177505493, "learning_rate": 8.225092625990047e-06, "loss": 0.5736, "step": 1626 }, { "epoch": 1.0483247422680413, "grad_norm": 0.5933573246002197, "learning_rate": 8.222226921128047e-06, "loss": 0.5765, "step": 1627 }, { "epoch": 1.0489690721649485, "grad_norm": 0.6378272175788879, "learning_rate": 8.219359404809999e-06, "loss": 0.5721, "step": 1628 }, { "epoch": 1.0496134020618557, "grad_norm": 0.6188105344772339, "learning_rate": 8.216490078647952e-06, "loss": 0.5936, "step": 1629 }, { "epoch": 1.050257731958763, "grad_norm": 0.6529179215431213, "learning_rate": 8.213618944254967e-06, "loss": 0.5719, "step": 1630 }, { "epoch": 1.0509020618556701, "grad_norm": 0.5707443356513977, "learning_rate": 8.210746003245126e-06, "loss": 0.5645, "step": 1631 }, { "epoch": 1.0515463917525774, "grad_norm": 0.6388113498687744, "learning_rate": 8.207871257233524e-06, "loss": 0.5739, "step": 1632 }, { "epoch": 1.0521907216494846, "grad_norm": 0.601162850856781, "learning_rate": 8.204994707836273e-06, "loss": 0.5877, "step": 1633 }, { "epoch": 1.0528350515463918, "grad_norm": 0.5794463157653809, "learning_rate": 8.202116356670498e-06, "loss": 0.5687, "step": 1634 }, { "epoch": 1.053479381443299, "grad_norm": 0.6841793060302734, "learning_rate": 8.199236205354337e-06, "loss": 0.5784, "step": 1635 }, { "epoch": 1.0541237113402062, "grad_norm": 0.6103260517120361, "learning_rate": 8.196354255506937e-06, "loss": 0.5878, "step": 1636 }, { "epoch": 1.0547680412371134, "grad_norm": 0.5457279086112976, "learning_rate": 8.193470508748463e-06, "loss": 0.5578, "step": 1637 }, { "epoch": 1.0554123711340206, "grad_norm": 0.7124139070510864, "learning_rate": 8.190584966700082e-06, "loss": 0.5509, "step": 1638 }, { "epoch": 1.0560567010309279, "grad_norm": 0.6309696435928345, "learning_rate": 8.187697630983975e-06, "loss": 0.5855, "step": 1639 }, { "epoch": 1.056701030927835, "grad_norm": 0.5825565457344055, "learning_rate": 8.184808503223331e-06, "loss": 0.5815, "step": 1640 }, { "epoch": 1.0573453608247423, "grad_norm": 0.5331952571868896, "learning_rate": 8.181917585042347e-06, "loss": 0.5589, "step": 1641 }, { "epoch": 1.0579896907216495, "grad_norm": 0.6160768270492554, "learning_rate": 8.179024878066223e-06, "loss": 0.5718, "step": 1642 }, { "epoch": 1.0586340206185567, "grad_norm": 0.6158849000930786, "learning_rate": 8.176130383921173e-06, "loss": 0.5738, "step": 1643 }, { "epoch": 1.059278350515464, "grad_norm": 0.5685930252075195, "learning_rate": 8.173234104234402e-06, "loss": 0.5665, "step": 1644 }, { "epoch": 1.0599226804123711, "grad_norm": 0.6989744901657104, "learning_rate": 8.17033604063413e-06, "loss": 0.5685, "step": 1645 }, { "epoch": 1.0605670103092784, "grad_norm": 0.7962100505828857, "learning_rate": 8.167436194749576e-06, "loss": 0.5717, "step": 1646 }, { "epoch": 1.0612113402061856, "grad_norm": 0.5195800065994263, "learning_rate": 8.164534568210964e-06, "loss": 0.5805, "step": 1647 }, { "epoch": 1.0618556701030928, "grad_norm": 0.7147637605667114, "learning_rate": 8.161631162649513e-06, "loss": 0.5535, "step": 1648 }, { "epoch": 1.0625, "grad_norm": 0.781481146812439, "learning_rate": 8.158725979697448e-06, "loss": 0.5722, "step": 1649 }, { "epoch": 1.0631443298969072, "grad_norm": 0.6806411147117615, "learning_rate": 8.15581902098799e-06, "loss": 0.5756, "step": 1650 }, { "epoch": 1.0637886597938144, "grad_norm": 0.6264119148254395, "learning_rate": 8.152910288155357e-06, "loss": 0.5526, "step": 1651 }, { "epoch": 1.0644329896907216, "grad_norm": 0.6693496108055115, "learning_rate": 8.14999978283477e-06, "loss": 0.5728, "step": 1652 }, { "epoch": 1.0650773195876289, "grad_norm": 0.5795789361000061, "learning_rate": 8.14708750666244e-06, "loss": 0.58, "step": 1653 }, { "epoch": 1.065721649484536, "grad_norm": 0.5916622281074524, "learning_rate": 8.14417346127558e-06, "loss": 0.6089, "step": 1654 }, { "epoch": 1.0663659793814433, "grad_norm": 0.5914902091026306, "learning_rate": 8.141257648312389e-06, "loss": 0.5815, "step": 1655 }, { "epoch": 1.0670103092783505, "grad_norm": 0.589816689491272, "learning_rate": 8.138340069412069e-06, "loss": 0.5665, "step": 1656 }, { "epoch": 1.0676546391752577, "grad_norm": 0.5838284492492676, "learning_rate": 8.135420726214808e-06, "loss": 0.5608, "step": 1657 }, { "epoch": 1.068298969072165, "grad_norm": 0.5986723303794861, "learning_rate": 8.13249962036179e-06, "loss": 0.5624, "step": 1658 }, { "epoch": 1.0689432989690721, "grad_norm": 0.5775933861732483, "learning_rate": 8.129576753495187e-06, "loss": 0.5655, "step": 1659 }, { "epoch": 1.0695876288659794, "grad_norm": 0.5839813351631165, "learning_rate": 8.126652127258164e-06, "loss": 0.577, "step": 1660 }, { "epoch": 1.0702319587628866, "grad_norm": 0.5542320013046265, "learning_rate": 8.12372574329487e-06, "loss": 0.5539, "step": 1661 }, { "epoch": 1.0708762886597938, "grad_norm": 0.5619986057281494, "learning_rate": 8.120797603250448e-06, "loss": 0.5743, "step": 1662 }, { "epoch": 1.071520618556701, "grad_norm": 0.5550148487091064, "learning_rate": 8.117867708771026e-06, "loss": 0.5727, "step": 1663 }, { "epoch": 1.0721649484536082, "grad_norm": 0.6201096177101135, "learning_rate": 8.114936061503717e-06, "loss": 0.5982, "step": 1664 }, { "epoch": 1.0728092783505154, "grad_norm": 0.6036857962608337, "learning_rate": 8.11200266309662e-06, "loss": 0.5686, "step": 1665 }, { "epoch": 1.0734536082474226, "grad_norm": 0.5542135238647461, "learning_rate": 8.109067515198822e-06, "loss": 0.5769, "step": 1666 }, { "epoch": 1.0740979381443299, "grad_norm": 0.6311356425285339, "learning_rate": 8.106130619460384e-06, "loss": 0.5685, "step": 1667 }, { "epoch": 1.074742268041237, "grad_norm": 0.5752009749412537, "learning_rate": 8.103191977532362e-06, "loss": 0.5775, "step": 1668 }, { "epoch": 1.0753865979381443, "grad_norm": 0.6080076694488525, "learning_rate": 8.100251591066784e-06, "loss": 0.5979, "step": 1669 }, { "epoch": 1.0760309278350515, "grad_norm": 0.6381897330284119, "learning_rate": 8.097309461716664e-06, "loss": 0.5807, "step": 1670 }, { "epoch": 1.0766752577319587, "grad_norm": 0.5368266105651855, "learning_rate": 8.094365591135996e-06, "loss": 0.5547, "step": 1671 }, { "epoch": 1.077319587628866, "grad_norm": 0.6196686625480652, "learning_rate": 8.091419980979746e-06, "loss": 0.575, "step": 1672 }, { "epoch": 1.0779639175257731, "grad_norm": 0.624686062335968, "learning_rate": 8.088472632903866e-06, "loss": 0.5798, "step": 1673 }, { "epoch": 1.0786082474226804, "grad_norm": 0.6162389516830444, "learning_rate": 8.085523548565283e-06, "loss": 0.5733, "step": 1674 }, { "epoch": 1.0792525773195876, "grad_norm": 0.6223421096801758, "learning_rate": 8.082572729621897e-06, "loss": 0.5776, "step": 1675 }, { "epoch": 1.0798969072164948, "grad_norm": 0.5665858387947083, "learning_rate": 8.079620177732587e-06, "loss": 0.5722, "step": 1676 }, { "epoch": 1.080541237113402, "grad_norm": 0.6715860366821289, "learning_rate": 8.076665894557202e-06, "loss": 0.5793, "step": 1677 }, { "epoch": 1.0811855670103092, "grad_norm": 0.5477104783058167, "learning_rate": 8.073709881756568e-06, "loss": 0.5809, "step": 1678 }, { "epoch": 1.0818298969072164, "grad_norm": 0.532974362373352, "learning_rate": 8.070752140992484e-06, "loss": 0.5645, "step": 1679 }, { "epoch": 1.0824742268041236, "grad_norm": 0.6037116050720215, "learning_rate": 8.067792673927716e-06, "loss": 0.5974, "step": 1680 }, { "epoch": 1.0831185567010309, "grad_norm": 0.575616180896759, "learning_rate": 8.064831482226004e-06, "loss": 0.5854, "step": 1681 }, { "epoch": 1.083762886597938, "grad_norm": 0.5680326819419861, "learning_rate": 8.061868567552055e-06, "loss": 0.563, "step": 1682 }, { "epoch": 1.0844072164948453, "grad_norm": 0.6410309076309204, "learning_rate": 8.058903931571548e-06, "loss": 0.5679, "step": 1683 }, { "epoch": 1.0850515463917525, "grad_norm": 0.6233634352684021, "learning_rate": 8.055937575951126e-06, "loss": 0.5691, "step": 1684 }, { "epoch": 1.0856958762886597, "grad_norm": 0.6438683867454529, "learning_rate": 8.052969502358405e-06, "loss": 0.5845, "step": 1685 }, { "epoch": 1.086340206185567, "grad_norm": 0.5698937773704529, "learning_rate": 8.049999712461956e-06, "loss": 0.5792, "step": 1686 }, { "epoch": 1.0869845360824741, "grad_norm": 0.619781494140625, "learning_rate": 8.047028207931326e-06, "loss": 0.5503, "step": 1687 }, { "epoch": 1.0876288659793814, "grad_norm": 0.5782028436660767, "learning_rate": 8.044054990437018e-06, "loss": 0.5848, "step": 1688 }, { "epoch": 1.0882731958762886, "grad_norm": 0.5991863012313843, "learning_rate": 8.041080061650504e-06, "loss": 0.5651, "step": 1689 }, { "epoch": 1.0889175257731958, "grad_norm": 0.5564179420471191, "learning_rate": 8.038103423244212e-06, "loss": 0.5768, "step": 1690 }, { "epoch": 1.089561855670103, "grad_norm": 0.6740025877952576, "learning_rate": 8.035125076891537e-06, "loss": 0.5502, "step": 1691 }, { "epoch": 1.0902061855670102, "grad_norm": 0.5896285176277161, "learning_rate": 8.032145024266832e-06, "loss": 0.569, "step": 1692 }, { "epoch": 1.0908505154639174, "grad_norm": 0.5947343707084656, "learning_rate": 8.029163267045405e-06, "loss": 0.5866, "step": 1693 }, { "epoch": 1.0914948453608246, "grad_norm": 0.5607129335403442, "learning_rate": 8.026179806903528e-06, "loss": 0.5821, "step": 1694 }, { "epoch": 1.0921391752577319, "grad_norm": 0.5299738049507141, "learning_rate": 8.02319464551843e-06, "loss": 0.5526, "step": 1695 }, { "epoch": 1.0927835051546393, "grad_norm": 0.5900497436523438, "learning_rate": 8.020207784568293e-06, "loss": 0.5554, "step": 1696 }, { "epoch": 1.0934278350515463, "grad_norm": 0.6226323246955872, "learning_rate": 8.017219225732255e-06, "loss": 0.5735, "step": 1697 }, { "epoch": 1.0940721649484537, "grad_norm": 0.5552526116371155, "learning_rate": 8.01422897069041e-06, "loss": 0.5791, "step": 1698 }, { "epoch": 1.0947164948453607, "grad_norm": 0.5886931419372559, "learning_rate": 8.011237021123807e-06, "loss": 0.5771, "step": 1699 }, { "epoch": 1.0953608247422681, "grad_norm": 0.6178483366966248, "learning_rate": 8.008243378714445e-06, "loss": 0.569, "step": 1700 }, { "epoch": 1.0960051546391754, "grad_norm": 1.7713351249694824, "learning_rate": 8.005248045145276e-06, "loss": 0.6438, "step": 1701 }, { "epoch": 1.0966494845360826, "grad_norm": 0.6597208976745605, "learning_rate": 8.0022510221002e-06, "loss": 0.5882, "step": 1702 }, { "epoch": 1.0972938144329898, "grad_norm": 0.6257805228233337, "learning_rate": 7.999252311264071e-06, "loss": 0.5823, "step": 1703 }, { "epoch": 1.097938144329897, "grad_norm": 0.6056534051895142, "learning_rate": 7.99625191432269e-06, "loss": 0.5979, "step": 1704 }, { "epoch": 1.0985824742268042, "grad_norm": 0.6608222723007202, "learning_rate": 7.993249832962804e-06, "loss": 0.5821, "step": 1705 }, { "epoch": 1.0992268041237114, "grad_norm": 0.6068564653396606, "learning_rate": 7.990246068872111e-06, "loss": 0.5908, "step": 1706 }, { "epoch": 1.0998711340206186, "grad_norm": 0.6525301337242126, "learning_rate": 7.987240623739251e-06, "loss": 0.572, "step": 1707 }, { "epoch": 1.1005154639175259, "grad_norm": 0.6569459438323975, "learning_rate": 7.98423349925381e-06, "loss": 0.5781, "step": 1708 }, { "epoch": 1.101159793814433, "grad_norm": 0.7363823652267456, "learning_rate": 7.981224697106323e-06, "loss": 0.5618, "step": 1709 }, { "epoch": 1.1018041237113403, "grad_norm": 0.5923417806625366, "learning_rate": 7.97821421898826e-06, "loss": 0.5673, "step": 1710 }, { "epoch": 1.1024484536082475, "grad_norm": 0.7146652340888977, "learning_rate": 7.975202066592038e-06, "loss": 0.5852, "step": 1711 }, { "epoch": 1.1030927835051547, "grad_norm": 0.6243744492530823, "learning_rate": 7.972188241611014e-06, "loss": 0.5759, "step": 1712 }, { "epoch": 1.103737113402062, "grad_norm": 0.5584855079650879, "learning_rate": 7.969172745739486e-06, "loss": 0.6013, "step": 1713 }, { "epoch": 1.1043814432989691, "grad_norm": 0.6535565853118896, "learning_rate": 7.96615558067269e-06, "loss": 0.5734, "step": 1714 }, { "epoch": 1.1050257731958764, "grad_norm": 0.6076893210411072, "learning_rate": 7.963136748106804e-06, "loss": 0.5729, "step": 1715 }, { "epoch": 1.1056701030927836, "grad_norm": 0.5445538759231567, "learning_rate": 7.960116249738939e-06, "loss": 0.5693, "step": 1716 }, { "epoch": 1.1063144329896908, "grad_norm": 0.7051904201507568, "learning_rate": 7.957094087267144e-06, "loss": 0.5777, "step": 1717 }, { "epoch": 1.106958762886598, "grad_norm": 0.5875561833381653, "learning_rate": 7.954070262390404e-06, "loss": 0.5708, "step": 1718 }, { "epoch": 1.1076030927835052, "grad_norm": 0.6419768929481506, "learning_rate": 7.951044776808638e-06, "loss": 0.5695, "step": 1719 }, { "epoch": 1.1082474226804124, "grad_norm": 0.5418486595153809, "learning_rate": 7.9480176322227e-06, "loss": 0.5575, "step": 1720 }, { "epoch": 1.1088917525773196, "grad_norm": 0.5898576974868774, "learning_rate": 7.944988830334377e-06, "loss": 0.5715, "step": 1721 }, { "epoch": 1.1095360824742269, "grad_norm": 0.5845350027084351, "learning_rate": 7.941958372846383e-06, "loss": 0.5695, "step": 1722 }, { "epoch": 1.110180412371134, "grad_norm": 0.613530158996582, "learning_rate": 7.938926261462366e-06, "loss": 0.559, "step": 1723 }, { "epoch": 1.1108247422680413, "grad_norm": 0.6073035597801208, "learning_rate": 7.935892497886908e-06, "loss": 0.572, "step": 1724 }, { "epoch": 1.1114690721649485, "grad_norm": 0.6385327577590942, "learning_rate": 7.932857083825512e-06, "loss": 0.583, "step": 1725 }, { "epoch": 1.1121134020618557, "grad_norm": 0.5657716393470764, "learning_rate": 7.92982002098461e-06, "loss": 0.5454, "step": 1726 }, { "epoch": 1.112757731958763, "grad_norm": 0.5977925062179565, "learning_rate": 7.92678131107157e-06, "loss": 0.5801, "step": 1727 }, { "epoch": 1.1134020618556701, "grad_norm": 0.5900896787643433, "learning_rate": 7.923740955794676e-06, "loss": 0.5657, "step": 1728 }, { "epoch": 1.1140463917525774, "grad_norm": 0.564995288848877, "learning_rate": 7.92069895686314e-06, "loss": 0.5701, "step": 1729 }, { "epoch": 1.1146907216494846, "grad_norm": 0.6209300756454468, "learning_rate": 7.9176553159871e-06, "loss": 0.5895, "step": 1730 }, { "epoch": 1.1153350515463918, "grad_norm": 0.5701628923416138, "learning_rate": 7.914610034877613e-06, "loss": 0.5616, "step": 1731 }, { "epoch": 1.115979381443299, "grad_norm": 0.6076413989067078, "learning_rate": 7.91156311524666e-06, "loss": 0.5719, "step": 1732 }, { "epoch": 1.1166237113402062, "grad_norm": 0.5848098397254944, "learning_rate": 7.908514558807148e-06, "loss": 0.5798, "step": 1733 }, { "epoch": 1.1172680412371134, "grad_norm": 0.5207605957984924, "learning_rate": 7.905464367272897e-06, "loss": 0.5505, "step": 1734 }, { "epoch": 1.1179123711340206, "grad_norm": 0.5907694101333618, "learning_rate": 7.902412542358647e-06, "loss": 0.5689, "step": 1735 }, { "epoch": 1.1185567010309279, "grad_norm": 0.6055130958557129, "learning_rate": 7.899359085780062e-06, "loss": 0.5891, "step": 1736 }, { "epoch": 1.119201030927835, "grad_norm": 0.5186037421226501, "learning_rate": 7.896303999253718e-06, "loss": 0.5646, "step": 1737 }, { "epoch": 1.1198453608247423, "grad_norm": 0.6468600034713745, "learning_rate": 7.893247284497108e-06, "loss": 0.5843, "step": 1738 }, { "epoch": 1.1204896907216495, "grad_norm": 0.5830214023590088, "learning_rate": 7.890188943228643e-06, "loss": 0.5568, "step": 1739 }, { "epoch": 1.1211340206185567, "grad_norm": 0.6374555230140686, "learning_rate": 7.887128977167645e-06, "loss": 0.5816, "step": 1740 }, { "epoch": 1.121778350515464, "grad_norm": 0.5781762599945068, "learning_rate": 7.884067388034351e-06, "loss": 0.5978, "step": 1741 }, { "epoch": 1.1224226804123711, "grad_norm": 0.5457374453544617, "learning_rate": 7.881004177549912e-06, "loss": 0.5756, "step": 1742 }, { "epoch": 1.1230670103092784, "grad_norm": 0.595295250415802, "learning_rate": 7.877939347436387e-06, "loss": 0.574, "step": 1743 }, { "epoch": 1.1237113402061856, "grad_norm": 0.528535783290863, "learning_rate": 7.87487289941675e-06, "loss": 0.5901, "step": 1744 }, { "epoch": 1.1243556701030928, "grad_norm": 0.5907325148582458, "learning_rate": 7.87180483521488e-06, "loss": 0.5714, "step": 1745 }, { "epoch": 1.125, "grad_norm": 0.5261876583099365, "learning_rate": 7.868735156555567e-06, "loss": 0.5916, "step": 1746 }, { "epoch": 1.1256443298969072, "grad_norm": 0.5692242383956909, "learning_rate": 7.865663865164508e-06, "loss": 0.5695, "step": 1747 }, { "epoch": 1.1262886597938144, "grad_norm": 0.5626437664031982, "learning_rate": 7.862590962768308e-06, "loss": 0.571, "step": 1748 }, { "epoch": 1.1269329896907216, "grad_norm": 0.6156812906265259, "learning_rate": 7.859516451094476e-06, "loss": 0.584, "step": 1749 }, { "epoch": 1.1275773195876289, "grad_norm": 0.5854163765907288, "learning_rate": 7.856440331871425e-06, "loss": 0.5866, "step": 1750 }, { "epoch": 1.128221649484536, "grad_norm": 0.615242600440979, "learning_rate": 7.853362606828474e-06, "loss": 0.5705, "step": 1751 }, { "epoch": 1.1288659793814433, "grad_norm": 0.6125758290290833, "learning_rate": 7.850283277695842e-06, "loss": 0.5737, "step": 1752 }, { "epoch": 1.1295103092783505, "grad_norm": 0.5516372919082642, "learning_rate": 7.847202346204655e-06, "loss": 0.5854, "step": 1753 }, { "epoch": 1.1301546391752577, "grad_norm": 0.630854070186615, "learning_rate": 7.844119814086932e-06, "loss": 0.5599, "step": 1754 }, { "epoch": 1.130798969072165, "grad_norm": 0.6526719927787781, "learning_rate": 7.8410356830756e-06, "loss": 0.5761, "step": 1755 }, { "epoch": 1.1314432989690721, "grad_norm": 0.5448639392852783, "learning_rate": 7.83794995490448e-06, "loss": 0.5775, "step": 1756 }, { "epoch": 1.1320876288659794, "grad_norm": 0.6779550313949585, "learning_rate": 7.834862631308288e-06, "loss": 0.5633, "step": 1757 }, { "epoch": 1.1327319587628866, "grad_norm": 0.6387903094291687, "learning_rate": 7.831773714022647e-06, "loss": 0.5725, "step": 1758 }, { "epoch": 1.1333762886597938, "grad_norm": 0.5421297550201416, "learning_rate": 7.828683204784064e-06, "loss": 0.5848, "step": 1759 }, { "epoch": 1.134020618556701, "grad_norm": 0.7366899251937866, "learning_rate": 7.825591105329952e-06, "loss": 0.5932, "step": 1760 }, { "epoch": 1.1346649484536082, "grad_norm": 0.7821252942085266, "learning_rate": 7.822497417398607e-06, "loss": 0.5926, "step": 1761 }, { "epoch": 1.1353092783505154, "grad_norm": 0.6638327240943909, "learning_rate": 7.819402142729226e-06, "loss": 0.5808, "step": 1762 }, { "epoch": 1.1359536082474226, "grad_norm": 0.8765934705734253, "learning_rate": 7.816305283061898e-06, "loss": 0.5742, "step": 1763 }, { "epoch": 1.1365979381443299, "grad_norm": 0.7599136233329773, "learning_rate": 7.813206840137599e-06, "loss": 0.5476, "step": 1764 }, { "epoch": 1.137242268041237, "grad_norm": 0.7483077049255371, "learning_rate": 7.810106815698195e-06, "loss": 0.5708, "step": 1765 }, { "epoch": 1.1378865979381443, "grad_norm": 0.8938262462615967, "learning_rate": 7.807005211486445e-06, "loss": 0.5861, "step": 1766 }, { "epoch": 1.1385309278350515, "grad_norm": 0.7024272680282593, "learning_rate": 7.803902029245993e-06, "loss": 0.5733, "step": 1767 }, { "epoch": 1.1391752577319587, "grad_norm": 0.6438678503036499, "learning_rate": 7.800797270721373e-06, "loss": 0.5777, "step": 1768 }, { "epoch": 1.139819587628866, "grad_norm": 0.9886898994445801, "learning_rate": 7.797690937657998e-06, "loss": 0.5747, "step": 1769 }, { "epoch": 1.1404639175257731, "grad_norm": 0.7548224329948425, "learning_rate": 7.794583031802176e-06, "loss": 0.5829, "step": 1770 }, { "epoch": 1.1411082474226804, "grad_norm": 0.8245490193367004, "learning_rate": 7.791473554901095e-06, "loss": 0.574, "step": 1771 }, { "epoch": 1.1417525773195876, "grad_norm": 0.7582380771636963, "learning_rate": 7.788362508702822e-06, "loss": 0.567, "step": 1772 }, { "epoch": 1.1423969072164948, "grad_norm": 0.7521668672561646, "learning_rate": 7.785249894956315e-06, "loss": 0.5686, "step": 1773 }, { "epoch": 1.143041237113402, "grad_norm": 0.737846314907074, "learning_rate": 7.782135715411405e-06, "loss": 0.5817, "step": 1774 }, { "epoch": 1.1436855670103092, "grad_norm": 0.744872510433197, "learning_rate": 7.779019971818807e-06, "loss": 0.5698, "step": 1775 }, { "epoch": 1.1443298969072164, "grad_norm": 0.6963609457015991, "learning_rate": 7.775902665930114e-06, "loss": 0.5703, "step": 1776 }, { "epoch": 1.1449742268041236, "grad_norm": 0.6479573845863342, "learning_rate": 7.772783799497798e-06, "loss": 0.5575, "step": 1777 }, { "epoch": 1.1456185567010309, "grad_norm": 0.721855640411377, "learning_rate": 7.769663374275211e-06, "loss": 0.5657, "step": 1778 }, { "epoch": 1.146262886597938, "grad_norm": 0.7050710320472717, "learning_rate": 7.766541392016575e-06, "loss": 0.5687, "step": 1779 }, { "epoch": 1.1469072164948453, "grad_norm": 0.6693905591964722, "learning_rate": 7.763417854476994e-06, "loss": 0.5776, "step": 1780 }, { "epoch": 1.1475515463917525, "grad_norm": 0.6165105700492859, "learning_rate": 7.76029276341244e-06, "loss": 0.5703, "step": 1781 }, { "epoch": 1.1481958762886597, "grad_norm": 0.6124728918075562, "learning_rate": 7.757166120579768e-06, "loss": 0.5916, "step": 1782 }, { "epoch": 1.148840206185567, "grad_norm": 0.6497332453727722, "learning_rate": 7.754037927736693e-06, "loss": 0.5676, "step": 1783 }, { "epoch": 1.1494845360824741, "grad_norm": 0.6448075771331787, "learning_rate": 7.750908186641808e-06, "loss": 0.5834, "step": 1784 }, { "epoch": 1.1501288659793814, "grad_norm": 0.6209561228752136, "learning_rate": 7.747776899054578e-06, "loss": 0.5718, "step": 1785 }, { "epoch": 1.1507731958762886, "grad_norm": 0.673071563243866, "learning_rate": 7.744644066735335e-06, "loss": 0.6009, "step": 1786 }, { "epoch": 1.1514175257731958, "grad_norm": 0.6220034956932068, "learning_rate": 7.741509691445278e-06, "loss": 0.5611, "step": 1787 }, { "epoch": 1.152061855670103, "grad_norm": 0.7180672883987427, "learning_rate": 7.738373774946476e-06, "loss": 0.5545, "step": 1788 }, { "epoch": 1.1527061855670104, "grad_norm": 0.7007160186767578, "learning_rate": 7.735236319001863e-06, "loss": 0.5688, "step": 1789 }, { "epoch": 1.1533505154639174, "grad_norm": 0.6026685833930969, "learning_rate": 7.73209732537524e-06, "loss": 0.5785, "step": 1790 }, { "epoch": 1.1539948453608249, "grad_norm": 0.6774597764015198, "learning_rate": 7.72895679583127e-06, "loss": 0.572, "step": 1791 }, { "epoch": 1.1546391752577319, "grad_norm": 0.5667352080345154, "learning_rate": 7.725814732135484e-06, "loss": 0.5695, "step": 1792 }, { "epoch": 1.1552835051546393, "grad_norm": 0.594508707523346, "learning_rate": 7.722671136054268e-06, "loss": 0.592, "step": 1793 }, { "epoch": 1.1559278350515463, "grad_norm": 0.6542758941650391, "learning_rate": 7.719526009354875e-06, "loss": 0.5743, "step": 1794 }, { "epoch": 1.1565721649484537, "grad_norm": 0.5780284404754639, "learning_rate": 7.716379353805419e-06, "loss": 0.5718, "step": 1795 }, { "epoch": 1.1572164948453607, "grad_norm": 0.6058859825134277, "learning_rate": 7.713231171174868e-06, "loss": 0.5667, "step": 1796 }, { "epoch": 1.1578608247422681, "grad_norm": 0.6840323805809021, "learning_rate": 7.710081463233056e-06, "loss": 0.5865, "step": 1797 }, { "epoch": 1.1585051546391751, "grad_norm": 0.5763911008834839, "learning_rate": 7.70693023175067e-06, "loss": 0.575, "step": 1798 }, { "epoch": 1.1591494845360826, "grad_norm": 0.6470733880996704, "learning_rate": 7.703777478499252e-06, "loss": 0.5715, "step": 1799 }, { "epoch": 1.1597938144329896, "grad_norm": 0.6214786171913147, "learning_rate": 7.700623205251206e-06, "loss": 0.5649, "step": 1800 }, { "epoch": 1.160438144329897, "grad_norm": 0.551515519618988, "learning_rate": 7.697467413779781e-06, "loss": 0.5563, "step": 1801 }, { "epoch": 1.1610824742268042, "grad_norm": 0.6152409911155701, "learning_rate": 7.694310105859087e-06, "loss": 0.5731, "step": 1802 }, { "epoch": 1.1617268041237114, "grad_norm": 0.6466968655586243, "learning_rate": 7.691151283264085e-06, "loss": 0.5694, "step": 1803 }, { "epoch": 1.1623711340206186, "grad_norm": 0.8145484924316406, "learning_rate": 7.687990947770585e-06, "loss": 0.5755, "step": 1804 }, { "epoch": 1.1630154639175259, "grad_norm": 0.6188675761222839, "learning_rate": 7.684829101155252e-06, "loss": 0.5612, "step": 1805 }, { "epoch": 1.163659793814433, "grad_norm": 0.7016410231590271, "learning_rate": 7.681665745195593e-06, "loss": 0.5648, "step": 1806 }, { "epoch": 1.1643041237113403, "grad_norm": 0.6102988719940186, "learning_rate": 7.678500881669972e-06, "loss": 0.5715, "step": 1807 }, { "epoch": 1.1649484536082475, "grad_norm": 0.6002679467201233, "learning_rate": 7.675334512357598e-06, "loss": 0.5877, "step": 1808 }, { "epoch": 1.1655927835051547, "grad_norm": 0.6926239132881165, "learning_rate": 7.672166639038521e-06, "loss": 0.5886, "step": 1809 }, { "epoch": 1.166237113402062, "grad_norm": 0.5459895730018616, "learning_rate": 7.668997263493643e-06, "loss": 0.5488, "step": 1810 }, { "epoch": 1.1668814432989691, "grad_norm": 0.584542453289032, "learning_rate": 7.665826387504708e-06, "loss": 0.5966, "step": 1811 }, { "epoch": 1.1675257731958764, "grad_norm": 0.5579724907875061, "learning_rate": 7.662654012854306e-06, "loss": 0.5612, "step": 1812 }, { "epoch": 1.1681701030927836, "grad_norm": 0.5972083210945129, "learning_rate": 7.659480141325864e-06, "loss": 0.5877, "step": 1813 }, { "epoch": 1.1688144329896908, "grad_norm": 0.5673110485076904, "learning_rate": 7.656304774703656e-06, "loss": 0.57, "step": 1814 }, { "epoch": 1.169458762886598, "grad_norm": 0.6643372774124146, "learning_rate": 7.653127914772793e-06, "loss": 0.5681, "step": 1815 }, { "epoch": 1.1701030927835052, "grad_norm": 0.6632299423217773, "learning_rate": 7.649949563319228e-06, "loss": 0.5625, "step": 1816 }, { "epoch": 1.1707474226804124, "grad_norm": 0.6024297475814819, "learning_rate": 7.64676972212975e-06, "loss": 0.5645, "step": 1817 }, { "epoch": 1.1713917525773196, "grad_norm": 0.6235643625259399, "learning_rate": 7.64358839299199e-06, "loss": 0.5895, "step": 1818 }, { "epoch": 1.1720360824742269, "grad_norm": 0.5492936372756958, "learning_rate": 7.640405577694408e-06, "loss": 0.583, "step": 1819 }, { "epoch": 1.172680412371134, "grad_norm": 0.6146360039710999, "learning_rate": 7.637221278026306e-06, "loss": 0.5738, "step": 1820 }, { "epoch": 1.1733247422680413, "grad_norm": 0.5621262192726135, "learning_rate": 7.634035495777819e-06, "loss": 0.5575, "step": 1821 }, { "epoch": 1.1739690721649485, "grad_norm": 0.6018881797790527, "learning_rate": 7.630848232739912e-06, "loss": 0.5456, "step": 1822 }, { "epoch": 1.1746134020618557, "grad_norm": 0.602499783039093, "learning_rate": 7.6276594907043876e-06, "loss": 0.5911, "step": 1823 }, { "epoch": 1.175257731958763, "grad_norm": 0.633616030216217, "learning_rate": 7.624469271463878e-06, "loss": 0.5671, "step": 1824 }, { "epoch": 1.1759020618556701, "grad_norm": 0.5846896171569824, "learning_rate": 7.621277576811844e-06, "loss": 0.5562, "step": 1825 }, { "epoch": 1.1765463917525774, "grad_norm": 0.5698741674423218, "learning_rate": 7.618084408542576e-06, "loss": 0.558, "step": 1826 }, { "epoch": 1.1771907216494846, "grad_norm": 0.6319502592086792, "learning_rate": 7.6148897684511965e-06, "loss": 0.5913, "step": 1827 }, { "epoch": 1.1778350515463918, "grad_norm": 0.6660181879997253, "learning_rate": 7.611693658333652e-06, "loss": 0.5664, "step": 1828 }, { "epoch": 1.178479381443299, "grad_norm": 0.6200321912765503, "learning_rate": 7.608496079986718e-06, "loss": 0.5674, "step": 1829 }, { "epoch": 1.1791237113402062, "grad_norm": 0.5971856117248535, "learning_rate": 7.605297035207989e-06, "loss": 0.5471, "step": 1830 }, { "epoch": 1.1797680412371134, "grad_norm": 0.6492776274681091, "learning_rate": 7.602096525795892e-06, "loss": 0.5653, "step": 1831 }, { "epoch": 1.1804123711340206, "grad_norm": 0.6144292950630188, "learning_rate": 7.598894553549676e-06, "loss": 0.5818, "step": 1832 }, { "epoch": 1.1810567010309279, "grad_norm": 0.6731366515159607, "learning_rate": 7.595691120269406e-06, "loss": 0.5691, "step": 1833 }, { "epoch": 1.181701030927835, "grad_norm": 0.6270490884780884, "learning_rate": 7.592486227755974e-06, "loss": 0.5801, "step": 1834 }, { "epoch": 1.1823453608247423, "grad_norm": 0.6081077456474304, "learning_rate": 7.589279877811091e-06, "loss": 0.5729, "step": 1835 }, { "epoch": 1.1829896907216495, "grad_norm": 0.7470238208770752, "learning_rate": 7.586072072237291e-06, "loss": 0.5837, "step": 1836 }, { "epoch": 1.1836340206185567, "grad_norm": 0.6267186403274536, "learning_rate": 7.582862812837919e-06, "loss": 0.5759, "step": 1837 }, { "epoch": 1.184278350515464, "grad_norm": 0.6680978536605835, "learning_rate": 7.579652101417139e-06, "loss": 0.5777, "step": 1838 }, { "epoch": 1.1849226804123711, "grad_norm": 0.6897329092025757, "learning_rate": 7.5764399397799405e-06, "loss": 0.5751, "step": 1839 }, { "epoch": 1.1855670103092784, "grad_norm": 0.6773098707199097, "learning_rate": 7.573226329732114e-06, "loss": 0.5776, "step": 1840 }, { "epoch": 1.1862113402061856, "grad_norm": 0.570708155632019, "learning_rate": 7.570011273080274e-06, "loss": 0.5743, "step": 1841 }, { "epoch": 1.1868556701030928, "grad_norm": 0.6059150695800781, "learning_rate": 7.566794771631847e-06, "loss": 0.5592, "step": 1842 }, { "epoch": 1.1875, "grad_norm": 0.6192939877510071, "learning_rate": 7.56357682719507e-06, "loss": 0.5872, "step": 1843 }, { "epoch": 1.1881443298969072, "grad_norm": 0.6065688729286194, "learning_rate": 7.560357441578987e-06, "loss": 0.5697, "step": 1844 }, { "epoch": 1.1887886597938144, "grad_norm": 0.5642627477645874, "learning_rate": 7.557136616593463e-06, "loss": 0.5653, "step": 1845 }, { "epoch": 1.1894329896907216, "grad_norm": 0.6173583269119263, "learning_rate": 7.5539143540491635e-06, "loss": 0.5631, "step": 1846 }, { "epoch": 1.1900773195876289, "grad_norm": 0.572210967540741, "learning_rate": 7.5506906557575635e-06, "loss": 0.5742, "step": 1847 }, { "epoch": 1.190721649484536, "grad_norm": 0.6059232354164124, "learning_rate": 7.547465523530945e-06, "loss": 0.5705, "step": 1848 }, { "epoch": 1.1913659793814433, "grad_norm": 0.607420802116394, "learning_rate": 7.544238959182403e-06, "loss": 0.5853, "step": 1849 }, { "epoch": 1.1920103092783505, "grad_norm": 0.5885394811630249, "learning_rate": 7.541010964525828e-06, "loss": 0.5588, "step": 1850 }, { "epoch": 1.1926546391752577, "grad_norm": 0.6172522306442261, "learning_rate": 7.537781541375917e-06, "loss": 0.5734, "step": 1851 }, { "epoch": 1.193298969072165, "grad_norm": 0.6126635670661926, "learning_rate": 7.534550691548173e-06, "loss": 0.5596, "step": 1852 }, { "epoch": 1.1939432989690721, "grad_norm": 0.6323084831237793, "learning_rate": 7.531318416858902e-06, "loss": 0.5628, "step": 1853 }, { "epoch": 1.1945876288659794, "grad_norm": 0.565899133682251, "learning_rate": 7.528084719125206e-06, "loss": 0.5872, "step": 1854 }, { "epoch": 1.1952319587628866, "grad_norm": 0.5521238446235657, "learning_rate": 7.524849600164989e-06, "loss": 0.5834, "step": 1855 }, { "epoch": 1.1958762886597938, "grad_norm": 0.6802346706390381, "learning_rate": 7.521613061796957e-06, "loss": 0.5661, "step": 1856 }, { "epoch": 1.196520618556701, "grad_norm": 0.5785925388336182, "learning_rate": 7.518375105840612e-06, "loss": 0.5699, "step": 1857 }, { "epoch": 1.1971649484536082, "grad_norm": 0.6164043545722961, "learning_rate": 7.51513573411625e-06, "loss": 0.5863, "step": 1858 }, { "epoch": 1.1978092783505154, "grad_norm": 0.5662636756896973, "learning_rate": 7.511894948444967e-06, "loss": 0.591, "step": 1859 }, { "epoch": 1.1984536082474226, "grad_norm": 0.6181192398071289, "learning_rate": 7.508652750648651e-06, "loss": 0.5862, "step": 1860 }, { "epoch": 1.1990979381443299, "grad_norm": 0.5194783806800842, "learning_rate": 7.505409142549987e-06, "loss": 0.5691, "step": 1861 }, { "epoch": 1.199742268041237, "grad_norm": 0.5715290904045105, "learning_rate": 7.502164125972449e-06, "loss": 0.5898, "step": 1862 }, { "epoch": 1.2003865979381443, "grad_norm": 0.5946175456047058, "learning_rate": 7.498917702740305e-06, "loss": 0.5594, "step": 1863 }, { "epoch": 1.2010309278350515, "grad_norm": 0.5648502707481384, "learning_rate": 7.495669874678616e-06, "loss": 0.574, "step": 1864 }, { "epoch": 1.2016752577319587, "grad_norm": 0.6160831451416016, "learning_rate": 7.49242064361323e-06, "loss": 0.5572, "step": 1865 }, { "epoch": 1.202319587628866, "grad_norm": 0.6111457943916321, "learning_rate": 7.48917001137078e-06, "loss": 0.5818, "step": 1866 }, { "epoch": 1.2029639175257731, "grad_norm": 0.535476803779602, "learning_rate": 7.485917979778696e-06, "loss": 0.5747, "step": 1867 }, { "epoch": 1.2036082474226804, "grad_norm": 0.5801500678062439, "learning_rate": 7.482664550665186e-06, "loss": 0.5888, "step": 1868 }, { "epoch": 1.2042525773195876, "grad_norm": 0.5323453545570374, "learning_rate": 7.479409725859249e-06, "loss": 0.5726, "step": 1869 }, { "epoch": 1.2048969072164948, "grad_norm": 0.6037577390670776, "learning_rate": 7.476153507190667e-06, "loss": 0.5753, "step": 1870 }, { "epoch": 1.205541237113402, "grad_norm": 0.5818812251091003, "learning_rate": 7.472895896490003e-06, "loss": 0.5761, "step": 1871 }, { "epoch": 1.2061855670103092, "grad_norm": 0.613215982913971, "learning_rate": 7.4696368955886076e-06, "loss": 0.5674, "step": 1872 }, { "epoch": 1.2068298969072164, "grad_norm": 0.548999011516571, "learning_rate": 7.466376506318609e-06, "loss": 0.5737, "step": 1873 }, { "epoch": 1.2074742268041236, "grad_norm": 0.5786558389663696, "learning_rate": 7.463114730512917e-06, "loss": 0.5506, "step": 1874 }, { "epoch": 1.2081185567010309, "grad_norm": 0.6421105265617371, "learning_rate": 7.459851570005222e-06, "loss": 0.5802, "step": 1875 }, { "epoch": 1.208762886597938, "grad_norm": 0.6085805296897888, "learning_rate": 7.456587026629991e-06, "loss": 0.5731, "step": 1876 }, { "epoch": 1.2094072164948453, "grad_norm": 0.7688948512077332, "learning_rate": 7.45332110222247e-06, "loss": 0.5568, "step": 1877 }, { "epoch": 1.2100515463917525, "grad_norm": 0.6081840991973877, "learning_rate": 7.450053798618679e-06, "loss": 0.5672, "step": 1878 }, { "epoch": 1.2106958762886597, "grad_norm": 0.5908514857292175, "learning_rate": 7.446785117655418e-06, "loss": 0.5594, "step": 1879 }, { "epoch": 1.211340206185567, "grad_norm": 0.7457608580589294, "learning_rate": 7.443515061170255e-06, "loss": 0.5431, "step": 1880 }, { "epoch": 1.2119845360824741, "grad_norm": 0.7039989829063416, "learning_rate": 7.440243631001536e-06, "loss": 0.5701, "step": 1881 }, { "epoch": 1.2126288659793814, "grad_norm": 0.6005958914756775, "learning_rate": 7.436970828988379e-06, "loss": 0.5779, "step": 1882 }, { "epoch": 1.2132731958762886, "grad_norm": 0.8835837244987488, "learning_rate": 7.433696656970673e-06, "loss": 0.5874, "step": 1883 }, { "epoch": 1.2139175257731958, "grad_norm": 0.5845495462417603, "learning_rate": 7.430421116789073e-06, "loss": 0.5778, "step": 1884 }, { "epoch": 1.214561855670103, "grad_norm": 0.7629231214523315, "learning_rate": 7.427144210285008e-06, "loss": 0.5782, "step": 1885 }, { "epoch": 1.2152061855670104, "grad_norm": 0.6457379460334778, "learning_rate": 7.423865939300674e-06, "loss": 0.5748, "step": 1886 }, { "epoch": 1.2158505154639174, "grad_norm": 0.583372950553894, "learning_rate": 7.420586305679032e-06, "loss": 0.5684, "step": 1887 }, { "epoch": 1.2164948453608249, "grad_norm": 0.6637343764305115, "learning_rate": 7.417305311263814e-06, "loss": 0.5838, "step": 1888 }, { "epoch": 1.2171391752577319, "grad_norm": 0.6337183117866516, "learning_rate": 7.414022957899509e-06, "loss": 0.5742, "step": 1889 }, { "epoch": 1.2177835051546393, "grad_norm": 0.5949130654335022, "learning_rate": 7.41073924743138e-06, "loss": 0.5531, "step": 1890 }, { "epoch": 1.2184278350515463, "grad_norm": 0.6296405792236328, "learning_rate": 7.4074541817054424e-06, "loss": 0.5863, "step": 1891 }, { "epoch": 1.2190721649484537, "grad_norm": 0.5354436039924622, "learning_rate": 7.404167762568483e-06, "loss": 0.5642, "step": 1892 }, { "epoch": 1.2197164948453607, "grad_norm": 0.6545103788375854, "learning_rate": 7.400879991868042e-06, "loss": 0.5787, "step": 1893 }, { "epoch": 1.2203608247422681, "grad_norm": 0.6422922611236572, "learning_rate": 7.397590871452425e-06, "loss": 0.5743, "step": 1894 }, { "epoch": 1.2210051546391751, "grad_norm": 0.5768683552742004, "learning_rate": 7.39430040317069e-06, "loss": 0.5544, "step": 1895 }, { "epoch": 1.2216494845360826, "grad_norm": 0.7317813634872437, "learning_rate": 7.391008588872661e-06, "loss": 0.568, "step": 1896 }, { "epoch": 1.2222938144329896, "grad_norm": 0.7040566205978394, "learning_rate": 7.387715430408913e-06, "loss": 0.573, "step": 1897 }, { "epoch": 1.222938144329897, "grad_norm": 0.5632281303405762, "learning_rate": 7.384420929630777e-06, "loss": 0.5618, "step": 1898 }, { "epoch": 1.2235824742268042, "grad_norm": 0.7030654549598694, "learning_rate": 7.38112508839034e-06, "loss": 0.5585, "step": 1899 }, { "epoch": 1.2242268041237114, "grad_norm": 0.6108115315437317, "learning_rate": 7.3778279085404425e-06, "loss": 0.5742, "step": 1900 }, { "epoch": 1.2248711340206186, "grad_norm": 0.5567194223403931, "learning_rate": 7.374529391934677e-06, "loss": 0.5745, "step": 1901 }, { "epoch": 1.2255154639175259, "grad_norm": 0.6026881337165833, "learning_rate": 7.371229540427388e-06, "loss": 0.5771, "step": 1902 }, { "epoch": 1.226159793814433, "grad_norm": 0.6576079726219177, "learning_rate": 7.367928355873671e-06, "loss": 0.5814, "step": 1903 }, { "epoch": 1.2268041237113403, "grad_norm": 0.6144108176231384, "learning_rate": 7.364625840129369e-06, "loss": 0.5668, "step": 1904 }, { "epoch": 1.2274484536082475, "grad_norm": 0.6987878084182739, "learning_rate": 7.361321995051073e-06, "loss": 0.5664, "step": 1905 }, { "epoch": 1.2280927835051547, "grad_norm": 0.5554780960083008, "learning_rate": 7.358016822496126e-06, "loss": 0.5813, "step": 1906 }, { "epoch": 1.228737113402062, "grad_norm": 0.5665278434753418, "learning_rate": 7.35471032432261e-06, "loss": 0.5799, "step": 1907 }, { "epoch": 1.2293814432989691, "grad_norm": 0.591500461101532, "learning_rate": 7.351402502389361e-06, "loss": 0.5619, "step": 1908 }, { "epoch": 1.2300257731958764, "grad_norm": 0.6596826910972595, "learning_rate": 7.34809335855595e-06, "loss": 0.5587, "step": 1909 }, { "epoch": 1.2306701030927836, "grad_norm": 0.6188045740127563, "learning_rate": 7.344782894682698e-06, "loss": 0.5757, "step": 1910 }, { "epoch": 1.2313144329896908, "grad_norm": 0.6493483781814575, "learning_rate": 7.341471112630667e-06, "loss": 0.5674, "step": 1911 }, { "epoch": 1.231958762886598, "grad_norm": 0.5654747486114502, "learning_rate": 7.338158014261656e-06, "loss": 0.5802, "step": 1912 }, { "epoch": 1.2326030927835052, "grad_norm": 0.5978965163230896, "learning_rate": 7.334843601438206e-06, "loss": 0.5583, "step": 1913 }, { "epoch": 1.2332474226804124, "grad_norm": 0.6230491995811462, "learning_rate": 7.331527876023602e-06, "loss": 0.5705, "step": 1914 }, { "epoch": 1.2338917525773196, "grad_norm": 0.6529566049575806, "learning_rate": 7.32821083988186e-06, "loss": 0.5621, "step": 1915 }, { "epoch": 1.2345360824742269, "grad_norm": 0.5428295731544495, "learning_rate": 7.324892494877734e-06, "loss": 0.5588, "step": 1916 }, { "epoch": 1.235180412371134, "grad_norm": 0.7511632442474365, "learning_rate": 7.321572842876715e-06, "loss": 0.5886, "step": 1917 }, { "epoch": 1.2358247422680413, "grad_norm": 0.5520836710929871, "learning_rate": 7.318251885745032e-06, "loss": 0.5598, "step": 1918 }, { "epoch": 1.2364690721649485, "grad_norm": 0.652940034866333, "learning_rate": 7.314929625349645e-06, "loss": 0.5787, "step": 1919 }, { "epoch": 1.2371134020618557, "grad_norm": 0.6092385649681091, "learning_rate": 7.311606063558243e-06, "loss": 0.5643, "step": 1920 }, { "epoch": 1.237757731958763, "grad_norm": 0.5927242636680603, "learning_rate": 7.308281202239252e-06, "loss": 0.5743, "step": 1921 }, { "epoch": 1.2384020618556701, "grad_norm": 0.5935609936714172, "learning_rate": 7.304955043261828e-06, "loss": 0.5645, "step": 1922 }, { "epoch": 1.2390463917525774, "grad_norm": 0.5664253830909729, "learning_rate": 7.301627588495851e-06, "loss": 0.5751, "step": 1923 }, { "epoch": 1.2396907216494846, "grad_norm": 0.6951310634613037, "learning_rate": 7.298298839811937e-06, "loss": 0.5701, "step": 1924 }, { "epoch": 1.2403350515463918, "grad_norm": 0.5603342652320862, "learning_rate": 7.294968799081423e-06, "loss": 0.5537, "step": 1925 }, { "epoch": 1.240979381443299, "grad_norm": 0.7179574966430664, "learning_rate": 7.29163746817638e-06, "loss": 0.5593, "step": 1926 }, { "epoch": 1.2416237113402062, "grad_norm": 0.5959005355834961, "learning_rate": 7.288304848969595e-06, "loss": 0.5578, "step": 1927 }, { "epoch": 1.2422680412371134, "grad_norm": 0.5333371758460999, "learning_rate": 7.284970943334586e-06, "loss": 0.571, "step": 1928 }, { "epoch": 1.2429123711340206, "grad_norm": 0.623350203037262, "learning_rate": 7.2816357531455925e-06, "loss": 0.5677, "step": 1929 }, { "epoch": 1.2435567010309279, "grad_norm": 0.6536988019943237, "learning_rate": 7.2782992802775754e-06, "loss": 0.5726, "step": 1930 }, { "epoch": 1.244201030927835, "grad_norm": 0.5610514283180237, "learning_rate": 7.274961526606215e-06, "loss": 0.5648, "step": 1931 }, { "epoch": 1.2448453608247423, "grad_norm": 0.5913098454475403, "learning_rate": 7.271622494007915e-06, "loss": 0.5608, "step": 1932 }, { "epoch": 1.2454896907216495, "grad_norm": 0.6304022073745728, "learning_rate": 7.268282184359798e-06, "loss": 0.5675, "step": 1933 }, { "epoch": 1.2461340206185567, "grad_norm": 0.6113458275794983, "learning_rate": 7.264940599539699e-06, "loss": 0.588, "step": 1934 }, { "epoch": 1.246778350515464, "grad_norm": 0.5569747686386108, "learning_rate": 7.26159774142618e-06, "loss": 0.5577, "step": 1935 }, { "epoch": 1.2474226804123711, "grad_norm": 0.7260902523994446, "learning_rate": 7.258253611898509e-06, "loss": 0.5611, "step": 1936 }, { "epoch": 1.2480670103092784, "grad_norm": 0.5902770161628723, "learning_rate": 7.254908212836673e-06, "loss": 0.5521, "step": 1937 }, { "epoch": 1.2487113402061856, "grad_norm": 0.5756947994232178, "learning_rate": 7.251561546121374e-06, "loss": 0.5742, "step": 1938 }, { "epoch": 1.2493556701030928, "grad_norm": 0.646477222442627, "learning_rate": 7.248213613634024e-06, "loss": 0.583, "step": 1939 }, { "epoch": 1.25, "grad_norm": 0.6173707842826843, "learning_rate": 7.244864417256747e-06, "loss": 0.5646, "step": 1940 }, { "epoch": 1.2506443298969072, "grad_norm": 0.5922741293907166, "learning_rate": 7.24151395887238e-06, "loss": 0.5642, "step": 1941 }, { "epoch": 1.2512886597938144, "grad_norm": 0.6101442575454712, "learning_rate": 7.238162240364465e-06, "loss": 0.5808, "step": 1942 }, { "epoch": 1.2519329896907216, "grad_norm": 0.6636485457420349, "learning_rate": 7.2348092636172575e-06, "loss": 0.5762, "step": 1943 }, { "epoch": 1.2525773195876289, "grad_norm": 0.5528267621994019, "learning_rate": 7.231455030515716e-06, "loss": 0.5789, "step": 1944 }, { "epoch": 1.253221649484536, "grad_norm": 0.584479808807373, "learning_rate": 7.228099542945507e-06, "loss": 0.5636, "step": 1945 }, { "epoch": 1.2538659793814433, "grad_norm": 0.5881912112236023, "learning_rate": 7.224742802793005e-06, "loss": 0.5951, "step": 1946 }, { "epoch": 1.2545103092783505, "grad_norm": 0.5662875771522522, "learning_rate": 7.221384811945284e-06, "loss": 0.5745, "step": 1947 }, { "epoch": 1.2551546391752577, "grad_norm": 0.5859895944595337, "learning_rate": 7.218025572290123e-06, "loss": 0.6003, "step": 1948 }, { "epoch": 1.255798969072165, "grad_norm": 0.5672014951705933, "learning_rate": 7.214665085716004e-06, "loss": 0.5704, "step": 1949 }, { "epoch": 1.2564432989690721, "grad_norm": 0.5690368413925171, "learning_rate": 7.211303354112108e-06, "loss": 0.5829, "step": 1950 }, { "epoch": 1.2570876288659794, "grad_norm": 0.5711697936058044, "learning_rate": 7.207940379368319e-06, "loss": 0.5839, "step": 1951 }, { "epoch": 1.2577319587628866, "grad_norm": 0.638728678226471, "learning_rate": 7.204576163375217e-06, "loss": 0.5704, "step": 1952 }, { "epoch": 1.2583762886597938, "grad_norm": 0.6120306253433228, "learning_rate": 7.201210708024078e-06, "loss": 0.5872, "step": 1953 }, { "epoch": 1.259020618556701, "grad_norm": 0.5573914647102356, "learning_rate": 7.197844015206883e-06, "loss": 0.5575, "step": 1954 }, { "epoch": 1.2596649484536082, "grad_norm": 0.6534932255744934, "learning_rate": 7.1944760868163e-06, "loss": 0.556, "step": 1955 }, { "epoch": 1.2603092783505154, "grad_norm": 0.5716878175735474, "learning_rate": 7.191106924745695e-06, "loss": 0.566, "step": 1956 }, { "epoch": 1.2609536082474226, "grad_norm": 0.6283462643623352, "learning_rate": 7.187736530889129e-06, "loss": 0.5736, "step": 1957 }, { "epoch": 1.2615979381443299, "grad_norm": 0.6502797603607178, "learning_rate": 7.184364907141351e-06, "loss": 0.5516, "step": 1958 }, { "epoch": 1.262242268041237, "grad_norm": 0.6316445469856262, "learning_rate": 7.180992055397805e-06, "loss": 0.5804, "step": 1959 }, { "epoch": 1.2628865979381443, "grad_norm": 0.5600289702415466, "learning_rate": 7.177617977554626e-06, "loss": 0.5462, "step": 1960 }, { "epoch": 1.2635309278350515, "grad_norm": 0.6329917311668396, "learning_rate": 7.174242675508636e-06, "loss": 0.5925, "step": 1961 }, { "epoch": 1.2641752577319587, "grad_norm": 0.6861703991889954, "learning_rate": 7.170866151157347e-06, "loss": 0.5699, "step": 1962 }, { "epoch": 1.264819587628866, "grad_norm": 0.6537634134292603, "learning_rate": 7.167488406398955e-06, "loss": 0.5544, "step": 1963 }, { "epoch": 1.2654639175257731, "grad_norm": 0.6253606677055359, "learning_rate": 7.1641094431323466e-06, "loss": 0.5689, "step": 1964 }, { "epoch": 1.2661082474226804, "grad_norm": 0.6050354838371277, "learning_rate": 7.160729263257091e-06, "loss": 0.5846, "step": 1965 }, { "epoch": 1.2667525773195876, "grad_norm": 0.6200399994850159, "learning_rate": 7.157347868673441e-06, "loss": 0.5892, "step": 1966 }, { "epoch": 1.2673969072164948, "grad_norm": 0.652214527130127, "learning_rate": 7.153965261282329e-06, "loss": 0.5798, "step": 1967 }, { "epoch": 1.268041237113402, "grad_norm": 0.5878923535346985, "learning_rate": 7.1505814429853805e-06, "loss": 0.5594, "step": 1968 }, { "epoch": 1.2686855670103092, "grad_norm": 0.6642172336578369, "learning_rate": 7.14719641568489e-06, "loss": 0.5711, "step": 1969 }, { "epoch": 1.2693298969072164, "grad_norm": 0.6486514806747437, "learning_rate": 7.143810181283834e-06, "loss": 0.5985, "step": 1970 }, { "epoch": 1.2699742268041236, "grad_norm": 0.5530940294265747, "learning_rate": 7.140422741685871e-06, "loss": 0.5859, "step": 1971 }, { "epoch": 1.2706185567010309, "grad_norm": 0.641739010810852, "learning_rate": 7.137034098795338e-06, "loss": 0.5792, "step": 1972 }, { "epoch": 1.271262886597938, "grad_norm": 0.6619211435317993, "learning_rate": 7.133644254517244e-06, "loss": 0.5744, "step": 1973 }, { "epoch": 1.2719072164948453, "grad_norm": 0.597416877746582, "learning_rate": 7.130253210757274e-06, "loss": 0.556, "step": 1974 }, { "epoch": 1.2725515463917525, "grad_norm": 0.6604141592979431, "learning_rate": 7.126860969421792e-06, "loss": 0.5655, "step": 1975 }, { "epoch": 1.2731958762886597, "grad_norm": 0.6420284509658813, "learning_rate": 7.1234675324178295e-06, "loss": 0.5745, "step": 1976 }, { "epoch": 1.2738402061855671, "grad_norm": 0.6427865624427795, "learning_rate": 7.120072901653091e-06, "loss": 0.5888, "step": 1977 }, { "epoch": 1.2744845360824741, "grad_norm": 0.6111615896224976, "learning_rate": 7.116677079035959e-06, "loss": 0.5721, "step": 1978 }, { "epoch": 1.2751288659793816, "grad_norm": 0.6245876550674438, "learning_rate": 7.113280066475476e-06, "loss": 0.5463, "step": 1979 }, { "epoch": 1.2757731958762886, "grad_norm": 0.598772406578064, "learning_rate": 7.10988186588136e-06, "loss": 0.5811, "step": 1980 }, { "epoch": 1.276417525773196, "grad_norm": 0.5607665777206421, "learning_rate": 7.106482479163994e-06, "loss": 0.5777, "step": 1981 }, { "epoch": 1.277061855670103, "grad_norm": 0.6919242739677429, "learning_rate": 7.103081908234431e-06, "loss": 0.5613, "step": 1982 }, { "epoch": 1.2777061855670104, "grad_norm": 0.5523735880851746, "learning_rate": 7.099680155004386e-06, "loss": 0.5695, "step": 1983 }, { "epoch": 1.2783505154639174, "grad_norm": 0.614302933216095, "learning_rate": 7.096277221386241e-06, "loss": 0.568, "step": 1984 }, { "epoch": 1.2789948453608249, "grad_norm": 0.6643432974815369, "learning_rate": 7.0928731092930404e-06, "loss": 0.5915, "step": 1985 }, { "epoch": 1.2796391752577319, "grad_norm": 0.5741017460823059, "learning_rate": 7.089467820638491e-06, "loss": 0.56, "step": 1986 }, { "epoch": 1.2802835051546393, "grad_norm": 0.5396902561187744, "learning_rate": 7.086061357336964e-06, "loss": 0.5675, "step": 1987 }, { "epoch": 1.2809278350515463, "grad_norm": 0.6370605230331421, "learning_rate": 7.082653721303485e-06, "loss": 0.5934, "step": 1988 }, { "epoch": 1.2815721649484537, "grad_norm": 0.6014312505722046, "learning_rate": 7.0792449144537445e-06, "loss": 0.574, "step": 1989 }, { "epoch": 1.2822164948453607, "grad_norm": 0.5979504585266113, "learning_rate": 7.075834938704091e-06, "loss": 0.5696, "step": 1990 }, { "epoch": 1.2828608247422681, "grad_norm": 0.617558479309082, "learning_rate": 7.072423795971526e-06, "loss": 0.5765, "step": 1991 }, { "epoch": 1.2835051546391751, "grad_norm": 0.5584973692893982, "learning_rate": 7.069011488173707e-06, "loss": 0.5796, "step": 1992 }, { "epoch": 1.2841494845360826, "grad_norm": 0.6024462580680847, "learning_rate": 7.065598017228954e-06, "loss": 0.5458, "step": 1993 }, { "epoch": 1.2847938144329896, "grad_norm": 0.5837967991828918, "learning_rate": 7.062183385056233e-06, "loss": 0.5882, "step": 1994 }, { "epoch": 1.285438144329897, "grad_norm": 0.6014271974563599, "learning_rate": 7.058767593575164e-06, "loss": 0.5701, "step": 1995 }, { "epoch": 1.286082474226804, "grad_norm": 0.5940923690795898, "learning_rate": 7.055350644706023e-06, "loss": 0.5797, "step": 1996 }, { "epoch": 1.2867268041237114, "grad_norm": 0.5609149932861328, "learning_rate": 7.0519325403697325e-06, "loss": 0.5725, "step": 1997 }, { "epoch": 1.2873711340206184, "grad_norm": 0.6353087425231934, "learning_rate": 7.0485132824878665e-06, "loss": 0.5653, "step": 1998 }, { "epoch": 1.2880154639175259, "grad_norm": 0.6124659180641174, "learning_rate": 7.045092872982646e-06, "loss": 0.5816, "step": 1999 }, { "epoch": 1.2886597938144329, "grad_norm": 0.5862791538238525, "learning_rate": 7.041671313776942e-06, "loss": 0.5678, "step": 2000 }, { "epoch": 1.2893041237113403, "grad_norm": 0.590772271156311, "learning_rate": 7.038248606794269e-06, "loss": 0.5782, "step": 2001 }, { "epoch": 1.2899484536082475, "grad_norm": 0.6315123438835144, "learning_rate": 7.034824753958789e-06, "loss": 0.5437, "step": 2002 }, { "epoch": 1.2905927835051547, "grad_norm": 0.5891753435134888, "learning_rate": 7.0313997571953055e-06, "loss": 0.5621, "step": 2003 }, { "epoch": 1.291237113402062, "grad_norm": 0.593658983707428, "learning_rate": 7.027973618429268e-06, "loss": 0.5709, "step": 2004 }, { "epoch": 1.2918814432989691, "grad_norm": 0.5413796305656433, "learning_rate": 7.024546339586767e-06, "loss": 0.5573, "step": 2005 }, { "epoch": 1.2925257731958764, "grad_norm": 0.6724035739898682, "learning_rate": 7.021117922594532e-06, "loss": 0.5739, "step": 2006 }, { "epoch": 1.2931701030927836, "grad_norm": 0.6257018446922302, "learning_rate": 7.0176883693799335e-06, "loss": 0.5638, "step": 2007 }, { "epoch": 1.2938144329896908, "grad_norm": 0.5654808878898621, "learning_rate": 7.014257681870985e-06, "loss": 0.5553, "step": 2008 }, { "epoch": 1.294458762886598, "grad_norm": 0.666853666305542, "learning_rate": 7.010825861996332e-06, "loss": 0.5673, "step": 2009 }, { "epoch": 1.2951030927835052, "grad_norm": 0.6589028239250183, "learning_rate": 7.007392911685257e-06, "loss": 0.5769, "step": 2010 }, { "epoch": 1.2957474226804124, "grad_norm": 0.5994409322738647, "learning_rate": 7.003958832867681e-06, "loss": 0.5561, "step": 2011 }, { "epoch": 1.2963917525773196, "grad_norm": 0.5318855047225952, "learning_rate": 7.0005236274741584e-06, "loss": 0.5487, "step": 2012 }, { "epoch": 1.2970360824742269, "grad_norm": 0.6418132781982422, "learning_rate": 6.997087297435874e-06, "loss": 0.5683, "step": 2013 }, { "epoch": 1.297680412371134, "grad_norm": 0.6177991628646851, "learning_rate": 6.99364984468465e-06, "loss": 0.5625, "step": 2014 }, { "epoch": 1.2983247422680413, "grad_norm": 0.594058096408844, "learning_rate": 6.9902112711529355e-06, "loss": 0.5671, "step": 2015 }, { "epoch": 1.2989690721649485, "grad_norm": 0.5748534798622131, "learning_rate": 6.986771578773812e-06, "loss": 0.5847, "step": 2016 }, { "epoch": 1.2996134020618557, "grad_norm": 0.5586365461349487, "learning_rate": 6.983330769480989e-06, "loss": 0.5558, "step": 2017 }, { "epoch": 1.300257731958763, "grad_norm": 0.607371985912323, "learning_rate": 6.979888845208805e-06, "loss": 0.5843, "step": 2018 }, { "epoch": 1.3009020618556701, "grad_norm": 0.5698740482330322, "learning_rate": 6.9764458078922235e-06, "loss": 0.5655, "step": 2019 }, { "epoch": 1.3015463917525774, "grad_norm": 0.5857729911804199, "learning_rate": 6.973001659466835e-06, "loss": 0.5658, "step": 2020 }, { "epoch": 1.3021907216494846, "grad_norm": 0.606586754322052, "learning_rate": 6.969556401868852e-06, "loss": 0.5603, "step": 2021 }, { "epoch": 1.3028350515463918, "grad_norm": 0.5570346713066101, "learning_rate": 6.966110037035118e-06, "loss": 0.5779, "step": 2022 }, { "epoch": 1.303479381443299, "grad_norm": 0.5575758218765259, "learning_rate": 6.962662566903089e-06, "loss": 0.5768, "step": 2023 }, { "epoch": 1.3041237113402062, "grad_norm": 0.5727883577346802, "learning_rate": 6.959213993410848e-06, "loss": 0.5675, "step": 2024 }, { "epoch": 1.3047680412371134, "grad_norm": 0.6129904389381409, "learning_rate": 6.955764318497099e-06, "loss": 0.5582, "step": 2025 }, { "epoch": 1.3054123711340206, "grad_norm": 0.5582519769668579, "learning_rate": 6.952313544101165e-06, "loss": 0.5561, "step": 2026 }, { "epoch": 1.3060567010309279, "grad_norm": 0.5822880864143372, "learning_rate": 6.948861672162983e-06, "loss": 0.5637, "step": 2027 }, { "epoch": 1.306701030927835, "grad_norm": 0.6168795824050903, "learning_rate": 6.945408704623109e-06, "loss": 0.5557, "step": 2028 }, { "epoch": 1.3073453608247423, "grad_norm": 0.5765210390090942, "learning_rate": 6.941954643422719e-06, "loss": 0.5774, "step": 2029 }, { "epoch": 1.3079896907216495, "grad_norm": 0.6217520236968994, "learning_rate": 6.9384994905035986e-06, "loss": 0.575, "step": 2030 }, { "epoch": 1.3086340206185567, "grad_norm": 0.6104972958564758, "learning_rate": 6.935043247808149e-06, "loss": 0.5496, "step": 2031 }, { "epoch": 1.309278350515464, "grad_norm": 0.5756781101226807, "learning_rate": 6.931585917279386e-06, "loss": 0.5567, "step": 2032 }, { "epoch": 1.3099226804123711, "grad_norm": 0.5611582398414612, "learning_rate": 6.92812750086093e-06, "loss": 0.5813, "step": 2033 }, { "epoch": 1.3105670103092784, "grad_norm": 0.6425173282623291, "learning_rate": 6.9246680004970245e-06, "loss": 0.5786, "step": 2034 }, { "epoch": 1.3112113402061856, "grad_norm": 0.5688796043395996, "learning_rate": 6.9212074181325085e-06, "loss": 0.5827, "step": 2035 }, { "epoch": 1.3118556701030928, "grad_norm": 0.5794715881347656, "learning_rate": 6.917745755712839e-06, "loss": 0.5724, "step": 2036 }, { "epoch": 1.3125, "grad_norm": 0.6376932859420776, "learning_rate": 6.9142830151840754e-06, "loss": 0.5833, "step": 2037 }, { "epoch": 1.3131443298969072, "grad_norm": 0.5732958912849426, "learning_rate": 6.910819198492884e-06, "loss": 0.5621, "step": 2038 }, { "epoch": 1.3137886597938144, "grad_norm": 0.5818691849708557, "learning_rate": 6.907354307586537e-06, "loss": 0.5763, "step": 2039 }, { "epoch": 1.3144329896907216, "grad_norm": 0.6242313385009766, "learning_rate": 6.903888344412911e-06, "loss": 0.5919, "step": 2040 }, { "epoch": 1.3150773195876289, "grad_norm": 0.6244560480117798, "learning_rate": 6.900421310920484e-06, "loss": 0.5426, "step": 2041 }, { "epoch": 1.315721649484536, "grad_norm": 0.5210911631584167, "learning_rate": 6.896953209058337e-06, "loss": 0.5813, "step": 2042 }, { "epoch": 1.3163659793814433, "grad_norm": 0.6142969727516174, "learning_rate": 6.893484040776148e-06, "loss": 0.5936, "step": 2043 }, { "epoch": 1.3170103092783505, "grad_norm": 0.5935072302818298, "learning_rate": 6.890013808024201e-06, "loss": 0.5483, "step": 2044 }, { "epoch": 1.3176546391752577, "grad_norm": 0.5600674152374268, "learning_rate": 6.886542512753373e-06, "loss": 0.5648, "step": 2045 }, { "epoch": 1.318298969072165, "grad_norm": 0.6163585782051086, "learning_rate": 6.8830701569151394e-06, "loss": 0.5875, "step": 2046 }, { "epoch": 1.3189432989690721, "grad_norm": 0.5813161134719849, "learning_rate": 6.879596742461574e-06, "loss": 0.5749, "step": 2047 }, { "epoch": 1.3195876288659794, "grad_norm": 0.6103264093399048, "learning_rate": 6.876122271345345e-06, "loss": 0.5642, "step": 2048 }, { "epoch": 1.3202319587628866, "grad_norm": 0.597240149974823, "learning_rate": 6.87264674551971e-06, "loss": 0.5806, "step": 2049 }, { "epoch": 1.3208762886597938, "grad_norm": 0.5751457214355469, "learning_rate": 6.869170166938528e-06, "loss": 0.5495, "step": 2050 }, { "epoch": 1.321520618556701, "grad_norm": 0.5440584421157837, "learning_rate": 6.865692537556241e-06, "loss": 0.5711, "step": 2051 }, { "epoch": 1.3221649484536082, "grad_norm": 0.5629772543907166, "learning_rate": 6.8622138593278895e-06, "loss": 0.568, "step": 2052 }, { "epoch": 1.3228092783505154, "grad_norm": 0.5181488990783691, "learning_rate": 6.858734134209097e-06, "loss": 0.5644, "step": 2053 }, { "epoch": 1.3234536082474226, "grad_norm": 0.6398852467536926, "learning_rate": 6.855253364156083e-06, "loss": 0.5677, "step": 2054 }, { "epoch": 1.3240979381443299, "grad_norm": 0.6233486533164978, "learning_rate": 6.851771551125646e-06, "loss": 0.5705, "step": 2055 }, { "epoch": 1.324742268041237, "grad_norm": 0.608259379863739, "learning_rate": 6.8482886970751785e-06, "loss": 0.5576, "step": 2056 }, { "epoch": 1.3253865979381443, "grad_norm": 0.6154826879501343, "learning_rate": 6.84480480396265e-06, "loss": 0.5798, "step": 2057 }, { "epoch": 1.3260309278350515, "grad_norm": 0.649985134601593, "learning_rate": 6.841319873746623e-06, "loss": 0.577, "step": 2058 }, { "epoch": 1.3266752577319587, "grad_norm": 0.5663392543792725, "learning_rate": 6.837833908386236e-06, "loss": 0.5722, "step": 2059 }, { "epoch": 1.327319587628866, "grad_norm": 0.5436640381813049, "learning_rate": 6.834346909841213e-06, "loss": 0.5732, "step": 2060 }, { "epoch": 1.3279639175257731, "grad_norm": 0.5217045545578003, "learning_rate": 6.830858880071859e-06, "loss": 0.5505, "step": 2061 }, { "epoch": 1.3286082474226804, "grad_norm": 0.5479525327682495, "learning_rate": 6.827369821039057e-06, "loss": 0.581, "step": 2062 }, { "epoch": 1.3292525773195876, "grad_norm": 0.6139518618583679, "learning_rate": 6.823879734704271e-06, "loss": 0.5918, "step": 2063 }, { "epoch": 1.3298969072164948, "grad_norm": 0.5549580454826355, "learning_rate": 6.820388623029538e-06, "loss": 0.5659, "step": 2064 }, { "epoch": 1.330541237113402, "grad_norm": 0.5647253394126892, "learning_rate": 6.816896487977478e-06, "loss": 0.5669, "step": 2065 }, { "epoch": 1.3311855670103092, "grad_norm": 0.5602126121520996, "learning_rate": 6.81340333151128e-06, "loss": 0.5742, "step": 2066 }, { "epoch": 1.3318298969072164, "grad_norm": 0.5850037336349487, "learning_rate": 6.8099091555947095e-06, "loss": 0.5692, "step": 2067 }, { "epoch": 1.3324742268041236, "grad_norm": 0.5803621411323547, "learning_rate": 6.806413962192109e-06, "loss": 0.5701, "step": 2068 }, { "epoch": 1.3331185567010309, "grad_norm": 0.6022758483886719, "learning_rate": 6.802917753268386e-06, "loss": 0.5939, "step": 2069 }, { "epoch": 1.333762886597938, "grad_norm": 0.6018967032432556, "learning_rate": 6.799420530789024e-06, "loss": 0.5722, "step": 2070 }, { "epoch": 1.3344072164948453, "grad_norm": 0.529085636138916, "learning_rate": 6.795922296720074e-06, "loss": 0.5878, "step": 2071 }, { "epoch": 1.3350515463917525, "grad_norm": 0.5613335371017456, "learning_rate": 6.792423053028156e-06, "loss": 0.591, "step": 2072 }, { "epoch": 1.3356958762886597, "grad_norm": 0.603859007358551, "learning_rate": 6.788922801680461e-06, "loss": 0.5657, "step": 2073 }, { "epoch": 1.3363402061855671, "grad_norm": 0.5399255156517029, "learning_rate": 6.785421544644742e-06, "loss": 0.5565, "step": 2074 }, { "epoch": 1.3369845360824741, "grad_norm": 0.5494346618652344, "learning_rate": 6.781919283889314e-06, "loss": 0.5738, "step": 2075 }, { "epoch": 1.3376288659793816, "grad_norm": 0.6141780018806458, "learning_rate": 6.7784160213830696e-06, "loss": 0.5719, "step": 2076 }, { "epoch": 1.3382731958762886, "grad_norm": 0.5502314567565918, "learning_rate": 6.7749117590954506e-06, "loss": 0.5721, "step": 2077 }, { "epoch": 1.338917525773196, "grad_norm": 0.6284996867179871, "learning_rate": 6.771406498996465e-06, "loss": 0.5631, "step": 2078 }, { "epoch": 1.339561855670103, "grad_norm": 0.6330024003982544, "learning_rate": 6.767900243056688e-06, "loss": 0.5722, "step": 2079 }, { "epoch": 1.3402061855670104, "grad_norm": 0.5579198002815247, "learning_rate": 6.7643929932472475e-06, "loss": 0.5711, "step": 2080 }, { "epoch": 1.3408505154639174, "grad_norm": 0.5927435755729675, "learning_rate": 6.760884751539831e-06, "loss": 0.5769, "step": 2081 }, { "epoch": 1.3414948453608249, "grad_norm": 0.6358541250228882, "learning_rate": 6.757375519906687e-06, "loss": 0.5573, "step": 2082 }, { "epoch": 1.3421391752577319, "grad_norm": 0.5892969369888306, "learning_rate": 6.753865300320618e-06, "loss": 0.5715, "step": 2083 }, { "epoch": 1.3427835051546393, "grad_norm": 0.6171767711639404, "learning_rate": 6.750354094754982e-06, "loss": 0.5742, "step": 2084 }, { "epoch": 1.3434278350515463, "grad_norm": 0.5769767761230469, "learning_rate": 6.7468419051836895e-06, "loss": 0.5511, "step": 2085 }, { "epoch": 1.3440721649484537, "grad_norm": 0.553989589214325, "learning_rate": 6.743328733581211e-06, "loss": 0.5843, "step": 2086 }, { "epoch": 1.3447164948453607, "grad_norm": 0.5887576937675476, "learning_rate": 6.739814581922562e-06, "loss": 0.5733, "step": 2087 }, { "epoch": 1.3453608247422681, "grad_norm": 0.6372261047363281, "learning_rate": 6.73629945218331e-06, "loss": 0.5853, "step": 2088 }, { "epoch": 1.3460051546391751, "grad_norm": 0.5865439176559448, "learning_rate": 6.732783346339575e-06, "loss": 0.5446, "step": 2089 }, { "epoch": 1.3466494845360826, "grad_norm": 0.6068906188011169, "learning_rate": 6.729266266368026e-06, "loss": 0.5781, "step": 2090 }, { "epoch": 1.3472938144329896, "grad_norm": 0.6300415992736816, "learning_rate": 6.725748214245876e-06, "loss": 0.5774, "step": 2091 }, { "epoch": 1.347938144329897, "grad_norm": 0.5384080410003662, "learning_rate": 6.722229191950887e-06, "loss": 0.5759, "step": 2092 }, { "epoch": 1.348582474226804, "grad_norm": 0.5510519742965698, "learning_rate": 6.7187092014613665e-06, "loss": 0.5711, "step": 2093 }, { "epoch": 1.3492268041237114, "grad_norm": 0.6873962879180908, "learning_rate": 6.7151882447561664e-06, "loss": 0.5516, "step": 2094 }, { "epoch": 1.3498711340206184, "grad_norm": 0.5698704719543457, "learning_rate": 6.711666323814681e-06, "loss": 0.5687, "step": 2095 }, { "epoch": 1.3505154639175259, "grad_norm": 0.6262398958206177, "learning_rate": 6.708143440616845e-06, "loss": 0.5592, "step": 2096 }, { "epoch": 1.3511597938144329, "grad_norm": 0.5507794618606567, "learning_rate": 6.7046195971431375e-06, "loss": 0.5712, "step": 2097 }, { "epoch": 1.3518041237113403, "grad_norm": 0.5709766745567322, "learning_rate": 6.7010947953745766e-06, "loss": 0.5689, "step": 2098 }, { "epoch": 1.3524484536082475, "grad_norm": 0.5939895510673523, "learning_rate": 6.697569037292717e-06, "loss": 0.5635, "step": 2099 }, { "epoch": 1.3530927835051547, "grad_norm": 0.5749164819717407, "learning_rate": 6.694042324879653e-06, "loss": 0.5675, "step": 2100 }, { "epoch": 1.353737113402062, "grad_norm": 0.6403311491012573, "learning_rate": 6.690514660118016e-06, "loss": 0.5795, "step": 2101 }, { "epoch": 1.3543814432989691, "grad_norm": 0.6052985787391663, "learning_rate": 6.68698604499097e-06, "loss": 0.5815, "step": 2102 }, { "epoch": 1.3550257731958764, "grad_norm": 0.6518397927284241, "learning_rate": 6.683456481482215e-06, "loss": 0.5793, "step": 2103 }, { "epoch": 1.3556701030927836, "grad_norm": 0.5896772742271423, "learning_rate": 6.679925971575985e-06, "loss": 0.6028, "step": 2104 }, { "epoch": 1.3563144329896908, "grad_norm": 0.6057361960411072, "learning_rate": 6.676394517257044e-06, "loss": 0.5882, "step": 2105 }, { "epoch": 1.356958762886598, "grad_norm": 0.6701005101203918, "learning_rate": 6.672862120510688e-06, "loss": 0.5734, "step": 2106 }, { "epoch": 1.3576030927835052, "grad_norm": 0.5927100777626038, "learning_rate": 6.669328783322744e-06, "loss": 0.5632, "step": 2107 }, { "epoch": 1.3582474226804124, "grad_norm": 0.5736852884292603, "learning_rate": 6.665794507679566e-06, "loss": 0.5862, "step": 2108 }, { "epoch": 1.3588917525773196, "grad_norm": 0.562252938747406, "learning_rate": 6.662259295568036e-06, "loss": 0.5661, "step": 2109 }, { "epoch": 1.3595360824742269, "grad_norm": 0.6337783336639404, "learning_rate": 6.658723148975563e-06, "loss": 0.5658, "step": 2110 }, { "epoch": 1.360180412371134, "grad_norm": 0.6357012987136841, "learning_rate": 6.6551860698900804e-06, "loss": 0.5709, "step": 2111 }, { "epoch": 1.3608247422680413, "grad_norm": 0.6229046583175659, "learning_rate": 6.6516480603000464e-06, "loss": 0.5597, "step": 2112 }, { "epoch": 1.3614690721649485, "grad_norm": 0.6252800226211548, "learning_rate": 6.648109122194443e-06, "loss": 0.5653, "step": 2113 }, { "epoch": 1.3621134020618557, "grad_norm": 0.6487570405006409, "learning_rate": 6.644569257562771e-06, "loss": 0.5652, "step": 2114 }, { "epoch": 1.362757731958763, "grad_norm": 0.5861597061157227, "learning_rate": 6.641028468395055e-06, "loss": 0.5686, "step": 2115 }, { "epoch": 1.3634020618556701, "grad_norm": 0.6169604659080505, "learning_rate": 6.637486756681843e-06, "loss": 0.5532, "step": 2116 }, { "epoch": 1.3640463917525774, "grad_norm": 0.6757826805114746, "learning_rate": 6.633944124414194e-06, "loss": 0.568, "step": 2117 }, { "epoch": 1.3646907216494846, "grad_norm": 0.6993438005447388, "learning_rate": 6.6304005735836865e-06, "loss": 0.57, "step": 2118 }, { "epoch": 1.3653350515463918, "grad_norm": 0.5546597242355347, "learning_rate": 6.626856106182421e-06, "loss": 0.58, "step": 2119 }, { "epoch": 1.365979381443299, "grad_norm": 0.5638024210929871, "learning_rate": 6.623310724203006e-06, "loss": 0.5646, "step": 2120 }, { "epoch": 1.3666237113402062, "grad_norm": 0.6241981387138367, "learning_rate": 6.619764429638567e-06, "loss": 0.5889, "step": 2121 }, { "epoch": 1.3672680412371134, "grad_norm": 0.642500638961792, "learning_rate": 6.616217224482746e-06, "loss": 0.5806, "step": 2122 }, { "epoch": 1.3679123711340206, "grad_norm": 0.5710388422012329, "learning_rate": 6.612669110729693e-06, "loss": 0.5789, "step": 2123 }, { "epoch": 1.3685567010309279, "grad_norm": 0.5975913405418396, "learning_rate": 6.609120090374065e-06, "loss": 0.5563, "step": 2124 }, { "epoch": 1.369201030927835, "grad_norm": 0.5452626943588257, "learning_rate": 6.605570165411038e-06, "loss": 0.5667, "step": 2125 }, { "epoch": 1.3698453608247423, "grad_norm": 0.5990256071090698, "learning_rate": 6.602019337836291e-06, "loss": 0.5549, "step": 2126 }, { "epoch": 1.3704896907216495, "grad_norm": 0.617241621017456, "learning_rate": 6.598467609646011e-06, "loss": 0.5957, "step": 2127 }, { "epoch": 1.3711340206185567, "grad_norm": 0.5741942524909973, "learning_rate": 6.5949149828368885e-06, "loss": 0.5525, "step": 2128 }, { "epoch": 1.371778350515464, "grad_norm": 0.6105802655220032, "learning_rate": 6.591361459406127e-06, "loss": 0.5678, "step": 2129 }, { "epoch": 1.3724226804123711, "grad_norm": 0.6270091533660889, "learning_rate": 6.5878070413514275e-06, "loss": 0.5888, "step": 2130 }, { "epoch": 1.3730670103092784, "grad_norm": 0.5665619969367981, "learning_rate": 6.584251730670994e-06, "loss": 0.5723, "step": 2131 }, { "epoch": 1.3737113402061856, "grad_norm": 0.5916158556938171, "learning_rate": 6.580695529363534e-06, "loss": 0.5673, "step": 2132 }, { "epoch": 1.3743556701030928, "grad_norm": 0.5557854175567627, "learning_rate": 6.577138439428258e-06, "loss": 0.565, "step": 2133 }, { "epoch": 1.375, "grad_norm": 0.5594346523284912, "learning_rate": 6.573580462864873e-06, "loss": 0.551, "step": 2134 }, { "epoch": 1.3756443298969072, "grad_norm": 0.6235532164573669, "learning_rate": 6.570021601673584e-06, "loss": 0.5752, "step": 2135 }, { "epoch": 1.3762886597938144, "grad_norm": 0.5828513503074646, "learning_rate": 6.566461857855096e-06, "loss": 0.5716, "step": 2136 }, { "epoch": 1.3769329896907216, "grad_norm": 0.5773240327835083, "learning_rate": 6.562901233410606e-06, "loss": 0.5706, "step": 2137 }, { "epoch": 1.3775773195876289, "grad_norm": 0.6201355457305908, "learning_rate": 6.559339730341812e-06, "loss": 0.5594, "step": 2138 }, { "epoch": 1.378221649484536, "grad_norm": 0.5988901257514954, "learning_rate": 6.5557773506509e-06, "loss": 0.5853, "step": 2139 }, { "epoch": 1.3788659793814433, "grad_norm": 0.5601123571395874, "learning_rate": 6.552214096340555e-06, "loss": 0.5571, "step": 2140 }, { "epoch": 1.3795103092783505, "grad_norm": 0.6576920747756958, "learning_rate": 6.548649969413945e-06, "loss": 0.5778, "step": 2141 }, { "epoch": 1.3801546391752577, "grad_norm": 0.5872002243995667, "learning_rate": 6.545084971874738e-06, "loss": 0.5854, "step": 2142 }, { "epoch": 1.380798969072165, "grad_norm": 0.6057130694389343, "learning_rate": 6.541519105727086e-06, "loss": 0.5756, "step": 2143 }, { "epoch": 1.3814432989690721, "grad_norm": 0.6306664943695068, "learning_rate": 6.537952372975629e-06, "loss": 0.5698, "step": 2144 }, { "epoch": 1.3820876288659794, "grad_norm": 0.5763765573501587, "learning_rate": 6.5343847756255e-06, "loss": 0.5729, "step": 2145 }, { "epoch": 1.3827319587628866, "grad_norm": 0.6002846360206604, "learning_rate": 6.5308163156823064e-06, "loss": 0.5677, "step": 2146 }, { "epoch": 1.3833762886597938, "grad_norm": 0.575271725654602, "learning_rate": 6.527246995152157e-06, "loss": 0.5671, "step": 2147 }, { "epoch": 1.384020618556701, "grad_norm": 0.6150466203689575, "learning_rate": 6.523676816041627e-06, "loss": 0.5802, "step": 2148 }, { "epoch": 1.3846649484536082, "grad_norm": 0.6165627837181091, "learning_rate": 6.520105780357788e-06, "loss": 0.5758, "step": 2149 }, { "epoch": 1.3853092783505154, "grad_norm": 0.548194408416748, "learning_rate": 6.516533890108184e-06, "loss": 0.5687, "step": 2150 }, { "epoch": 1.3859536082474226, "grad_norm": 0.5699905753135681, "learning_rate": 6.512961147300846e-06, "loss": 0.5712, "step": 2151 }, { "epoch": 1.3865979381443299, "grad_norm": 0.5820119380950928, "learning_rate": 6.50938755394428e-06, "loss": 0.5596, "step": 2152 }, { "epoch": 1.387242268041237, "grad_norm": 0.5686125755310059, "learning_rate": 6.505813112047472e-06, "loss": 0.5648, "step": 2153 }, { "epoch": 1.3878865979381443, "grad_norm": 0.5726298689842224, "learning_rate": 6.502237823619883e-06, "loss": 0.5569, "step": 2154 }, { "epoch": 1.3885309278350515, "grad_norm": 0.6043676733970642, "learning_rate": 6.498661690671454e-06, "loss": 0.5887, "step": 2155 }, { "epoch": 1.3891752577319587, "grad_norm": 0.6354026198387146, "learning_rate": 6.495084715212597e-06, "loss": 0.5906, "step": 2156 }, { "epoch": 1.389819587628866, "grad_norm": 0.5917797684669495, "learning_rate": 6.491506899254198e-06, "loss": 0.5846, "step": 2157 }, { "epoch": 1.3904639175257731, "grad_norm": 0.705446183681488, "learning_rate": 6.4879282448076185e-06, "loss": 0.5729, "step": 2158 }, { "epoch": 1.3911082474226804, "grad_norm": 0.5939332842826843, "learning_rate": 6.484348753884689e-06, "loss": 0.5737, "step": 2159 }, { "epoch": 1.3917525773195876, "grad_norm": 0.5548120141029358, "learning_rate": 6.480768428497706e-06, "loss": 0.5691, "step": 2160 }, { "epoch": 1.3923969072164948, "grad_norm": 0.5753044486045837, "learning_rate": 6.477187270659444e-06, "loss": 0.5776, "step": 2161 }, { "epoch": 1.393041237113402, "grad_norm": 0.6489563584327698, "learning_rate": 6.473605282383142e-06, "loss": 0.5641, "step": 2162 }, { "epoch": 1.3936855670103092, "grad_norm": 0.547673225402832, "learning_rate": 6.470022465682503e-06, "loss": 0.5721, "step": 2163 }, { "epoch": 1.3943298969072164, "grad_norm": 0.6014969944953918, "learning_rate": 6.4664388225716965e-06, "loss": 0.6024, "step": 2164 }, { "epoch": 1.3949742268041236, "grad_norm": 0.6771320104598999, "learning_rate": 6.46285435506536e-06, "loss": 0.583, "step": 2165 }, { "epoch": 1.3956185567010309, "grad_norm": 0.5489358305931091, "learning_rate": 6.459269065178592e-06, "loss": 0.5481, "step": 2166 }, { "epoch": 1.396262886597938, "grad_norm": 0.6003953814506531, "learning_rate": 6.455682954926953e-06, "loss": 0.591, "step": 2167 }, { "epoch": 1.3969072164948453, "grad_norm": 0.6595608592033386, "learning_rate": 6.452096026326463e-06, "loss": 0.5605, "step": 2168 }, { "epoch": 1.3975515463917525, "grad_norm": 0.619280219078064, "learning_rate": 6.448508281393607e-06, "loss": 0.5659, "step": 2169 }, { "epoch": 1.3981958762886597, "grad_norm": 0.5611604452133179, "learning_rate": 6.444919722145327e-06, "loss": 0.5634, "step": 2170 }, { "epoch": 1.3988402061855671, "grad_norm": 0.6909818053245544, "learning_rate": 6.441330350599021e-06, "loss": 0.5916, "step": 2171 }, { "epoch": 1.3994845360824741, "grad_norm": 0.6425567269325256, "learning_rate": 6.4377401687725426e-06, "loss": 0.5508, "step": 2172 }, { "epoch": 1.4001288659793816, "grad_norm": 0.5924858450889587, "learning_rate": 6.4341491786842055e-06, "loss": 0.5687, "step": 2173 }, { "epoch": 1.4007731958762886, "grad_norm": 0.6679426431655884, "learning_rate": 6.430557382352775e-06, "loss": 0.5804, "step": 2174 }, { "epoch": 1.401417525773196, "grad_norm": 0.5949759483337402, "learning_rate": 6.426964781797468e-06, "loss": 0.5677, "step": 2175 }, { "epoch": 1.402061855670103, "grad_norm": 0.6090109348297119, "learning_rate": 6.423371379037957e-06, "loss": 0.5707, "step": 2176 }, { "epoch": 1.4027061855670104, "grad_norm": 0.62115079164505, "learning_rate": 6.419777176094365e-06, "loss": 0.5591, "step": 2177 }, { "epoch": 1.4033505154639174, "grad_norm": 0.5420624613761902, "learning_rate": 6.41618217498726e-06, "loss": 0.5801, "step": 2178 }, { "epoch": 1.4039948453608249, "grad_norm": 0.6740012168884277, "learning_rate": 6.412586377737666e-06, "loss": 0.5738, "step": 2179 }, { "epoch": 1.4046391752577319, "grad_norm": 0.5683395862579346, "learning_rate": 6.408989786367051e-06, "loss": 0.583, "step": 2180 }, { "epoch": 1.4052835051546393, "grad_norm": 0.5988467931747437, "learning_rate": 6.405392402897327e-06, "loss": 0.5616, "step": 2181 }, { "epoch": 1.4059278350515463, "grad_norm": 0.5757729411125183, "learning_rate": 6.401794229350855e-06, "loss": 0.576, "step": 2182 }, { "epoch": 1.4065721649484537, "grad_norm": 0.697544276714325, "learning_rate": 6.398195267750439e-06, "loss": 0.5739, "step": 2183 }, { "epoch": 1.4072164948453607, "grad_norm": 0.6235569715499878, "learning_rate": 6.394595520119326e-06, "loss": 0.569, "step": 2184 }, { "epoch": 1.4078608247422681, "grad_norm": 0.614134669303894, "learning_rate": 6.390994988481205e-06, "loss": 0.5504, "step": 2185 }, { "epoch": 1.4085051546391751, "grad_norm": 0.5959309935569763, "learning_rate": 6.387393674860205e-06, "loss": 0.564, "step": 2186 }, { "epoch": 1.4091494845360826, "grad_norm": 0.5773481726646423, "learning_rate": 6.383791581280895e-06, "loss": 0.5724, "step": 2187 }, { "epoch": 1.4097938144329896, "grad_norm": 0.6968047618865967, "learning_rate": 6.380188709768283e-06, "loss": 0.5792, "step": 2188 }, { "epoch": 1.410438144329897, "grad_norm": 0.6231539845466614, "learning_rate": 6.376585062347816e-06, "loss": 0.5603, "step": 2189 }, { "epoch": 1.411082474226804, "grad_norm": 0.5863984823226929, "learning_rate": 6.372980641045371e-06, "loss": 0.5603, "step": 2190 }, { "epoch": 1.4117268041237114, "grad_norm": 0.6275905966758728, "learning_rate": 6.369375447887269e-06, "loss": 0.582, "step": 2191 }, { "epoch": 1.4123711340206184, "grad_norm": 0.5436961054801941, "learning_rate": 6.365769484900256e-06, "loss": 0.5714, "step": 2192 }, { "epoch": 1.4130154639175259, "grad_norm": 0.6650251746177673, "learning_rate": 6.362162754111516e-06, "loss": 0.5555, "step": 2193 }, { "epoch": 1.4136597938144329, "grad_norm": 0.5670956969261169, "learning_rate": 6.358555257548667e-06, "loss": 0.5736, "step": 2194 }, { "epoch": 1.4143041237113403, "grad_norm": 0.6057481169700623, "learning_rate": 6.354946997239749e-06, "loss": 0.5831, "step": 2195 }, { "epoch": 1.4149484536082475, "grad_norm": 0.6890610456466675, "learning_rate": 6.351337975213239e-06, "loss": 0.5767, "step": 2196 }, { "epoch": 1.4155927835051547, "grad_norm": 0.6425281167030334, "learning_rate": 6.34772819349804e-06, "loss": 0.5766, "step": 2197 }, { "epoch": 1.416237113402062, "grad_norm": 0.6397491097450256, "learning_rate": 6.34411765412348e-06, "loss": 0.5671, "step": 2198 }, { "epoch": 1.4168814432989691, "grad_norm": 0.6880882382392883, "learning_rate": 6.3405063591193186e-06, "loss": 0.5574, "step": 2199 }, { "epoch": 1.4175257731958764, "grad_norm": 0.5278183817863464, "learning_rate": 6.33689431051573e-06, "loss": 0.5876, "step": 2200 }, { "epoch": 1.4181701030927836, "grad_norm": 0.5402125120162964, "learning_rate": 6.3332815103433245e-06, "loss": 0.5733, "step": 2201 }, { "epoch": 1.4188144329896908, "grad_norm": 0.5976809859275818, "learning_rate": 6.329667960633125e-06, "loss": 0.5773, "step": 2202 }, { "epoch": 1.419458762886598, "grad_norm": 0.6188656687736511, "learning_rate": 6.326053663416581e-06, "loss": 0.5749, "step": 2203 }, { "epoch": 1.4201030927835052, "grad_norm": 0.5613150000572205, "learning_rate": 6.3224386207255605e-06, "loss": 0.5588, "step": 2204 }, { "epoch": 1.4207474226804124, "grad_norm": 0.5557404160499573, "learning_rate": 6.31882283459235e-06, "loss": 0.58, "step": 2205 }, { "epoch": 1.4213917525773196, "grad_norm": 0.684718132019043, "learning_rate": 6.315206307049656e-06, "loss": 0.5477, "step": 2206 }, { "epoch": 1.4220360824742269, "grad_norm": 0.5605690479278564, "learning_rate": 6.3115890401306005e-06, "loss": 0.5707, "step": 2207 }, { "epoch": 1.422680412371134, "grad_norm": 0.606815755367279, "learning_rate": 6.30797103586872e-06, "loss": 0.5686, "step": 2208 }, { "epoch": 1.4233247422680413, "grad_norm": 0.6261694431304932, "learning_rate": 6.3043522962979675e-06, "loss": 0.568, "step": 2209 }, { "epoch": 1.4239690721649485, "grad_norm": 0.5703265070915222, "learning_rate": 6.3007328234527084e-06, "loss": 0.559, "step": 2210 }, { "epoch": 1.4246134020618557, "grad_norm": 0.5849876403808594, "learning_rate": 6.297112619367721e-06, "loss": 0.5568, "step": 2211 }, { "epoch": 1.425257731958763, "grad_norm": 0.610697329044342, "learning_rate": 6.293491686078194e-06, "loss": 0.5571, "step": 2212 }, { "epoch": 1.4259020618556701, "grad_norm": 0.5583590865135193, "learning_rate": 6.289870025619729e-06, "loss": 0.5603, "step": 2213 }, { "epoch": 1.4265463917525774, "grad_norm": 0.6998155117034912, "learning_rate": 6.2862476400283265e-06, "loss": 0.5708, "step": 2214 }, { "epoch": 1.4271907216494846, "grad_norm": 0.5908988118171692, "learning_rate": 6.282624531340409e-06, "loss": 0.5639, "step": 2215 }, { "epoch": 1.4278350515463918, "grad_norm": 0.6591269969940186, "learning_rate": 6.2790007015927946e-06, "loss": 0.5637, "step": 2216 }, { "epoch": 1.428479381443299, "grad_norm": 0.5577413439750671, "learning_rate": 6.275376152822712e-06, "loss": 0.5657, "step": 2217 }, { "epoch": 1.4291237113402062, "grad_norm": 0.6307860612869263, "learning_rate": 6.271750887067792e-06, "loss": 0.5869, "step": 2218 }, { "epoch": 1.4297680412371134, "grad_norm": 0.6133348941802979, "learning_rate": 6.2681249063660695e-06, "loss": 0.5838, "step": 2219 }, { "epoch": 1.4304123711340206, "grad_norm": 0.5639550685882568, "learning_rate": 6.264498212755981e-06, "loss": 0.5676, "step": 2220 }, { "epoch": 1.4310567010309279, "grad_norm": 0.7488969564437866, "learning_rate": 6.260870808276364e-06, "loss": 0.571, "step": 2221 }, { "epoch": 1.431701030927835, "grad_norm": 0.6306519508361816, "learning_rate": 6.257242694966452e-06, "loss": 0.588, "step": 2222 }, { "epoch": 1.4323453608247423, "grad_norm": 0.6267641186714172, "learning_rate": 6.253613874865884e-06, "loss": 0.5741, "step": 2223 }, { "epoch": 1.4329896907216495, "grad_norm": 0.7122153639793396, "learning_rate": 6.249984350014693e-06, "loss": 0.577, "step": 2224 }, { "epoch": 1.4336340206185567, "grad_norm": 0.5704350471496582, "learning_rate": 6.246354122453305e-06, "loss": 0.5607, "step": 2225 }, { "epoch": 1.434278350515464, "grad_norm": 0.5815285444259644, "learning_rate": 6.242723194222546e-06, "loss": 0.5672, "step": 2226 }, { "epoch": 1.4349226804123711, "grad_norm": 0.641434907913208, "learning_rate": 6.2390915673636335e-06, "loss": 0.5727, "step": 2227 }, { "epoch": 1.4355670103092784, "grad_norm": 0.571875274181366, "learning_rate": 6.235459243918177e-06, "loss": 0.5893, "step": 2228 }, { "epoch": 1.4362113402061856, "grad_norm": 0.569794774055481, "learning_rate": 6.2318262259281795e-06, "loss": 0.56, "step": 2229 }, { "epoch": 1.4368556701030928, "grad_norm": 0.602528989315033, "learning_rate": 6.228192515436032e-06, "loss": 0.5693, "step": 2230 }, { "epoch": 1.4375, "grad_norm": 0.6553901433944702, "learning_rate": 6.224558114484521e-06, "loss": 0.5623, "step": 2231 }, { "epoch": 1.4381443298969072, "grad_norm": 0.5833603739738464, "learning_rate": 6.220923025116809e-06, "loss": 0.5771, "step": 2232 }, { "epoch": 1.4387886597938144, "grad_norm": 0.6706432700157166, "learning_rate": 6.21728724937646e-06, "loss": 0.5645, "step": 2233 }, { "epoch": 1.4394329896907216, "grad_norm": 0.6310835480690002, "learning_rate": 6.213650789307415e-06, "loss": 0.5951, "step": 2234 }, { "epoch": 1.4400773195876289, "grad_norm": 0.5840377807617188, "learning_rate": 6.210013646954001e-06, "loss": 0.5658, "step": 2235 }, { "epoch": 1.440721649484536, "grad_norm": 0.678183913230896, "learning_rate": 6.2063758243609275e-06, "loss": 0.564, "step": 2236 }, { "epoch": 1.4413659793814433, "grad_norm": 0.5970022082328796, "learning_rate": 6.202737323573291e-06, "loss": 0.5621, "step": 2237 }, { "epoch": 1.4420103092783505, "grad_norm": 0.6173033714294434, "learning_rate": 6.199098146636565e-06, "loss": 0.5843, "step": 2238 }, { "epoch": 1.4426546391752577, "grad_norm": 0.5985862612724304, "learning_rate": 6.195458295596606e-06, "loss": 0.5579, "step": 2239 }, { "epoch": 1.443298969072165, "grad_norm": 0.619403064250946, "learning_rate": 6.191817772499644e-06, "loss": 0.5801, "step": 2240 }, { "epoch": 1.4439432989690721, "grad_norm": 0.6007285118103027, "learning_rate": 6.188176579392294e-06, "loss": 0.5611, "step": 2241 }, { "epoch": 1.4445876288659794, "grad_norm": 0.6586670279502869, "learning_rate": 6.184534718321545e-06, "loss": 0.5787, "step": 2242 }, { "epoch": 1.4452319587628866, "grad_norm": 0.6849499344825745, "learning_rate": 6.180892191334761e-06, "loss": 0.5804, "step": 2243 }, { "epoch": 1.4458762886597938, "grad_norm": 0.6148064732551575, "learning_rate": 6.177249000479677e-06, "loss": 0.5883, "step": 2244 }, { "epoch": 1.446520618556701, "grad_norm": 0.630891740322113, "learning_rate": 6.173605147804409e-06, "loss": 0.5581, "step": 2245 }, { "epoch": 1.4471649484536082, "grad_norm": 0.6338483095169067, "learning_rate": 6.169960635357437e-06, "loss": 0.5581, "step": 2246 }, { "epoch": 1.4478092783505154, "grad_norm": 0.6377773880958557, "learning_rate": 6.166315465187616e-06, "loss": 0.5831, "step": 2247 }, { "epoch": 1.4484536082474226, "grad_norm": 0.6423456072807312, "learning_rate": 6.162669639344172e-06, "loss": 0.5716, "step": 2248 }, { "epoch": 1.4490979381443299, "grad_norm": 0.6101641654968262, "learning_rate": 6.159023159876698e-06, "loss": 0.5539, "step": 2249 }, { "epoch": 1.449742268041237, "grad_norm": 0.6728474497795105, "learning_rate": 6.155376028835152e-06, "loss": 0.5988, "step": 2250 }, { "epoch": 1.4503865979381443, "grad_norm": 0.6324322819709778, "learning_rate": 6.15172824826986e-06, "loss": 0.5595, "step": 2251 }, { "epoch": 1.4510309278350515, "grad_norm": 0.5861817002296448, "learning_rate": 6.148079820231518e-06, "loss": 0.5905, "step": 2252 }, { "epoch": 1.4516752577319587, "grad_norm": 0.6090129017829895, "learning_rate": 6.14443074677118e-06, "loss": 0.5576, "step": 2253 }, { "epoch": 1.452319587628866, "grad_norm": 0.6426385045051575, "learning_rate": 6.140781029940263e-06, "loss": 0.5824, "step": 2254 }, { "epoch": 1.4529639175257731, "grad_norm": 0.6545850038528442, "learning_rate": 6.137130671790548e-06, "loss": 0.5792, "step": 2255 }, { "epoch": 1.4536082474226804, "grad_norm": 0.6058359146118164, "learning_rate": 6.133479674374176e-06, "loss": 0.5859, "step": 2256 }, { "epoch": 1.4542525773195876, "grad_norm": 0.6880110502243042, "learning_rate": 6.129828039743647e-06, "loss": 0.5697, "step": 2257 }, { "epoch": 1.4548969072164948, "grad_norm": 0.6989340782165527, "learning_rate": 6.12617576995182e-06, "loss": 0.5604, "step": 2258 }, { "epoch": 1.455541237113402, "grad_norm": 0.5820856690406799, "learning_rate": 6.122522867051909e-06, "loss": 0.5727, "step": 2259 }, { "epoch": 1.4561855670103092, "grad_norm": 0.6580266356468201, "learning_rate": 6.118869333097488e-06, "loss": 0.57, "step": 2260 }, { "epoch": 1.4568298969072164, "grad_norm": 0.6585583686828613, "learning_rate": 6.115215170142481e-06, "loss": 0.5737, "step": 2261 }, { "epoch": 1.4574742268041236, "grad_norm": 0.5526707768440247, "learning_rate": 6.111560380241167e-06, "loss": 0.5707, "step": 2262 }, { "epoch": 1.4581185567010309, "grad_norm": 0.669322669506073, "learning_rate": 6.107904965448182e-06, "loss": 0.566, "step": 2263 }, { "epoch": 1.458762886597938, "grad_norm": 0.6471865177154541, "learning_rate": 6.104248927818506e-06, "loss": 0.568, "step": 2264 }, { "epoch": 1.4594072164948453, "grad_norm": 0.5455098152160645, "learning_rate": 6.100592269407474e-06, "loss": 0.578, "step": 2265 }, { "epoch": 1.4600515463917525, "grad_norm": 0.6297515034675598, "learning_rate": 6.0969349922707675e-06, "loss": 0.5876, "step": 2266 }, { "epoch": 1.4606958762886597, "grad_norm": 0.5765472054481506, "learning_rate": 6.09327709846442e-06, "loss": 0.5831, "step": 2267 }, { "epoch": 1.4613402061855671, "grad_norm": 0.540901243686676, "learning_rate": 6.089618590044805e-06, "loss": 0.5608, "step": 2268 }, { "epoch": 1.4619845360824741, "grad_norm": 0.6672888398170471, "learning_rate": 6.085959469068646e-06, "loss": 0.5639, "step": 2269 }, { "epoch": 1.4626288659793816, "grad_norm": 0.6555026769638062, "learning_rate": 6.082299737593012e-06, "loss": 0.5699, "step": 2270 }, { "epoch": 1.4632731958762886, "grad_norm": 0.6175703406333923, "learning_rate": 6.078639397675313e-06, "loss": 0.5543, "step": 2271 }, { "epoch": 1.463917525773196, "grad_norm": 0.7411698698997498, "learning_rate": 6.074978451373299e-06, "loss": 0.604, "step": 2272 }, { "epoch": 1.464561855670103, "grad_norm": 0.6104111075401306, "learning_rate": 6.071316900745066e-06, "loss": 0.5694, "step": 2273 }, { "epoch": 1.4652061855670104, "grad_norm": 0.6564821004867554, "learning_rate": 6.067654747849045e-06, "loss": 0.5447, "step": 2274 }, { "epoch": 1.4658505154639174, "grad_norm": 0.5827586054801941, "learning_rate": 6.063991994744009e-06, "loss": 0.5572, "step": 2275 }, { "epoch": 1.4664948453608249, "grad_norm": 0.5664418935775757, "learning_rate": 6.060328643489064e-06, "loss": 0.5617, "step": 2276 }, { "epoch": 1.4671391752577319, "grad_norm": 0.6508159637451172, "learning_rate": 6.0566646961436574e-06, "loss": 0.5524, "step": 2277 }, { "epoch": 1.4677835051546393, "grad_norm": 0.5928805470466614, "learning_rate": 6.05300015476757e-06, "loss": 0.5703, "step": 2278 }, { "epoch": 1.4684278350515463, "grad_norm": 0.6203755736351013, "learning_rate": 6.049335021420916e-06, "loss": 0.5944, "step": 2279 }, { "epoch": 1.4690721649484537, "grad_norm": 0.6793735027313232, "learning_rate": 6.04566929816414e-06, "loss": 0.5646, "step": 2280 }, { "epoch": 1.4697164948453607, "grad_norm": 0.5717807412147522, "learning_rate": 6.042002987058024e-06, "loss": 0.5612, "step": 2281 }, { "epoch": 1.4703608247422681, "grad_norm": 0.6920600533485413, "learning_rate": 6.038336090163677e-06, "loss": 0.5573, "step": 2282 }, { "epoch": 1.4710051546391751, "grad_norm": 0.6167482137680054, "learning_rate": 6.034668609542533e-06, "loss": 0.5686, "step": 2283 }, { "epoch": 1.4716494845360826, "grad_norm": 0.5690513849258423, "learning_rate": 6.031000547256364e-06, "loss": 0.5782, "step": 2284 }, { "epoch": 1.4722938144329896, "grad_norm": 0.6361425518989563, "learning_rate": 6.0273319053672615e-06, "loss": 0.5741, "step": 2285 }, { "epoch": 1.472938144329897, "grad_norm": 0.5793992877006531, "learning_rate": 6.023662685937643e-06, "loss": 0.5758, "step": 2286 }, { "epoch": 1.473582474226804, "grad_norm": 0.6722201704978943, "learning_rate": 6.019992891030255e-06, "loss": 0.5567, "step": 2287 }, { "epoch": 1.4742268041237114, "grad_norm": 0.579777181148529, "learning_rate": 6.016322522708167e-06, "loss": 0.5778, "step": 2288 }, { "epoch": 1.4748711340206184, "grad_norm": 0.6219563484191895, "learning_rate": 6.012651583034765e-06, "loss": 0.5737, "step": 2289 }, { "epoch": 1.4755154639175259, "grad_norm": 0.6367224454879761, "learning_rate": 6.008980074073761e-06, "loss": 0.5568, "step": 2290 }, { "epoch": 1.4761597938144329, "grad_norm": 0.6670817136764526, "learning_rate": 6.005307997889189e-06, "loss": 0.5806, "step": 2291 }, { "epoch": 1.4768041237113403, "grad_norm": 0.8409967422485352, "learning_rate": 6.001635356545398e-06, "loss": 0.5881, "step": 2292 }, { "epoch": 1.4774484536082475, "grad_norm": 0.621327817440033, "learning_rate": 5.997962152107054e-06, "loss": 0.564, "step": 2293 }, { "epoch": 1.4780927835051547, "grad_norm": 0.7037537693977356, "learning_rate": 5.994288386639142e-06, "loss": 0.57, "step": 2294 }, { "epoch": 1.478737113402062, "grad_norm": 0.7231361865997314, "learning_rate": 5.9906140622069635e-06, "loss": 0.5587, "step": 2295 }, { "epoch": 1.4793814432989691, "grad_norm": 0.5630394220352173, "learning_rate": 5.9869391808761315e-06, "loss": 0.5652, "step": 2296 }, { "epoch": 1.4800257731958764, "grad_norm": 0.6634013056755066, "learning_rate": 5.983263744712573e-06, "loss": 0.5461, "step": 2297 }, { "epoch": 1.4806701030927836, "grad_norm": 0.6819922924041748, "learning_rate": 5.979587755782526e-06, "loss": 0.5713, "step": 2298 }, { "epoch": 1.4813144329896908, "grad_norm": 0.5720377564430237, "learning_rate": 5.975911216152544e-06, "loss": 0.5784, "step": 2299 }, { "epoch": 1.481958762886598, "grad_norm": 0.6043102145195007, "learning_rate": 5.972234127889483e-06, "loss": 0.5796, "step": 2300 }, { "epoch": 1.4826030927835052, "grad_norm": 0.5881932973861694, "learning_rate": 5.96855649306051e-06, "loss": 0.5484, "step": 2301 }, { "epoch": 1.4832474226804124, "grad_norm": 0.6583842635154724, "learning_rate": 5.9648783137331035e-06, "loss": 0.5832, "step": 2302 }, { "epoch": 1.4838917525773196, "grad_norm": 0.5925966501235962, "learning_rate": 5.961199591975043e-06, "loss": 0.5736, "step": 2303 }, { "epoch": 1.4845360824742269, "grad_norm": 0.5386655926704407, "learning_rate": 5.957520329854413e-06, "loss": 0.5763, "step": 2304 }, { "epoch": 1.485180412371134, "grad_norm": 0.6180065274238586, "learning_rate": 5.9538405294396055e-06, "loss": 0.5632, "step": 2305 }, { "epoch": 1.4858247422680413, "grad_norm": 0.6635538935661316, "learning_rate": 5.9501601927993135e-06, "loss": 0.563, "step": 2306 }, { "epoch": 1.4864690721649485, "grad_norm": 0.5391649603843689, "learning_rate": 5.94647932200253e-06, "loss": 0.5623, "step": 2307 }, { "epoch": 1.4871134020618557, "grad_norm": 0.5836979746818542, "learning_rate": 5.942797919118547e-06, "loss": 0.5636, "step": 2308 }, { "epoch": 1.487757731958763, "grad_norm": 0.6347827911376953, "learning_rate": 5.939115986216962e-06, "loss": 0.5513, "step": 2309 }, { "epoch": 1.4884020618556701, "grad_norm": 0.5896915197372437, "learning_rate": 5.935433525367665e-06, "loss": 0.5485, "step": 2310 }, { "epoch": 1.4890463917525774, "grad_norm": 0.6254266500473022, "learning_rate": 5.9317505386408436e-06, "loss": 0.5675, "step": 2311 }, { "epoch": 1.4896907216494846, "grad_norm": 0.5700423717498779, "learning_rate": 5.92806702810698e-06, "loss": 0.5579, "step": 2312 }, { "epoch": 1.4903350515463918, "grad_norm": 0.6338189244270325, "learning_rate": 5.924382995836854e-06, "loss": 0.5873, "step": 2313 }, { "epoch": 1.490979381443299, "grad_norm": 0.637526273727417, "learning_rate": 5.920698443901538e-06, "loss": 0.5787, "step": 2314 }, { "epoch": 1.4916237113402062, "grad_norm": 0.6439713835716248, "learning_rate": 5.917013374372395e-06, "loss": 0.5677, "step": 2315 }, { "epoch": 1.4922680412371134, "grad_norm": 0.6487382650375366, "learning_rate": 5.9133277893210785e-06, "loss": 0.5516, "step": 2316 }, { "epoch": 1.4929123711340206, "grad_norm": 0.596684992313385, "learning_rate": 5.909641690819535e-06, "loss": 0.5732, "step": 2317 }, { "epoch": 1.4935567010309279, "grad_norm": 0.6364780068397522, "learning_rate": 5.905955080939998e-06, "loss": 0.5549, "step": 2318 }, { "epoch": 1.494201030927835, "grad_norm": 0.6057417988777161, "learning_rate": 5.902267961754984e-06, "loss": 0.555, "step": 2319 }, { "epoch": 1.4948453608247423, "grad_norm": 0.6117101907730103, "learning_rate": 5.898580335337305e-06, "loss": 0.5595, "step": 2320 }, { "epoch": 1.4954896907216495, "grad_norm": 0.6413109302520752, "learning_rate": 5.89489220376005e-06, "loss": 0.5699, "step": 2321 }, { "epoch": 1.4961340206185567, "grad_norm": 0.6915721297264099, "learning_rate": 5.891203569096595e-06, "loss": 0.5861, "step": 2322 }, { "epoch": 1.496778350515464, "grad_norm": 0.637198805809021, "learning_rate": 5.8875144334206e-06, "loss": 0.5802, "step": 2323 }, { "epoch": 1.4974226804123711, "grad_norm": 0.68109130859375, "learning_rate": 5.883824798806007e-06, "loss": 0.5651, "step": 2324 }, { "epoch": 1.4980670103092784, "grad_norm": 0.6878547668457031, "learning_rate": 5.8801346673270355e-06, "loss": 0.5734, "step": 2325 }, { "epoch": 1.4987113402061856, "grad_norm": 0.5909542441368103, "learning_rate": 5.8764440410581846e-06, "loss": 0.5771, "step": 2326 }, { "epoch": 1.4993556701030928, "grad_norm": 0.5877557396888733, "learning_rate": 5.872752922074235e-06, "loss": 0.5694, "step": 2327 }, { "epoch": 1.5, "grad_norm": 0.6129542589187622, "learning_rate": 5.869061312450242e-06, "loss": 0.568, "step": 2328 }, { "epoch": 1.5006443298969072, "grad_norm": 0.6371265053749084, "learning_rate": 5.865369214261535e-06, "loss": 0.5784, "step": 2329 }, { "epoch": 1.5012886597938144, "grad_norm": 0.6082968711853027, "learning_rate": 5.861676629583722e-06, "loss": 0.5601, "step": 2330 }, { "epoch": 1.5019329896907216, "grad_norm": 0.635674774646759, "learning_rate": 5.85798356049268e-06, "loss": 0.5616, "step": 2331 }, { "epoch": 1.5025773195876289, "grad_norm": 0.6448675990104675, "learning_rate": 5.854290009064565e-06, "loss": 0.5722, "step": 2332 }, { "epoch": 1.503221649484536, "grad_norm": 0.6337045431137085, "learning_rate": 5.850595977375797e-06, "loss": 0.5714, "step": 2333 }, { "epoch": 1.5038659793814433, "grad_norm": 0.6073354482650757, "learning_rate": 5.84690146750307e-06, "loss": 0.5662, "step": 2334 }, { "epoch": 1.5045103092783505, "grad_norm": 0.6594905257225037, "learning_rate": 5.843206481523345e-06, "loss": 0.5566, "step": 2335 }, { "epoch": 1.5051546391752577, "grad_norm": 0.6073745489120483, "learning_rate": 5.839511021513853e-06, "loss": 0.5692, "step": 2336 }, { "epoch": 1.505798969072165, "grad_norm": 0.6517214775085449, "learning_rate": 5.835815089552086e-06, "loss": 0.5901, "step": 2337 }, { "epoch": 1.5064432989690721, "grad_norm": 0.5605165958404541, "learning_rate": 5.832118687715811e-06, "loss": 0.5672, "step": 2338 }, { "epoch": 1.5070876288659794, "grad_norm": 0.5625775456428528, "learning_rate": 5.82842181808305e-06, "loss": 0.5815, "step": 2339 }, { "epoch": 1.5077319587628866, "grad_norm": 0.5972201228141785, "learning_rate": 5.824724482732091e-06, "loss": 0.5872, "step": 2340 }, { "epoch": 1.5083762886597938, "grad_norm": 0.5701768398284912, "learning_rate": 5.821026683741485e-06, "loss": 0.5705, "step": 2341 }, { "epoch": 1.509020618556701, "grad_norm": 0.583584189414978, "learning_rate": 5.817328423190044e-06, "loss": 0.572, "step": 2342 }, { "epoch": 1.5096649484536082, "grad_norm": 0.5868527293205261, "learning_rate": 5.813629703156838e-06, "loss": 0.5776, "step": 2343 }, { "epoch": 1.5103092783505154, "grad_norm": 0.5532597303390503, "learning_rate": 5.809930525721193e-06, "loss": 0.5817, "step": 2344 }, { "epoch": 1.5109536082474226, "grad_norm": 0.6332736611366272, "learning_rate": 5.806230892962699e-06, "loss": 0.5761, "step": 2345 }, { "epoch": 1.5115979381443299, "grad_norm": 0.5602304339408875, "learning_rate": 5.802530806961195e-06, "loss": 0.5825, "step": 2346 }, { "epoch": 1.512242268041237, "grad_norm": 0.662944495677948, "learning_rate": 5.798830269796778e-06, "loss": 0.5912, "step": 2347 }, { "epoch": 1.5128865979381443, "grad_norm": 0.6494640111923218, "learning_rate": 5.795129283549798e-06, "loss": 0.5522, "step": 2348 }, { "epoch": 1.5135309278350515, "grad_norm": 0.5753223299980164, "learning_rate": 5.7914278503008595e-06, "loss": 0.5586, "step": 2349 }, { "epoch": 1.5141752577319587, "grad_norm": 0.5959009528160095, "learning_rate": 5.7877259721308135e-06, "loss": 0.5606, "step": 2350 }, { "epoch": 1.514819587628866, "grad_norm": 0.5576042532920837, "learning_rate": 5.784023651120766e-06, "loss": 0.5481, "step": 2351 }, { "epoch": 1.5154639175257731, "grad_norm": 0.6131908893585205, "learning_rate": 5.78032088935207e-06, "loss": 0.5733, "step": 2352 }, { "epoch": 1.5161082474226806, "grad_norm": 0.5523278713226318, "learning_rate": 5.776617688906327e-06, "loss": 0.5797, "step": 2353 }, { "epoch": 1.5167525773195876, "grad_norm": 0.5951313972473145, "learning_rate": 5.772914051865382e-06, "loss": 0.5678, "step": 2354 }, { "epoch": 1.517396907216495, "grad_norm": 0.6576697826385498, "learning_rate": 5.769209980311329e-06, "loss": 0.5841, "step": 2355 }, { "epoch": 1.518041237113402, "grad_norm": 0.6101453900337219, "learning_rate": 5.765505476326505e-06, "loss": 0.5791, "step": 2356 }, { "epoch": 1.5186855670103094, "grad_norm": 0.6726386547088623, "learning_rate": 5.761800541993493e-06, "loss": 0.5532, "step": 2357 }, { "epoch": 1.5193298969072164, "grad_norm": 0.7179418802261353, "learning_rate": 5.75809517939511e-06, "loss": 0.5786, "step": 2358 }, { "epoch": 1.5199742268041239, "grad_norm": 0.5796752572059631, "learning_rate": 5.7543893906144235e-06, "loss": 0.5847, "step": 2359 }, { "epoch": 1.5206185567010309, "grad_norm": 0.6951767206192017, "learning_rate": 5.7506831777347334e-06, "loss": 0.5569, "step": 2360 }, { "epoch": 1.5212628865979383, "grad_norm": 0.6242327690124512, "learning_rate": 5.746976542839583e-06, "loss": 0.5623, "step": 2361 }, { "epoch": 1.5219072164948453, "grad_norm": 0.5439872145652771, "learning_rate": 5.743269488012747e-06, "loss": 0.563, "step": 2362 }, { "epoch": 1.5225515463917527, "grad_norm": 0.6080238819122314, "learning_rate": 5.739562015338244e-06, "loss": 0.5855, "step": 2363 }, { "epoch": 1.5231958762886597, "grad_norm": 0.5154092311859131, "learning_rate": 5.73585412690032e-06, "loss": 0.5749, "step": 2364 }, { "epoch": 1.5238402061855671, "grad_norm": 0.5903816819190979, "learning_rate": 5.732145824783458e-06, "loss": 0.5748, "step": 2365 }, { "epoch": 1.5244845360824741, "grad_norm": 0.6343150734901428, "learning_rate": 5.728437111072376e-06, "loss": 0.5745, "step": 2366 }, { "epoch": 1.5251288659793816, "grad_norm": 0.5970695614814758, "learning_rate": 5.7247279878520165e-06, "loss": 0.562, "step": 2367 }, { "epoch": 1.5257731958762886, "grad_norm": 0.6921707391738892, "learning_rate": 5.7210184572075615e-06, "loss": 0.5762, "step": 2368 }, { "epoch": 1.526417525773196, "grad_norm": 0.6604434847831726, "learning_rate": 5.717308521224414e-06, "loss": 0.5692, "step": 2369 }, { "epoch": 1.527061855670103, "grad_norm": 0.6803650259971619, "learning_rate": 5.7135981819882095e-06, "loss": 0.5771, "step": 2370 }, { "epoch": 1.5277061855670104, "grad_norm": 0.6425766348838806, "learning_rate": 5.7098874415848075e-06, "loss": 0.5727, "step": 2371 }, { "epoch": 1.5283505154639174, "grad_norm": 0.5974029302597046, "learning_rate": 5.706176302100295e-06, "loss": 0.5636, "step": 2372 }, { "epoch": 1.5289948453608249, "grad_norm": 0.619191586971283, "learning_rate": 5.702464765620981e-06, "loss": 0.5478, "step": 2373 }, { "epoch": 1.5296391752577319, "grad_norm": 0.6919388771057129, "learning_rate": 5.698752834233401e-06, "loss": 0.5789, "step": 2374 }, { "epoch": 1.5302835051546393, "grad_norm": 0.6133217811584473, "learning_rate": 5.695040510024311e-06, "loss": 0.577, "step": 2375 }, { "epoch": 1.5309278350515463, "grad_norm": 0.5972933173179626, "learning_rate": 5.691327795080685e-06, "loss": 0.5729, "step": 2376 }, { "epoch": 1.5315721649484537, "grad_norm": 0.6228527426719666, "learning_rate": 5.687614691489723e-06, "loss": 0.5787, "step": 2377 }, { "epoch": 1.5322164948453607, "grad_norm": 0.6757509112358093, "learning_rate": 5.683901201338837e-06, "loss": 0.5828, "step": 2378 }, { "epoch": 1.5328608247422681, "grad_norm": 0.6096044778823853, "learning_rate": 5.680187326715662e-06, "loss": 0.5588, "step": 2379 }, { "epoch": 1.5335051546391751, "grad_norm": 0.5970376133918762, "learning_rate": 5.67647306970804e-06, "loss": 0.5739, "step": 2380 }, { "epoch": 1.5341494845360826, "grad_norm": 0.7548148036003113, "learning_rate": 5.672758432404042e-06, "loss": 0.5624, "step": 2381 }, { "epoch": 1.5347938144329896, "grad_norm": 0.5876861214637756, "learning_rate": 5.6690434168919395e-06, "loss": 0.5648, "step": 2382 }, { "epoch": 1.535438144329897, "grad_norm": 0.6344348788261414, "learning_rate": 5.665328025260223e-06, "loss": 0.5734, "step": 2383 }, { "epoch": 1.536082474226804, "grad_norm": 0.7062456011772156, "learning_rate": 5.661612259597595e-06, "loss": 0.5738, "step": 2384 }, { "epoch": 1.5367268041237114, "grad_norm": 0.5853224992752075, "learning_rate": 5.657896121992966e-06, "loss": 0.5835, "step": 2385 }, { "epoch": 1.5373711340206184, "grad_norm": 0.7017791867256165, "learning_rate": 5.654179614535457e-06, "loss": 0.581, "step": 2386 }, { "epoch": 1.5380154639175259, "grad_norm": 0.6157275438308716, "learning_rate": 5.650462739314394e-06, "loss": 0.5889, "step": 2387 }, { "epoch": 1.5386597938144329, "grad_norm": 0.7312247157096863, "learning_rate": 5.646745498419316e-06, "loss": 0.5737, "step": 2388 }, { "epoch": 1.5393041237113403, "grad_norm": 0.7091370224952698, "learning_rate": 5.64302789393996e-06, "loss": 0.5749, "step": 2389 }, { "epoch": 1.5399484536082473, "grad_norm": 0.5973899960517883, "learning_rate": 5.6393099279662745e-06, "loss": 0.5653, "step": 2390 }, { "epoch": 1.5405927835051547, "grad_norm": 0.7482192516326904, "learning_rate": 5.635591602588403e-06, "loss": 0.5637, "step": 2391 }, { "epoch": 1.5412371134020617, "grad_norm": 0.7342258095741272, "learning_rate": 5.6318729198966995e-06, "loss": 0.5814, "step": 2392 }, { "epoch": 1.5418814432989691, "grad_norm": 0.6146312355995178, "learning_rate": 5.628153881981714e-06, "loss": 0.5675, "step": 2393 }, { "epoch": 1.5425257731958761, "grad_norm": 0.7134588956832886, "learning_rate": 5.624434490934197e-06, "loss": 0.5723, "step": 2394 }, { "epoch": 1.5431701030927836, "grad_norm": 0.6117458939552307, "learning_rate": 5.620714748845098e-06, "loss": 0.5443, "step": 2395 }, { "epoch": 1.5438144329896906, "grad_norm": 0.6293690204620361, "learning_rate": 5.616994657805566e-06, "loss": 0.5716, "step": 2396 }, { "epoch": 1.544458762886598, "grad_norm": 0.6737545132637024, "learning_rate": 5.613274219906941e-06, "loss": 0.5671, "step": 2397 }, { "epoch": 1.545103092783505, "grad_norm": 0.5964216589927673, "learning_rate": 5.6095534372407615e-06, "loss": 0.5604, "step": 2398 }, { "epoch": 1.5457474226804124, "grad_norm": 0.6087165474891663, "learning_rate": 5.60583231189876e-06, "loss": 0.5775, "step": 2399 }, { "epoch": 1.5463917525773194, "grad_norm": 0.5938271880149841, "learning_rate": 5.602110845972861e-06, "loss": 0.5554, "step": 2400 }, { "epoch": 1.5470360824742269, "grad_norm": 0.5817025303840637, "learning_rate": 5.598389041555179e-06, "loss": 0.5669, "step": 2401 }, { "epoch": 1.547680412371134, "grad_norm": 0.612062931060791, "learning_rate": 5.594666900738021e-06, "loss": 0.5837, "step": 2402 }, { "epoch": 1.5483247422680413, "grad_norm": 0.5917054414749146, "learning_rate": 5.590944425613883e-06, "loss": 0.551, "step": 2403 }, { "epoch": 1.5489690721649485, "grad_norm": 0.5419139266014099, "learning_rate": 5.5872216182754466e-06, "loss": 0.5821, "step": 2404 }, { "epoch": 1.5496134020618557, "grad_norm": 0.6062477231025696, "learning_rate": 5.583498480815582e-06, "loss": 0.5667, "step": 2405 }, { "epoch": 1.550257731958763, "grad_norm": 0.6021741628646851, "learning_rate": 5.579775015327347e-06, "loss": 0.5817, "step": 2406 }, { "epoch": 1.5509020618556701, "grad_norm": 0.5728664994239807, "learning_rate": 5.576051223903979e-06, "loss": 0.5807, "step": 2407 }, { "epoch": 1.5515463917525774, "grad_norm": 0.6068059802055359, "learning_rate": 5.572327108638902e-06, "loss": 0.5741, "step": 2408 }, { "epoch": 1.5521907216494846, "grad_norm": 0.6533156633377075, "learning_rate": 5.568602671625721e-06, "loss": 0.586, "step": 2409 }, { "epoch": 1.5528350515463918, "grad_norm": 0.5846585631370544, "learning_rate": 5.564877914958222e-06, "loss": 0.5804, "step": 2410 }, { "epoch": 1.553479381443299, "grad_norm": 0.674182116985321, "learning_rate": 5.561152840730371e-06, "loss": 0.5624, "step": 2411 }, { "epoch": 1.5541237113402062, "grad_norm": 0.6686476469039917, "learning_rate": 5.557427451036312e-06, "loss": 0.5685, "step": 2412 }, { "epoch": 1.5547680412371134, "grad_norm": 0.5918810963630676, "learning_rate": 5.553701747970368e-06, "loss": 0.5744, "step": 2413 }, { "epoch": 1.5554123711340206, "grad_norm": 0.6345512866973877, "learning_rate": 5.549975733627036e-06, "loss": 0.5738, "step": 2414 }, { "epoch": 1.5560567010309279, "grad_norm": 0.6151610612869263, "learning_rate": 5.546249410100989e-06, "loss": 0.5562, "step": 2415 }, { "epoch": 1.556701030927835, "grad_norm": 0.5562717318534851, "learning_rate": 5.5425227794870715e-06, "loss": 0.5705, "step": 2416 }, { "epoch": 1.5573453608247423, "grad_norm": 0.5614816546440125, "learning_rate": 5.538795843880307e-06, "loss": 0.5851, "step": 2417 }, { "epoch": 1.5579896907216495, "grad_norm": 0.6077166199684143, "learning_rate": 5.535068605375884e-06, "loss": 0.5987, "step": 2418 }, { "epoch": 1.5586340206185567, "grad_norm": 0.5747889280319214, "learning_rate": 5.5313410660691615e-06, "loss": 0.5766, "step": 2419 }, { "epoch": 1.559278350515464, "grad_norm": 0.5716685056686401, "learning_rate": 5.5276132280556725e-06, "loss": 0.5638, "step": 2420 }, { "epoch": 1.5599226804123711, "grad_norm": 0.6083840131759644, "learning_rate": 5.523885093431116e-06, "loss": 0.579, "step": 2421 }, { "epoch": 1.5605670103092784, "grad_norm": 0.626989483833313, "learning_rate": 5.520156664291352e-06, "loss": 0.5742, "step": 2422 }, { "epoch": 1.5612113402061856, "grad_norm": 0.6322778463363647, "learning_rate": 5.516427942732415e-06, "loss": 0.5862, "step": 2423 }, { "epoch": 1.5618556701030928, "grad_norm": 0.573638916015625, "learning_rate": 5.5126989308505e-06, "loss": 0.5903, "step": 2424 }, { "epoch": 1.5625, "grad_norm": 0.688402533531189, "learning_rate": 5.508969630741963e-06, "loss": 0.574, "step": 2425 }, { "epoch": 1.5631443298969072, "grad_norm": 0.6005432605743408, "learning_rate": 5.505240044503324e-06, "loss": 0.5747, "step": 2426 }, { "epoch": 1.5637886597938144, "grad_norm": 0.599851667881012, "learning_rate": 5.501510174231264e-06, "loss": 0.571, "step": 2427 }, { "epoch": 1.5644329896907216, "grad_norm": 0.5714493989944458, "learning_rate": 5.497780022022625e-06, "loss": 0.5488, "step": 2428 }, { "epoch": 1.5650773195876289, "grad_norm": 0.6712982058525085, "learning_rate": 5.494049589974406e-06, "loss": 0.5925, "step": 2429 }, { "epoch": 1.565721649484536, "grad_norm": 0.5594907999038696, "learning_rate": 5.49031888018376e-06, "loss": 0.5836, "step": 2430 }, { "epoch": 1.5663659793814433, "grad_norm": 0.6012742519378662, "learning_rate": 5.486587894748003e-06, "loss": 0.5726, "step": 2431 }, { "epoch": 1.5670103092783505, "grad_norm": 0.6134519577026367, "learning_rate": 5.482856635764603e-06, "loss": 0.5613, "step": 2432 }, { "epoch": 1.5676546391752577, "grad_norm": 0.589475691318512, "learning_rate": 5.4791251053311785e-06, "loss": 0.5708, "step": 2433 }, { "epoch": 1.568298969072165, "grad_norm": 0.5831257104873657, "learning_rate": 5.475393305545504e-06, "loss": 0.5668, "step": 2434 }, { "epoch": 1.5689432989690721, "grad_norm": 0.6110215783119202, "learning_rate": 5.4716612385055066e-06, "loss": 0.5725, "step": 2435 }, { "epoch": 1.5695876288659794, "grad_norm": 0.5446265935897827, "learning_rate": 5.46792890630926e-06, "loss": 0.5696, "step": 2436 }, { "epoch": 1.5702319587628866, "grad_norm": 0.580930769443512, "learning_rate": 5.4641963110549885e-06, "loss": 0.5713, "step": 2437 }, { "epoch": 1.5708762886597938, "grad_norm": 0.5448374152183533, "learning_rate": 5.460463454841066e-06, "loss": 0.5731, "step": 2438 }, { "epoch": 1.571520618556701, "grad_norm": 0.5850638151168823, "learning_rate": 5.456730339766012e-06, "loss": 0.5682, "step": 2439 }, { "epoch": 1.5721649484536082, "grad_norm": 0.5638377666473389, "learning_rate": 5.4529969679284885e-06, "loss": 0.5682, "step": 2440 }, { "epoch": 1.5728092783505154, "grad_norm": 0.5564789772033691, "learning_rate": 5.4492633414273065e-06, "loss": 0.5728, "step": 2441 }, { "epoch": 1.5734536082474226, "grad_norm": 0.572659969329834, "learning_rate": 5.44552946236142e-06, "loss": 0.5778, "step": 2442 }, { "epoch": 1.5740979381443299, "grad_norm": 0.5400278568267822, "learning_rate": 5.44179533282992e-06, "loss": 0.5624, "step": 2443 }, { "epoch": 1.574742268041237, "grad_norm": 0.5710208415985107, "learning_rate": 5.438060954932043e-06, "loss": 0.5679, "step": 2444 }, { "epoch": 1.5753865979381443, "grad_norm": 0.5702251195907593, "learning_rate": 5.434326330767161e-06, "loss": 0.5627, "step": 2445 }, { "epoch": 1.5760309278350515, "grad_norm": 0.5531743764877319, "learning_rate": 5.430591462434792e-06, "loss": 0.5553, "step": 2446 }, { "epoch": 1.5766752577319587, "grad_norm": 0.5226263403892517, "learning_rate": 5.426856352034583e-06, "loss": 0.5806, "step": 2447 }, { "epoch": 1.577319587628866, "grad_norm": 0.5504281520843506, "learning_rate": 5.42312100166632e-06, "loss": 0.5745, "step": 2448 }, { "epoch": 1.5779639175257731, "grad_norm": 0.5548949241638184, "learning_rate": 5.4193854134299265e-06, "loss": 0.5549, "step": 2449 }, { "epoch": 1.5786082474226806, "grad_norm": 0.560558021068573, "learning_rate": 5.415649589425456e-06, "loss": 0.5798, "step": 2450 }, { "epoch": 1.5792525773195876, "grad_norm": 0.5808137059211731, "learning_rate": 5.411913531753098e-06, "loss": 0.5624, "step": 2451 }, { "epoch": 1.579896907216495, "grad_norm": 0.5527358055114746, "learning_rate": 5.408177242513168e-06, "loss": 0.556, "step": 2452 }, { "epoch": 1.580541237113402, "grad_norm": 0.5723777413368225, "learning_rate": 5.404440723806118e-06, "loss": 0.5745, "step": 2453 }, { "epoch": 1.5811855670103094, "grad_norm": 0.597320020198822, "learning_rate": 5.400703977732526e-06, "loss": 0.5858, "step": 2454 }, { "epoch": 1.5818298969072164, "grad_norm": 0.5348289608955383, "learning_rate": 5.396967006393095e-06, "loss": 0.5713, "step": 2455 }, { "epoch": 1.5824742268041239, "grad_norm": 0.6205667853355408, "learning_rate": 5.393229811888663e-06, "loss": 0.5617, "step": 2456 }, { "epoch": 1.5831185567010309, "grad_norm": 0.5948175191879272, "learning_rate": 5.3894923963201824e-06, "loss": 0.5799, "step": 2457 }, { "epoch": 1.5837628865979383, "grad_norm": 0.6362664103507996, "learning_rate": 5.385754761788739e-06, "loss": 0.5764, "step": 2458 }, { "epoch": 1.5844072164948453, "grad_norm": 0.5841876268386841, "learning_rate": 5.3820169103955365e-06, "loss": 0.5922, "step": 2459 }, { "epoch": 1.5850515463917527, "grad_norm": 0.578324556350708, "learning_rate": 5.378278844241904e-06, "loss": 0.5914, "step": 2460 }, { "epoch": 1.5856958762886597, "grad_norm": 0.5969411730766296, "learning_rate": 5.374540565429288e-06, "loss": 0.5748, "step": 2461 }, { "epoch": 1.5863402061855671, "grad_norm": 0.6339635848999023, "learning_rate": 5.370802076059257e-06, "loss": 0.581, "step": 2462 }, { "epoch": 1.5869845360824741, "grad_norm": 0.5583421587944031, "learning_rate": 5.367063378233495e-06, "loss": 0.5883, "step": 2463 }, { "epoch": 1.5876288659793816, "grad_norm": 0.5328691005706787, "learning_rate": 5.363324474053809e-06, "loss": 0.5883, "step": 2464 }, { "epoch": 1.5882731958762886, "grad_norm": 0.6327545046806335, "learning_rate": 5.3595853656221165e-06, "loss": 0.5505, "step": 2465 }, { "epoch": 1.588917525773196, "grad_norm": 0.572603702545166, "learning_rate": 5.355846055040449e-06, "loss": 0.5456, "step": 2466 }, { "epoch": 1.589561855670103, "grad_norm": 0.5737975239753723, "learning_rate": 5.352106544410956e-06, "loss": 0.5858, "step": 2467 }, { "epoch": 1.5902061855670104, "grad_norm": 0.5971495509147644, "learning_rate": 5.348366835835898e-06, "loss": 0.5768, "step": 2468 }, { "epoch": 1.5908505154639174, "grad_norm": 0.5504598617553711, "learning_rate": 5.344626931417648e-06, "loss": 0.5562, "step": 2469 }, { "epoch": 1.5914948453608249, "grad_norm": 0.5645929574966431, "learning_rate": 5.340886833258684e-06, "loss": 0.5674, "step": 2470 }, { "epoch": 1.5921391752577319, "grad_norm": 0.5594881772994995, "learning_rate": 5.337146543461599e-06, "loss": 0.5715, "step": 2471 }, { "epoch": 1.5927835051546393, "grad_norm": 0.5603945255279541, "learning_rate": 5.33340606412909e-06, "loss": 0.5474, "step": 2472 }, { "epoch": 1.5934278350515463, "grad_norm": 0.662563681602478, "learning_rate": 5.329665397363961e-06, "loss": 0.5782, "step": 2473 }, { "epoch": 1.5940721649484537, "grad_norm": 0.6720561981201172, "learning_rate": 5.325924545269123e-06, "loss": 0.5796, "step": 2474 }, { "epoch": 1.5947164948453607, "grad_norm": 0.707204282283783, "learning_rate": 5.322183509947592e-06, "loss": 0.5692, "step": 2475 }, { "epoch": 1.5953608247422681, "grad_norm": 0.5770460963249207, "learning_rate": 5.318442293502482e-06, "loss": 0.5512, "step": 2476 }, { "epoch": 1.5960051546391751, "grad_norm": 0.5911105871200562, "learning_rate": 5.314700898037013e-06, "loss": 0.5795, "step": 2477 }, { "epoch": 1.5966494845360826, "grad_norm": 0.6167894601821899, "learning_rate": 5.310959325654507e-06, "loss": 0.5815, "step": 2478 }, { "epoch": 1.5972938144329896, "grad_norm": 0.5733704566955566, "learning_rate": 5.307217578458381e-06, "loss": 0.583, "step": 2479 }, { "epoch": 1.597938144329897, "grad_norm": 0.5597540736198425, "learning_rate": 5.303475658552153e-06, "loss": 0.5818, "step": 2480 }, { "epoch": 1.598582474226804, "grad_norm": 0.606522262096405, "learning_rate": 5.299733568039437e-06, "loss": 0.5767, "step": 2481 }, { "epoch": 1.5992268041237114, "grad_norm": 0.7065243721008301, "learning_rate": 5.295991309023944e-06, "loss": 0.5672, "step": 2482 }, { "epoch": 1.5998711340206184, "grad_norm": 0.5558124780654907, "learning_rate": 5.292248883609481e-06, "loss": 0.5842, "step": 2483 }, { "epoch": 1.6005154639175259, "grad_norm": 0.6716904640197754, "learning_rate": 5.288506293899942e-06, "loss": 0.5629, "step": 2484 }, { "epoch": 1.6011597938144329, "grad_norm": 0.6840060949325562, "learning_rate": 5.284763541999321e-06, "loss": 0.5746, "step": 2485 }, { "epoch": 1.6018041237113403, "grad_norm": 0.6474680304527283, "learning_rate": 5.281020630011703e-06, "loss": 0.5615, "step": 2486 }, { "epoch": 1.6024484536082473, "grad_norm": 0.6998871564865112, "learning_rate": 5.277277560041257e-06, "loss": 0.5824, "step": 2487 }, { "epoch": 1.6030927835051547, "grad_norm": 0.6475796103477478, "learning_rate": 5.273534334192243e-06, "loss": 0.5808, "step": 2488 }, { "epoch": 1.6037371134020617, "grad_norm": 0.5912460088729858, "learning_rate": 5.269790954569012e-06, "loss": 0.5729, "step": 2489 }, { "epoch": 1.6043814432989691, "grad_norm": 0.6340388655662537, "learning_rate": 5.266047423276e-06, "loss": 0.5777, "step": 2490 }, { "epoch": 1.6050257731958761, "grad_norm": 0.5734379887580872, "learning_rate": 5.262303742417723e-06, "loss": 0.5754, "step": 2491 }, { "epoch": 1.6056701030927836, "grad_norm": 0.5868168473243713, "learning_rate": 5.258559914098791e-06, "loss": 0.5596, "step": 2492 }, { "epoch": 1.6063144329896906, "grad_norm": 0.6242101192474365, "learning_rate": 5.254815940423888e-06, "loss": 0.5637, "step": 2493 }, { "epoch": 1.606958762886598, "grad_norm": 0.7274353504180908, "learning_rate": 5.251071823497784e-06, "loss": 0.5826, "step": 2494 }, { "epoch": 1.607603092783505, "grad_norm": 0.6243786811828613, "learning_rate": 5.247327565425329e-06, "loss": 0.5705, "step": 2495 }, { "epoch": 1.6082474226804124, "grad_norm": 0.6449264287948608, "learning_rate": 5.2435831683114515e-06, "loss": 0.5832, "step": 2496 }, { "epoch": 1.6088917525773194, "grad_norm": 0.6325680613517761, "learning_rate": 5.239838634261159e-06, "loss": 0.5631, "step": 2497 }, { "epoch": 1.6095360824742269, "grad_norm": 0.5790272355079651, "learning_rate": 5.236093965379537e-06, "loss": 0.5773, "step": 2498 }, { "epoch": 1.610180412371134, "grad_norm": 0.7743120789527893, "learning_rate": 5.232349163771741e-06, "loss": 0.5727, "step": 2499 }, { "epoch": 1.6108247422680413, "grad_norm": 0.5306834578514099, "learning_rate": 5.228604231543011e-06, "loss": 0.5701, "step": 2500 }, { "epoch": 1.6114690721649485, "grad_norm": 0.5839548707008362, "learning_rate": 5.224859170798652e-06, "loss": 0.573, "step": 2501 }, { "epoch": 1.6121134020618557, "grad_norm": 0.7540789842605591, "learning_rate": 5.2211139836440435e-06, "loss": 0.5918, "step": 2502 }, { "epoch": 1.612757731958763, "grad_norm": 0.5795901417732239, "learning_rate": 5.217368672184638e-06, "loss": 0.5734, "step": 2503 }, { "epoch": 1.6134020618556701, "grad_norm": 0.6494381427764893, "learning_rate": 5.2136232385259585e-06, "loss": 0.5549, "step": 2504 }, { "epoch": 1.6140463917525774, "grad_norm": 0.5452098846435547, "learning_rate": 5.209877684773593e-06, "loss": 0.5566, "step": 2505 }, { "epoch": 1.6146907216494846, "grad_norm": 0.5427628755569458, "learning_rate": 5.206132013033199e-06, "loss": 0.564, "step": 2506 }, { "epoch": 1.6153350515463918, "grad_norm": 0.5426459312438965, "learning_rate": 5.202386225410501e-06, "loss": 0.5625, "step": 2507 }, { "epoch": 1.615979381443299, "grad_norm": 0.6104823350906372, "learning_rate": 5.198640324011287e-06, "loss": 0.5668, "step": 2508 }, { "epoch": 1.6166237113402062, "grad_norm": 0.6106782555580139, "learning_rate": 5.1948943109414084e-06, "loss": 0.5804, "step": 2509 }, { "epoch": 1.6172680412371134, "grad_norm": 0.6496805548667908, "learning_rate": 5.191148188306785e-06, "loss": 0.5818, "step": 2510 }, { "epoch": 1.6179123711340206, "grad_norm": 0.5541388392448425, "learning_rate": 5.187401958213391e-06, "loss": 0.5682, "step": 2511 }, { "epoch": 1.6185567010309279, "grad_norm": 0.6322898864746094, "learning_rate": 5.183655622767263e-06, "loss": 0.5629, "step": 2512 }, { "epoch": 1.619201030927835, "grad_norm": 0.6547199487686157, "learning_rate": 5.1799091840744995e-06, "loss": 0.5688, "step": 2513 }, { "epoch": 1.6198453608247423, "grad_norm": 0.5836328268051147, "learning_rate": 5.176162644241257e-06, "loss": 0.5409, "step": 2514 }, { "epoch": 1.6204896907216495, "grad_norm": 0.5521203279495239, "learning_rate": 5.172416005373746e-06, "loss": 0.5909, "step": 2515 }, { "epoch": 1.6211340206185567, "grad_norm": 0.6780145168304443, "learning_rate": 5.1686692695782325e-06, "loss": 0.5758, "step": 2516 }, { "epoch": 1.621778350515464, "grad_norm": 0.5796098709106445, "learning_rate": 5.16492243896104e-06, "loss": 0.5617, "step": 2517 }, { "epoch": 1.6224226804123711, "grad_norm": 0.5720744729042053, "learning_rate": 5.1611755156285425e-06, "loss": 0.5831, "step": 2518 }, { "epoch": 1.6230670103092784, "grad_norm": 0.6169164180755615, "learning_rate": 5.15742850168717e-06, "loss": 0.5521, "step": 2519 }, { "epoch": 1.6237113402061856, "grad_norm": 0.6308625936508179, "learning_rate": 5.153681399243397e-06, "loss": 0.567, "step": 2520 }, { "epoch": 1.6243556701030928, "grad_norm": 0.5662408471107483, "learning_rate": 5.149934210403753e-06, "loss": 0.5776, "step": 2521 }, { "epoch": 1.625, "grad_norm": 0.5509442090988159, "learning_rate": 5.146186937274816e-06, "loss": 0.5769, "step": 2522 }, { "epoch": 1.6256443298969072, "grad_norm": 0.650338888168335, "learning_rate": 5.142439581963209e-06, "loss": 0.5687, "step": 2523 }, { "epoch": 1.6262886597938144, "grad_norm": 0.5777653455734253, "learning_rate": 5.138692146575603e-06, "loss": 0.5861, "step": 2524 }, { "epoch": 1.6269329896907216, "grad_norm": 0.5911526679992676, "learning_rate": 5.134944633218712e-06, "loss": 0.5589, "step": 2525 }, { "epoch": 1.6275773195876289, "grad_norm": 0.6390924453735352, "learning_rate": 5.131197043999294e-06, "loss": 0.5779, "step": 2526 }, { "epoch": 1.628221649484536, "grad_norm": 0.5111465454101562, "learning_rate": 5.127449381024153e-06, "loss": 0.5579, "step": 2527 }, { "epoch": 1.6288659793814433, "grad_norm": 0.6700563430786133, "learning_rate": 5.1237016464001324e-06, "loss": 0.555, "step": 2528 }, { "epoch": 1.6295103092783505, "grad_norm": 0.5829666256904602, "learning_rate": 5.119953842234114e-06, "loss": 0.5539, "step": 2529 }, { "epoch": 1.6301546391752577, "grad_norm": 0.6501736640930176, "learning_rate": 5.116205970633021e-06, "loss": 0.5751, "step": 2530 }, { "epoch": 1.630798969072165, "grad_norm": 0.6356644630432129, "learning_rate": 5.1124580337038135e-06, "loss": 0.5802, "step": 2531 }, { "epoch": 1.6314432989690721, "grad_norm": 0.5880082249641418, "learning_rate": 5.10871003355349e-06, "loss": 0.5578, "step": 2532 }, { "epoch": 1.6320876288659794, "grad_norm": 0.6934760808944702, "learning_rate": 5.104961972289082e-06, "loss": 0.5646, "step": 2533 }, { "epoch": 1.6327319587628866, "grad_norm": 0.6609493494033813, "learning_rate": 5.101213852017656e-06, "loss": 0.5839, "step": 2534 }, { "epoch": 1.6333762886597938, "grad_norm": 0.6138235926628113, "learning_rate": 5.097465674846313e-06, "loss": 0.5772, "step": 2535 }, { "epoch": 1.634020618556701, "grad_norm": 0.6089770793914795, "learning_rate": 5.093717442882185e-06, "loss": 0.5858, "step": 2536 }, { "epoch": 1.6346649484536082, "grad_norm": 0.5361588001251221, "learning_rate": 5.089969158232437e-06, "loss": 0.5709, "step": 2537 }, { "epoch": 1.6353092783505154, "grad_norm": 0.6211996674537659, "learning_rate": 5.086220823004258e-06, "loss": 0.5748, "step": 2538 }, { "epoch": 1.6359536082474226, "grad_norm": 0.551705539226532, "learning_rate": 5.082472439304872e-06, "loss": 0.5785, "step": 2539 }, { "epoch": 1.6365979381443299, "grad_norm": 0.5460169911384583, "learning_rate": 5.078724009241526e-06, "loss": 0.5607, "step": 2540 }, { "epoch": 1.637242268041237, "grad_norm": 0.5848313570022583, "learning_rate": 5.074975534921496e-06, "loss": 0.5834, "step": 2541 }, { "epoch": 1.6378865979381443, "grad_norm": 0.5924203991889954, "learning_rate": 5.071227018452077e-06, "loss": 0.5693, "step": 2542 }, { "epoch": 1.6385309278350515, "grad_norm": 0.6382013559341431, "learning_rate": 5.067478461940598e-06, "loss": 0.5649, "step": 2543 }, { "epoch": 1.6391752577319587, "grad_norm": 0.5207600593566895, "learning_rate": 5.063729867494402e-06, "loss": 0.5447, "step": 2544 }, { "epoch": 1.639819587628866, "grad_norm": 0.6731232404708862, "learning_rate": 5.059981237220852e-06, "loss": 0.596, "step": 2545 }, { "epoch": 1.6404639175257731, "grad_norm": 0.6249750852584839, "learning_rate": 5.0562325732273405e-06, "loss": 0.5627, "step": 2546 }, { "epoch": 1.6411082474226806, "grad_norm": 0.6246306300163269, "learning_rate": 5.052483877621272e-06, "loss": 0.5832, "step": 2547 }, { "epoch": 1.6417525773195876, "grad_norm": 0.4886535406112671, "learning_rate": 5.048735152510067e-06, "loss": 0.5637, "step": 2548 }, { "epoch": 1.642396907216495, "grad_norm": 0.6020257472991943, "learning_rate": 5.044986400001169e-06, "loss": 0.5665, "step": 2549 }, { "epoch": 1.643041237113402, "grad_norm": 0.6264144778251648, "learning_rate": 5.041237622202033e-06, "loss": 0.5808, "step": 2550 }, { "epoch": 1.6436855670103094, "grad_norm": 0.615664005279541, "learning_rate": 5.03748882122013e-06, "loss": 0.564, "step": 2551 }, { "epoch": 1.6443298969072164, "grad_norm": 0.7042937874794006, "learning_rate": 5.03373999916294e-06, "loss": 0.5644, "step": 2552 }, { "epoch": 1.6449742268041239, "grad_norm": 0.5629335641860962, "learning_rate": 5.029991158137959e-06, "loss": 0.563, "step": 2553 }, { "epoch": 1.6456185567010309, "grad_norm": 0.6183993220329285, "learning_rate": 5.026242300252695e-06, "loss": 0.5528, "step": 2554 }, { "epoch": 1.6462628865979383, "grad_norm": 0.6102872490882874, "learning_rate": 5.022493427614659e-06, "loss": 0.5671, "step": 2555 }, { "epoch": 1.6469072164948453, "grad_norm": 0.6410378813743591, "learning_rate": 5.018744542331376e-06, "loss": 0.573, "step": 2556 }, { "epoch": 1.6475515463917527, "grad_norm": 0.5479141473770142, "learning_rate": 5.014995646510376e-06, "loss": 0.5792, "step": 2557 }, { "epoch": 1.6481958762886597, "grad_norm": 0.618513822555542, "learning_rate": 5.011246742259197e-06, "loss": 0.578, "step": 2558 }, { "epoch": 1.6488402061855671, "grad_norm": 0.6147071123123169, "learning_rate": 5.007497831685377e-06, "loss": 0.5694, "step": 2559 }, { "epoch": 1.6494845360824741, "grad_norm": 0.565677285194397, "learning_rate": 5.0037489168964624e-06, "loss": 0.5952, "step": 2560 }, { "epoch": 1.6501288659793816, "grad_norm": 0.6193034648895264, "learning_rate": 5e-06, "loss": 0.5765, "step": 2561 }, { "epoch": 1.6507731958762886, "grad_norm": 0.5769947171211243, "learning_rate": 4.996251083103538e-06, "loss": 0.5879, "step": 2562 }, { "epoch": 1.651417525773196, "grad_norm": 0.5633503794670105, "learning_rate": 4.992502168314624e-06, "loss": 0.5629, "step": 2563 }, { "epoch": 1.652061855670103, "grad_norm": 0.6208115220069885, "learning_rate": 4.988753257740805e-06, "loss": 0.5722, "step": 2564 }, { "epoch": 1.6527061855670104, "grad_norm": 0.5484943985939026, "learning_rate": 4.985004353489625e-06, "loss": 0.5729, "step": 2565 }, { "epoch": 1.6533505154639174, "grad_norm": 0.5583198070526123, "learning_rate": 4.981255457668625e-06, "loss": 0.5789, "step": 2566 }, { "epoch": 1.6539948453608249, "grad_norm": 0.5633841156959534, "learning_rate": 4.977506572385343e-06, "loss": 0.576, "step": 2567 }, { "epoch": 1.6546391752577319, "grad_norm": 0.5589533448219299, "learning_rate": 4.973757699747308e-06, "loss": 0.5618, "step": 2568 }, { "epoch": 1.6552835051546393, "grad_norm": 0.6012706756591797, "learning_rate": 4.970008841862043e-06, "loss": 0.5732, "step": 2569 }, { "epoch": 1.6559278350515463, "grad_norm": 0.6086344718933105, "learning_rate": 4.9662600008370625e-06, "loss": 0.5627, "step": 2570 }, { "epoch": 1.6565721649484537, "grad_norm": 0.6072671413421631, "learning_rate": 4.962511178779871e-06, "loss": 0.5706, "step": 2571 }, { "epoch": 1.6572164948453607, "grad_norm": 0.5317866802215576, "learning_rate": 4.958762377797967e-06, "loss": 0.5758, "step": 2572 }, { "epoch": 1.6578608247422681, "grad_norm": 0.512733519077301, "learning_rate": 4.955013599998831e-06, "loss": 0.5821, "step": 2573 }, { "epoch": 1.6585051546391751, "grad_norm": 0.5739752054214478, "learning_rate": 4.9512648474899344e-06, "loss": 0.5504, "step": 2574 }, { "epoch": 1.6591494845360826, "grad_norm": 0.567739725112915, "learning_rate": 4.94751612237873e-06, "loss": 0.5787, "step": 2575 }, { "epoch": 1.6597938144329896, "grad_norm": 0.48787903785705566, "learning_rate": 4.94376742677266e-06, "loss": 0.5832, "step": 2576 }, { "epoch": 1.660438144329897, "grad_norm": 0.5481950640678406, "learning_rate": 4.940018762779149e-06, "loss": 0.5774, "step": 2577 }, { "epoch": 1.661082474226804, "grad_norm": 0.5634629726409912, "learning_rate": 4.936270132505601e-06, "loss": 0.5721, "step": 2578 }, { "epoch": 1.6617268041237114, "grad_norm": 0.6715635657310486, "learning_rate": 4.932521538059403e-06, "loss": 0.5737, "step": 2579 }, { "epoch": 1.6623711340206184, "grad_norm": 0.5822291374206543, "learning_rate": 4.928772981547924e-06, "loss": 0.5849, "step": 2580 }, { "epoch": 1.6630154639175259, "grad_norm": 0.6195212006568909, "learning_rate": 4.925024465078506e-06, "loss": 0.57, "step": 2581 }, { "epoch": 1.6636597938144329, "grad_norm": 0.5709973573684692, "learning_rate": 4.921275990758476e-06, "loss": 0.5705, "step": 2582 }, { "epoch": 1.6643041237113403, "grad_norm": 0.5742570757865906, "learning_rate": 4.91752756069513e-06, "loss": 0.5561, "step": 2583 }, { "epoch": 1.6649484536082473, "grad_norm": 0.5978179574012756, "learning_rate": 4.913779176995743e-06, "loss": 0.5728, "step": 2584 }, { "epoch": 1.6655927835051547, "grad_norm": 0.5400027632713318, "learning_rate": 4.910030841767565e-06, "loss": 0.5744, "step": 2585 }, { "epoch": 1.6662371134020617, "grad_norm": 0.5768642425537109, "learning_rate": 4.906282557117817e-06, "loss": 0.5629, "step": 2586 }, { "epoch": 1.6668814432989691, "grad_norm": 0.6322346925735474, "learning_rate": 4.902534325153689e-06, "loss": 0.5533, "step": 2587 }, { "epoch": 1.6675257731958761, "grad_norm": 0.6008243560791016, "learning_rate": 4.898786147982346e-06, "loss": 0.5512, "step": 2588 }, { "epoch": 1.6681701030927836, "grad_norm": 0.5516601204872131, "learning_rate": 4.895038027710919e-06, "loss": 0.5511, "step": 2589 }, { "epoch": 1.6688144329896906, "grad_norm": 0.6047698855400085, "learning_rate": 4.891289966446511e-06, "loss": 0.5623, "step": 2590 }, { "epoch": 1.669458762886598, "grad_norm": 0.6072372794151306, "learning_rate": 4.887541966296187e-06, "loss": 0.56, "step": 2591 }, { "epoch": 1.670103092783505, "grad_norm": 0.525249183177948, "learning_rate": 4.8837940293669795e-06, "loss": 0.5867, "step": 2592 }, { "epoch": 1.6707474226804124, "grad_norm": 0.5463447570800781, "learning_rate": 4.880046157765887e-06, "loss": 0.5815, "step": 2593 }, { "epoch": 1.6713917525773194, "grad_norm": 0.6461141705513, "learning_rate": 4.876298353599868e-06, "loss": 0.558, "step": 2594 }, { "epoch": 1.6720360824742269, "grad_norm": 0.5734620094299316, "learning_rate": 4.872550618975848e-06, "loss": 0.5918, "step": 2595 }, { "epoch": 1.672680412371134, "grad_norm": 0.5866066217422485, "learning_rate": 4.8688029560007064e-06, "loss": 0.5454, "step": 2596 }, { "epoch": 1.6733247422680413, "grad_norm": 0.5820803046226501, "learning_rate": 4.865055366781291e-06, "loss": 0.5604, "step": 2597 }, { "epoch": 1.6739690721649485, "grad_norm": 0.6103181838989258, "learning_rate": 4.8613078534244e-06, "loss": 0.5766, "step": 2598 }, { "epoch": 1.6746134020618557, "grad_norm": 0.6379008293151855, "learning_rate": 4.857560418036792e-06, "loss": 0.5692, "step": 2599 }, { "epoch": 1.675257731958763, "grad_norm": 0.5892603993415833, "learning_rate": 4.853813062725186e-06, "loss": 0.574, "step": 2600 }, { "epoch": 1.6759020618556701, "grad_norm": 0.61902916431427, "learning_rate": 4.850065789596249e-06, "loss": 0.5642, "step": 2601 }, { "epoch": 1.6765463917525774, "grad_norm": 0.5464742183685303, "learning_rate": 4.846318600756605e-06, "loss": 0.5896, "step": 2602 }, { "epoch": 1.6771907216494846, "grad_norm": 0.5853357315063477, "learning_rate": 4.842571498312833e-06, "loss": 0.586, "step": 2603 }, { "epoch": 1.6778350515463918, "grad_norm": 0.5869336128234863, "learning_rate": 4.838824484371459e-06, "loss": 0.5818, "step": 2604 }, { "epoch": 1.678479381443299, "grad_norm": 0.6063686013221741, "learning_rate": 4.8350775610389626e-06, "loss": 0.5672, "step": 2605 }, { "epoch": 1.6791237113402062, "grad_norm": 0.5745872259140015, "learning_rate": 4.831330730421769e-06, "loss": 0.5733, "step": 2606 }, { "epoch": 1.6797680412371134, "grad_norm": 0.5656169056892395, "learning_rate": 4.827583994626254e-06, "loss": 0.5597, "step": 2607 }, { "epoch": 1.6804123711340206, "grad_norm": 0.5770779252052307, "learning_rate": 4.823837355758743e-06, "loss": 0.5605, "step": 2608 }, { "epoch": 1.6810567010309279, "grad_norm": 0.5399340391159058, "learning_rate": 4.8200908159255004e-06, "loss": 0.5573, "step": 2609 }, { "epoch": 1.681701030927835, "grad_norm": 0.5613174438476562, "learning_rate": 4.816344377232737e-06, "loss": 0.5696, "step": 2610 }, { "epoch": 1.6823453608247423, "grad_norm": 0.6286584734916687, "learning_rate": 4.812598041786611e-06, "loss": 0.5578, "step": 2611 }, { "epoch": 1.6829896907216495, "grad_norm": 0.5467870235443115, "learning_rate": 4.808851811693217e-06, "loss": 0.5672, "step": 2612 }, { "epoch": 1.6836340206185567, "grad_norm": 0.5605920553207397, "learning_rate": 4.805105689058593e-06, "loss": 0.5625, "step": 2613 }, { "epoch": 1.684278350515464, "grad_norm": 0.6260461211204529, "learning_rate": 4.801359675988715e-06, "loss": 0.5566, "step": 2614 }, { "epoch": 1.6849226804123711, "grad_norm": 0.586699366569519, "learning_rate": 4.797613774589501e-06, "loss": 0.5717, "step": 2615 }, { "epoch": 1.6855670103092784, "grad_norm": 0.5790180563926697, "learning_rate": 4.793867986966802e-06, "loss": 0.5715, "step": 2616 }, { "epoch": 1.6862113402061856, "grad_norm": 0.5498654842376709, "learning_rate": 4.790122315226408e-06, "loss": 0.5825, "step": 2617 }, { "epoch": 1.6868556701030928, "grad_norm": 0.5632086396217346, "learning_rate": 4.786376761474042e-06, "loss": 0.5551, "step": 2618 }, { "epoch": 1.6875, "grad_norm": 0.5331572890281677, "learning_rate": 4.782631327815363e-06, "loss": 0.5752, "step": 2619 }, { "epoch": 1.6881443298969072, "grad_norm": 0.5676560997962952, "learning_rate": 4.778886016355958e-06, "loss": 0.5723, "step": 2620 }, { "epoch": 1.6887886597938144, "grad_norm": 0.612542450428009, "learning_rate": 4.775140829201351e-06, "loss": 0.5514, "step": 2621 }, { "epoch": 1.6894329896907216, "grad_norm": 0.5722492933273315, "learning_rate": 4.7713957684569926e-06, "loss": 0.5709, "step": 2622 }, { "epoch": 1.6900773195876289, "grad_norm": 0.5344690084457397, "learning_rate": 4.767650836228261e-06, "loss": 0.5696, "step": 2623 }, { "epoch": 1.690721649484536, "grad_norm": 0.672766923904419, "learning_rate": 4.763906034620466e-06, "loss": 0.5996, "step": 2624 }, { "epoch": 1.6913659793814433, "grad_norm": 0.5352060794830322, "learning_rate": 4.7601613657388415e-06, "loss": 0.5475, "step": 2625 }, { "epoch": 1.6920103092783505, "grad_norm": 0.5878434181213379, "learning_rate": 4.756416831688549e-06, "loss": 0.5647, "step": 2626 }, { "epoch": 1.6926546391752577, "grad_norm": 0.5802128911018372, "learning_rate": 4.752672434574672e-06, "loss": 0.5659, "step": 2627 }, { "epoch": 1.693298969072165, "grad_norm": 0.5820068717002869, "learning_rate": 4.748928176502216e-06, "loss": 0.5694, "step": 2628 }, { "epoch": 1.6939432989690721, "grad_norm": 0.5800421833992004, "learning_rate": 4.745184059576113e-06, "loss": 0.5961, "step": 2629 }, { "epoch": 1.6945876288659794, "grad_norm": 0.5976626873016357, "learning_rate": 4.74144008590121e-06, "loss": 0.5611, "step": 2630 }, { "epoch": 1.6952319587628866, "grad_norm": 0.642803430557251, "learning_rate": 4.737696257582277e-06, "loss": 0.5741, "step": 2631 }, { "epoch": 1.6958762886597938, "grad_norm": 0.5535104870796204, "learning_rate": 4.733952576724002e-06, "loss": 0.5784, "step": 2632 }, { "epoch": 1.696520618556701, "grad_norm": 0.5979894399642944, "learning_rate": 4.730209045430989e-06, "loss": 0.5816, "step": 2633 }, { "epoch": 1.6971649484536082, "grad_norm": 0.6116093993186951, "learning_rate": 4.726465665807759e-06, "loss": 0.5628, "step": 2634 }, { "epoch": 1.6978092783505154, "grad_norm": 0.6394659876823425, "learning_rate": 4.722722439958745e-06, "loss": 0.5786, "step": 2635 }, { "epoch": 1.6984536082474226, "grad_norm": 0.5464919805526733, "learning_rate": 4.718979369988299e-06, "loss": 0.5819, "step": 2636 }, { "epoch": 1.6990979381443299, "grad_norm": 0.5884142518043518, "learning_rate": 4.71523645800068e-06, "loss": 0.5918, "step": 2637 }, { "epoch": 1.699742268041237, "grad_norm": 0.5647599101066589, "learning_rate": 4.7114937061000595e-06, "loss": 0.5757, "step": 2638 }, { "epoch": 1.7003865979381443, "grad_norm": 0.6377708315849304, "learning_rate": 4.707751116390522e-06, "loss": 0.5671, "step": 2639 }, { "epoch": 1.7010309278350515, "grad_norm": 0.6433849334716797, "learning_rate": 4.704008690976058e-06, "loss": 0.5926, "step": 2640 }, { "epoch": 1.7016752577319587, "grad_norm": 0.6344012022018433, "learning_rate": 4.7002664319605654e-06, "loss": 0.5834, "step": 2641 }, { "epoch": 1.702319587628866, "grad_norm": 0.7104661464691162, "learning_rate": 4.696524341447849e-06, "loss": 0.5786, "step": 2642 }, { "epoch": 1.7029639175257731, "grad_norm": 0.591728925704956, "learning_rate": 4.692782421541619e-06, "loss": 0.5646, "step": 2643 }, { "epoch": 1.7036082474226806, "grad_norm": 0.6541820168495178, "learning_rate": 4.689040674345494e-06, "loss": 0.565, "step": 2644 }, { "epoch": 1.7042525773195876, "grad_norm": 0.5952646136283875, "learning_rate": 4.685299101962987e-06, "loss": 0.5626, "step": 2645 }, { "epoch": 1.704896907216495, "grad_norm": 0.5713399648666382, "learning_rate": 4.681557706497519e-06, "loss": 0.5702, "step": 2646 }, { "epoch": 1.705541237113402, "grad_norm": 0.5416255593299866, "learning_rate": 4.677816490052409e-06, "loss": 0.5958, "step": 2647 }, { "epoch": 1.7061855670103094, "grad_norm": 0.5888693332672119, "learning_rate": 4.674075454730878e-06, "loss": 0.5642, "step": 2648 }, { "epoch": 1.7068298969072164, "grad_norm": 0.5879565477371216, "learning_rate": 4.67033460263604e-06, "loss": 0.5733, "step": 2649 }, { "epoch": 1.7074742268041239, "grad_norm": 0.555097758769989, "learning_rate": 4.666593935870912e-06, "loss": 0.5622, "step": 2650 }, { "epoch": 1.7081185567010309, "grad_norm": 0.591090202331543, "learning_rate": 4.662853456538403e-06, "loss": 0.5538, "step": 2651 }, { "epoch": 1.7087628865979383, "grad_norm": 0.5573540329933167, "learning_rate": 4.659113166741318e-06, "loss": 0.575, "step": 2652 }, { "epoch": 1.7094072164948453, "grad_norm": 0.6191607713699341, "learning_rate": 4.6553730685823534e-06, "loss": 0.5606, "step": 2653 }, { "epoch": 1.7100515463917527, "grad_norm": 0.6047584414482117, "learning_rate": 4.6516331641641026e-06, "loss": 0.5796, "step": 2654 }, { "epoch": 1.7106958762886597, "grad_norm": 0.6288405656814575, "learning_rate": 4.647893455589046e-06, "loss": 0.562, "step": 2655 }, { "epoch": 1.7113402061855671, "grad_norm": 0.5679271221160889, "learning_rate": 4.644153944959553e-06, "loss": 0.5716, "step": 2656 }, { "epoch": 1.7119845360824741, "grad_norm": 0.569719672203064, "learning_rate": 4.640414634377887e-06, "loss": 0.5945, "step": 2657 }, { "epoch": 1.7126288659793816, "grad_norm": 0.5804483294487, "learning_rate": 4.636675525946193e-06, "loss": 0.5869, "step": 2658 }, { "epoch": 1.7132731958762886, "grad_norm": 0.6337373852729797, "learning_rate": 4.632936621766506e-06, "loss": 0.5727, "step": 2659 }, { "epoch": 1.713917525773196, "grad_norm": 0.5533632636070251, "learning_rate": 4.629197923940745e-06, "loss": 0.5756, "step": 2660 }, { "epoch": 1.714561855670103, "grad_norm": 0.6014413237571716, "learning_rate": 4.625459434570712e-06, "loss": 0.5667, "step": 2661 }, { "epoch": 1.7152061855670104, "grad_norm": 0.6278036832809448, "learning_rate": 4.621721155758097e-06, "loss": 0.5562, "step": 2662 }, { "epoch": 1.7158505154639174, "grad_norm": 0.5508294105529785, "learning_rate": 4.6179830896044635e-06, "loss": 0.5783, "step": 2663 }, { "epoch": 1.7164948453608249, "grad_norm": 0.5993038415908813, "learning_rate": 4.614245238211262e-06, "loss": 0.5664, "step": 2664 }, { "epoch": 1.7171391752577319, "grad_norm": 0.6044915318489075, "learning_rate": 4.610507603679818e-06, "loss": 0.5797, "step": 2665 }, { "epoch": 1.7177835051546393, "grad_norm": 0.5934248566627502, "learning_rate": 4.606770188111339e-06, "loss": 0.5869, "step": 2666 }, { "epoch": 1.7184278350515463, "grad_norm": 0.5219321250915527, "learning_rate": 4.603032993606905e-06, "loss": 0.5712, "step": 2667 }, { "epoch": 1.7190721649484537, "grad_norm": 0.5772237181663513, "learning_rate": 4.5992960222674756e-06, "loss": 0.5774, "step": 2668 }, { "epoch": 1.7197164948453607, "grad_norm": 0.5810965895652771, "learning_rate": 4.595559276193883e-06, "loss": 0.5673, "step": 2669 }, { "epoch": 1.7203608247422681, "grad_norm": 0.5527998208999634, "learning_rate": 4.591822757486833e-06, "loss": 0.5662, "step": 2670 }, { "epoch": 1.7210051546391751, "grad_norm": 0.6488362550735474, "learning_rate": 4.588086468246904e-06, "loss": 0.5536, "step": 2671 }, { "epoch": 1.7216494845360826, "grad_norm": 0.624985933303833, "learning_rate": 4.584350410574545e-06, "loss": 0.5622, "step": 2672 }, { "epoch": 1.7222938144329896, "grad_norm": 0.632620632648468, "learning_rate": 4.580614586570075e-06, "loss": 0.5517, "step": 2673 }, { "epoch": 1.722938144329897, "grad_norm": 0.5424154996871948, "learning_rate": 4.576878998333681e-06, "loss": 0.5523, "step": 2674 }, { "epoch": 1.723582474226804, "grad_norm": 0.6706408262252808, "learning_rate": 4.573143647965419e-06, "loss": 0.5681, "step": 2675 }, { "epoch": 1.7242268041237114, "grad_norm": 0.6516699194908142, "learning_rate": 4.5694085375652105e-06, "loss": 0.5878, "step": 2676 }, { "epoch": 1.7248711340206184, "grad_norm": 0.6535937786102295, "learning_rate": 4.565673669232841e-06, "loss": 0.5844, "step": 2677 }, { "epoch": 1.7255154639175259, "grad_norm": 0.6568611264228821, "learning_rate": 4.56193904506796e-06, "loss": 0.5801, "step": 2678 }, { "epoch": 1.7261597938144329, "grad_norm": 0.6565068960189819, "learning_rate": 4.558204667170081e-06, "loss": 0.5665, "step": 2679 }, { "epoch": 1.7268041237113403, "grad_norm": 0.6474490165710449, "learning_rate": 4.554470537638581e-06, "loss": 0.5586, "step": 2680 }, { "epoch": 1.7274484536082473, "grad_norm": 0.7198812365531921, "learning_rate": 4.5507366585726935e-06, "loss": 0.5637, "step": 2681 }, { "epoch": 1.7280927835051547, "grad_norm": 0.685660719871521, "learning_rate": 4.5470030320715114e-06, "loss": 0.572, "step": 2682 }, { "epoch": 1.7287371134020617, "grad_norm": 0.6694272756576538, "learning_rate": 4.543269660233989e-06, "loss": 0.5589, "step": 2683 }, { "epoch": 1.7293814432989691, "grad_norm": 0.5473025441169739, "learning_rate": 4.539536545158935e-06, "loss": 0.5537, "step": 2684 }, { "epoch": 1.7300257731958761, "grad_norm": 0.6066479682922363, "learning_rate": 4.535803688945012e-06, "loss": 0.5758, "step": 2685 }, { "epoch": 1.7306701030927836, "grad_norm": 0.6119387745857239, "learning_rate": 4.532071093690741e-06, "loss": 0.5674, "step": 2686 }, { "epoch": 1.7313144329896906, "grad_norm": 0.5819975137710571, "learning_rate": 4.528338761494495e-06, "loss": 0.5816, "step": 2687 }, { "epoch": 1.731958762886598, "grad_norm": 0.5992373824119568, "learning_rate": 4.524606694454497e-06, "loss": 0.5673, "step": 2688 }, { "epoch": 1.732603092783505, "grad_norm": 0.5858043432235718, "learning_rate": 4.520874894668822e-06, "loss": 0.5653, "step": 2689 }, { "epoch": 1.7332474226804124, "grad_norm": 0.5261032581329346, "learning_rate": 4.517143364235399e-06, "loss": 0.5762, "step": 2690 }, { "epoch": 1.7338917525773194, "grad_norm": 0.5684706568717957, "learning_rate": 4.513412105251998e-06, "loss": 0.5562, "step": 2691 }, { "epoch": 1.7345360824742269, "grad_norm": 0.7387073040008545, "learning_rate": 4.509681119816241e-06, "loss": 0.5887, "step": 2692 }, { "epoch": 1.735180412371134, "grad_norm": 0.5535421371459961, "learning_rate": 4.5059504100255965e-06, "loss": 0.5812, "step": 2693 }, { "epoch": 1.7358247422680413, "grad_norm": 0.6771222949028015, "learning_rate": 4.502219977977377e-06, "loss": 0.5705, "step": 2694 }, { "epoch": 1.7364690721649485, "grad_norm": 0.6787487268447876, "learning_rate": 4.498489825768738e-06, "loss": 0.5859, "step": 2695 }, { "epoch": 1.7371134020618557, "grad_norm": 0.5667046904563904, "learning_rate": 4.494759955496678e-06, "loss": 0.5569, "step": 2696 }, { "epoch": 1.737757731958763, "grad_norm": 0.7517831921577454, "learning_rate": 4.491030369258038e-06, "loss": 0.5593, "step": 2697 }, { "epoch": 1.7384020618556701, "grad_norm": 0.6950629353523254, "learning_rate": 4.487301069149501e-06, "loss": 0.566, "step": 2698 }, { "epoch": 1.7390463917525774, "grad_norm": 0.6444603204727173, "learning_rate": 4.483572057267586e-06, "loss": 0.5741, "step": 2699 }, { "epoch": 1.7396907216494846, "grad_norm": 0.664344310760498, "learning_rate": 4.479843335708648e-06, "loss": 0.5686, "step": 2700 }, { "epoch": 1.7403350515463918, "grad_norm": 0.5849413275718689, "learning_rate": 4.476114906568886e-06, "loss": 0.5806, "step": 2701 }, { "epoch": 1.740979381443299, "grad_norm": 0.5611298680305481, "learning_rate": 4.472386771944328e-06, "loss": 0.5724, "step": 2702 }, { "epoch": 1.7416237113402062, "grad_norm": 0.6091529726982117, "learning_rate": 4.46865893393084e-06, "loss": 0.5791, "step": 2703 }, { "epoch": 1.7422680412371134, "grad_norm": 0.5561123490333557, "learning_rate": 4.464931394624118e-06, "loss": 0.5745, "step": 2704 }, { "epoch": 1.7429123711340206, "grad_norm": 0.5851315855979919, "learning_rate": 4.461204156119694e-06, "loss": 0.5898, "step": 2705 }, { "epoch": 1.7435567010309279, "grad_norm": 0.645096480846405, "learning_rate": 4.457477220512929e-06, "loss": 0.594, "step": 2706 }, { "epoch": 1.744201030927835, "grad_norm": 0.6325176954269409, "learning_rate": 4.453750589899013e-06, "loss": 0.5819, "step": 2707 }, { "epoch": 1.7448453608247423, "grad_norm": 0.6620206236839294, "learning_rate": 4.450024266372965e-06, "loss": 0.5659, "step": 2708 }, { "epoch": 1.7454896907216495, "grad_norm": 0.7081724405288696, "learning_rate": 4.4462982520296336e-06, "loss": 0.568, "step": 2709 }, { "epoch": 1.7461340206185567, "grad_norm": 0.5690625309944153, "learning_rate": 4.442572548963688e-06, "loss": 0.5947, "step": 2710 }, { "epoch": 1.746778350515464, "grad_norm": 0.6309353113174438, "learning_rate": 4.438847159269631e-06, "loss": 0.5516, "step": 2711 }, { "epoch": 1.7474226804123711, "grad_norm": 0.6303577423095703, "learning_rate": 4.43512208504178e-06, "loss": 0.5545, "step": 2712 }, { "epoch": 1.7480670103092784, "grad_norm": 0.6212533116340637, "learning_rate": 4.431397328374282e-06, "loss": 0.5518, "step": 2713 }, { "epoch": 1.7487113402061856, "grad_norm": 0.5665958523750305, "learning_rate": 4.4276728913611e-06, "loss": 0.5819, "step": 2714 }, { "epoch": 1.7493556701030928, "grad_norm": 0.5828904509544373, "learning_rate": 4.423948776096021e-06, "loss": 0.5709, "step": 2715 }, { "epoch": 1.75, "grad_norm": 0.5338139533996582, "learning_rate": 4.420224984672654e-06, "loss": 0.549, "step": 2716 }, { "epoch": 1.7506443298969072, "grad_norm": 0.5911694169044495, "learning_rate": 4.416501519184418e-06, "loss": 0.5648, "step": 2717 }, { "epoch": 1.7512886597938144, "grad_norm": 0.5545548796653748, "learning_rate": 4.412778381724553e-06, "loss": 0.5757, "step": 2718 }, { "epoch": 1.7519329896907216, "grad_norm": 0.5479432940483093, "learning_rate": 4.409055574386118e-06, "loss": 0.5703, "step": 2719 }, { "epoch": 1.7525773195876289, "grad_norm": 0.5440157055854797, "learning_rate": 4.4053330992619795e-06, "loss": 0.5681, "step": 2720 }, { "epoch": 1.753221649484536, "grad_norm": 0.6193472146987915, "learning_rate": 4.401610958444822e-06, "loss": 0.5814, "step": 2721 }, { "epoch": 1.7538659793814433, "grad_norm": 0.5735450983047485, "learning_rate": 4.39788915402714e-06, "loss": 0.5779, "step": 2722 }, { "epoch": 1.7545103092783505, "grad_norm": 0.5814749002456665, "learning_rate": 4.3941676881012415e-06, "loss": 0.5805, "step": 2723 }, { "epoch": 1.7551546391752577, "grad_norm": 0.6089169979095459, "learning_rate": 4.39044656275924e-06, "loss": 0.5717, "step": 2724 }, { "epoch": 1.755798969072165, "grad_norm": 0.6927036046981812, "learning_rate": 4.38672578009306e-06, "loss": 0.5628, "step": 2725 }, { "epoch": 1.7564432989690721, "grad_norm": 0.6161792874336243, "learning_rate": 4.383005342194436e-06, "loss": 0.5642, "step": 2726 }, { "epoch": 1.7570876288659794, "grad_norm": 0.6846563220024109, "learning_rate": 4.379285251154903e-06, "loss": 0.5607, "step": 2727 }, { "epoch": 1.7577319587628866, "grad_norm": 0.6080223917961121, "learning_rate": 4.3755655090658045e-06, "loss": 0.5829, "step": 2728 }, { "epoch": 1.7583762886597938, "grad_norm": 0.6134759187698364, "learning_rate": 4.371846118018288e-06, "loss": 0.5736, "step": 2729 }, { "epoch": 1.759020618556701, "grad_norm": 0.6868017911911011, "learning_rate": 4.368127080103303e-06, "loss": 0.5766, "step": 2730 }, { "epoch": 1.7596649484536082, "grad_norm": 0.5687644481658936, "learning_rate": 4.3644083974116e-06, "loss": 0.5803, "step": 2731 }, { "epoch": 1.7603092783505154, "grad_norm": 0.5915237069129944, "learning_rate": 4.360690072033729e-06, "loss": 0.57, "step": 2732 }, { "epoch": 1.7609536082474226, "grad_norm": 0.7398574948310852, "learning_rate": 4.3569721060600404e-06, "loss": 0.6014, "step": 2733 }, { "epoch": 1.7615979381443299, "grad_norm": 0.5262463092803955, "learning_rate": 4.353254501580685e-06, "loss": 0.5767, "step": 2734 }, { "epoch": 1.762242268041237, "grad_norm": 0.6216572523117065, "learning_rate": 4.349537260685606e-06, "loss": 0.5742, "step": 2735 }, { "epoch": 1.7628865979381443, "grad_norm": 0.6996229290962219, "learning_rate": 4.345820385464543e-06, "loss": 0.5621, "step": 2736 }, { "epoch": 1.7635309278350515, "grad_norm": 0.5435518622398376, "learning_rate": 4.342103878007035e-06, "loss": 0.5691, "step": 2737 }, { "epoch": 1.7641752577319587, "grad_norm": 0.5825214385986328, "learning_rate": 4.338387740402406e-06, "loss": 0.5613, "step": 2738 }, { "epoch": 1.764819587628866, "grad_norm": 0.583271324634552, "learning_rate": 4.334671974739778e-06, "loss": 0.5609, "step": 2739 }, { "epoch": 1.7654639175257731, "grad_norm": 0.5987973809242249, "learning_rate": 4.330956583108062e-06, "loss": 0.5665, "step": 2740 }, { "epoch": 1.7661082474226806, "grad_norm": 0.5757119655609131, "learning_rate": 4.32724156759596e-06, "loss": 0.5733, "step": 2741 }, { "epoch": 1.7667525773195876, "grad_norm": 0.5447065830230713, "learning_rate": 4.323526930291961e-06, "loss": 0.5756, "step": 2742 }, { "epoch": 1.767396907216495, "grad_norm": 0.6177639365196228, "learning_rate": 4.319812673284341e-06, "loss": 0.578, "step": 2743 }, { "epoch": 1.768041237113402, "grad_norm": 0.6059287786483765, "learning_rate": 4.3160987986611636e-06, "loss": 0.5624, "step": 2744 }, { "epoch": 1.7686855670103094, "grad_norm": 0.6235113143920898, "learning_rate": 4.312385308510278e-06, "loss": 0.5636, "step": 2745 }, { "epoch": 1.7693298969072164, "grad_norm": 0.5871730446815491, "learning_rate": 4.308672204919316e-06, "loss": 0.5809, "step": 2746 }, { "epoch": 1.7699742268041239, "grad_norm": 0.5943889021873474, "learning_rate": 4.304959489975691e-06, "loss": 0.5568, "step": 2747 }, { "epoch": 1.7706185567010309, "grad_norm": 0.680832028388977, "learning_rate": 4.301247165766601e-06, "loss": 0.56, "step": 2748 }, { "epoch": 1.7712628865979383, "grad_norm": 0.6777725219726562, "learning_rate": 4.2975352343790215e-06, "loss": 0.5716, "step": 2749 }, { "epoch": 1.7719072164948453, "grad_norm": 0.6909160017967224, "learning_rate": 4.2938236978997076e-06, "loss": 0.5591, "step": 2750 }, { "epoch": 1.7725515463917527, "grad_norm": 0.5914145708084106, "learning_rate": 4.290112558415193e-06, "loss": 0.5642, "step": 2751 }, { "epoch": 1.7731958762886597, "grad_norm": 0.6330007314682007, "learning_rate": 4.286401818011791e-06, "loss": 0.5891, "step": 2752 }, { "epoch": 1.7738402061855671, "grad_norm": 0.655286431312561, "learning_rate": 4.282691478775587e-06, "loss": 0.5841, "step": 2753 }, { "epoch": 1.7744845360824741, "grad_norm": 0.6868444085121155, "learning_rate": 4.2789815427924385e-06, "loss": 0.5751, "step": 2754 }, { "epoch": 1.7751288659793816, "grad_norm": 0.5675916075706482, "learning_rate": 4.2752720121479835e-06, "loss": 0.5794, "step": 2755 }, { "epoch": 1.7757731958762886, "grad_norm": 0.7160640954971313, "learning_rate": 4.271562888927626e-06, "loss": 0.5863, "step": 2756 }, { "epoch": 1.776417525773196, "grad_norm": 0.6528710126876831, "learning_rate": 4.267854175216543e-06, "loss": 0.581, "step": 2757 }, { "epoch": 1.777061855670103, "grad_norm": 0.6195478439331055, "learning_rate": 4.264145873099681e-06, "loss": 0.5539, "step": 2758 }, { "epoch": 1.7777061855670104, "grad_norm": 0.522448718547821, "learning_rate": 4.260437984661758e-06, "loss": 0.5636, "step": 2759 }, { "epoch": 1.7783505154639174, "grad_norm": 0.6648046970367432, "learning_rate": 4.256730511987254e-06, "loss": 0.566, "step": 2760 }, { "epoch": 1.7789948453608249, "grad_norm": 0.6658997535705566, "learning_rate": 4.253023457160418e-06, "loss": 0.5628, "step": 2761 }, { "epoch": 1.7796391752577319, "grad_norm": 0.5952633619308472, "learning_rate": 4.249316822265267e-06, "loss": 0.5793, "step": 2762 }, { "epoch": 1.7802835051546393, "grad_norm": 0.5630367994308472, "learning_rate": 4.245610609385579e-06, "loss": 0.5779, "step": 2763 }, { "epoch": 1.7809278350515463, "grad_norm": 0.5479838252067566, "learning_rate": 4.241904820604891e-06, "loss": 0.5764, "step": 2764 }, { "epoch": 1.7815721649484537, "grad_norm": 0.5988810062408447, "learning_rate": 4.23819945800651e-06, "loss": 0.5567, "step": 2765 }, { "epoch": 1.7822164948453607, "grad_norm": 0.5609938502311707, "learning_rate": 4.234494523673497e-06, "loss": 0.5992, "step": 2766 }, { "epoch": 1.7828608247422681, "grad_norm": 0.5700583457946777, "learning_rate": 4.230790019688673e-06, "loss": 0.5818, "step": 2767 }, { "epoch": 1.7835051546391751, "grad_norm": 0.5923458933830261, "learning_rate": 4.227085948134621e-06, "loss": 0.5884, "step": 2768 }, { "epoch": 1.7841494845360826, "grad_norm": 0.5662264823913574, "learning_rate": 4.223382311093674e-06, "loss": 0.5804, "step": 2769 }, { "epoch": 1.7847938144329896, "grad_norm": 0.6398096084594727, "learning_rate": 4.2196791106479305e-06, "loss": 0.5724, "step": 2770 }, { "epoch": 1.785438144329897, "grad_norm": 0.5714444518089294, "learning_rate": 4.215976348879235e-06, "loss": 0.5808, "step": 2771 }, { "epoch": 1.786082474226804, "grad_norm": 0.5401431322097778, "learning_rate": 4.2122740278691865e-06, "loss": 0.5582, "step": 2772 }, { "epoch": 1.7867268041237114, "grad_norm": 0.5713866949081421, "learning_rate": 4.208572149699141e-06, "loss": 0.5719, "step": 2773 }, { "epoch": 1.7873711340206184, "grad_norm": 0.5974870920181274, "learning_rate": 4.2048707164502025e-06, "loss": 0.5607, "step": 2774 }, { "epoch": 1.7880154639175259, "grad_norm": 0.5361068248748779, "learning_rate": 4.201169730203223e-06, "loss": 0.5462, "step": 2775 }, { "epoch": 1.7886597938144329, "grad_norm": 0.5184583067893982, "learning_rate": 4.1974691930388055e-06, "loss": 0.5749, "step": 2776 }, { "epoch": 1.7893041237113403, "grad_norm": 0.6876800060272217, "learning_rate": 4.193769107037302e-06, "loss": 0.5718, "step": 2777 }, { "epoch": 1.7899484536082473, "grad_norm": 0.6487565636634827, "learning_rate": 4.190069474278808e-06, "loss": 0.5757, "step": 2778 }, { "epoch": 1.7905927835051547, "grad_norm": 0.5934075713157654, "learning_rate": 4.186370296843164e-06, "loss": 0.5839, "step": 2779 }, { "epoch": 1.7912371134020617, "grad_norm": 0.7641155123710632, "learning_rate": 4.182671576809957e-06, "loss": 0.563, "step": 2780 }, { "epoch": 1.7918814432989691, "grad_norm": 0.6969650387763977, "learning_rate": 4.1789733162585165e-06, "loss": 0.5655, "step": 2781 }, { "epoch": 1.7925257731958761, "grad_norm": 0.7335861325263977, "learning_rate": 4.17527551726791e-06, "loss": 0.5862, "step": 2782 }, { "epoch": 1.7931701030927836, "grad_norm": 0.6572924852371216, "learning_rate": 4.1715781819169525e-06, "loss": 0.561, "step": 2783 }, { "epoch": 1.7938144329896906, "grad_norm": 0.6245686411857605, "learning_rate": 4.167881312284192e-06, "loss": 0.571, "step": 2784 }, { "epoch": 1.794458762886598, "grad_norm": 0.631572425365448, "learning_rate": 4.164184910447915e-06, "loss": 0.5525, "step": 2785 }, { "epoch": 1.795103092783505, "grad_norm": 0.7144885659217834, "learning_rate": 4.16048897848615e-06, "loss": 0.5795, "step": 2786 }, { "epoch": 1.7957474226804124, "grad_norm": 0.7426769137382507, "learning_rate": 4.156793518476655e-06, "loss": 0.5708, "step": 2787 }, { "epoch": 1.7963917525773194, "grad_norm": 0.6087492108345032, "learning_rate": 4.153098532496931e-06, "loss": 0.5684, "step": 2788 }, { "epoch": 1.7970360824742269, "grad_norm": 0.603992223739624, "learning_rate": 4.149404022624204e-06, "loss": 0.5743, "step": 2789 }, { "epoch": 1.797680412371134, "grad_norm": 0.67958003282547, "learning_rate": 4.145709990935434e-06, "loss": 0.5445, "step": 2790 }, { "epoch": 1.7983247422680413, "grad_norm": 0.5533589124679565, "learning_rate": 4.14201643950732e-06, "loss": 0.5695, "step": 2791 }, { "epoch": 1.7989690721649485, "grad_norm": 0.5693312287330627, "learning_rate": 4.138323370416279e-06, "loss": 0.5568, "step": 2792 }, { "epoch": 1.7996134020618557, "grad_norm": 0.6333770155906677, "learning_rate": 4.134630785738467e-06, "loss": 0.5698, "step": 2793 }, { "epoch": 1.800257731958763, "grad_norm": 0.6461651921272278, "learning_rate": 4.13093868754976e-06, "loss": 0.5846, "step": 2794 }, { "epoch": 1.8009020618556701, "grad_norm": 0.6083168387413025, "learning_rate": 4.127247077925766e-06, "loss": 0.5751, "step": 2795 }, { "epoch": 1.8015463917525774, "grad_norm": 0.6204668283462524, "learning_rate": 4.123555958941817e-06, "loss": 0.5749, "step": 2796 }, { "epoch": 1.8021907216494846, "grad_norm": 0.6975253224372864, "learning_rate": 4.119865332672965e-06, "loss": 0.5732, "step": 2797 }, { "epoch": 1.8028350515463918, "grad_norm": 0.6527647376060486, "learning_rate": 4.116175201193994e-06, "loss": 0.5714, "step": 2798 }, { "epoch": 1.803479381443299, "grad_norm": 0.6494725942611694, "learning_rate": 4.1124855665794006e-06, "loss": 0.5714, "step": 2799 }, { "epoch": 1.8041237113402062, "grad_norm": 0.756386935710907, "learning_rate": 4.108796430903406e-06, "loss": 0.555, "step": 2800 }, { "epoch": 1.8047680412371134, "grad_norm": 0.7319086194038391, "learning_rate": 4.105107796239952e-06, "loss": 0.5537, "step": 2801 }, { "epoch": 1.8054123711340206, "grad_norm": 0.6667985916137695, "learning_rate": 4.101419664662698e-06, "loss": 0.5718, "step": 2802 }, { "epoch": 1.8060567010309279, "grad_norm": 0.7130371928215027, "learning_rate": 4.097732038245017e-06, "loss": 0.5612, "step": 2803 }, { "epoch": 1.806701030927835, "grad_norm": 0.7019771933555603, "learning_rate": 4.094044919060005e-06, "loss": 0.5788, "step": 2804 }, { "epoch": 1.8073453608247423, "grad_norm": 0.6469071507453918, "learning_rate": 4.090358309180465e-06, "loss": 0.5778, "step": 2805 }, { "epoch": 1.8079896907216495, "grad_norm": 0.6443684697151184, "learning_rate": 4.0866722106789214e-06, "loss": 0.5539, "step": 2806 }, { "epoch": 1.8086340206185567, "grad_norm": 0.6952950358390808, "learning_rate": 4.082986625627606e-06, "loss": 0.5797, "step": 2807 }, { "epoch": 1.809278350515464, "grad_norm": 0.6123279929161072, "learning_rate": 4.079301556098462e-06, "loss": 0.5548, "step": 2808 }, { "epoch": 1.8099226804123711, "grad_norm": 0.6153736710548401, "learning_rate": 4.075617004163147e-06, "loss": 0.5801, "step": 2809 }, { "epoch": 1.8105670103092784, "grad_norm": 0.6588917374610901, "learning_rate": 4.071932971893022e-06, "loss": 0.57, "step": 2810 }, { "epoch": 1.8112113402061856, "grad_norm": 0.6365850567817688, "learning_rate": 4.068249461359158e-06, "loss": 0.571, "step": 2811 }, { "epoch": 1.8118556701030928, "grad_norm": 0.5756806135177612, "learning_rate": 4.0645664746323365e-06, "loss": 0.5647, "step": 2812 }, { "epoch": 1.8125, "grad_norm": 0.6428361535072327, "learning_rate": 4.0608840137830395e-06, "loss": 0.5619, "step": 2813 }, { "epoch": 1.8131443298969072, "grad_norm": 0.6551625728607178, "learning_rate": 4.057202080881454e-06, "loss": 0.5625, "step": 2814 }, { "epoch": 1.8137886597938144, "grad_norm": 0.6869320869445801, "learning_rate": 4.0535206779974725e-06, "loss": 0.5447, "step": 2815 }, { "epoch": 1.8144329896907216, "grad_norm": 0.5578991770744324, "learning_rate": 4.049839807200688e-06, "loss": 0.567, "step": 2816 }, { "epoch": 1.8150773195876289, "grad_norm": 0.6826474666595459, "learning_rate": 4.046159470560396e-06, "loss": 0.579, "step": 2817 }, { "epoch": 1.815721649484536, "grad_norm": 0.693675696849823, "learning_rate": 4.042479670145588e-06, "loss": 0.5547, "step": 2818 }, { "epoch": 1.8163659793814433, "grad_norm": 0.6688114404678345, "learning_rate": 4.0388004080249595e-06, "loss": 0.5867, "step": 2819 }, { "epoch": 1.8170103092783505, "grad_norm": 0.5978458523750305, "learning_rate": 4.035121686266899e-06, "loss": 0.5724, "step": 2820 }, { "epoch": 1.8176546391752577, "grad_norm": 0.6737707257270813, "learning_rate": 4.031443506939492e-06, "loss": 0.5884, "step": 2821 }, { "epoch": 1.818298969072165, "grad_norm": 0.6601100564002991, "learning_rate": 4.02776587211052e-06, "loss": 0.5837, "step": 2822 }, { "epoch": 1.8189432989690721, "grad_norm": 0.5907653570175171, "learning_rate": 4.024088783847457e-06, "loss": 0.5699, "step": 2823 }, { "epoch": 1.8195876288659794, "grad_norm": 0.7241925597190857, "learning_rate": 4.020412244217474e-06, "loss": 0.5669, "step": 2824 }, { "epoch": 1.8202319587628866, "grad_norm": 0.6565781831741333, "learning_rate": 4.016736255287429e-06, "loss": 0.5864, "step": 2825 }, { "epoch": 1.8208762886597938, "grad_norm": 0.7306413650512695, "learning_rate": 4.013060819123869e-06, "loss": 0.5784, "step": 2826 }, { "epoch": 1.821520618556701, "grad_norm": 0.6010239720344543, "learning_rate": 4.009385937793037e-06, "loss": 0.5595, "step": 2827 }, { "epoch": 1.8221649484536082, "grad_norm": 0.5784388780593872, "learning_rate": 4.005711613360859e-06, "loss": 0.5601, "step": 2828 }, { "epoch": 1.8228092783505154, "grad_norm": 0.6895382404327393, "learning_rate": 4.002037847892947e-06, "loss": 0.5644, "step": 2829 }, { "epoch": 1.8234536082474226, "grad_norm": 0.6320266723632812, "learning_rate": 3.998364643454604e-06, "loss": 0.578, "step": 2830 }, { "epoch": 1.8240979381443299, "grad_norm": 0.6296818256378174, "learning_rate": 3.994692002110812e-06, "loss": 0.5684, "step": 2831 }, { "epoch": 1.824742268041237, "grad_norm": 0.6949408054351807, "learning_rate": 3.99101992592624e-06, "loss": 0.5534, "step": 2832 }, { "epoch": 1.8253865979381443, "grad_norm": 0.630831241607666, "learning_rate": 3.987348416965237e-06, "loss": 0.5273, "step": 2833 }, { "epoch": 1.8260309278350515, "grad_norm": 0.6404091715812683, "learning_rate": 3.983677477291835e-06, "loss": 0.5811, "step": 2834 }, { "epoch": 1.8266752577319587, "grad_norm": 0.797935962677002, "learning_rate": 3.980007108969746e-06, "loss": 0.5668, "step": 2835 }, { "epoch": 1.827319587628866, "grad_norm": 0.7964347004890442, "learning_rate": 3.976337314062358e-06, "loss": 0.5502, "step": 2836 }, { "epoch": 1.8279639175257731, "grad_norm": 0.5584090948104858, "learning_rate": 3.972668094632741e-06, "loss": 0.557, "step": 2837 }, { "epoch": 1.8286082474226806, "grad_norm": 0.727648138999939, "learning_rate": 3.968999452743638e-06, "loss": 0.5862, "step": 2838 }, { "epoch": 1.8292525773195876, "grad_norm": 0.6226491928100586, "learning_rate": 3.965331390457468e-06, "loss": 0.5644, "step": 2839 }, { "epoch": 1.829896907216495, "grad_norm": 0.5664494037628174, "learning_rate": 3.961663909836326e-06, "loss": 0.5719, "step": 2840 }, { "epoch": 1.830541237113402, "grad_norm": 0.649645209312439, "learning_rate": 3.957997012941976e-06, "loss": 0.5647, "step": 2841 }, { "epoch": 1.8311855670103094, "grad_norm": 0.6424176692962646, "learning_rate": 3.9543307018358604e-06, "loss": 0.5498, "step": 2842 }, { "epoch": 1.8318298969072164, "grad_norm": 0.5963993072509766, "learning_rate": 3.950664978579086e-06, "loss": 0.5688, "step": 2843 }, { "epoch": 1.8324742268041239, "grad_norm": 0.7034690976142883, "learning_rate": 3.946999845232431e-06, "loss": 0.5835, "step": 2844 }, { "epoch": 1.8331185567010309, "grad_norm": 0.6531211137771606, "learning_rate": 3.943335303856343e-06, "loss": 0.5578, "step": 2845 }, { "epoch": 1.8337628865979383, "grad_norm": 0.6883847713470459, "learning_rate": 3.9396713565109375e-06, "loss": 0.5596, "step": 2846 }, { "epoch": 1.8344072164948453, "grad_norm": 0.8375775218009949, "learning_rate": 3.936008005255993e-06, "loss": 0.5744, "step": 2847 }, { "epoch": 1.8350515463917527, "grad_norm": 0.7019558548927307, "learning_rate": 3.932345252150956e-06, "loss": 0.5867, "step": 2848 }, { "epoch": 1.8356958762886597, "grad_norm": 0.7355852127075195, "learning_rate": 3.928683099254936e-06, "loss": 0.5692, "step": 2849 }, { "epoch": 1.8363402061855671, "grad_norm": 0.7643498778343201, "learning_rate": 3.925021548626702e-06, "loss": 0.5883, "step": 2850 }, { "epoch": 1.8369845360824741, "grad_norm": 0.5355536937713623, "learning_rate": 3.921360602324688e-06, "loss": 0.5865, "step": 2851 }, { "epoch": 1.8376288659793816, "grad_norm": 0.7007696628570557, "learning_rate": 3.917700262406988e-06, "loss": 0.5778, "step": 2852 }, { "epoch": 1.8382731958762886, "grad_norm": 0.640404224395752, "learning_rate": 3.9140405309313555e-06, "loss": 0.5656, "step": 2853 }, { "epoch": 1.838917525773196, "grad_norm": 0.5226154923439026, "learning_rate": 3.910381409955197e-06, "loss": 0.573, "step": 2854 }, { "epoch": 1.839561855670103, "grad_norm": 0.6087594628334045, "learning_rate": 3.9067229015355825e-06, "loss": 0.5754, "step": 2855 }, { "epoch": 1.8402061855670104, "grad_norm": 0.5688955783843994, "learning_rate": 3.903065007729234e-06, "loss": 0.5615, "step": 2856 }, { "epoch": 1.8408505154639174, "grad_norm": 0.5595744252204895, "learning_rate": 3.8994077305925285e-06, "loss": 0.5595, "step": 2857 }, { "epoch": 1.8414948453608249, "grad_norm": 0.5974722504615784, "learning_rate": 3.895751072181496e-06, "loss": 0.5823, "step": 2858 }, { "epoch": 1.8421391752577319, "grad_norm": 0.5206627249717712, "learning_rate": 3.892095034551819e-06, "loss": 0.5802, "step": 2859 }, { "epoch": 1.8427835051546393, "grad_norm": 0.5773821473121643, "learning_rate": 3.888439619758833e-06, "loss": 0.5773, "step": 2860 }, { "epoch": 1.8434278350515463, "grad_norm": 0.5465962290763855, "learning_rate": 3.8847848298575205e-06, "loss": 0.5657, "step": 2861 }, { "epoch": 1.8440721649484537, "grad_norm": 0.5602813959121704, "learning_rate": 3.881130666902513e-06, "loss": 0.561, "step": 2862 }, { "epoch": 1.8447164948453607, "grad_norm": 0.5731761455535889, "learning_rate": 3.877477132948092e-06, "loss": 0.5676, "step": 2863 }, { "epoch": 1.8453608247422681, "grad_norm": 0.5247682929039001, "learning_rate": 3.873824230048182e-06, "loss": 0.5872, "step": 2864 }, { "epoch": 1.8460051546391751, "grad_norm": 0.5829671025276184, "learning_rate": 3.870171960256353e-06, "loss": 0.5482, "step": 2865 }, { "epoch": 1.8466494845360826, "grad_norm": 0.5797039866447449, "learning_rate": 3.866520325625825e-06, "loss": 0.5589, "step": 2866 }, { "epoch": 1.8472938144329896, "grad_norm": 0.5680426359176636, "learning_rate": 3.862869328209454e-06, "loss": 0.5707, "step": 2867 }, { "epoch": 1.847938144329897, "grad_norm": 0.5798651576042175, "learning_rate": 3.859218970059739e-06, "loss": 0.5755, "step": 2868 }, { "epoch": 1.848582474226804, "grad_norm": 0.616807222366333, "learning_rate": 3.855569253228821e-06, "loss": 0.5797, "step": 2869 }, { "epoch": 1.8492268041237114, "grad_norm": 0.6231767535209656, "learning_rate": 3.8519201797684825e-06, "loss": 0.565, "step": 2870 }, { "epoch": 1.8498711340206184, "grad_norm": 0.5925441980361938, "learning_rate": 3.8482717517301405e-06, "loss": 0.5677, "step": 2871 }, { "epoch": 1.8505154639175259, "grad_norm": 0.5870957970619202, "learning_rate": 3.844623971164851e-06, "loss": 0.5783, "step": 2872 }, { "epoch": 1.8511597938144329, "grad_norm": 0.6288034319877625, "learning_rate": 3.840976840123304e-06, "loss": 0.5873, "step": 2873 }, { "epoch": 1.8518041237113403, "grad_norm": 0.5655624866485596, "learning_rate": 3.83733036065583e-06, "loss": 0.5826, "step": 2874 }, { "epoch": 1.8524484536082473, "grad_norm": 0.6403019428253174, "learning_rate": 3.833684534812385e-06, "loss": 0.5885, "step": 2875 }, { "epoch": 1.8530927835051547, "grad_norm": 0.5900413990020752, "learning_rate": 3.830039364642566e-06, "loss": 0.5948, "step": 2876 }, { "epoch": 1.8537371134020617, "grad_norm": 0.6103615164756775, "learning_rate": 3.826394852195594e-06, "loss": 0.5785, "step": 2877 }, { "epoch": 1.8543814432989691, "grad_norm": 0.6125506162643433, "learning_rate": 3.822750999520323e-06, "loss": 0.5613, "step": 2878 }, { "epoch": 1.8550257731958761, "grad_norm": 0.6317910552024841, "learning_rate": 3.81910780866524e-06, "loss": 0.57, "step": 2879 }, { "epoch": 1.8556701030927836, "grad_norm": 0.5720580816268921, "learning_rate": 3.815465281678455e-06, "loss": 0.5504, "step": 2880 }, { "epoch": 1.8563144329896906, "grad_norm": 0.6535496711730957, "learning_rate": 3.811823420607706e-06, "loss": 0.5727, "step": 2881 }, { "epoch": 1.856958762886598, "grad_norm": 0.5344809293746948, "learning_rate": 3.808182227500357e-06, "loss": 0.5612, "step": 2882 }, { "epoch": 1.857603092783505, "grad_norm": 0.613595187664032, "learning_rate": 3.8045417044033963e-06, "loss": 0.55, "step": 2883 }, { "epoch": 1.8582474226804124, "grad_norm": 0.5629225373268127, "learning_rate": 3.800901853363436e-06, "loss": 0.5734, "step": 2884 }, { "epoch": 1.8588917525773194, "grad_norm": 0.5034288167953491, "learning_rate": 3.797262676426711e-06, "loss": 0.5836, "step": 2885 }, { "epoch": 1.8595360824742269, "grad_norm": 0.5992762446403503, "learning_rate": 3.7936241756390746e-06, "loss": 0.5681, "step": 2886 }, { "epoch": 1.860180412371134, "grad_norm": 0.6669167280197144, "learning_rate": 3.789986353046001e-06, "loss": 0.5707, "step": 2887 }, { "epoch": 1.8608247422680413, "grad_norm": 0.5786616206169128, "learning_rate": 3.786349210692587e-06, "loss": 0.5657, "step": 2888 }, { "epoch": 1.8614690721649485, "grad_norm": 0.6450387239456177, "learning_rate": 3.7827127506235413e-06, "loss": 0.5724, "step": 2889 }, { "epoch": 1.8621134020618557, "grad_norm": 0.6181351542472839, "learning_rate": 3.779076974883191e-06, "loss": 0.5788, "step": 2890 }, { "epoch": 1.862757731958763, "grad_norm": 0.6475846171379089, "learning_rate": 3.7754418855154823e-06, "loss": 0.5712, "step": 2891 }, { "epoch": 1.8634020618556701, "grad_norm": 0.5473456978797913, "learning_rate": 3.7718074845639694e-06, "loss": 0.5647, "step": 2892 }, { "epoch": 1.8640463917525774, "grad_norm": 0.7396250367164612, "learning_rate": 3.7681737740718226e-06, "loss": 0.5793, "step": 2893 }, { "epoch": 1.8646907216494846, "grad_norm": 0.639931321144104, "learning_rate": 3.764540756081826e-06, "loss": 0.5877, "step": 2894 }, { "epoch": 1.8653350515463918, "grad_norm": 0.7182340621948242, "learning_rate": 3.76090843263637e-06, "loss": 0.5551, "step": 2895 }, { "epoch": 1.865979381443299, "grad_norm": 0.7341027855873108, "learning_rate": 3.7572768057774543e-06, "loss": 0.5641, "step": 2896 }, { "epoch": 1.8666237113402062, "grad_norm": 0.6858738660812378, "learning_rate": 3.753645877546696e-06, "loss": 0.5657, "step": 2897 }, { "epoch": 1.8672680412371134, "grad_norm": 0.8969218730926514, "learning_rate": 3.750015649985308e-06, "loss": 0.5642, "step": 2898 }, { "epoch": 1.8679123711340206, "grad_norm": 0.7517695426940918, "learning_rate": 3.746386125134116e-06, "loss": 0.5611, "step": 2899 }, { "epoch": 1.8685567010309279, "grad_norm": 0.6960027813911438, "learning_rate": 3.7427573050335493e-06, "loss": 0.5733, "step": 2900 }, { "epoch": 1.869201030927835, "grad_norm": 0.7326213121414185, "learning_rate": 3.7391291917236385e-06, "loss": 0.5941, "step": 2901 }, { "epoch": 1.8698453608247423, "grad_norm": 0.6369088292121887, "learning_rate": 3.7355017872440203e-06, "loss": 0.5609, "step": 2902 }, { "epoch": 1.8704896907216495, "grad_norm": 0.6879265308380127, "learning_rate": 3.7318750936339317e-06, "loss": 0.5628, "step": 2903 }, { "epoch": 1.8711340206185567, "grad_norm": 0.6020103693008423, "learning_rate": 3.7282491129322097e-06, "loss": 0.5578, "step": 2904 }, { "epoch": 1.871778350515464, "grad_norm": 0.7473760843276978, "learning_rate": 3.7246238471772887e-06, "loss": 0.558, "step": 2905 }, { "epoch": 1.8724226804123711, "grad_norm": 0.6284377574920654, "learning_rate": 3.7209992984072062e-06, "loss": 0.5583, "step": 2906 }, { "epoch": 1.8730670103092784, "grad_norm": 0.7013309597969055, "learning_rate": 3.717375468659593e-06, "loss": 0.5625, "step": 2907 }, { "epoch": 1.8737113402061856, "grad_norm": 0.7521691918373108, "learning_rate": 3.713752359971674e-06, "loss": 0.5648, "step": 2908 }, { "epoch": 1.8743556701030928, "grad_norm": 0.5844619870185852, "learning_rate": 3.7101299743802745e-06, "loss": 0.5667, "step": 2909 }, { "epoch": 1.875, "grad_norm": 0.6186590194702148, "learning_rate": 3.7065083139218073e-06, "loss": 0.5709, "step": 2910 }, { "epoch": 1.8756443298969072, "grad_norm": 0.6949865818023682, "learning_rate": 3.7028873806322797e-06, "loss": 0.5843, "step": 2911 }, { "epoch": 1.8762886597938144, "grad_norm": 0.5843291282653809, "learning_rate": 3.6992671765472932e-06, "loss": 0.5592, "step": 2912 }, { "epoch": 1.8769329896907216, "grad_norm": 0.622700035572052, "learning_rate": 3.695647703702035e-06, "loss": 0.5417, "step": 2913 }, { "epoch": 1.8775773195876289, "grad_norm": 0.6248645186424255, "learning_rate": 3.6920289641312812e-06, "loss": 0.5502, "step": 2914 }, { "epoch": 1.878221649484536, "grad_norm": 0.6045897006988525, "learning_rate": 3.688410959869401e-06, "loss": 0.5753, "step": 2915 }, { "epoch": 1.8788659793814433, "grad_norm": 0.6737159490585327, "learning_rate": 3.6847936929503446e-06, "loss": 0.5696, "step": 2916 }, { "epoch": 1.8795103092783505, "grad_norm": 0.6536122560501099, "learning_rate": 3.6811771654076508e-06, "loss": 0.5644, "step": 2917 }, { "epoch": 1.8801546391752577, "grad_norm": 0.5907912254333496, "learning_rate": 3.677561379274441e-06, "loss": 0.5803, "step": 2918 }, { "epoch": 1.880798969072165, "grad_norm": 0.6342718005180359, "learning_rate": 3.6739463365834194e-06, "loss": 0.5584, "step": 2919 }, { "epoch": 1.8814432989690721, "grad_norm": 0.5782403945922852, "learning_rate": 3.6703320393668763e-06, "loss": 0.5796, "step": 2920 }, { "epoch": 1.8820876288659794, "grad_norm": 0.5586979985237122, "learning_rate": 3.666718489656677e-06, "loss": 0.5524, "step": 2921 }, { "epoch": 1.8827319587628866, "grad_norm": 0.6115497350692749, "learning_rate": 3.663105689484271e-06, "loss": 0.5691, "step": 2922 }, { "epoch": 1.8833762886597938, "grad_norm": 0.6043757200241089, "learning_rate": 3.6594936408806835e-06, "loss": 0.5691, "step": 2923 }, { "epoch": 1.884020618556701, "grad_norm": 0.6012067794799805, "learning_rate": 3.6558823458765214e-06, "loss": 0.5887, "step": 2924 }, { "epoch": 1.8846649484536082, "grad_norm": 0.6255972385406494, "learning_rate": 3.6522718065019625e-06, "loss": 0.5657, "step": 2925 }, { "epoch": 1.8853092783505154, "grad_norm": 0.5661329627037048, "learning_rate": 3.6486620247867625e-06, "loss": 0.5482, "step": 2926 }, { "epoch": 1.8859536082474226, "grad_norm": 0.604391872882843, "learning_rate": 3.6450530027602536e-06, "loss": 0.5668, "step": 2927 }, { "epoch": 1.8865979381443299, "grad_norm": 0.6430517435073853, "learning_rate": 3.6414447424513367e-06, "loss": 0.5717, "step": 2928 }, { "epoch": 1.887242268041237, "grad_norm": 0.5622802972793579, "learning_rate": 3.6378372458884847e-06, "loss": 0.5659, "step": 2929 }, { "epoch": 1.8878865979381443, "grad_norm": 0.5962522029876709, "learning_rate": 3.6342305150997463e-06, "loss": 0.5377, "step": 2930 }, { "epoch": 1.8885309278350515, "grad_norm": 0.7491913437843323, "learning_rate": 3.6306245521127343e-06, "loss": 0.5594, "step": 2931 }, { "epoch": 1.8891752577319587, "grad_norm": 0.6561474204063416, "learning_rate": 3.627019358954629e-06, "loss": 0.5639, "step": 2932 }, { "epoch": 1.889819587628866, "grad_norm": 0.6665676832199097, "learning_rate": 3.6234149376521855e-06, "loss": 0.5703, "step": 2933 }, { "epoch": 1.8904639175257731, "grad_norm": 0.5928755402565002, "learning_rate": 3.6198112902317166e-06, "loss": 0.5577, "step": 2934 }, { "epoch": 1.8911082474226806, "grad_norm": 0.6737682819366455, "learning_rate": 3.616208418719106e-06, "loss": 0.5634, "step": 2935 }, { "epoch": 1.8917525773195876, "grad_norm": 0.6588972210884094, "learning_rate": 3.6126063251397968e-06, "loss": 0.572, "step": 2936 }, { "epoch": 1.892396907216495, "grad_norm": 0.5469744205474854, "learning_rate": 3.609005011518796e-06, "loss": 0.5564, "step": 2937 }, { "epoch": 1.893041237113402, "grad_norm": 0.5985351204872131, "learning_rate": 3.6054044798806755e-06, "loss": 0.5727, "step": 2938 }, { "epoch": 1.8936855670103094, "grad_norm": 0.6712562441825867, "learning_rate": 3.6018047322495624e-06, "loss": 0.5735, "step": 2939 }, { "epoch": 1.8943298969072164, "grad_norm": 0.5498131513595581, "learning_rate": 3.598205770649147e-06, "loss": 0.5594, "step": 2940 }, { "epoch": 1.8949742268041239, "grad_norm": 0.6197175979614258, "learning_rate": 3.5946075971026746e-06, "loss": 0.5559, "step": 2941 }, { "epoch": 1.8956185567010309, "grad_norm": 0.6255816221237183, "learning_rate": 3.591010213632951e-06, "loss": 0.5784, "step": 2942 }, { "epoch": 1.8962628865979383, "grad_norm": 0.6416733264923096, "learning_rate": 3.5874136222623357e-06, "loss": 0.579, "step": 2943 }, { "epoch": 1.8969072164948453, "grad_norm": 0.5862115025520325, "learning_rate": 3.5838178250127407e-06, "loss": 0.5877, "step": 2944 }, { "epoch": 1.8975515463917527, "grad_norm": 0.5828849077224731, "learning_rate": 3.5802228239056374e-06, "loss": 0.5827, "step": 2945 }, { "epoch": 1.8981958762886597, "grad_norm": 0.6208138465881348, "learning_rate": 3.576628620962045e-06, "loss": 0.5622, "step": 2946 }, { "epoch": 1.8988402061855671, "grad_norm": 0.6862756609916687, "learning_rate": 3.5730352182025336e-06, "loss": 0.5664, "step": 2947 }, { "epoch": 1.8994845360824741, "grad_norm": 0.6346303820610046, "learning_rate": 3.5694426176472276e-06, "loss": 0.5755, "step": 2948 }, { "epoch": 1.9001288659793816, "grad_norm": 0.6011556386947632, "learning_rate": 3.5658508213157966e-06, "loss": 0.5589, "step": 2949 }, { "epoch": 1.9007731958762886, "grad_norm": 0.5938005447387695, "learning_rate": 3.5622598312274583e-06, "loss": 0.5442, "step": 2950 }, { "epoch": 1.901417525773196, "grad_norm": 0.6295154690742493, "learning_rate": 3.558669649400981e-06, "loss": 0.5514, "step": 2951 }, { "epoch": 1.902061855670103, "grad_norm": 0.6183854341506958, "learning_rate": 3.555080277854673e-06, "loss": 0.5727, "step": 2952 }, { "epoch": 1.9027061855670104, "grad_norm": 0.5930750966072083, "learning_rate": 3.5514917186063926e-06, "loss": 0.5724, "step": 2953 }, { "epoch": 1.9033505154639174, "grad_norm": 0.6359767317771912, "learning_rate": 3.5479039736735376e-06, "loss": 0.5651, "step": 2954 }, { "epoch": 1.9039948453608249, "grad_norm": 0.630477249622345, "learning_rate": 3.544317045073049e-06, "loss": 0.5781, "step": 2955 }, { "epoch": 1.9046391752577319, "grad_norm": 0.6364520192146301, "learning_rate": 3.5407309348214094e-06, "loss": 0.5612, "step": 2956 }, { "epoch": 1.9052835051546393, "grad_norm": 0.6050917506217957, "learning_rate": 3.5371456449346407e-06, "loss": 0.576, "step": 2957 }, { "epoch": 1.9059278350515463, "grad_norm": 0.6281921863555908, "learning_rate": 3.533561177428305e-06, "loss": 0.5737, "step": 2958 }, { "epoch": 1.9065721649484537, "grad_norm": 0.5901297330856323, "learning_rate": 3.529977534317499e-06, "loss": 0.5641, "step": 2959 }, { "epoch": 1.9072164948453607, "grad_norm": 0.5906472206115723, "learning_rate": 3.5263947176168602e-06, "loss": 0.561, "step": 2960 }, { "epoch": 1.9078608247422681, "grad_norm": 0.6476042866706848, "learning_rate": 3.522812729340557e-06, "loss": 0.5845, "step": 2961 }, { "epoch": 1.9085051546391751, "grad_norm": 0.7147461771965027, "learning_rate": 3.5192315715022956e-06, "loss": 0.5889, "step": 2962 }, { "epoch": 1.9091494845360826, "grad_norm": 0.7037899494171143, "learning_rate": 3.5156512461153147e-06, "loss": 0.5816, "step": 2963 }, { "epoch": 1.9097938144329896, "grad_norm": 0.6095551252365112, "learning_rate": 3.5120717551923844e-06, "loss": 0.5644, "step": 2964 }, { "epoch": 1.910438144329897, "grad_norm": 0.7169997096061707, "learning_rate": 3.5084931007458035e-06, "loss": 0.5615, "step": 2965 }, { "epoch": 1.911082474226804, "grad_norm": 0.5706173181533813, "learning_rate": 3.5049152847874053e-06, "loss": 0.5919, "step": 2966 }, { "epoch": 1.9117268041237114, "grad_norm": 0.6190486550331116, "learning_rate": 3.501338309328548e-06, "loss": 0.573, "step": 2967 }, { "epoch": 1.9123711340206184, "grad_norm": 0.6370425820350647, "learning_rate": 3.4977621763801173e-06, "loss": 0.5646, "step": 2968 }, { "epoch": 1.9130154639175259, "grad_norm": 0.5744853019714355, "learning_rate": 3.49418688795253e-06, "loss": 0.5656, "step": 2969 }, { "epoch": 1.9136597938144329, "grad_norm": 0.6448116898536682, "learning_rate": 3.49061244605572e-06, "loss": 0.5691, "step": 2970 }, { "epoch": 1.9143041237113403, "grad_norm": 0.5669074058532715, "learning_rate": 3.487038852699155e-06, "loss": 0.5595, "step": 2971 }, { "epoch": 1.9149484536082473, "grad_norm": 0.6046828031539917, "learning_rate": 3.4834661098918164e-06, "loss": 0.5573, "step": 2972 }, { "epoch": 1.9155927835051547, "grad_norm": 0.7353982329368591, "learning_rate": 3.479894219642213e-06, "loss": 0.5448, "step": 2973 }, { "epoch": 1.9162371134020617, "grad_norm": 0.5973274111747742, "learning_rate": 3.4763231839583733e-06, "loss": 0.5508, "step": 2974 }, { "epoch": 1.9168814432989691, "grad_norm": 0.6752418279647827, "learning_rate": 3.4727530048478455e-06, "loss": 0.5817, "step": 2975 }, { "epoch": 1.9175257731958761, "grad_norm": 0.5370710492134094, "learning_rate": 3.469183684317694e-06, "loss": 0.5694, "step": 2976 }, { "epoch": 1.9181701030927836, "grad_norm": 0.7213362455368042, "learning_rate": 3.465615224374502e-06, "loss": 0.5614, "step": 2977 }, { "epoch": 1.9188144329896906, "grad_norm": 0.5863215327262878, "learning_rate": 3.462047627024372e-06, "loss": 0.5689, "step": 2978 }, { "epoch": 1.919458762886598, "grad_norm": 0.6264986395835876, "learning_rate": 3.4584808942729165e-06, "loss": 0.573, "step": 2979 }, { "epoch": 1.920103092783505, "grad_norm": 0.6613036394119263, "learning_rate": 3.4549150281252635e-06, "loss": 0.5783, "step": 2980 }, { "epoch": 1.9207474226804124, "grad_norm": 0.7003908753395081, "learning_rate": 3.451350030586057e-06, "loss": 0.5739, "step": 2981 }, { "epoch": 1.9213917525773194, "grad_norm": 0.6633438467979431, "learning_rate": 3.4477859036594487e-06, "loss": 0.5609, "step": 2982 }, { "epoch": 1.9220360824742269, "grad_norm": 0.6075451374053955, "learning_rate": 3.4442226493491015e-06, "loss": 0.5623, "step": 2983 }, { "epoch": 1.922680412371134, "grad_norm": 0.5333400368690491, "learning_rate": 3.4406602696581905e-06, "loss": 0.5756, "step": 2984 }, { "epoch": 1.9233247422680413, "grad_norm": 0.5642880797386169, "learning_rate": 3.437098766589396e-06, "loss": 0.572, "step": 2985 }, { "epoch": 1.9239690721649485, "grad_norm": 0.5168780088424683, "learning_rate": 3.4335381421449056e-06, "loss": 0.5673, "step": 2986 }, { "epoch": 1.9246134020618557, "grad_norm": 0.6351976990699768, "learning_rate": 3.429978398326417e-06, "loss": 0.5625, "step": 2987 }, { "epoch": 1.925257731958763, "grad_norm": 0.560809314250946, "learning_rate": 3.426419537135127e-06, "loss": 0.5591, "step": 2988 }, { "epoch": 1.9259020618556701, "grad_norm": 0.6231833696365356, "learning_rate": 3.422861560571742e-06, "loss": 0.594, "step": 2989 }, { "epoch": 1.9265463917525774, "grad_norm": 0.5107647776603699, "learning_rate": 3.4193044706364664e-06, "loss": 0.564, "step": 2990 }, { "epoch": 1.9271907216494846, "grad_norm": 0.6139330267906189, "learning_rate": 3.415748269329007e-06, "loss": 0.5529, "step": 2991 }, { "epoch": 1.9278350515463918, "grad_norm": 0.5991519093513489, "learning_rate": 3.4121929586485738e-06, "loss": 0.5767, "step": 2992 }, { "epoch": 1.928479381443299, "grad_norm": 0.5618698596954346, "learning_rate": 3.4086385405938738e-06, "loss": 0.58, "step": 2993 }, { "epoch": 1.9291237113402062, "grad_norm": 0.5879883766174316, "learning_rate": 3.4050850171631123e-06, "loss": 0.5596, "step": 2994 }, { "epoch": 1.9297680412371134, "grad_norm": 0.6425175666809082, "learning_rate": 3.401532390353991e-06, "loss": 0.5706, "step": 2995 }, { "epoch": 1.9304123711340206, "grad_norm": 0.5969499349594116, "learning_rate": 3.39798066216371e-06, "loss": 0.558, "step": 2996 }, { "epoch": 1.9310567010309279, "grad_norm": 0.6213477849960327, "learning_rate": 3.3944298345889637e-06, "loss": 0.5675, "step": 2997 }, { "epoch": 1.931701030927835, "grad_norm": 0.6278191804885864, "learning_rate": 3.390879909625936e-06, "loss": 0.5732, "step": 2998 }, { "epoch": 1.9323453608247423, "grad_norm": 0.5371977090835571, "learning_rate": 3.3873308892703105e-06, "loss": 0.5658, "step": 2999 }, { "epoch": 1.9329896907216495, "grad_norm": 0.6341403126716614, "learning_rate": 3.383782775517256e-06, "loss": 0.5481, "step": 3000 }, { "epoch": 1.9336340206185567, "grad_norm": 0.7055769562721252, "learning_rate": 3.3802355703614335e-06, "loss": 0.5846, "step": 3001 }, { "epoch": 1.934278350515464, "grad_norm": 0.6069998741149902, "learning_rate": 3.376689275796996e-06, "loss": 0.5692, "step": 3002 }, { "epoch": 1.9349226804123711, "grad_norm": 0.7225788235664368, "learning_rate": 3.373143893817582e-06, "loss": 0.5593, "step": 3003 }, { "epoch": 1.9355670103092784, "grad_norm": 0.6093934178352356, "learning_rate": 3.3695994264163134e-06, "loss": 0.5528, "step": 3004 }, { "epoch": 1.9362113402061856, "grad_norm": 0.6550713777542114, "learning_rate": 3.3660558755858075e-06, "loss": 0.5509, "step": 3005 }, { "epoch": 1.9368556701030928, "grad_norm": 0.6599929928779602, "learning_rate": 3.3625132433181573e-06, "loss": 0.5609, "step": 3006 }, { "epoch": 1.9375, "grad_norm": 0.5893612504005432, "learning_rate": 3.3589715316049443e-06, "loss": 0.569, "step": 3007 }, { "epoch": 1.9381443298969072, "grad_norm": 0.6570191383361816, "learning_rate": 3.3554307424372312e-06, "loss": 0.5772, "step": 3008 }, { "epoch": 1.9387886597938144, "grad_norm": 0.5857716798782349, "learning_rate": 3.351890877805559e-06, "loss": 0.5607, "step": 3009 }, { "epoch": 1.9394329896907216, "grad_norm": 0.607175350189209, "learning_rate": 3.348351939699955e-06, "loss": 0.5442, "step": 3010 }, { "epoch": 1.9400773195876289, "grad_norm": 0.5684579610824585, "learning_rate": 3.3448139301099212e-06, "loss": 0.5699, "step": 3011 }, { "epoch": 1.940721649484536, "grad_norm": 0.6330509185791016, "learning_rate": 3.3412768510244387e-06, "loss": 0.5602, "step": 3012 }, { "epoch": 1.9413659793814433, "grad_norm": 0.6467970013618469, "learning_rate": 3.3377407044319645e-06, "loss": 0.5646, "step": 3013 }, { "epoch": 1.9420103092783505, "grad_norm": 0.5769273638725281, "learning_rate": 3.334205492320435e-06, "loss": 0.5676, "step": 3014 }, { "epoch": 1.9426546391752577, "grad_norm": 0.6083638072013855, "learning_rate": 3.3306712166772574e-06, "loss": 0.5712, "step": 3015 }, { "epoch": 1.943298969072165, "grad_norm": 0.5437398552894592, "learning_rate": 3.327137879489313e-06, "loss": 0.5726, "step": 3016 }, { "epoch": 1.9439432989690721, "grad_norm": 0.5831213593482971, "learning_rate": 3.323605482742959e-06, "loss": 0.56, "step": 3017 }, { "epoch": 1.9445876288659794, "grad_norm": 0.5771275758743286, "learning_rate": 3.320074028424018e-06, "loss": 0.562, "step": 3018 }, { "epoch": 1.9452319587628866, "grad_norm": 0.6104055047035217, "learning_rate": 3.316543518517787e-06, "loss": 0.5604, "step": 3019 }, { "epoch": 1.9458762886597938, "grad_norm": 0.631687343120575, "learning_rate": 3.3130139550090325e-06, "loss": 0.557, "step": 3020 }, { "epoch": 1.946520618556701, "grad_norm": 0.559636116027832, "learning_rate": 3.3094853398819863e-06, "loss": 0.5761, "step": 3021 }, { "epoch": 1.9471649484536082, "grad_norm": 0.7011544108390808, "learning_rate": 3.305957675120347e-06, "loss": 0.562, "step": 3022 }, { "epoch": 1.9478092783505154, "grad_norm": 0.5833136439323425, "learning_rate": 3.3024309627072837e-06, "loss": 0.5768, "step": 3023 }, { "epoch": 1.9484536082474226, "grad_norm": 0.5448467135429382, "learning_rate": 3.298905204625424e-06, "loss": 0.5692, "step": 3024 }, { "epoch": 1.9490979381443299, "grad_norm": 0.6346337199211121, "learning_rate": 3.2953804028568633e-06, "loss": 0.571, "step": 3025 }, { "epoch": 1.949742268041237, "grad_norm": 0.6229522824287415, "learning_rate": 3.2918565593831565e-06, "loss": 0.5695, "step": 3026 }, { "epoch": 1.9503865979381443, "grad_norm": 0.5937408208847046, "learning_rate": 3.2883336761853203e-06, "loss": 0.562, "step": 3027 }, { "epoch": 1.9510309278350515, "grad_norm": 0.597442090511322, "learning_rate": 3.284811755243834e-06, "loss": 0.559, "step": 3028 }, { "epoch": 1.9516752577319587, "grad_norm": 0.5803118944168091, "learning_rate": 3.2812907985386343e-06, "loss": 0.5688, "step": 3029 }, { "epoch": 1.952319587628866, "grad_norm": 0.6128878593444824, "learning_rate": 3.2777708080491133e-06, "loss": 0.5716, "step": 3030 }, { "epoch": 1.9529639175257731, "grad_norm": 0.6236315965652466, "learning_rate": 3.2742517857541258e-06, "loss": 0.569, "step": 3031 }, { "epoch": 1.9536082474226806, "grad_norm": 0.5778292417526245, "learning_rate": 3.2707337336319757e-06, "loss": 0.5594, "step": 3032 }, { "epoch": 1.9542525773195876, "grad_norm": 0.576391339302063, "learning_rate": 3.2672166536604265e-06, "loss": 0.5642, "step": 3033 }, { "epoch": 1.954896907216495, "grad_norm": 0.5682873725891113, "learning_rate": 3.2637005478166915e-06, "loss": 0.5732, "step": 3034 }, { "epoch": 1.955541237113402, "grad_norm": 0.5709570646286011, "learning_rate": 3.260185418077441e-06, "loss": 0.5639, "step": 3035 }, { "epoch": 1.9561855670103094, "grad_norm": 0.6205390095710754, "learning_rate": 3.2566712664187907e-06, "loss": 0.5762, "step": 3036 }, { "epoch": 1.9568298969072164, "grad_norm": 0.641661524772644, "learning_rate": 3.253158094816311e-06, "loss": 0.5771, "step": 3037 }, { "epoch": 1.9574742268041239, "grad_norm": 0.5997269153594971, "learning_rate": 3.24964590524502e-06, "loss": 0.5583, "step": 3038 }, { "epoch": 1.9581185567010309, "grad_norm": 0.5892099142074585, "learning_rate": 3.2461346996793843e-06, "loss": 0.5567, "step": 3039 }, { "epoch": 1.9587628865979383, "grad_norm": 0.6477048397064209, "learning_rate": 3.2426244800933133e-06, "loss": 0.568, "step": 3040 }, { "epoch": 1.9594072164948453, "grad_norm": 0.557009756565094, "learning_rate": 3.2391152484601696e-06, "loss": 0.5505, "step": 3041 }, { "epoch": 1.9600515463917527, "grad_norm": 0.5889344215393066, "learning_rate": 3.2356070067527533e-06, "loss": 0.5851, "step": 3042 }, { "epoch": 1.9606958762886597, "grad_norm": 0.5734573006629944, "learning_rate": 3.2320997569433125e-06, "loss": 0.5691, "step": 3043 }, { "epoch": 1.9613402061855671, "grad_norm": 0.5899697542190552, "learning_rate": 3.228593501003536e-06, "loss": 0.5674, "step": 3044 }, { "epoch": 1.9619845360824741, "grad_norm": 0.5743888020515442, "learning_rate": 3.2250882409045515e-06, "loss": 0.5624, "step": 3045 }, { "epoch": 1.9626288659793816, "grad_norm": 0.5512078404426575, "learning_rate": 3.221583978616932e-06, "loss": 0.575, "step": 3046 }, { "epoch": 1.9632731958762886, "grad_norm": 0.5667924284934998, "learning_rate": 3.2180807161106865e-06, "loss": 0.5612, "step": 3047 }, { "epoch": 1.963917525773196, "grad_norm": 0.6319229602813721, "learning_rate": 3.2145784553552604e-06, "loss": 0.5703, "step": 3048 }, { "epoch": 1.964561855670103, "grad_norm": 0.5728392004966736, "learning_rate": 3.2110771983195404e-06, "loss": 0.5591, "step": 3049 }, { "epoch": 1.9652061855670104, "grad_norm": 0.6335326433181763, "learning_rate": 3.207576946971845e-06, "loss": 0.5853, "step": 3050 }, { "epoch": 1.9658505154639174, "grad_norm": 0.5846401453018188, "learning_rate": 3.2040777032799288e-06, "loss": 0.5734, "step": 3051 }, { "epoch": 1.9664948453608249, "grad_norm": 0.6446945667266846, "learning_rate": 3.200579469210978e-06, "loss": 0.5664, "step": 3052 }, { "epoch": 1.9671391752577319, "grad_norm": 0.6255974173545837, "learning_rate": 3.1970822467316167e-06, "loss": 0.5995, "step": 3053 }, { "epoch": 1.9677835051546393, "grad_norm": 0.5689506530761719, "learning_rate": 3.1935860378078935e-06, "loss": 0.5547, "step": 3054 }, { "epoch": 1.9684278350515463, "grad_norm": 0.6446697115898132, "learning_rate": 3.1900908444052913e-06, "loss": 0.5762, "step": 3055 }, { "epoch": 1.9690721649484537, "grad_norm": 0.6906946301460266, "learning_rate": 3.1865966684887222e-06, "loss": 0.5641, "step": 3056 }, { "epoch": 1.9697164948453607, "grad_norm": 0.5808166265487671, "learning_rate": 3.1831035120225244e-06, "loss": 0.5674, "step": 3057 }, { "epoch": 1.9703608247422681, "grad_norm": 0.6548766493797302, "learning_rate": 3.179611376970462e-06, "loss": 0.5748, "step": 3058 }, { "epoch": 1.9710051546391751, "grad_norm": 0.6507375836372375, "learning_rate": 3.1761202652957302e-06, "loss": 0.5554, "step": 3059 }, { "epoch": 1.9716494845360826, "grad_norm": 0.5633305907249451, "learning_rate": 3.1726301789609427e-06, "loss": 0.5948, "step": 3060 }, { "epoch": 1.9722938144329896, "grad_norm": 0.7386074662208557, "learning_rate": 3.169141119928142e-06, "loss": 0.5816, "step": 3061 }, { "epoch": 1.972938144329897, "grad_norm": 0.6929582953453064, "learning_rate": 3.1656530901587885e-06, "loss": 0.5785, "step": 3062 }, { "epoch": 1.973582474226804, "grad_norm": 0.5781306624412537, "learning_rate": 3.1621660916137655e-06, "loss": 0.5764, "step": 3063 }, { "epoch": 1.9742268041237114, "grad_norm": 0.7445975542068481, "learning_rate": 3.158680126253379e-06, "loss": 0.5672, "step": 3064 }, { "epoch": 1.9748711340206184, "grad_norm": 0.6940599083900452, "learning_rate": 3.155195196037352e-06, "loss": 0.5703, "step": 3065 }, { "epoch": 1.9755154639175259, "grad_norm": 0.6025962829589844, "learning_rate": 3.1517113029248236e-06, "loss": 0.5633, "step": 3066 }, { "epoch": 1.9761597938144329, "grad_norm": 0.7622149586677551, "learning_rate": 3.148228448874354e-06, "loss": 0.5751, "step": 3067 }, { "epoch": 1.9768041237113403, "grad_norm": 0.6574704647064209, "learning_rate": 3.1447466358439184e-06, "loss": 0.5804, "step": 3068 }, { "epoch": 1.9774484536082473, "grad_norm": 0.646478533744812, "learning_rate": 3.1412658657909035e-06, "loss": 0.575, "step": 3069 }, { "epoch": 1.9780927835051547, "grad_norm": 0.6944490075111389, "learning_rate": 3.137786140672112e-06, "loss": 0.5581, "step": 3070 }, { "epoch": 1.9787371134020617, "grad_norm": 0.5357581973075867, "learning_rate": 3.134307462443761e-06, "loss": 0.5721, "step": 3071 }, { "epoch": 1.9793814432989691, "grad_norm": 0.6844817996025085, "learning_rate": 3.130829833061475e-06, "loss": 0.5733, "step": 3072 }, { "epoch": 1.9800257731958761, "grad_norm": 0.6380484104156494, "learning_rate": 3.1273532544802914e-06, "loss": 0.5561, "step": 3073 }, { "epoch": 1.9806701030927836, "grad_norm": 0.681602954864502, "learning_rate": 3.1238777286546577e-06, "loss": 0.5954, "step": 3074 }, { "epoch": 1.9813144329896906, "grad_norm": 0.6421112418174744, "learning_rate": 3.120403257538428e-06, "loss": 0.5678, "step": 3075 }, { "epoch": 1.981958762886598, "grad_norm": 0.5675474405288696, "learning_rate": 3.1169298430848605e-06, "loss": 0.5682, "step": 3076 }, { "epoch": 1.982603092783505, "grad_norm": 0.7062885165214539, "learning_rate": 3.1134574872466284e-06, "loss": 0.5606, "step": 3077 }, { "epoch": 1.9832474226804124, "grad_norm": 0.6121830940246582, "learning_rate": 3.109986191975799e-06, "loss": 0.5658, "step": 3078 }, { "epoch": 1.9838917525773194, "grad_norm": 0.5681074261665344, "learning_rate": 3.1065159592238524e-06, "loss": 0.5718, "step": 3079 }, { "epoch": 1.9845360824742269, "grad_norm": 0.5459615588188171, "learning_rate": 3.103046790941665e-06, "loss": 0.5705, "step": 3080 }, { "epoch": 1.985180412371134, "grad_norm": 0.6053488850593567, "learning_rate": 3.0995786890795165e-06, "loss": 0.5727, "step": 3081 }, { "epoch": 1.9858247422680413, "grad_norm": 0.5228463411331177, "learning_rate": 3.0961116555870895e-06, "loss": 0.598, "step": 3082 }, { "epoch": 1.9864690721649485, "grad_norm": 0.5424851775169373, "learning_rate": 3.092645692413464e-06, "loss": 0.579, "step": 3083 }, { "epoch": 1.9871134020618557, "grad_norm": 0.6140269637107849, "learning_rate": 3.0891808015071166e-06, "loss": 0.5759, "step": 3084 }, { "epoch": 1.987757731958763, "grad_norm": 0.5462747812271118, "learning_rate": 3.0857169848159262e-06, "loss": 0.5541, "step": 3085 }, { "epoch": 1.9884020618556701, "grad_norm": 0.5603780150413513, "learning_rate": 3.082254244287163e-06, "loss": 0.5623, "step": 3086 }, { "epoch": 1.9890463917525774, "grad_norm": 0.5834332704544067, "learning_rate": 3.0787925818674928e-06, "loss": 0.5972, "step": 3087 }, { "epoch": 1.9896907216494846, "grad_norm": 0.5461161732673645, "learning_rate": 3.0753319995029772e-06, "loss": 0.5541, "step": 3088 }, { "epoch": 1.9903350515463918, "grad_norm": 0.5278072357177734, "learning_rate": 3.07187249913907e-06, "loss": 0.5742, "step": 3089 }, { "epoch": 1.990979381443299, "grad_norm": 0.5495733618736267, "learning_rate": 3.0684140827206164e-06, "loss": 0.5596, "step": 3090 }, { "epoch": 1.9916237113402062, "grad_norm": 0.673107385635376, "learning_rate": 3.0649567521918522e-06, "loss": 0.5729, "step": 3091 }, { "epoch": 1.9922680412371134, "grad_norm": 0.5311851501464844, "learning_rate": 3.0615005094964035e-06, "loss": 0.545, "step": 3092 }, { "epoch": 1.9929123711340206, "grad_norm": 0.5295006632804871, "learning_rate": 3.0580453565772834e-06, "loss": 0.5543, "step": 3093 }, { "epoch": 1.9935567010309279, "grad_norm": 0.6038599014282227, "learning_rate": 3.054591295376891e-06, "loss": 0.5774, "step": 3094 }, { "epoch": 1.994201030927835, "grad_norm": 0.6591808199882507, "learning_rate": 3.0511383278370186e-06, "loss": 0.5584, "step": 3095 }, { "epoch": 1.9948453608247423, "grad_norm": 0.5865058898925781, "learning_rate": 3.0476864558988364e-06, "loss": 0.5723, "step": 3096 }, { "epoch": 1.9954896907216495, "grad_norm": 0.5835967063903809, "learning_rate": 3.044235681502901e-06, "loss": 0.5572, "step": 3097 }, { "epoch": 1.9961340206185567, "grad_norm": 0.5974211096763611, "learning_rate": 3.040786006589153e-06, "loss": 0.5654, "step": 3098 }, { "epoch": 1.996778350515464, "grad_norm": 0.5425552129745483, "learning_rate": 3.037337433096913e-06, "loss": 0.5694, "step": 3099 }, { "epoch": 1.9974226804123711, "grad_norm": 0.517967939376831, "learning_rate": 3.0338899629648845e-06, "loss": 0.5851, "step": 3100 }, { "epoch": 1.9980670103092784, "grad_norm": 0.567091166973114, "learning_rate": 3.0304435981311484e-06, "loss": 0.5826, "step": 3101 }, { "epoch": 1.9987113402061856, "grad_norm": 0.5537819862365723, "learning_rate": 3.026998340533167e-06, "loss": 0.5745, "step": 3102 }, { "epoch": 1.9993556701030928, "grad_norm": 0.6055207252502441, "learning_rate": 3.023554192107778e-06, "loss": 0.5846, "step": 3103 }, { "epoch": 2.0, "grad_norm": 0.5422649383544922, "learning_rate": 3.0201111547911967e-06, "loss": 0.5589, "step": 3104 }, { "epoch": 2.0006443298969074, "grad_norm": 0.6168543696403503, "learning_rate": 3.0166692305190125e-06, "loss": 0.5172, "step": 3105 }, { "epoch": 2.0012886597938144, "grad_norm": 0.6243107318878174, "learning_rate": 3.0132284212261886e-06, "loss": 0.5117, "step": 3106 }, { "epoch": 2.001932989690722, "grad_norm": 0.5921964645385742, "learning_rate": 3.009788728847066e-06, "loss": 0.5166, "step": 3107 }, { "epoch": 2.002577319587629, "grad_norm": 0.6694928407669067, "learning_rate": 3.0063501553153523e-06, "loss": 0.4988, "step": 3108 }, { "epoch": 2.0032216494845363, "grad_norm": 0.6406036019325256, "learning_rate": 3.0029127025641275e-06, "loss": 0.5231, "step": 3109 }, { "epoch": 2.0038659793814433, "grad_norm": 0.6457499861717224, "learning_rate": 2.9994763725258445e-06, "loss": 0.4946, "step": 3110 }, { "epoch": 2.0045103092783507, "grad_norm": 0.7345821261405945, "learning_rate": 2.9960411671323215e-06, "loss": 0.504, "step": 3111 }, { "epoch": 2.0051546391752577, "grad_norm": 0.6717120409011841, "learning_rate": 2.992607088314744e-06, "loss": 0.5068, "step": 3112 }, { "epoch": 2.005798969072165, "grad_norm": 0.659115731716156, "learning_rate": 2.98917413800367e-06, "loss": 0.5034, "step": 3113 }, { "epoch": 2.006443298969072, "grad_norm": 0.8044353127479553, "learning_rate": 2.985742318129015e-06, "loss": 0.4891, "step": 3114 }, { "epoch": 2.0070876288659796, "grad_norm": 0.637010395526886, "learning_rate": 2.982311630620066e-06, "loss": 0.4991, "step": 3115 }, { "epoch": 2.0077319587628866, "grad_norm": 0.6876194477081299, "learning_rate": 2.97888207740547e-06, "loss": 0.4902, "step": 3116 }, { "epoch": 2.008376288659794, "grad_norm": 0.6199772953987122, "learning_rate": 2.9754536604132345e-06, "loss": 0.4994, "step": 3117 }, { "epoch": 2.009020618556701, "grad_norm": 0.6285339593887329, "learning_rate": 2.972026381570733e-06, "loss": 0.5019, "step": 3118 }, { "epoch": 2.0096649484536084, "grad_norm": 0.6567961573600769, "learning_rate": 2.9686002428046958e-06, "loss": 0.5194, "step": 3119 }, { "epoch": 2.0103092783505154, "grad_norm": 0.5958118438720703, "learning_rate": 2.965175246041212e-06, "loss": 0.4929, "step": 3120 }, { "epoch": 2.010953608247423, "grad_norm": 0.5741519927978516, "learning_rate": 2.961751393205732e-06, "loss": 0.4823, "step": 3121 }, { "epoch": 2.01159793814433, "grad_norm": 0.6423365473747253, "learning_rate": 2.9583286862230587e-06, "loss": 0.503, "step": 3122 }, { "epoch": 2.0122422680412373, "grad_norm": 0.6438117623329163, "learning_rate": 2.9549071270173553e-06, "loss": 0.5031, "step": 3123 }, { "epoch": 2.0128865979381443, "grad_norm": 0.5740032196044922, "learning_rate": 2.9514867175121348e-06, "loss": 0.5239, "step": 3124 }, { "epoch": 2.0135309278350517, "grad_norm": 0.6626701951026917, "learning_rate": 2.9480674596302687e-06, "loss": 0.4872, "step": 3125 }, { "epoch": 2.0141752577319587, "grad_norm": 0.5923926830291748, "learning_rate": 2.944649355293979e-06, "loss": 0.5008, "step": 3126 }, { "epoch": 2.014819587628866, "grad_norm": 0.6832872629165649, "learning_rate": 2.9412324064248378e-06, "loss": 0.4865, "step": 3127 }, { "epoch": 2.015463917525773, "grad_norm": 0.5794115662574768, "learning_rate": 2.937816614943769e-06, "loss": 0.5127, "step": 3128 }, { "epoch": 2.0161082474226806, "grad_norm": 0.649507999420166, "learning_rate": 2.9344019827710486e-06, "loss": 0.5327, "step": 3129 }, { "epoch": 2.0167525773195876, "grad_norm": 0.6320545077323914, "learning_rate": 2.930988511826293e-06, "loss": 0.4983, "step": 3130 }, { "epoch": 2.017396907216495, "grad_norm": 0.5850167274475098, "learning_rate": 2.927576204028476e-06, "loss": 0.493, "step": 3131 }, { "epoch": 2.018041237113402, "grad_norm": 0.5762223601341248, "learning_rate": 2.9241650612959094e-06, "loss": 0.4855, "step": 3132 }, { "epoch": 2.0186855670103094, "grad_norm": 0.6347513198852539, "learning_rate": 2.920755085546255e-06, "loss": 0.482, "step": 3133 }, { "epoch": 2.0193298969072164, "grad_norm": 0.6021965146064758, "learning_rate": 2.9173462786965156e-06, "loss": 0.4985, "step": 3134 }, { "epoch": 2.019974226804124, "grad_norm": 0.587089478969574, "learning_rate": 2.9139386426630374e-06, "loss": 0.4759, "step": 3135 }, { "epoch": 2.020618556701031, "grad_norm": 0.5973417162895203, "learning_rate": 2.9105321793615106e-06, "loss": 0.4989, "step": 3136 }, { "epoch": 2.0212628865979383, "grad_norm": 0.5425488352775574, "learning_rate": 2.9071268907069617e-06, "loss": 0.4813, "step": 3137 }, { "epoch": 2.0219072164948453, "grad_norm": 0.5895681977272034, "learning_rate": 2.90372277861376e-06, "loss": 0.5172, "step": 3138 }, { "epoch": 2.0225515463917527, "grad_norm": 0.5571269989013672, "learning_rate": 2.900319844995616e-06, "loss": 0.5079, "step": 3139 }, { "epoch": 2.0231958762886597, "grad_norm": 0.6275810599327087, "learning_rate": 2.8969180917655702e-06, "loss": 0.5193, "step": 3140 }, { "epoch": 2.023840206185567, "grad_norm": 0.6244996190071106, "learning_rate": 2.8935175208360066e-06, "loss": 0.5124, "step": 3141 }, { "epoch": 2.024484536082474, "grad_norm": 0.6672137379646301, "learning_rate": 2.890118134118643e-06, "loss": 0.5352, "step": 3142 }, { "epoch": 2.0251288659793816, "grad_norm": 0.6159517765045166, "learning_rate": 2.8867199335245264e-06, "loss": 0.5131, "step": 3143 }, { "epoch": 2.0257731958762886, "grad_norm": 0.622498631477356, "learning_rate": 2.883322920964043e-06, "loss": 0.4998, "step": 3144 }, { "epoch": 2.026417525773196, "grad_norm": 0.5853153467178345, "learning_rate": 2.879927098346911e-06, "loss": 0.5055, "step": 3145 }, { "epoch": 2.027061855670103, "grad_norm": 0.5357677936553955, "learning_rate": 2.876532467582174e-06, "loss": 0.5186, "step": 3146 }, { "epoch": 2.0277061855670104, "grad_norm": 0.573798418045044, "learning_rate": 2.8731390305782105e-06, "loss": 0.5043, "step": 3147 }, { "epoch": 2.0283505154639174, "grad_norm": 0.6274280548095703, "learning_rate": 2.8697467892427266e-06, "loss": 0.4829, "step": 3148 }, { "epoch": 2.028994845360825, "grad_norm": 0.5605049133300781, "learning_rate": 2.8663557454827568e-06, "loss": 0.4827, "step": 3149 }, { "epoch": 2.029639175257732, "grad_norm": 0.5456294417381287, "learning_rate": 2.8629659012046633e-06, "loss": 0.5034, "step": 3150 }, { "epoch": 2.0302835051546393, "grad_norm": 0.5501642823219299, "learning_rate": 2.859577258314129e-06, "loss": 0.4943, "step": 3151 }, { "epoch": 2.0309278350515463, "grad_norm": 0.6313206553459167, "learning_rate": 2.8561898187161676e-06, "loss": 0.519, "step": 3152 }, { "epoch": 2.0315721649484537, "grad_norm": 0.5998279452323914, "learning_rate": 2.852803584315113e-06, "loss": 0.4949, "step": 3153 }, { "epoch": 2.0322164948453607, "grad_norm": 0.5472701787948608, "learning_rate": 2.8494185570146203e-06, "loss": 0.4907, "step": 3154 }, { "epoch": 2.032860824742268, "grad_norm": 0.6760470867156982, "learning_rate": 2.8460347387176702e-06, "loss": 0.5027, "step": 3155 }, { "epoch": 2.033505154639175, "grad_norm": 0.6328722238540649, "learning_rate": 2.842652131326562e-06, "loss": 0.4968, "step": 3156 }, { "epoch": 2.0341494845360826, "grad_norm": 0.567051112651825, "learning_rate": 2.839270736742911e-06, "loss": 0.5146, "step": 3157 }, { "epoch": 2.0347938144329896, "grad_norm": 0.6494664549827576, "learning_rate": 2.835890556867653e-06, "loss": 0.5097, "step": 3158 }, { "epoch": 2.035438144329897, "grad_norm": 0.6444433927536011, "learning_rate": 2.832511593601046e-06, "loss": 0.5114, "step": 3159 }, { "epoch": 2.036082474226804, "grad_norm": 0.5874865055084229, "learning_rate": 2.8291338488426543e-06, "loss": 0.4963, "step": 3160 }, { "epoch": 2.0367268041237114, "grad_norm": 0.567441463470459, "learning_rate": 2.8257573244913637e-06, "loss": 0.4795, "step": 3161 }, { "epoch": 2.0373711340206184, "grad_norm": 0.6117042899131775, "learning_rate": 2.822382022445376e-06, "loss": 0.4992, "step": 3162 }, { "epoch": 2.038015463917526, "grad_norm": 0.5416954755783081, "learning_rate": 2.819007944602196e-06, "loss": 0.5075, "step": 3163 }, { "epoch": 2.038659793814433, "grad_norm": 0.5603544116020203, "learning_rate": 2.815635092858653e-06, "loss": 0.5273, "step": 3164 }, { "epoch": 2.0393041237113403, "grad_norm": 0.5522147417068481, "learning_rate": 2.812263469110874e-06, "loss": 0.5153, "step": 3165 }, { "epoch": 2.0399484536082473, "grad_norm": 0.5174918174743652, "learning_rate": 2.8088930752543063e-06, "loss": 0.5172, "step": 3166 }, { "epoch": 2.0405927835051547, "grad_norm": 0.5938911437988281, "learning_rate": 2.805523913183701e-06, "loss": 0.5096, "step": 3167 }, { "epoch": 2.0412371134020617, "grad_norm": 0.5469308495521545, "learning_rate": 2.802155984793117e-06, "loss": 0.4986, "step": 3168 }, { "epoch": 2.041881443298969, "grad_norm": 0.5258004665374756, "learning_rate": 2.798789291975921e-06, "loss": 0.5033, "step": 3169 }, { "epoch": 2.042525773195876, "grad_norm": 0.5692205429077148, "learning_rate": 2.7954238366247855e-06, "loss": 0.5142, "step": 3170 }, { "epoch": 2.0431701030927836, "grad_norm": 0.5466060638427734, "learning_rate": 2.7920596206316824e-06, "loss": 0.5116, "step": 3171 }, { "epoch": 2.0438144329896906, "grad_norm": 0.5754587054252625, "learning_rate": 2.788696645887894e-06, "loss": 0.4967, "step": 3172 }, { "epoch": 2.044458762886598, "grad_norm": 0.5700514912605286, "learning_rate": 2.785334914283998e-06, "loss": 0.5075, "step": 3173 }, { "epoch": 2.045103092783505, "grad_norm": 0.5673998594284058, "learning_rate": 2.7819744277098783e-06, "loss": 0.504, "step": 3174 }, { "epoch": 2.0457474226804124, "grad_norm": 0.5845622420310974, "learning_rate": 2.7786151880547183e-06, "loss": 0.5034, "step": 3175 }, { "epoch": 2.0463917525773194, "grad_norm": 0.6105824708938599, "learning_rate": 2.775257197206996e-06, "loss": 0.5006, "step": 3176 }, { "epoch": 2.047036082474227, "grad_norm": 0.6045058965682983, "learning_rate": 2.7719004570544934e-06, "loss": 0.4803, "step": 3177 }, { "epoch": 2.047680412371134, "grad_norm": 0.6526871919631958, "learning_rate": 2.768544969484287e-06, "loss": 0.5068, "step": 3178 }, { "epoch": 2.0483247422680413, "grad_norm": 0.6354962587356567, "learning_rate": 2.7651907363827446e-06, "loss": 0.5098, "step": 3179 }, { "epoch": 2.0489690721649483, "grad_norm": 0.5361290574073792, "learning_rate": 2.7618377596355363e-06, "loss": 0.5147, "step": 3180 }, { "epoch": 2.0496134020618557, "grad_norm": 0.6181796789169312, "learning_rate": 2.7584860411276225e-06, "loss": 0.5, "step": 3181 }, { "epoch": 2.0502577319587627, "grad_norm": 0.6402423977851868, "learning_rate": 2.7551355827432547e-06, "loss": 0.5082, "step": 3182 }, { "epoch": 2.05090206185567, "grad_norm": 0.5623791813850403, "learning_rate": 2.751786386365977e-06, "loss": 0.4984, "step": 3183 }, { "epoch": 2.051546391752577, "grad_norm": 0.5379309058189392, "learning_rate": 2.7484384538786284e-06, "loss": 0.509, "step": 3184 }, { "epoch": 2.0521907216494846, "grad_norm": 0.6344433426856995, "learning_rate": 2.7450917871633263e-06, "loss": 0.5287, "step": 3185 }, { "epoch": 2.0528350515463916, "grad_norm": 0.5769842863082886, "learning_rate": 2.741746388101493e-06, "loss": 0.5041, "step": 3186 }, { "epoch": 2.053479381443299, "grad_norm": 0.5342798829078674, "learning_rate": 2.7384022585738215e-06, "loss": 0.5289, "step": 3187 }, { "epoch": 2.054123711340206, "grad_norm": 0.5542866587638855, "learning_rate": 2.7350594004603005e-06, "loss": 0.5352, "step": 3188 }, { "epoch": 2.0547680412371134, "grad_norm": 0.5922492742538452, "learning_rate": 2.731717815640205e-06, "loss": 0.5064, "step": 3189 }, { "epoch": 2.0554123711340204, "grad_norm": 0.5511579513549805, "learning_rate": 2.7283775059920865e-06, "loss": 0.5241, "step": 3190 }, { "epoch": 2.056056701030928, "grad_norm": 0.6450681090354919, "learning_rate": 2.7250384733937863e-06, "loss": 0.5001, "step": 3191 }, { "epoch": 2.056701030927835, "grad_norm": 0.6334452629089355, "learning_rate": 2.7217007197224275e-06, "loss": 0.5174, "step": 3192 }, { "epoch": 2.0573453608247423, "grad_norm": 0.6042719483375549, "learning_rate": 2.718364246854408e-06, "loss": 0.5186, "step": 3193 }, { "epoch": 2.0579896907216493, "grad_norm": 0.5315778851509094, "learning_rate": 2.7150290566654136e-06, "loss": 0.503, "step": 3194 }, { "epoch": 2.0586340206185567, "grad_norm": 0.604465901851654, "learning_rate": 2.711695151030406e-06, "loss": 0.5119, "step": 3195 }, { "epoch": 2.0592783505154637, "grad_norm": 0.6110650897026062, "learning_rate": 2.7083625318236213e-06, "loss": 0.4973, "step": 3196 }, { "epoch": 2.059922680412371, "grad_norm": 0.5122670531272888, "learning_rate": 2.7050312009185763e-06, "loss": 0.5125, "step": 3197 }, { "epoch": 2.060567010309278, "grad_norm": 0.648507833480835, "learning_rate": 2.7017011601880653e-06, "loss": 0.5025, "step": 3198 }, { "epoch": 2.0612113402061856, "grad_norm": 0.5753565430641174, "learning_rate": 2.698372411504151e-06, "loss": 0.5062, "step": 3199 }, { "epoch": 2.0618556701030926, "grad_norm": 0.5659112334251404, "learning_rate": 2.6950449567381766e-06, "loss": 0.5045, "step": 3200 }, { "epoch": 2.0625, "grad_norm": 0.5934152007102966, "learning_rate": 2.6917187977607502e-06, "loss": 0.5137, "step": 3201 }, { "epoch": 2.0631443298969074, "grad_norm": 0.5883473753929138, "learning_rate": 2.688393936441759e-06, "loss": 0.5034, "step": 3202 }, { "epoch": 2.0637886597938144, "grad_norm": 0.6052191853523254, "learning_rate": 2.6850703746503572e-06, "loss": 0.4979, "step": 3203 }, { "epoch": 2.0644329896907214, "grad_norm": 0.6086553931236267, "learning_rate": 2.681748114254968e-06, "loss": 0.496, "step": 3204 }, { "epoch": 2.065077319587629, "grad_norm": 0.6120154857635498, "learning_rate": 2.6784271571232846e-06, "loss": 0.4921, "step": 3205 }, { "epoch": 2.0657216494845363, "grad_norm": 0.5308817625045776, "learning_rate": 2.6751075051222684e-06, "loss": 0.5078, "step": 3206 }, { "epoch": 2.0663659793814433, "grad_norm": 0.5931569933891296, "learning_rate": 2.671789160118142e-06, "loss": 0.5126, "step": 3207 }, { "epoch": 2.0670103092783507, "grad_norm": 0.5734588503837585, "learning_rate": 2.6684721239764e-06, "loss": 0.5013, "step": 3208 }, { "epoch": 2.0676546391752577, "grad_norm": 0.5575089454650879, "learning_rate": 2.6651563985617945e-06, "loss": 0.4723, "step": 3209 }, { "epoch": 2.068298969072165, "grad_norm": 0.6209414005279541, "learning_rate": 2.6618419857383453e-06, "loss": 0.5145, "step": 3210 }, { "epoch": 2.068943298969072, "grad_norm": 0.6479102373123169, "learning_rate": 2.6585288873693355e-06, "loss": 0.5131, "step": 3211 }, { "epoch": 2.0695876288659796, "grad_norm": 0.625268816947937, "learning_rate": 2.6552171053173025e-06, "loss": 0.5032, "step": 3212 }, { "epoch": 2.0702319587628866, "grad_norm": 0.6561115980148315, "learning_rate": 2.651906641444051e-06, "loss": 0.5045, "step": 3213 }, { "epoch": 2.070876288659794, "grad_norm": 0.6045131087303162, "learning_rate": 2.648597497610642e-06, "loss": 0.5281, "step": 3214 }, { "epoch": 2.071520618556701, "grad_norm": 0.5403143763542175, "learning_rate": 2.645289675677391e-06, "loss": 0.5102, "step": 3215 }, { "epoch": 2.0721649484536084, "grad_norm": 0.603255033493042, "learning_rate": 2.6419831775038763e-06, "loss": 0.5288, "step": 3216 }, { "epoch": 2.0728092783505154, "grad_norm": 0.5688672065734863, "learning_rate": 2.638678004948929e-06, "loss": 0.5108, "step": 3217 }, { "epoch": 2.073453608247423, "grad_norm": 0.6071426272392273, "learning_rate": 2.635374159870634e-06, "loss": 0.52, "step": 3218 }, { "epoch": 2.07409793814433, "grad_norm": 0.6250527501106262, "learning_rate": 2.6320716441263305e-06, "loss": 0.5108, "step": 3219 }, { "epoch": 2.0747422680412373, "grad_norm": 0.5689556002616882, "learning_rate": 2.628770459572614e-06, "loss": 0.4761, "step": 3220 }, { "epoch": 2.0753865979381443, "grad_norm": 0.589923083782196, "learning_rate": 2.6254706080653225e-06, "loss": 0.4846, "step": 3221 }, { "epoch": 2.0760309278350517, "grad_norm": 0.6015509366989136, "learning_rate": 2.6221720914595587e-06, "loss": 0.511, "step": 3222 }, { "epoch": 2.0766752577319587, "grad_norm": 0.5903323888778687, "learning_rate": 2.618874911609661e-06, "loss": 0.4894, "step": 3223 }, { "epoch": 2.077319587628866, "grad_norm": 0.5695286393165588, "learning_rate": 2.6155790703692234e-06, "loss": 0.4893, "step": 3224 }, { "epoch": 2.077963917525773, "grad_norm": 0.6069237589836121, "learning_rate": 2.612284569591089e-06, "loss": 0.4973, "step": 3225 }, { "epoch": 2.0786082474226806, "grad_norm": 0.5966874361038208, "learning_rate": 2.6089914111273398e-06, "loss": 0.5036, "step": 3226 }, { "epoch": 2.0792525773195876, "grad_norm": 0.5178881883621216, "learning_rate": 2.6056995968293097e-06, "loss": 0.5267, "step": 3227 }, { "epoch": 2.079896907216495, "grad_norm": 0.5514748692512512, "learning_rate": 2.6024091285475784e-06, "loss": 0.5219, "step": 3228 }, { "epoch": 2.080541237113402, "grad_norm": 0.5479579567909241, "learning_rate": 2.5991200081319597e-06, "loss": 0.4934, "step": 3229 }, { "epoch": 2.0811855670103094, "grad_norm": 0.5592732429504395, "learning_rate": 2.595832237431518e-06, "loss": 0.4798, "step": 3230 }, { "epoch": 2.0818298969072164, "grad_norm": 0.5339449048042297, "learning_rate": 2.5925458182945597e-06, "loss": 0.515, "step": 3231 }, { "epoch": 2.082474226804124, "grad_norm": 0.5735135674476624, "learning_rate": 2.589260752568622e-06, "loss": 0.5187, "step": 3232 }, { "epoch": 2.083118556701031, "grad_norm": 0.537899374961853, "learning_rate": 2.5859770421004914e-06, "loss": 0.5216, "step": 3233 }, { "epoch": 2.0837628865979383, "grad_norm": 0.5463421940803528, "learning_rate": 2.5826946887361894e-06, "loss": 0.5043, "step": 3234 }, { "epoch": 2.0844072164948453, "grad_norm": 0.6030306816101074, "learning_rate": 2.5794136943209693e-06, "loss": 0.5084, "step": 3235 }, { "epoch": 2.0850515463917527, "grad_norm": 0.5813928246498108, "learning_rate": 2.576134060699328e-06, "loss": 0.4898, "step": 3236 }, { "epoch": 2.0856958762886597, "grad_norm": 0.5270187854766846, "learning_rate": 2.572855789714995e-06, "loss": 0.4944, "step": 3237 }, { "epoch": 2.086340206185567, "grad_norm": 0.5610232353210449, "learning_rate": 2.569578883210929e-06, "loss": 0.5105, "step": 3238 }, { "epoch": 2.086984536082474, "grad_norm": 0.549145519733429, "learning_rate": 2.5663033430293288e-06, "loss": 0.4941, "step": 3239 }, { "epoch": 2.0876288659793816, "grad_norm": 0.746394157409668, "learning_rate": 2.5630291710116208e-06, "loss": 0.5214, "step": 3240 }, { "epoch": 2.0882731958762886, "grad_norm": 0.5425606966018677, "learning_rate": 2.5597563689984635e-06, "loss": 0.4897, "step": 3241 }, { "epoch": 2.088917525773196, "grad_norm": 0.6939836740493774, "learning_rate": 2.5564849388297465e-06, "loss": 0.4821, "step": 3242 }, { "epoch": 2.089561855670103, "grad_norm": 0.6509991884231567, "learning_rate": 2.5532148823445835e-06, "loss": 0.5019, "step": 3243 }, { "epoch": 2.0902061855670104, "grad_norm": 0.5955027937889099, "learning_rate": 2.549946201381323e-06, "loss": 0.5076, "step": 3244 }, { "epoch": 2.0908505154639174, "grad_norm": 0.6894629597663879, "learning_rate": 2.5466788977775323e-06, "loss": 0.493, "step": 3245 }, { "epoch": 2.091494845360825, "grad_norm": 0.6522104144096375, "learning_rate": 2.54341297337001e-06, "loss": 0.4967, "step": 3246 }, { "epoch": 2.092139175257732, "grad_norm": 0.5758423209190369, "learning_rate": 2.54014842999478e-06, "loss": 0.4997, "step": 3247 }, { "epoch": 2.0927835051546393, "grad_norm": 0.666552722454071, "learning_rate": 2.536885269487084e-06, "loss": 0.4977, "step": 3248 }, { "epoch": 2.0934278350515463, "grad_norm": 0.6543226838111877, "learning_rate": 2.533623493681392e-06, "loss": 0.5015, "step": 3249 }, { "epoch": 2.0940721649484537, "grad_norm": 0.5819869041442871, "learning_rate": 2.530363104411394e-06, "loss": 0.5105, "step": 3250 }, { "epoch": 2.0947164948453607, "grad_norm": 0.5606314539909363, "learning_rate": 2.527104103509998e-06, "loss": 0.4791, "step": 3251 }, { "epoch": 2.095360824742268, "grad_norm": 0.6825358271598816, "learning_rate": 2.5238464928093344e-06, "loss": 0.5155, "step": 3252 }, { "epoch": 2.096005154639175, "grad_norm": 0.5852670073509216, "learning_rate": 2.520590274140753e-06, "loss": 0.515, "step": 3253 }, { "epoch": 2.0966494845360826, "grad_norm": 0.5166373252868652, "learning_rate": 2.517335449334816e-06, "loss": 0.5079, "step": 3254 }, { "epoch": 2.0972938144329896, "grad_norm": 0.6964030265808105, "learning_rate": 2.5140820202213058e-06, "loss": 0.5085, "step": 3255 }, { "epoch": 2.097938144329897, "grad_norm": 0.5925313830375671, "learning_rate": 2.510829988629222e-06, "loss": 0.5064, "step": 3256 }, { "epoch": 2.098582474226804, "grad_norm": 0.5303962826728821, "learning_rate": 2.5075793563867713e-06, "loss": 0.4983, "step": 3257 }, { "epoch": 2.0992268041237114, "grad_norm": 0.652702271938324, "learning_rate": 2.5043301253213853e-06, "loss": 0.5099, "step": 3258 }, { "epoch": 2.0998711340206184, "grad_norm": 0.6199591159820557, "learning_rate": 2.5010822972596953e-06, "loss": 0.5179, "step": 3259 }, { "epoch": 2.100515463917526, "grad_norm": 0.6050835847854614, "learning_rate": 2.4978358740275522e-06, "loss": 0.4868, "step": 3260 }, { "epoch": 2.101159793814433, "grad_norm": 0.6189655065536499, "learning_rate": 2.494590857450016e-06, "loss": 0.5036, "step": 3261 }, { "epoch": 2.1018041237113403, "grad_norm": 0.5425646305084229, "learning_rate": 2.491347249351351e-06, "loss": 0.5211, "step": 3262 }, { "epoch": 2.1024484536082473, "grad_norm": 0.5372640490531921, "learning_rate": 2.488105051555034e-06, "loss": 0.5127, "step": 3263 }, { "epoch": 2.1030927835051547, "grad_norm": 0.5899726152420044, "learning_rate": 2.4848642658837522e-06, "loss": 0.4861, "step": 3264 }, { "epoch": 2.1037371134020617, "grad_norm": 0.6412032246589661, "learning_rate": 2.4816248941593896e-06, "loss": 0.5101, "step": 3265 }, { "epoch": 2.104381443298969, "grad_norm": 0.5141750574111938, "learning_rate": 2.478386938203043e-06, "loss": 0.5187, "step": 3266 }, { "epoch": 2.105025773195876, "grad_norm": 0.551234781742096, "learning_rate": 2.4751503998350125e-06, "loss": 0.506, "step": 3267 }, { "epoch": 2.1056701030927836, "grad_norm": 0.7070196866989136, "learning_rate": 2.471915280874796e-06, "loss": 0.4995, "step": 3268 }, { "epoch": 2.1063144329896906, "grad_norm": 0.59438157081604, "learning_rate": 2.468681583141099e-06, "loss": 0.4726, "step": 3269 }, { "epoch": 2.106958762886598, "grad_norm": 0.6116174459457397, "learning_rate": 2.465449308451829e-06, "loss": 0.5154, "step": 3270 }, { "epoch": 2.107603092783505, "grad_norm": 0.5807211995124817, "learning_rate": 2.462218458624085e-06, "loss": 0.5118, "step": 3271 }, { "epoch": 2.1082474226804124, "grad_norm": 0.5991030335426331, "learning_rate": 2.458989035474174e-06, "loss": 0.5199, "step": 3272 }, { "epoch": 2.1088917525773194, "grad_norm": 0.5999624133110046, "learning_rate": 2.455761040817599e-06, "loss": 0.5012, "step": 3273 }, { "epoch": 2.109536082474227, "grad_norm": 0.6278117299079895, "learning_rate": 2.452534476469055e-06, "loss": 0.498, "step": 3274 }, { "epoch": 2.110180412371134, "grad_norm": 0.5259156823158264, "learning_rate": 2.449309344242438e-06, "loss": 0.5229, "step": 3275 }, { "epoch": 2.1108247422680413, "grad_norm": 0.6432604193687439, "learning_rate": 2.4460856459508374e-06, "loss": 0.5226, "step": 3276 }, { "epoch": 2.1114690721649483, "grad_norm": 0.5693439245223999, "learning_rate": 2.4428633834065366e-06, "loss": 0.5144, "step": 3277 }, { "epoch": 2.1121134020618557, "grad_norm": 0.5774598121643066, "learning_rate": 2.439642558421014e-06, "loss": 0.5147, "step": 3278 }, { "epoch": 2.1127577319587627, "grad_norm": 0.5604068636894226, "learning_rate": 2.436423172804932e-06, "loss": 0.5188, "step": 3279 }, { "epoch": 2.11340206185567, "grad_norm": 0.5903475284576416, "learning_rate": 2.4332052283681533e-06, "loss": 0.5224, "step": 3280 }, { "epoch": 2.114046391752577, "grad_norm": 0.6124439239501953, "learning_rate": 2.429988726919727e-06, "loss": 0.4845, "step": 3281 }, { "epoch": 2.1146907216494846, "grad_norm": 0.579146146774292, "learning_rate": 2.4267736702678875e-06, "loss": 0.5144, "step": 3282 }, { "epoch": 2.1153350515463916, "grad_norm": 0.5883265137672424, "learning_rate": 2.4235600602200625e-06, "loss": 0.516, "step": 3283 }, { "epoch": 2.115979381443299, "grad_norm": 0.5851946473121643, "learning_rate": 2.4203478985828614e-06, "loss": 0.4961, "step": 3284 }, { "epoch": 2.116623711340206, "grad_norm": 0.6224655508995056, "learning_rate": 2.417137187162083e-06, "loss": 0.519, "step": 3285 }, { "epoch": 2.1172680412371134, "grad_norm": 0.5219900012016296, "learning_rate": 2.4139279277627113e-06, "loss": 0.5123, "step": 3286 }, { "epoch": 2.1179123711340204, "grad_norm": 0.5776611566543579, "learning_rate": 2.41072012218891e-06, "loss": 0.5083, "step": 3287 }, { "epoch": 2.118556701030928, "grad_norm": 0.6136059165000916, "learning_rate": 2.4075137722440275e-06, "loss": 0.5138, "step": 3288 }, { "epoch": 2.119201030927835, "grad_norm": 0.5960826277732849, "learning_rate": 2.4043088797305974e-06, "loss": 0.4941, "step": 3289 }, { "epoch": 2.1198453608247423, "grad_norm": 0.6153236031532288, "learning_rate": 2.401105446450327e-06, "loss": 0.4911, "step": 3290 }, { "epoch": 2.1204896907216493, "grad_norm": 0.5660428404808044, "learning_rate": 2.3979034742041083e-06, "loss": 0.5095, "step": 3291 }, { "epoch": 2.1211340206185567, "grad_norm": 0.5420465469360352, "learning_rate": 2.3947029647920134e-06, "loss": 0.5101, "step": 3292 }, { "epoch": 2.1217783505154637, "grad_norm": 0.6083401441574097, "learning_rate": 2.391503920013283e-06, "loss": 0.4934, "step": 3293 }, { "epoch": 2.122422680412371, "grad_norm": 0.6013513207435608, "learning_rate": 2.3883063416663486e-06, "loss": 0.5397, "step": 3294 }, { "epoch": 2.1230670103092786, "grad_norm": 0.5624054074287415, "learning_rate": 2.3851102315488034e-06, "loss": 0.5214, "step": 3295 }, { "epoch": 2.1237113402061856, "grad_norm": 0.5858538746833801, "learning_rate": 2.381915591457424e-06, "loss": 0.4985, "step": 3296 }, { "epoch": 2.1243556701030926, "grad_norm": 0.643890380859375, "learning_rate": 2.3787224231881577e-06, "loss": 0.4857, "step": 3297 }, { "epoch": 2.125, "grad_norm": 0.5655024647712708, "learning_rate": 2.3755307285361236e-06, "loss": 0.524, "step": 3298 }, { "epoch": 2.1256443298969074, "grad_norm": 0.5463210940361023, "learning_rate": 2.3723405092956124e-06, "loss": 0.5058, "step": 3299 }, { "epoch": 2.1262886597938144, "grad_norm": 0.6043874621391296, "learning_rate": 2.36915176726009e-06, "loss": 0.4901, "step": 3300 }, { "epoch": 2.1269329896907214, "grad_norm": 0.5824875831604004, "learning_rate": 2.3659645042221834e-06, "loss": 0.5014, "step": 3301 }, { "epoch": 2.127577319587629, "grad_norm": 0.6433515548706055, "learning_rate": 2.362778721973695e-06, "loss": 0.5038, "step": 3302 }, { "epoch": 2.1282216494845363, "grad_norm": 0.5705822110176086, "learning_rate": 2.3595944223055944e-06, "loss": 0.499, "step": 3303 }, { "epoch": 2.1288659793814433, "grad_norm": 0.5635325312614441, "learning_rate": 2.3564116070080124e-06, "loss": 0.5316, "step": 3304 }, { "epoch": 2.1295103092783507, "grad_norm": 0.8648751974105835, "learning_rate": 2.3532302778702503e-06, "loss": 0.498, "step": 3305 }, { "epoch": 2.1301546391752577, "grad_norm": 0.573313295841217, "learning_rate": 2.3500504366807743e-06, "loss": 0.4865, "step": 3306 }, { "epoch": 2.130798969072165, "grad_norm": 0.6861265301704407, "learning_rate": 2.346872085227208e-06, "loss": 0.5332, "step": 3307 }, { "epoch": 2.131443298969072, "grad_norm": 0.6224304437637329, "learning_rate": 2.343695225296345e-06, "loss": 0.4955, "step": 3308 }, { "epoch": 2.1320876288659796, "grad_norm": 0.5648608803749084, "learning_rate": 2.340519858674138e-06, "loss": 0.494, "step": 3309 }, { "epoch": 2.1327319587628866, "grad_norm": 0.7072269916534424, "learning_rate": 2.337345987145696e-06, "loss": 0.5214, "step": 3310 }, { "epoch": 2.133376288659794, "grad_norm": 0.5706368088722229, "learning_rate": 2.3341736124952924e-06, "loss": 0.5241, "step": 3311 }, { "epoch": 2.134020618556701, "grad_norm": 0.5667051076889038, "learning_rate": 2.3310027365063575e-06, "loss": 0.508, "step": 3312 }, { "epoch": 2.1346649484536084, "grad_norm": 0.6424936056137085, "learning_rate": 2.3278333609614797e-06, "loss": 0.5038, "step": 3313 }, { "epoch": 2.1353092783505154, "grad_norm": 0.6593139171600342, "learning_rate": 2.324665487642404e-06, "loss": 0.5174, "step": 3314 }, { "epoch": 2.135953608247423, "grad_norm": 0.5929504036903381, "learning_rate": 2.3214991183300283e-06, "loss": 0.491, "step": 3315 }, { "epoch": 2.13659793814433, "grad_norm": 0.667662501335144, "learning_rate": 2.3183342548044067e-06, "loss": 0.5107, "step": 3316 }, { "epoch": 2.1372422680412373, "grad_norm": 0.5778210759162903, "learning_rate": 2.3151708988447506e-06, "loss": 0.5012, "step": 3317 }, { "epoch": 2.1378865979381443, "grad_norm": 0.7320390939712524, "learning_rate": 2.3120090522294158e-06, "loss": 0.514, "step": 3318 }, { "epoch": 2.1385309278350517, "grad_norm": 0.65573650598526, "learning_rate": 2.3088487167359174e-06, "loss": 0.5092, "step": 3319 }, { "epoch": 2.1391752577319587, "grad_norm": 0.6834623217582703, "learning_rate": 2.3056898941409143e-06, "loss": 0.5259, "step": 3320 }, { "epoch": 2.139819587628866, "grad_norm": 0.6904722452163696, "learning_rate": 2.3025325862202203e-06, "loss": 0.5276, "step": 3321 }, { "epoch": 2.140463917525773, "grad_norm": 0.5838131308555603, "learning_rate": 2.2993767947487973e-06, "loss": 0.5278, "step": 3322 }, { "epoch": 2.1411082474226806, "grad_norm": 0.7113854289054871, "learning_rate": 2.296222521500749e-06, "loss": 0.519, "step": 3323 }, { "epoch": 2.1417525773195876, "grad_norm": 0.60853111743927, "learning_rate": 2.2930697682493306e-06, "loss": 0.5213, "step": 3324 }, { "epoch": 2.142396907216495, "grad_norm": 0.5694112181663513, "learning_rate": 2.2899185367669458e-06, "loss": 0.5196, "step": 3325 }, { "epoch": 2.143041237113402, "grad_norm": 0.6552335023880005, "learning_rate": 2.286768828825133e-06, "loss": 0.4859, "step": 3326 }, { "epoch": 2.1436855670103094, "grad_norm": 0.5681605935096741, "learning_rate": 2.2836206461945836e-06, "loss": 0.5101, "step": 3327 }, { "epoch": 2.1443298969072164, "grad_norm": 0.5769798755645752, "learning_rate": 2.2804739906451277e-06, "loss": 0.506, "step": 3328 }, { "epoch": 2.144974226804124, "grad_norm": 0.5870856642723083, "learning_rate": 2.277328863945733e-06, "loss": 0.5044, "step": 3329 }, { "epoch": 2.145618556701031, "grad_norm": 0.5802473425865173, "learning_rate": 2.274185267864518e-06, "loss": 0.5149, "step": 3330 }, { "epoch": 2.1462628865979383, "grad_norm": 0.6276204586029053, "learning_rate": 2.2710432041687298e-06, "loss": 0.5098, "step": 3331 }, { "epoch": 2.1469072164948453, "grad_norm": 0.5639940500259399, "learning_rate": 2.26790267462476e-06, "loss": 0.4878, "step": 3332 }, { "epoch": 2.1475515463917527, "grad_norm": 0.6051271557807922, "learning_rate": 2.2647636809981384e-06, "loss": 0.522, "step": 3333 }, { "epoch": 2.1481958762886597, "grad_norm": 0.5801188945770264, "learning_rate": 2.2616262250535255e-06, "loss": 0.506, "step": 3334 }, { "epoch": 2.148840206185567, "grad_norm": 0.6118499040603638, "learning_rate": 2.2584903085547228e-06, "loss": 0.5267, "step": 3335 }, { "epoch": 2.149484536082474, "grad_norm": 0.5623173117637634, "learning_rate": 2.2553559332646675e-06, "loss": 0.4761, "step": 3336 }, { "epoch": 2.1501288659793816, "grad_norm": 0.6459617018699646, "learning_rate": 2.252223100945423e-06, "loss": 0.498, "step": 3337 }, { "epoch": 2.1507731958762886, "grad_norm": 0.5993242859840393, "learning_rate": 2.2490918133581925e-06, "loss": 0.4959, "step": 3338 }, { "epoch": 2.151417525773196, "grad_norm": 0.6051180362701416, "learning_rate": 2.2459620722633097e-06, "loss": 0.5332, "step": 3339 }, { "epoch": 2.152061855670103, "grad_norm": 0.5973924994468689, "learning_rate": 2.2428338794202335e-06, "loss": 0.5089, "step": 3340 }, { "epoch": 2.1527061855670104, "grad_norm": 0.635106086730957, "learning_rate": 2.2397072365875587e-06, "loss": 0.494, "step": 3341 }, { "epoch": 2.1533505154639174, "grad_norm": 0.6637786030769348, "learning_rate": 2.236582145523008e-06, "loss": 0.5074, "step": 3342 }, { "epoch": 2.153994845360825, "grad_norm": 0.643333911895752, "learning_rate": 2.2334586079834266e-06, "loss": 0.5039, "step": 3343 }, { "epoch": 2.154639175257732, "grad_norm": 0.6320238709449768, "learning_rate": 2.2303366257247907e-06, "loss": 0.5108, "step": 3344 }, { "epoch": 2.1552835051546393, "grad_norm": 0.67747563123703, "learning_rate": 2.227216200502204e-06, "loss": 0.5136, "step": 3345 }, { "epoch": 2.1559278350515463, "grad_norm": 0.651117742061615, "learning_rate": 2.2240973340698886e-06, "loss": 0.5039, "step": 3346 }, { "epoch": 2.1565721649484537, "grad_norm": 0.6139840483665466, "learning_rate": 2.220980028181195e-06, "loss": 0.5005, "step": 3347 }, { "epoch": 2.1572164948453607, "grad_norm": 0.7104285359382629, "learning_rate": 2.217864284588596e-06, "loss": 0.518, "step": 3348 }, { "epoch": 2.157860824742268, "grad_norm": 0.5938101410865784, "learning_rate": 2.214750105043685e-06, "loss": 0.5012, "step": 3349 }, { "epoch": 2.158505154639175, "grad_norm": 0.679914116859436, "learning_rate": 2.2116374912971776e-06, "loss": 0.4984, "step": 3350 }, { "epoch": 2.1591494845360826, "grad_norm": 0.5426285266876221, "learning_rate": 2.2085264450989057e-06, "loss": 0.5203, "step": 3351 }, { "epoch": 2.1597938144329896, "grad_norm": 0.5874548554420471, "learning_rate": 2.205416968197823e-06, "loss": 0.5159, "step": 3352 }, { "epoch": 2.160438144329897, "grad_norm": 0.6199877858161926, "learning_rate": 2.2023090623420037e-06, "loss": 0.5045, "step": 3353 }, { "epoch": 2.161082474226804, "grad_norm": 0.5979757308959961, "learning_rate": 2.1992027292786295e-06, "loss": 0.5047, "step": 3354 }, { "epoch": 2.1617268041237114, "grad_norm": 0.5881633162498474, "learning_rate": 2.1960979707540093e-06, "loss": 0.5309, "step": 3355 }, { "epoch": 2.1623711340206184, "grad_norm": 0.5572503209114075, "learning_rate": 2.1929947885135567e-06, "loss": 0.5093, "step": 3356 }, { "epoch": 2.163015463917526, "grad_norm": 0.5663607716560364, "learning_rate": 2.1898931843018057e-06, "loss": 0.5113, "step": 3357 }, { "epoch": 2.163659793814433, "grad_norm": 0.6188644766807556, "learning_rate": 2.186793159862403e-06, "loss": 0.5077, "step": 3358 }, { "epoch": 2.1643041237113403, "grad_norm": 0.6114136576652527, "learning_rate": 2.183694716938103e-06, "loss": 0.5062, "step": 3359 }, { "epoch": 2.1649484536082473, "grad_norm": 0.5690320730209351, "learning_rate": 2.1805978572707737e-06, "loss": 0.5095, "step": 3360 }, { "epoch": 2.1655927835051547, "grad_norm": 0.5309311747550964, "learning_rate": 2.1775025826013953e-06, "loss": 0.5038, "step": 3361 }, { "epoch": 2.1662371134020617, "grad_norm": 0.5511360168457031, "learning_rate": 2.174408894670051e-06, "loss": 0.515, "step": 3362 }, { "epoch": 2.166881443298969, "grad_norm": 0.643571674823761, "learning_rate": 2.1713167952159366e-06, "loss": 0.5009, "step": 3363 }, { "epoch": 2.167525773195876, "grad_norm": 0.5902764797210693, "learning_rate": 2.1682262859773564e-06, "loss": 0.501, "step": 3364 }, { "epoch": 2.1681701030927836, "grad_norm": 0.6816482543945312, "learning_rate": 2.165137368691711e-06, "loss": 0.524, "step": 3365 }, { "epoch": 2.1688144329896906, "grad_norm": 0.5589942336082458, "learning_rate": 2.1620500450955224e-06, "loss": 0.5008, "step": 3366 }, { "epoch": 2.169458762886598, "grad_norm": 0.6705927848815918, "learning_rate": 2.1589643169244007e-06, "loss": 0.508, "step": 3367 }, { "epoch": 2.170103092783505, "grad_norm": 0.638234555721283, "learning_rate": 2.155880185913068e-06, "loss": 0.4809, "step": 3368 }, { "epoch": 2.1707474226804124, "grad_norm": 0.5963079929351807, "learning_rate": 2.1527976537953472e-06, "loss": 0.5124, "step": 3369 }, { "epoch": 2.1713917525773194, "grad_norm": 0.5643369555473328, "learning_rate": 2.1497167223041588e-06, "loss": 0.5232, "step": 3370 }, { "epoch": 2.172036082474227, "grad_norm": 0.6136000156402588, "learning_rate": 2.146637393171528e-06, "loss": 0.5054, "step": 3371 }, { "epoch": 2.172680412371134, "grad_norm": 0.6817814707756042, "learning_rate": 2.1435596681285774e-06, "loss": 0.5112, "step": 3372 }, { "epoch": 2.1733247422680413, "grad_norm": 0.5896363258361816, "learning_rate": 2.1404835489055258e-06, "loss": 0.5062, "step": 3373 }, { "epoch": 2.1739690721649483, "grad_norm": 0.5341305136680603, "learning_rate": 2.1374090372316924e-06, "loss": 0.5131, "step": 3374 }, { "epoch": 2.1746134020618557, "grad_norm": 0.6231070756912231, "learning_rate": 2.1343361348354934e-06, "loss": 0.5133, "step": 3375 }, { "epoch": 2.1752577319587627, "grad_norm": 0.6503050923347473, "learning_rate": 2.1312648434444342e-06, "loss": 0.513, "step": 3376 }, { "epoch": 2.17590206185567, "grad_norm": 0.6105585098266602, "learning_rate": 2.128195164785121e-06, "loss": 0.4922, "step": 3377 }, { "epoch": 2.176546391752577, "grad_norm": 0.6990000605583191, "learning_rate": 2.125127100583252e-06, "loss": 0.5217, "step": 3378 }, { "epoch": 2.1771907216494846, "grad_norm": 0.7196961045265198, "learning_rate": 2.1220606525636136e-06, "loss": 0.5118, "step": 3379 }, { "epoch": 2.1778350515463916, "grad_norm": 0.5390651226043701, "learning_rate": 2.118995822450089e-06, "loss": 0.5065, "step": 3380 }, { "epoch": 2.178479381443299, "grad_norm": 0.6123354434967041, "learning_rate": 2.1159326119656508e-06, "loss": 0.4932, "step": 3381 }, { "epoch": 2.179123711340206, "grad_norm": 0.6742780208587646, "learning_rate": 2.1128710228323567e-06, "loss": 0.5263, "step": 3382 }, { "epoch": 2.1797680412371134, "grad_norm": 0.5559479594230652, "learning_rate": 2.1098110567713586e-06, "loss": 0.5021, "step": 3383 }, { "epoch": 2.1804123711340204, "grad_norm": 0.665079653263092, "learning_rate": 2.1067527155028923e-06, "loss": 0.4971, "step": 3384 }, { "epoch": 2.181056701030928, "grad_norm": 0.7057538032531738, "learning_rate": 2.103696000746282e-06, "loss": 0.5055, "step": 3385 }, { "epoch": 2.181701030927835, "grad_norm": 0.6188464164733887, "learning_rate": 2.100640914219939e-06, "loss": 0.509, "step": 3386 }, { "epoch": 2.1823453608247423, "grad_norm": 0.6717378497123718, "learning_rate": 2.0975874576413528e-06, "loss": 0.5137, "step": 3387 }, { "epoch": 2.1829896907216493, "grad_norm": 0.7509697079658508, "learning_rate": 2.0945356327271038e-06, "loss": 0.5154, "step": 3388 }, { "epoch": 2.1836340206185567, "grad_norm": 0.6907312870025635, "learning_rate": 2.091485441192853e-06, "loss": 0.513, "step": 3389 }, { "epoch": 2.1842783505154637, "grad_norm": 0.7141199111938477, "learning_rate": 2.0884368847533394e-06, "loss": 0.5345, "step": 3390 }, { "epoch": 2.184922680412371, "grad_norm": 0.7159143686294556, "learning_rate": 2.08538996512239e-06, "loss": 0.5116, "step": 3391 }, { "epoch": 2.1855670103092786, "grad_norm": 0.6466333270072937, "learning_rate": 2.082344684012902e-06, "loss": 0.5113, "step": 3392 }, { "epoch": 2.1862113402061856, "grad_norm": 0.7207666039466858, "learning_rate": 2.0793010431368604e-06, "loss": 0.4893, "step": 3393 }, { "epoch": 2.1868556701030926, "grad_norm": 0.7013742327690125, "learning_rate": 2.0762590442053255e-06, "loss": 0.5213, "step": 3394 }, { "epoch": 2.1875, "grad_norm": 0.6925342679023743, "learning_rate": 2.0732186889284306e-06, "loss": 0.5304, "step": 3395 }, { "epoch": 2.1881443298969074, "grad_norm": 0.6032581925392151, "learning_rate": 2.0701799790153897e-06, "loss": 0.5137, "step": 3396 }, { "epoch": 2.1887886597938144, "grad_norm": 0.5974168181419373, "learning_rate": 2.067142916174492e-06, "loss": 0.5208, "step": 3397 }, { "epoch": 2.1894329896907214, "grad_norm": 0.6727386713027954, "learning_rate": 2.0641075021130947e-06, "loss": 0.5081, "step": 3398 }, { "epoch": 2.190077319587629, "grad_norm": 0.577372133731842, "learning_rate": 2.061073738537635e-06, "loss": 0.511, "step": 3399 }, { "epoch": 2.1907216494845363, "grad_norm": 0.5980327129364014, "learning_rate": 2.0580416271536204e-06, "loss": 0.5209, "step": 3400 }, { "epoch": 2.1913659793814433, "grad_norm": 0.6952008605003357, "learning_rate": 2.055011169665624e-06, "loss": 0.5305, "step": 3401 }, { "epoch": 2.1920103092783507, "grad_norm": 0.5668666362762451, "learning_rate": 2.051982367777301e-06, "loss": 0.4861, "step": 3402 }, { "epoch": 2.1926546391752577, "grad_norm": 0.5910234451293945, "learning_rate": 2.048955223191363e-06, "loss": 0.503, "step": 3403 }, { "epoch": 2.193298969072165, "grad_norm": 0.6050562262535095, "learning_rate": 2.0459297376095973e-06, "loss": 0.5316, "step": 3404 }, { "epoch": 2.193943298969072, "grad_norm": 0.5864976644515991, "learning_rate": 2.0429059127328586e-06, "loss": 0.5148, "step": 3405 }, { "epoch": 2.1945876288659796, "grad_norm": 0.5720738172531128, "learning_rate": 2.039883750261063e-06, "loss": 0.5117, "step": 3406 }, { "epoch": 2.1952319587628866, "grad_norm": 0.5664193630218506, "learning_rate": 2.0368632518931976e-06, "loss": 0.5163, "step": 3407 }, { "epoch": 2.195876288659794, "grad_norm": 0.6770726442337036, "learning_rate": 2.033844419327312e-06, "loss": 0.5133, "step": 3408 }, { "epoch": 2.196520618556701, "grad_norm": 0.5915381908416748, "learning_rate": 2.030827254260516e-06, "loss": 0.5207, "step": 3409 }, { "epoch": 2.1971649484536084, "grad_norm": 0.6074342727661133, "learning_rate": 2.027811758388988e-06, "loss": 0.4998, "step": 3410 }, { "epoch": 2.1978092783505154, "grad_norm": 0.6402825713157654, "learning_rate": 2.0247979334079646e-06, "loss": 0.5143, "step": 3411 }, { "epoch": 2.198453608247423, "grad_norm": 0.6262320876121521, "learning_rate": 2.021785781011742e-06, "loss": 0.5126, "step": 3412 }, { "epoch": 2.19909793814433, "grad_norm": 0.6778460144996643, "learning_rate": 2.018775302893678e-06, "loss": 0.4988, "step": 3413 }, { "epoch": 2.1997422680412373, "grad_norm": 0.5808110237121582, "learning_rate": 2.01576650074619e-06, "loss": 0.5057, "step": 3414 }, { "epoch": 2.2003865979381443, "grad_norm": 0.575802743434906, "learning_rate": 2.0127593762607507e-06, "loss": 0.5025, "step": 3415 }, { "epoch": 2.2010309278350517, "grad_norm": 0.5838847756385803, "learning_rate": 2.00975393112789e-06, "loss": 0.506, "step": 3416 }, { "epoch": 2.2016752577319587, "grad_norm": 0.6538989543914795, "learning_rate": 2.006750167037198e-06, "loss": 0.5348, "step": 3417 }, { "epoch": 2.202319587628866, "grad_norm": 0.5692601799964905, "learning_rate": 2.0037480856773123e-06, "loss": 0.5251, "step": 3418 }, { "epoch": 2.202963917525773, "grad_norm": 0.5434490442276001, "learning_rate": 2.0007476887359305e-06, "loss": 0.508, "step": 3419 }, { "epoch": 2.2036082474226806, "grad_norm": 0.7136861681938171, "learning_rate": 1.997748977899801e-06, "loss": 0.5118, "step": 3420 }, { "epoch": 2.2042525773195876, "grad_norm": 0.5609163641929626, "learning_rate": 1.9947519548547256e-06, "loss": 0.5014, "step": 3421 }, { "epoch": 2.204896907216495, "grad_norm": 0.6280122399330139, "learning_rate": 1.9917566212855566e-06, "loss": 0.5132, "step": 3422 }, { "epoch": 2.205541237113402, "grad_norm": 0.6028144955635071, "learning_rate": 1.988762978876194e-06, "loss": 0.4979, "step": 3423 }, { "epoch": 2.2061855670103094, "grad_norm": 0.5332035422325134, "learning_rate": 1.9857710293095905e-06, "loss": 0.4946, "step": 3424 }, { "epoch": 2.2068298969072164, "grad_norm": 0.6243787407875061, "learning_rate": 1.9827807742677474e-06, "loss": 0.5061, "step": 3425 }, { "epoch": 2.207474226804124, "grad_norm": 0.6252151131629944, "learning_rate": 1.979792215431709e-06, "loss": 0.5157, "step": 3426 }, { "epoch": 2.208118556701031, "grad_norm": 0.5630053877830505, "learning_rate": 1.9768053544815725e-06, "loss": 0.5162, "step": 3427 }, { "epoch": 2.2087628865979383, "grad_norm": 0.5502218008041382, "learning_rate": 1.9738201930964723e-06, "loss": 0.5134, "step": 3428 }, { "epoch": 2.2094072164948453, "grad_norm": 0.6867371797561646, "learning_rate": 1.9708367329545956e-06, "loss": 0.5014, "step": 3429 }, { "epoch": 2.2100515463917527, "grad_norm": 0.604263424873352, "learning_rate": 1.9678549757331706e-06, "loss": 0.5217, "step": 3430 }, { "epoch": 2.2106958762886597, "grad_norm": 0.687187910079956, "learning_rate": 1.9648749231084636e-06, "loss": 0.5124, "step": 3431 }, { "epoch": 2.211340206185567, "grad_norm": 0.6299733519554138, "learning_rate": 1.961896576755788e-06, "loss": 0.5162, "step": 3432 }, { "epoch": 2.211984536082474, "grad_norm": 0.5868656635284424, "learning_rate": 1.958919938349498e-06, "loss": 0.5166, "step": 3433 }, { "epoch": 2.2126288659793816, "grad_norm": 0.6427131295204163, "learning_rate": 1.955945009562983e-06, "loss": 0.489, "step": 3434 }, { "epoch": 2.2132731958762886, "grad_norm": 0.5712685585021973, "learning_rate": 1.9529717920686752e-06, "loss": 0.5203, "step": 3435 }, { "epoch": 2.213917525773196, "grad_norm": 0.5762919187545776, "learning_rate": 1.9500002875380458e-06, "loss": 0.5186, "step": 3436 }, { "epoch": 2.214561855670103, "grad_norm": 0.5806930661201477, "learning_rate": 1.9470304976415954e-06, "loss": 0.5038, "step": 3437 }, { "epoch": 2.2152061855670104, "grad_norm": 0.6341609358787537, "learning_rate": 1.9440624240488736e-06, "loss": 0.5149, "step": 3438 }, { "epoch": 2.2158505154639174, "grad_norm": 0.5956522226333618, "learning_rate": 1.9410960684284526e-06, "loss": 0.5127, "step": 3439 }, { "epoch": 2.216494845360825, "grad_norm": 0.633905827999115, "learning_rate": 1.938131432447945e-06, "loss": 0.512, "step": 3440 }, { "epoch": 2.217139175257732, "grad_norm": 0.5977440476417542, "learning_rate": 1.935168517773998e-06, "loss": 0.4963, "step": 3441 }, { "epoch": 2.2177835051546393, "grad_norm": 0.621911883354187, "learning_rate": 1.932207326072285e-06, "loss": 0.4979, "step": 3442 }, { "epoch": 2.2184278350515463, "grad_norm": 0.6055919528007507, "learning_rate": 1.9292478590075163e-06, "loss": 0.4933, "step": 3443 }, { "epoch": 2.2190721649484537, "grad_norm": 0.6518499851226807, "learning_rate": 1.9262901182434322e-06, "loss": 0.4968, "step": 3444 }, { "epoch": 2.2197164948453607, "grad_norm": 0.5453543663024902, "learning_rate": 1.923334105442799e-06, "loss": 0.5067, "step": 3445 }, { "epoch": 2.220360824742268, "grad_norm": 0.536137044429779, "learning_rate": 1.920379822267414e-06, "loss": 0.5037, "step": 3446 }, { "epoch": 2.221005154639175, "grad_norm": 0.5824775695800781, "learning_rate": 1.9174272703781046e-06, "loss": 0.5198, "step": 3447 }, { "epoch": 2.2216494845360826, "grad_norm": 0.6649184823036194, "learning_rate": 1.914476451434718e-06, "loss": 0.5374, "step": 3448 }, { "epoch": 2.2222938144329896, "grad_norm": 0.5331534147262573, "learning_rate": 1.911527367096134e-06, "loss": 0.5305, "step": 3449 }, { "epoch": 2.222938144329897, "grad_norm": 0.6791490316390991, "learning_rate": 1.9085800190202562e-06, "loss": 0.4921, "step": 3450 }, { "epoch": 2.223582474226804, "grad_norm": 0.685572624206543, "learning_rate": 1.9056344088640065e-06, "loss": 0.5066, "step": 3451 }, { "epoch": 2.2242268041237114, "grad_norm": 0.5287613868713379, "learning_rate": 1.9026905382833361e-06, "loss": 0.51, "step": 3452 }, { "epoch": 2.2248711340206184, "grad_norm": 0.5919535756111145, "learning_rate": 1.899748408933218e-06, "loss": 0.4894, "step": 3453 }, { "epoch": 2.225515463917526, "grad_norm": 0.7189034223556519, "learning_rate": 1.89680802246764e-06, "loss": 0.5013, "step": 3454 }, { "epoch": 2.226159793814433, "grad_norm": 0.5689953565597534, "learning_rate": 1.8938693805396168e-06, "loss": 0.5042, "step": 3455 }, { "epoch": 2.2268041237113403, "grad_norm": 0.6066180467605591, "learning_rate": 1.8909324848011802e-06, "loss": 0.5083, "step": 3456 }, { "epoch": 2.2274484536082473, "grad_norm": 0.6308591961860657, "learning_rate": 1.8879973369033793e-06, "loss": 0.5102, "step": 3457 }, { "epoch": 2.2280927835051547, "grad_norm": 0.5689573287963867, "learning_rate": 1.8850639384962842e-06, "loss": 0.52, "step": 3458 }, { "epoch": 2.2287371134020617, "grad_norm": 0.5774674415588379, "learning_rate": 1.8821322912289747e-06, "loss": 0.5023, "step": 3459 }, { "epoch": 2.229381443298969, "grad_norm": 0.5923420190811157, "learning_rate": 1.8792023967495521e-06, "loss": 0.5147, "step": 3460 }, { "epoch": 2.230025773195876, "grad_norm": 0.6494964957237244, "learning_rate": 1.8762742567051318e-06, "loss": 0.4908, "step": 3461 }, { "epoch": 2.2306701030927836, "grad_norm": 0.6703037023544312, "learning_rate": 1.873347872741838e-06, "loss": 0.5371, "step": 3462 }, { "epoch": 2.2313144329896906, "grad_norm": 0.5794981122016907, "learning_rate": 1.870423246504815e-06, "loss": 0.4801, "step": 3463 }, { "epoch": 2.231958762886598, "grad_norm": 0.7414759993553162, "learning_rate": 1.867500379638212e-06, "loss": 0.5148, "step": 3464 }, { "epoch": 2.232603092783505, "grad_norm": 0.6232085824012756, "learning_rate": 1.8645792737851925e-06, "loss": 0.4904, "step": 3465 }, { "epoch": 2.2332474226804124, "grad_norm": 0.6612986922264099, "learning_rate": 1.8616599305879334e-06, "loss": 0.5082, "step": 3466 }, { "epoch": 2.2338917525773194, "grad_norm": 0.5838163495063782, "learning_rate": 1.8587423516876123e-06, "loss": 0.5072, "step": 3467 }, { "epoch": 2.234536082474227, "grad_norm": 0.6869513988494873, "learning_rate": 1.8558265387244224e-06, "loss": 0.5063, "step": 3468 }, { "epoch": 2.235180412371134, "grad_norm": 0.6020177602767944, "learning_rate": 1.8529124933375619e-06, "loss": 0.5024, "step": 3469 }, { "epoch": 2.2358247422680413, "grad_norm": 0.6506085395812988, "learning_rate": 1.8500002171652321e-06, "loss": 0.5116, "step": 3470 }, { "epoch": 2.2364690721649483, "grad_norm": 0.7353200912475586, "learning_rate": 1.847089711844644e-06, "loss": 0.5084, "step": 3471 }, { "epoch": 2.2371134020618557, "grad_norm": 0.6158300042152405, "learning_rate": 1.8441809790120135e-06, "loss": 0.4886, "step": 3472 }, { "epoch": 2.2377577319587627, "grad_norm": 0.6294876337051392, "learning_rate": 1.8412740203025543e-06, "loss": 0.5068, "step": 3473 }, { "epoch": 2.23840206185567, "grad_norm": 0.6034902334213257, "learning_rate": 1.8383688373504882e-06, "loss": 0.4901, "step": 3474 }, { "epoch": 2.239046391752577, "grad_norm": 0.6029127836227417, "learning_rate": 1.835465431789037e-06, "loss": 0.5298, "step": 3475 }, { "epoch": 2.2396907216494846, "grad_norm": 0.6084356904029846, "learning_rate": 1.8325638052504235e-06, "loss": 0.51, "step": 3476 }, { "epoch": 2.2403350515463916, "grad_norm": 0.5563419461250305, "learning_rate": 1.8296639593658721e-06, "loss": 0.4923, "step": 3477 }, { "epoch": 2.240979381443299, "grad_norm": 0.565776526927948, "learning_rate": 1.8267658957656e-06, "loss": 0.4909, "step": 3478 }, { "epoch": 2.241623711340206, "grad_norm": 0.5582957863807678, "learning_rate": 1.8238696160788288e-06, "loss": 0.4941, "step": 3479 }, { "epoch": 2.2422680412371134, "grad_norm": 0.6267240643501282, "learning_rate": 1.8209751219337772e-06, "loss": 0.4987, "step": 3480 }, { "epoch": 2.2429123711340204, "grad_norm": 0.5832197666168213, "learning_rate": 1.8180824149576537e-06, "loss": 0.4977, "step": 3481 }, { "epoch": 2.243556701030928, "grad_norm": 0.5687676072120667, "learning_rate": 1.8151914967766694e-06, "loss": 0.5148, "step": 3482 }, { "epoch": 2.244201030927835, "grad_norm": 0.5695680379867554, "learning_rate": 1.812302369016027e-06, "loss": 0.529, "step": 3483 }, { "epoch": 2.2448453608247423, "grad_norm": 0.5990322232246399, "learning_rate": 1.80941503329992e-06, "loss": 0.5016, "step": 3484 }, { "epoch": 2.2454896907216493, "grad_norm": 0.6340898275375366, "learning_rate": 1.8065294912515386e-06, "loss": 0.5005, "step": 3485 }, { "epoch": 2.2461340206185567, "grad_norm": 0.6436617970466614, "learning_rate": 1.8036457444930643e-06, "loss": 0.5092, "step": 3486 }, { "epoch": 2.2467783505154637, "grad_norm": 0.5349907875061035, "learning_rate": 1.8007637946456652e-06, "loss": 0.4988, "step": 3487 }, { "epoch": 2.247422680412371, "grad_norm": 0.6219534873962402, "learning_rate": 1.7978836433295026e-06, "loss": 0.5176, "step": 3488 }, { "epoch": 2.2480670103092786, "grad_norm": 0.6141459941864014, "learning_rate": 1.795005292163729e-06, "loss": 0.4879, "step": 3489 }, { "epoch": 2.2487113402061856, "grad_norm": 0.5558739900588989, "learning_rate": 1.792128742766478e-06, "loss": 0.4964, "step": 3490 }, { "epoch": 2.2493556701030926, "grad_norm": 0.5290919542312622, "learning_rate": 1.7892539967548761e-06, "loss": 0.5085, "step": 3491 }, { "epoch": 2.25, "grad_norm": 0.557908833026886, "learning_rate": 1.7863810557450345e-06, "loss": 0.5114, "step": 3492 }, { "epoch": 2.2506443298969074, "grad_norm": 0.5661336183547974, "learning_rate": 1.783509921352049e-06, "loss": 0.5121, "step": 3493 }, { "epoch": 2.2512886597938144, "grad_norm": 0.5651065111160278, "learning_rate": 1.780640595190002e-06, "loss": 0.4938, "step": 3494 }, { "epoch": 2.2519329896907214, "grad_norm": 0.5230874419212341, "learning_rate": 1.7777730788719538e-06, "loss": 0.4923, "step": 3495 }, { "epoch": 2.252577319587629, "grad_norm": 0.5666623115539551, "learning_rate": 1.774907374009953e-06, "loss": 0.4807, "step": 3496 }, { "epoch": 2.2532216494845363, "grad_norm": 0.5791028738021851, "learning_rate": 1.7720434822150296e-06, "loss": 0.4984, "step": 3497 }, { "epoch": 2.2538659793814433, "grad_norm": 0.5673807859420776, "learning_rate": 1.769181405097189e-06, "loss": 0.4989, "step": 3498 }, { "epoch": 2.2545103092783503, "grad_norm": 0.5846549868583679, "learning_rate": 1.7663211442654216e-06, "loss": 0.5116, "step": 3499 }, { "epoch": 2.2551546391752577, "grad_norm": 0.5163290500640869, "learning_rate": 1.7634627013276968e-06, "loss": 0.4996, "step": 3500 }, { "epoch": 2.255798969072165, "grad_norm": 0.5661481022834778, "learning_rate": 1.760606077890956e-06, "loss": 0.5198, "step": 3501 }, { "epoch": 2.256443298969072, "grad_norm": 0.6128891110420227, "learning_rate": 1.7577512755611277e-06, "loss": 0.5056, "step": 3502 }, { "epoch": 2.257087628865979, "grad_norm": 0.5830926895141602, "learning_rate": 1.7548982959431055e-06, "loss": 0.5183, "step": 3503 }, { "epoch": 2.2577319587628866, "grad_norm": 0.6003444194793701, "learning_rate": 1.7520471406407658e-06, "loss": 0.5134, "step": 3504 }, { "epoch": 2.258376288659794, "grad_norm": 0.5934081077575684, "learning_rate": 1.7491978112569603e-06, "loss": 0.4821, "step": 3505 }, { "epoch": 2.259020618556701, "grad_norm": 0.6053898334503174, "learning_rate": 1.7463503093935063e-06, "loss": 0.5403, "step": 3506 }, { "epoch": 2.2596649484536084, "grad_norm": 0.6199234127998352, "learning_rate": 1.7435046366512015e-06, "loss": 0.5072, "step": 3507 }, { "epoch": 2.2603092783505154, "grad_norm": 0.5358789563179016, "learning_rate": 1.7406607946298137e-06, "loss": 0.4991, "step": 3508 }, { "epoch": 2.260953608247423, "grad_norm": 0.6582151651382446, "learning_rate": 1.7378187849280776e-06, "loss": 0.5034, "step": 3509 }, { "epoch": 2.26159793814433, "grad_norm": 0.5801000595092773, "learning_rate": 1.7349786091437016e-06, "loss": 0.5302, "step": 3510 }, { "epoch": 2.2622422680412373, "grad_norm": 0.6071609258651733, "learning_rate": 1.7321402688733618e-06, "loss": 0.5246, "step": 3511 }, { "epoch": 2.2628865979381443, "grad_norm": 0.6301447153091431, "learning_rate": 1.7293037657127038e-06, "loss": 0.496, "step": 3512 }, { "epoch": 2.2635309278350517, "grad_norm": 0.5833312273025513, "learning_rate": 1.72646910125634e-06, "loss": 0.5016, "step": 3513 }, { "epoch": 2.2641752577319587, "grad_norm": 0.6147417426109314, "learning_rate": 1.7236362770978453e-06, "loss": 0.4892, "step": 3514 }, { "epoch": 2.264819587628866, "grad_norm": 0.6170550584793091, "learning_rate": 1.7208052948297654e-06, "loss": 0.5412, "step": 3515 }, { "epoch": 2.265463917525773, "grad_norm": 0.6128168702125549, "learning_rate": 1.7179761560436097e-06, "loss": 0.496, "step": 3516 }, { "epoch": 2.2661082474226806, "grad_norm": 0.5914236307144165, "learning_rate": 1.7151488623298462e-06, "loss": 0.4891, "step": 3517 }, { "epoch": 2.2667525773195876, "grad_norm": 0.665953516960144, "learning_rate": 1.7123234152779111e-06, "loss": 0.5286, "step": 3518 }, { "epoch": 2.267396907216495, "grad_norm": 0.5690065026283264, "learning_rate": 1.7094998164762022e-06, "loss": 0.5025, "step": 3519 }, { "epoch": 2.268041237113402, "grad_norm": 0.5959851741790771, "learning_rate": 1.7066780675120736e-06, "loss": 0.53, "step": 3520 }, { "epoch": 2.2686855670103094, "grad_norm": 0.6313357949256897, "learning_rate": 1.7038581699718442e-06, "loss": 0.531, "step": 3521 }, { "epoch": 2.2693298969072164, "grad_norm": 0.5616005063056946, "learning_rate": 1.7010401254407922e-06, "loss": 0.4964, "step": 3522 }, { "epoch": 2.269974226804124, "grad_norm": 0.5553709268569946, "learning_rate": 1.6982239355031488e-06, "loss": 0.5038, "step": 3523 }, { "epoch": 2.270618556701031, "grad_norm": 0.6505584120750427, "learning_rate": 1.695409601742108e-06, "loss": 0.5084, "step": 3524 }, { "epoch": 2.2712628865979383, "grad_norm": 0.5786884427070618, "learning_rate": 1.69259712573982e-06, "loss": 0.4879, "step": 3525 }, { "epoch": 2.2719072164948453, "grad_norm": 0.5944621562957764, "learning_rate": 1.6897865090773858e-06, "loss": 0.5146, "step": 3526 }, { "epoch": 2.2725515463917527, "grad_norm": 0.6536762714385986, "learning_rate": 1.6869777533348651e-06, "loss": 0.5181, "step": 3527 }, { "epoch": 2.2731958762886597, "grad_norm": 0.5581928491592407, "learning_rate": 1.6841708600912725e-06, "loss": 0.5069, "step": 3528 }, { "epoch": 2.273840206185567, "grad_norm": 0.6318266987800598, "learning_rate": 1.6813658309245723e-06, "loss": 0.4926, "step": 3529 }, { "epoch": 2.274484536082474, "grad_norm": 0.6136743426322937, "learning_rate": 1.678562667411685e-06, "loss": 0.5229, "step": 3530 }, { "epoch": 2.2751288659793816, "grad_norm": 0.5932323932647705, "learning_rate": 1.6757613711284753e-06, "loss": 0.5018, "step": 3531 }, { "epoch": 2.2757731958762886, "grad_norm": 0.5297527313232422, "learning_rate": 1.6729619436497645e-06, "loss": 0.5116, "step": 3532 }, { "epoch": 2.276417525773196, "grad_norm": 0.5820354223251343, "learning_rate": 1.6701643865493234e-06, "loss": 0.4933, "step": 3533 }, { "epoch": 2.277061855670103, "grad_norm": 0.5573790073394775, "learning_rate": 1.6673687013998652e-06, "loss": 0.52, "step": 3534 }, { "epoch": 2.2777061855670104, "grad_norm": 0.5584697723388672, "learning_rate": 1.6645748897730569e-06, "loss": 0.5176, "step": 3535 }, { "epoch": 2.2783505154639174, "grad_norm": 0.5423673391342163, "learning_rate": 1.661782953239512e-06, "loss": 0.4985, "step": 3536 }, { "epoch": 2.278994845360825, "grad_norm": 0.6353598237037659, "learning_rate": 1.6589928933687838e-06, "loss": 0.5261, "step": 3537 }, { "epoch": 2.279639175257732, "grad_norm": 0.6256431341171265, "learning_rate": 1.6562047117293783e-06, "loss": 0.4908, "step": 3538 }, { "epoch": 2.2802835051546393, "grad_norm": 0.5597547292709351, "learning_rate": 1.653418409888739e-06, "loss": 0.5437, "step": 3539 }, { "epoch": 2.2809278350515463, "grad_norm": 0.5626111030578613, "learning_rate": 1.6506339894132577e-06, "loss": 0.5138, "step": 3540 }, { "epoch": 2.2815721649484537, "grad_norm": 0.5367644429206848, "learning_rate": 1.6478514518682675e-06, "loss": 0.5095, "step": 3541 }, { "epoch": 2.2822164948453607, "grad_norm": 0.5669991374015808, "learning_rate": 1.6450707988180387e-06, "loss": 0.5078, "step": 3542 }, { "epoch": 2.282860824742268, "grad_norm": 0.5783042907714844, "learning_rate": 1.6422920318257878e-06, "loss": 0.5225, "step": 3543 }, { "epoch": 2.283505154639175, "grad_norm": 0.5792542099952698, "learning_rate": 1.6395151524536696e-06, "loss": 0.5233, "step": 3544 }, { "epoch": 2.2841494845360826, "grad_norm": 0.5657278299331665, "learning_rate": 1.6367401622627737e-06, "loss": 0.51, "step": 3545 }, { "epoch": 2.2847938144329896, "grad_norm": 0.5379877090454102, "learning_rate": 1.6339670628131327e-06, "loss": 0.5223, "step": 3546 }, { "epoch": 2.285438144329897, "grad_norm": 0.5460789203643799, "learning_rate": 1.631195855663714e-06, "loss": 0.5118, "step": 3547 }, { "epoch": 2.286082474226804, "grad_norm": 0.6452895998954773, "learning_rate": 1.628426542372422e-06, "loss": 0.5132, "step": 3548 }, { "epoch": 2.2867268041237114, "grad_norm": 0.5368199348449707, "learning_rate": 1.6256591244960967e-06, "loss": 0.5048, "step": 3549 }, { "epoch": 2.2873711340206184, "grad_norm": 0.5997890830039978, "learning_rate": 1.6228936035905096e-06, "loss": 0.5342, "step": 3550 }, { "epoch": 2.288015463917526, "grad_norm": 0.5628935098648071, "learning_rate": 1.6201299812103682e-06, "loss": 0.5131, "step": 3551 }, { "epoch": 2.288659793814433, "grad_norm": 0.5648255944252014, "learning_rate": 1.6173682589093153e-06, "loss": 0.5108, "step": 3552 }, { "epoch": 2.2893041237113403, "grad_norm": 0.5763730406761169, "learning_rate": 1.614608438239919e-06, "loss": 0.5068, "step": 3553 }, { "epoch": 2.2899484536082473, "grad_norm": 0.5420230627059937, "learning_rate": 1.6118505207536838e-06, "loss": 0.5053, "step": 3554 }, { "epoch": 2.2905927835051547, "grad_norm": 0.5796075463294983, "learning_rate": 1.609094508001044e-06, "loss": 0.5235, "step": 3555 }, { "epoch": 2.2912371134020617, "grad_norm": 0.5570102334022522, "learning_rate": 1.6063404015313583e-06, "loss": 0.4975, "step": 3556 }, { "epoch": 2.291881443298969, "grad_norm": 0.6034559607505798, "learning_rate": 1.603588202892919e-06, "loss": 0.5034, "step": 3557 }, { "epoch": 2.292525773195876, "grad_norm": 0.579647421836853, "learning_rate": 1.6008379136329443e-06, "loss": 0.5114, "step": 3558 }, { "epoch": 2.2931701030927836, "grad_norm": 0.5557237863540649, "learning_rate": 1.5980895352975773e-06, "loss": 0.5203, "step": 3559 }, { "epoch": 2.2938144329896906, "grad_norm": 0.5770649313926697, "learning_rate": 1.5953430694318884e-06, "loss": 0.5074, "step": 3560 }, { "epoch": 2.294458762886598, "grad_norm": 0.5313156843185425, "learning_rate": 1.5925985175798741e-06, "loss": 0.5089, "step": 3561 }, { "epoch": 2.295103092783505, "grad_norm": 0.5332927703857422, "learning_rate": 1.5898558812844512e-06, "loss": 0.5177, "step": 3562 }, { "epoch": 2.2957474226804124, "grad_norm": 0.5441436767578125, "learning_rate": 1.587115162087462e-06, "loss": 0.5047, "step": 3563 }, { "epoch": 2.2963917525773194, "grad_norm": 0.5344541072845459, "learning_rate": 1.5843763615296725e-06, "loss": 0.5088, "step": 3564 }, { "epoch": 2.297036082474227, "grad_norm": 0.533200204372406, "learning_rate": 1.5816394811507668e-06, "loss": 0.5094, "step": 3565 }, { "epoch": 2.297680412371134, "grad_norm": 0.616710901260376, "learning_rate": 1.578904522489354e-06, "loss": 0.4951, "step": 3566 }, { "epoch": 2.2983247422680413, "grad_norm": 0.5368475317955017, "learning_rate": 1.5761714870829563e-06, "loss": 0.5084, "step": 3567 }, { "epoch": 2.2989690721649483, "grad_norm": 0.5656040906906128, "learning_rate": 1.57344037646802e-06, "loss": 0.503, "step": 3568 }, { "epoch": 2.2996134020618557, "grad_norm": 0.6240822672843933, "learning_rate": 1.5707111921799096e-06, "loss": 0.5025, "step": 3569 }, { "epoch": 2.3002577319587627, "grad_norm": 0.6051507592201233, "learning_rate": 1.5679839357529015e-06, "loss": 0.5069, "step": 3570 }, { "epoch": 2.30090206185567, "grad_norm": 0.4970721900463104, "learning_rate": 1.5652586087201926e-06, "loss": 0.5094, "step": 3571 }, { "epoch": 2.301546391752577, "grad_norm": 0.6566011309623718, "learning_rate": 1.5625352126138966e-06, "loss": 0.5037, "step": 3572 }, { "epoch": 2.3021907216494846, "grad_norm": 0.5913017988204956, "learning_rate": 1.5598137489650356e-06, "loss": 0.5183, "step": 3573 }, { "epoch": 2.3028350515463916, "grad_norm": 0.5657615661621094, "learning_rate": 1.5570942193035527e-06, "loss": 0.5077, "step": 3574 }, { "epoch": 2.303479381443299, "grad_norm": 0.5589684844017029, "learning_rate": 1.554376625158296e-06, "loss": 0.5274, "step": 3575 }, { "epoch": 2.304123711340206, "grad_norm": 0.5090001821517944, "learning_rate": 1.5516609680570316e-06, "loss": 0.5211, "step": 3576 }, { "epoch": 2.3047680412371134, "grad_norm": 0.574367105960846, "learning_rate": 1.5489472495264357e-06, "loss": 0.5147, "step": 3577 }, { "epoch": 2.305412371134021, "grad_norm": 0.5193124413490295, "learning_rate": 1.5462354710920913e-06, "loss": 0.5112, "step": 3578 }, { "epoch": 2.306056701030928, "grad_norm": 0.6310102343559265, "learning_rate": 1.5435256342784938e-06, "loss": 0.4852, "step": 3579 }, { "epoch": 2.306701030927835, "grad_norm": 0.5716999173164368, "learning_rate": 1.540817740609048e-06, "loss": 0.5167, "step": 3580 }, { "epoch": 2.3073453608247423, "grad_norm": 0.6648568511009216, "learning_rate": 1.5381117916060612e-06, "loss": 0.5085, "step": 3581 }, { "epoch": 2.3079896907216497, "grad_norm": 0.7128541469573975, "learning_rate": 1.5354077887907525e-06, "loss": 0.5148, "step": 3582 }, { "epoch": 2.3086340206185567, "grad_norm": 0.5870909094810486, "learning_rate": 1.5327057336832457e-06, "loss": 0.5214, "step": 3583 }, { "epoch": 2.3092783505154637, "grad_norm": 0.644723653793335, "learning_rate": 1.5300056278025687e-06, "loss": 0.4837, "step": 3584 }, { "epoch": 2.309922680412371, "grad_norm": 0.6553997993469238, "learning_rate": 1.5273074726666553e-06, "loss": 0.501, "step": 3585 }, { "epoch": 2.3105670103092786, "grad_norm": 0.5470686554908752, "learning_rate": 1.5246112697923389e-06, "loss": 0.5174, "step": 3586 }, { "epoch": 2.3112113402061856, "grad_norm": 0.634038507938385, "learning_rate": 1.5219170206953593e-06, "loss": 0.4972, "step": 3587 }, { "epoch": 2.3118556701030926, "grad_norm": 0.653601348400116, "learning_rate": 1.5192247268903575e-06, "loss": 0.4934, "step": 3588 }, { "epoch": 2.3125, "grad_norm": 0.5841736197471619, "learning_rate": 1.516534389890872e-06, "loss": 0.503, "step": 3589 }, { "epoch": 2.3131443298969074, "grad_norm": 0.6666666865348816, "learning_rate": 1.5138460112093446e-06, "loss": 0.533, "step": 3590 }, { "epoch": 2.3137886597938144, "grad_norm": 0.7628197073936462, "learning_rate": 1.5111595923571171e-06, "loss": 0.5029, "step": 3591 }, { "epoch": 2.3144329896907214, "grad_norm": 0.6202483773231506, "learning_rate": 1.5084751348444233e-06, "loss": 0.5242, "step": 3592 }, { "epoch": 2.315077319587629, "grad_norm": 0.5535804629325867, "learning_rate": 1.5057926401804017e-06, "loss": 0.527, "step": 3593 }, { "epoch": 2.3157216494845363, "grad_norm": 0.679093599319458, "learning_rate": 1.5031121098730843e-06, "loss": 0.5043, "step": 3594 }, { "epoch": 2.3163659793814433, "grad_norm": 0.6358084678649902, "learning_rate": 1.5004335454293973e-06, "loss": 0.5099, "step": 3595 }, { "epoch": 2.3170103092783503, "grad_norm": 0.5903222560882568, "learning_rate": 1.4977569483551634e-06, "loss": 0.4939, "step": 3596 }, { "epoch": 2.3176546391752577, "grad_norm": 0.6449636220932007, "learning_rate": 1.4950823201551012e-06, "loss": 0.4981, "step": 3597 }, { "epoch": 2.318298969072165, "grad_norm": 0.6159698963165283, "learning_rate": 1.4924096623328172e-06, "loss": 0.5333, "step": 3598 }, { "epoch": 2.318943298969072, "grad_norm": 0.5667354464530945, "learning_rate": 1.489738976390815e-06, "loss": 0.4957, "step": 3599 }, { "epoch": 2.319587628865979, "grad_norm": 0.7339819669723511, "learning_rate": 1.4870702638304874e-06, "loss": 0.5083, "step": 3600 }, { "epoch": 2.3202319587628866, "grad_norm": 0.6784106492996216, "learning_rate": 1.4844035261521189e-06, "loss": 0.515, "step": 3601 }, { "epoch": 2.320876288659794, "grad_norm": 0.564491868019104, "learning_rate": 1.4817387648548847e-06, "loss": 0.518, "step": 3602 }, { "epoch": 2.321520618556701, "grad_norm": 0.657211422920227, "learning_rate": 1.479075981436845e-06, "loss": 0.4988, "step": 3603 }, { "epoch": 2.3221649484536084, "grad_norm": 0.7521808743476868, "learning_rate": 1.476415177394951e-06, "loss": 0.5139, "step": 3604 }, { "epoch": 2.3228092783505154, "grad_norm": 0.5906467437744141, "learning_rate": 1.4737563542250427e-06, "loss": 0.5248, "step": 3605 }, { "epoch": 2.323453608247423, "grad_norm": 0.6515675783157349, "learning_rate": 1.471099513421842e-06, "loss": 0.5056, "step": 3606 }, { "epoch": 2.32409793814433, "grad_norm": 0.6868390440940857, "learning_rate": 1.46844465647896e-06, "loss": 0.5201, "step": 3607 }, { "epoch": 2.3247422680412373, "grad_norm": 0.6122994422912598, "learning_rate": 1.4657917848888925e-06, "loss": 0.5163, "step": 3608 }, { "epoch": 2.3253865979381443, "grad_norm": 0.7487451434135437, "learning_rate": 1.4631409001430157e-06, "loss": 0.4972, "step": 3609 }, { "epoch": 2.3260309278350517, "grad_norm": 0.6142949461936951, "learning_rate": 1.4604920037315946e-06, "loss": 0.5219, "step": 3610 }, { "epoch": 2.3266752577319587, "grad_norm": 0.5664282441139221, "learning_rate": 1.457845097143769e-06, "loss": 0.4981, "step": 3611 }, { "epoch": 2.327319587628866, "grad_norm": 0.7842673063278198, "learning_rate": 1.4552001818675665e-06, "loss": 0.5195, "step": 3612 }, { "epoch": 2.327963917525773, "grad_norm": 0.6869465708732605, "learning_rate": 1.4525572593898945e-06, "loss": 0.5108, "step": 3613 }, { "epoch": 2.3286082474226806, "grad_norm": 0.5708756446838379, "learning_rate": 1.4499163311965342e-06, "loss": 0.5133, "step": 3614 }, { "epoch": 2.3292525773195876, "grad_norm": 0.7388863563537598, "learning_rate": 1.4472773987721528e-06, "loss": 0.5133, "step": 3615 }, { "epoch": 2.329896907216495, "grad_norm": 0.6431736350059509, "learning_rate": 1.4446404636002931e-06, "loss": 0.5146, "step": 3616 }, { "epoch": 2.330541237113402, "grad_norm": 0.5489908456802368, "learning_rate": 1.442005527163372e-06, "loss": 0.5136, "step": 3617 }, { "epoch": 2.3311855670103094, "grad_norm": 0.6794812679290771, "learning_rate": 1.4393725909426876e-06, "loss": 0.5003, "step": 3618 }, { "epoch": 2.3318298969072164, "grad_norm": 0.6035128831863403, "learning_rate": 1.4367416564184106e-06, "loss": 0.5039, "step": 3619 }, { "epoch": 2.332474226804124, "grad_norm": 0.6277344226837158, "learning_rate": 1.4341127250695868e-06, "loss": 0.5179, "step": 3620 }, { "epoch": 2.333118556701031, "grad_norm": 0.6802962422370911, "learning_rate": 1.4314857983741386e-06, "loss": 0.5126, "step": 3621 }, { "epoch": 2.3337628865979383, "grad_norm": 0.5636953115463257, "learning_rate": 1.4288608778088557e-06, "loss": 0.5097, "step": 3622 }, { "epoch": 2.3344072164948453, "grad_norm": 0.6171088814735413, "learning_rate": 1.4262379648494045e-06, "loss": 0.4949, "step": 3623 }, { "epoch": 2.3350515463917527, "grad_norm": 0.6198208928108215, "learning_rate": 1.4236170609703237e-06, "loss": 0.506, "step": 3624 }, { "epoch": 2.3356958762886597, "grad_norm": 0.5904018878936768, "learning_rate": 1.420998167645017e-06, "loss": 0.499, "step": 3625 }, { "epoch": 2.336340206185567, "grad_norm": 0.7402067184448242, "learning_rate": 1.4183812863457624e-06, "loss": 0.5384, "step": 3626 }, { "epoch": 2.336984536082474, "grad_norm": 0.6628352403640747, "learning_rate": 1.4157664185437081e-06, "loss": 0.5067, "step": 3627 }, { "epoch": 2.3376288659793816, "grad_norm": 0.6809840798377991, "learning_rate": 1.4131535657088642e-06, "loss": 0.5197, "step": 3628 }, { "epoch": 2.3382731958762886, "grad_norm": 0.6787606477737427, "learning_rate": 1.4105427293101138e-06, "loss": 0.5203, "step": 3629 }, { "epoch": 2.338917525773196, "grad_norm": 0.6317927837371826, "learning_rate": 1.4079339108152051e-06, "loss": 0.5215, "step": 3630 }, { "epoch": 2.339561855670103, "grad_norm": 0.6520801782608032, "learning_rate": 1.4053271116907485e-06, "loss": 0.5195, "step": 3631 }, { "epoch": 2.3402061855670104, "grad_norm": 0.6307938694953918, "learning_rate": 1.402722333402224e-06, "loss": 0.5172, "step": 3632 }, { "epoch": 2.3408505154639174, "grad_norm": 0.6017380952835083, "learning_rate": 1.4001195774139737e-06, "loss": 0.5107, "step": 3633 }, { "epoch": 2.341494845360825, "grad_norm": 0.5602443814277649, "learning_rate": 1.3975188451892003e-06, "loss": 0.4913, "step": 3634 }, { "epoch": 2.342139175257732, "grad_norm": 0.5815846920013428, "learning_rate": 1.3949201381899719e-06, "loss": 0.523, "step": 3635 }, { "epoch": 2.3427835051546393, "grad_norm": 0.5873664021492004, "learning_rate": 1.3923234578772177e-06, "loss": 0.5085, "step": 3636 }, { "epoch": 2.3434278350515463, "grad_norm": 0.542645275592804, "learning_rate": 1.389728805710726e-06, "loss": 0.4996, "step": 3637 }, { "epoch": 2.3440721649484537, "grad_norm": 0.5939317941665649, "learning_rate": 1.3871361831491486e-06, "loss": 0.4884, "step": 3638 }, { "epoch": 2.3447164948453607, "grad_norm": 0.6279857158660889, "learning_rate": 1.3845455916499895e-06, "loss": 0.5022, "step": 3639 }, { "epoch": 2.345360824742268, "grad_norm": 0.5618649125099182, "learning_rate": 1.3819570326696175e-06, "loss": 0.5143, "step": 3640 }, { "epoch": 2.346005154639175, "grad_norm": 0.6094964742660522, "learning_rate": 1.3793705076632568e-06, "loss": 0.4979, "step": 3641 }, { "epoch": 2.3466494845360826, "grad_norm": 0.5655139088630676, "learning_rate": 1.3767860180849851e-06, "loss": 0.4964, "step": 3642 }, { "epoch": 2.3472938144329896, "grad_norm": 0.554371178150177, "learning_rate": 1.3742035653877395e-06, "loss": 0.5, "step": 3643 }, { "epoch": 2.347938144329897, "grad_norm": 0.5628513693809509, "learning_rate": 1.371623151023313e-06, "loss": 0.5122, "step": 3644 }, { "epoch": 2.348582474226804, "grad_norm": 0.5565637350082397, "learning_rate": 1.369044776442347e-06, "loss": 0.4949, "step": 3645 }, { "epoch": 2.3492268041237114, "grad_norm": 0.5739715695381165, "learning_rate": 1.3664684430943431e-06, "loss": 0.5218, "step": 3646 }, { "epoch": 2.3498711340206184, "grad_norm": 0.5961542129516602, "learning_rate": 1.3638941524276489e-06, "loss": 0.5086, "step": 3647 }, { "epoch": 2.350515463917526, "grad_norm": 0.5765420794487, "learning_rate": 1.3613219058894678e-06, "loss": 0.5085, "step": 3648 }, { "epoch": 2.351159793814433, "grad_norm": 0.5334039330482483, "learning_rate": 1.3587517049258548e-06, "loss": 0.5206, "step": 3649 }, { "epoch": 2.3518041237113403, "grad_norm": 0.5835526585578918, "learning_rate": 1.35618355098171e-06, "loss": 0.5004, "step": 3650 }, { "epoch": 2.3524484536082473, "grad_norm": 0.5902984142303467, "learning_rate": 1.3536174455007878e-06, "loss": 0.4972, "step": 3651 }, { "epoch": 2.3530927835051547, "grad_norm": 0.6169564127922058, "learning_rate": 1.3510533899256906e-06, "loss": 0.5155, "step": 3652 }, { "epoch": 2.3537371134020617, "grad_norm": 0.6197267174720764, "learning_rate": 1.3484913856978627e-06, "loss": 0.4981, "step": 3653 }, { "epoch": 2.354381443298969, "grad_norm": 0.6901814341545105, "learning_rate": 1.3459314342576024e-06, "loss": 0.5176, "step": 3654 }, { "epoch": 2.355025773195876, "grad_norm": 0.558114230632782, "learning_rate": 1.34337353704405e-06, "loss": 0.4992, "step": 3655 }, { "epoch": 2.3556701030927836, "grad_norm": 0.5637655854225159, "learning_rate": 1.3408176954951912e-06, "loss": 0.5186, "step": 3656 }, { "epoch": 2.3563144329896906, "grad_norm": 0.686194658279419, "learning_rate": 1.3382639110478595e-06, "loss": 0.5276, "step": 3657 }, { "epoch": 2.356958762886598, "grad_norm": 0.5993351936340332, "learning_rate": 1.3357121851377259e-06, "loss": 0.5039, "step": 3658 }, { "epoch": 2.357603092783505, "grad_norm": 0.5427981615066528, "learning_rate": 1.3331625191993086e-06, "loss": 0.5036, "step": 3659 }, { "epoch": 2.3582474226804124, "grad_norm": 0.6961241960525513, "learning_rate": 1.3306149146659681e-06, "loss": 0.5079, "step": 3660 }, { "epoch": 2.3588917525773194, "grad_norm": 0.6130250692367554, "learning_rate": 1.3280693729699018e-06, "loss": 0.5227, "step": 3661 }, { "epoch": 2.359536082474227, "grad_norm": 0.5750904679298401, "learning_rate": 1.3255258955421507e-06, "loss": 0.5218, "step": 3662 }, { "epoch": 2.360180412371134, "grad_norm": 0.5873813033103943, "learning_rate": 1.322984483812597e-06, "loss": 0.5054, "step": 3663 }, { "epoch": 2.3608247422680413, "grad_norm": 0.5072430372238159, "learning_rate": 1.3204451392099566e-06, "loss": 0.4926, "step": 3664 }, { "epoch": 2.3614690721649483, "grad_norm": 0.5889444947242737, "learning_rate": 1.317907863161787e-06, "loss": 0.526, "step": 3665 }, { "epoch": 2.3621134020618557, "grad_norm": 0.5165184736251831, "learning_rate": 1.315372657094483e-06, "loss": 0.5273, "step": 3666 }, { "epoch": 2.3627577319587627, "grad_norm": 0.5640048384666443, "learning_rate": 1.3128395224332723e-06, "loss": 0.5119, "step": 3667 }, { "epoch": 2.36340206185567, "grad_norm": 0.5346174240112305, "learning_rate": 1.310308460602222e-06, "loss": 0.5257, "step": 3668 }, { "epoch": 2.364046391752577, "grad_norm": 0.5909415483474731, "learning_rate": 1.3077794730242327e-06, "loss": 0.5003, "step": 3669 }, { "epoch": 2.3646907216494846, "grad_norm": 0.5503520369529724, "learning_rate": 1.3052525611210366e-06, "loss": 0.5038, "step": 3670 }, { "epoch": 2.3653350515463916, "grad_norm": 0.5523881316184998, "learning_rate": 1.3027277263132021e-06, "loss": 0.4846, "step": 3671 }, { "epoch": 2.365979381443299, "grad_norm": 0.5816909074783325, "learning_rate": 1.300204970020128e-06, "loss": 0.517, "step": 3672 }, { "epoch": 2.366623711340206, "grad_norm": 0.5465254783630371, "learning_rate": 1.2976842936600464e-06, "loss": 0.4775, "step": 3673 }, { "epoch": 2.3672680412371134, "grad_norm": 0.5887964963912964, "learning_rate": 1.29516569865002e-06, "loss": 0.4984, "step": 3674 }, { "epoch": 2.367912371134021, "grad_norm": 0.6285538673400879, "learning_rate": 1.292649186405937e-06, "loss": 0.5045, "step": 3675 }, { "epoch": 2.368556701030928, "grad_norm": 0.5346565842628479, "learning_rate": 1.29013475834252e-06, "loss": 0.5116, "step": 3676 }, { "epoch": 2.369201030927835, "grad_norm": 0.5765019655227661, "learning_rate": 1.2876224158733203e-06, "loss": 0.4908, "step": 3677 }, { "epoch": 2.3698453608247423, "grad_norm": 0.5487481951713562, "learning_rate": 1.2851121604107102e-06, "loss": 0.5087, "step": 3678 }, { "epoch": 2.3704896907216497, "grad_norm": 0.5605402588844299, "learning_rate": 1.2826039933658962e-06, "loss": 0.5092, "step": 3679 }, { "epoch": 2.3711340206185567, "grad_norm": 0.5584040284156799, "learning_rate": 1.2800979161489074e-06, "loss": 0.4883, "step": 3680 }, { "epoch": 2.3717783505154637, "grad_norm": 0.6176328659057617, "learning_rate": 1.2775939301685963e-06, "loss": 0.518, "step": 3681 }, { "epoch": 2.372422680412371, "grad_norm": 0.5237749814987183, "learning_rate": 1.2750920368326442e-06, "loss": 0.5216, "step": 3682 }, { "epoch": 2.3730670103092786, "grad_norm": 0.5446197390556335, "learning_rate": 1.2725922375475507e-06, "loss": 0.5148, "step": 3683 }, { "epoch": 2.3737113402061856, "grad_norm": 0.6159966588020325, "learning_rate": 1.2700945337186428e-06, "loss": 0.5183, "step": 3684 }, { "epoch": 2.3743556701030926, "grad_norm": 0.5705289244651794, "learning_rate": 1.2675989267500683e-06, "loss": 0.5143, "step": 3685 }, { "epoch": 2.375, "grad_norm": 0.6166693568229675, "learning_rate": 1.265105418044793e-06, "loss": 0.5156, "step": 3686 }, { "epoch": 2.3756443298969074, "grad_norm": 0.586311936378479, "learning_rate": 1.2626140090046073e-06, "loss": 0.5163, "step": 3687 }, { "epoch": 2.3762886597938144, "grad_norm": 0.5999664664268494, "learning_rate": 1.260124701030121e-06, "loss": 0.4921, "step": 3688 }, { "epoch": 2.3769329896907214, "grad_norm": 0.6165964603424072, "learning_rate": 1.2576374955207582e-06, "loss": 0.5048, "step": 3689 }, { "epoch": 2.377577319587629, "grad_norm": 0.5410439968109131, "learning_rate": 1.2551523938747657e-06, "loss": 0.5049, "step": 3690 }, { "epoch": 2.3782216494845363, "grad_norm": 0.5588263869285583, "learning_rate": 1.2526693974892057e-06, "loss": 0.5032, "step": 3691 }, { "epoch": 2.3788659793814433, "grad_norm": 0.604430079460144, "learning_rate": 1.2501885077599574e-06, "loss": 0.5007, "step": 3692 }, { "epoch": 2.3795103092783503, "grad_norm": 0.5743597745895386, "learning_rate": 1.2477097260817174e-06, "loss": 0.4952, "step": 3693 }, { "epoch": 2.3801546391752577, "grad_norm": 0.5411279797554016, "learning_rate": 1.2452330538479917e-06, "loss": 0.5002, "step": 3694 }, { "epoch": 2.380798969072165, "grad_norm": 0.5240528583526611, "learning_rate": 1.2427584924511054e-06, "loss": 0.4976, "step": 3695 }, { "epoch": 2.381443298969072, "grad_norm": 0.5206605792045593, "learning_rate": 1.2402860432821972e-06, "loss": 0.5117, "step": 3696 }, { "epoch": 2.382087628865979, "grad_norm": 0.554923415184021, "learning_rate": 1.2378157077312143e-06, "loss": 0.5158, "step": 3697 }, { "epoch": 2.3827319587628866, "grad_norm": 0.5524095892906189, "learning_rate": 1.2353474871869186e-06, "loss": 0.5005, "step": 3698 }, { "epoch": 2.383376288659794, "grad_norm": 0.5512387156486511, "learning_rate": 1.2328813830368847e-06, "loss": 0.4898, "step": 3699 }, { "epoch": 2.384020618556701, "grad_norm": 0.5201281309127808, "learning_rate": 1.2304173966674921e-06, "loss": 0.5278, "step": 3700 }, { "epoch": 2.3846649484536084, "grad_norm": 0.5235637426376343, "learning_rate": 1.2279555294639334e-06, "loss": 0.5126, "step": 3701 }, { "epoch": 2.3853092783505154, "grad_norm": 0.5240121483802795, "learning_rate": 1.2254957828102121e-06, "loss": 0.497, "step": 3702 }, { "epoch": 2.385953608247423, "grad_norm": 0.5414180159568787, "learning_rate": 1.2230381580891332e-06, "loss": 0.5082, "step": 3703 }, { "epoch": 2.38659793814433, "grad_norm": 0.5364102721214294, "learning_rate": 1.2205826566823136e-06, "loss": 0.5341, "step": 3704 }, { "epoch": 2.3872422680412373, "grad_norm": 0.529292106628418, "learning_rate": 1.2181292799701766e-06, "loss": 0.5022, "step": 3705 }, { "epoch": 2.3878865979381443, "grad_norm": 0.5327604413032532, "learning_rate": 1.2156780293319476e-06, "loss": 0.5069, "step": 3706 }, { "epoch": 2.3885309278350517, "grad_norm": 0.5826494693756104, "learning_rate": 1.2132289061456597e-06, "loss": 0.4967, "step": 3707 }, { "epoch": 2.3891752577319587, "grad_norm": 0.5423866510391235, "learning_rate": 1.2107819117881498e-06, "loss": 0.5127, "step": 3708 }, { "epoch": 2.389819587628866, "grad_norm": 0.5802697539329529, "learning_rate": 1.208337047635057e-06, "loss": 0.5076, "step": 3709 }, { "epoch": 2.390463917525773, "grad_norm": 0.5579074025154114, "learning_rate": 1.2058943150608244e-06, "loss": 0.5085, "step": 3710 }, { "epoch": 2.3911082474226806, "grad_norm": 0.5333154201507568, "learning_rate": 1.2034537154386933e-06, "loss": 0.5035, "step": 3711 }, { "epoch": 2.3917525773195876, "grad_norm": 0.5889834761619568, "learning_rate": 1.2010152501407096e-06, "loss": 0.5164, "step": 3712 }, { "epoch": 2.392396907216495, "grad_norm": 0.5261783003807068, "learning_rate": 1.1985789205377196e-06, "loss": 0.5087, "step": 3713 }, { "epoch": 2.393041237113402, "grad_norm": 0.571392297744751, "learning_rate": 1.1961447279993644e-06, "loss": 0.495, "step": 3714 }, { "epoch": 2.3936855670103094, "grad_norm": 0.6070022583007812, "learning_rate": 1.1937126738940873e-06, "loss": 0.519, "step": 3715 }, { "epoch": 2.3943298969072164, "grad_norm": 0.5963649153709412, "learning_rate": 1.1912827595891313e-06, "loss": 0.5199, "step": 3716 }, { "epoch": 2.394974226804124, "grad_norm": 0.5196818709373474, "learning_rate": 1.188854986450531e-06, "loss": 0.5129, "step": 3717 }, { "epoch": 2.395618556701031, "grad_norm": 0.5775653719902039, "learning_rate": 1.1864293558431229e-06, "loss": 0.494, "step": 3718 }, { "epoch": 2.3962628865979383, "grad_norm": 0.5872161388397217, "learning_rate": 1.1840058691305329e-06, "loss": 0.4987, "step": 3719 }, { "epoch": 2.3969072164948453, "grad_norm": 0.5491116642951965, "learning_rate": 1.1815845276751875e-06, "loss": 0.5251, "step": 3720 }, { "epoch": 2.3975515463917527, "grad_norm": 0.6473113298416138, "learning_rate": 1.1791653328383052e-06, "loss": 0.5091, "step": 3721 }, { "epoch": 2.3981958762886597, "grad_norm": 0.6531003713607788, "learning_rate": 1.1767482859798956e-06, "loss": 0.505, "step": 3722 }, { "epoch": 2.398840206185567, "grad_norm": 0.5314717888832092, "learning_rate": 1.1743333884587631e-06, "loss": 0.5095, "step": 3723 }, { "epoch": 2.399484536082474, "grad_norm": 0.5916634798049927, "learning_rate": 1.171920641632505e-06, "loss": 0.5095, "step": 3724 }, { "epoch": 2.4001288659793816, "grad_norm": 0.5959480404853821, "learning_rate": 1.1695100468575048e-06, "loss": 0.5085, "step": 3725 }, { "epoch": 2.4007731958762886, "grad_norm": 0.5882223844528198, "learning_rate": 1.1671016054889407e-06, "loss": 0.5314, "step": 3726 }, { "epoch": 2.401417525773196, "grad_norm": 0.603432834148407, "learning_rate": 1.1646953188807781e-06, "loss": 0.5141, "step": 3727 }, { "epoch": 2.402061855670103, "grad_norm": 0.6084831953048706, "learning_rate": 1.1622911883857717e-06, "loss": 0.506, "step": 3728 }, { "epoch": 2.4027061855670104, "grad_norm": 0.5965844988822937, "learning_rate": 1.1598892153554663e-06, "loss": 0.5067, "step": 3729 }, { "epoch": 2.4033505154639174, "grad_norm": 0.6001335978507996, "learning_rate": 1.1574894011401883e-06, "loss": 0.4913, "step": 3730 }, { "epoch": 2.403994845360825, "grad_norm": 0.6235634684562683, "learning_rate": 1.155091747089055e-06, "loss": 0.5493, "step": 3731 }, { "epoch": 2.404639175257732, "grad_norm": 0.6093147397041321, "learning_rate": 1.1526962545499693e-06, "loss": 0.5126, "step": 3732 }, { "epoch": 2.4052835051546393, "grad_norm": 0.6109543442726135, "learning_rate": 1.1503029248696147e-06, "loss": 0.5208, "step": 3733 }, { "epoch": 2.4059278350515463, "grad_norm": 0.6125878691673279, "learning_rate": 1.1479117593934636e-06, "loss": 0.493, "step": 3734 }, { "epoch": 2.4065721649484537, "grad_norm": 0.6256527304649353, "learning_rate": 1.1455227594657713e-06, "loss": 0.5182, "step": 3735 }, { "epoch": 2.4072164948453607, "grad_norm": 0.5605131983757019, "learning_rate": 1.1431359264295717e-06, "loss": 0.5255, "step": 3736 }, { "epoch": 2.407860824742268, "grad_norm": 0.622406542301178, "learning_rate": 1.1407512616266842e-06, "loss": 0.5061, "step": 3737 }, { "epoch": 2.408505154639175, "grad_norm": 0.6294195055961609, "learning_rate": 1.1383687663977095e-06, "loss": 0.5201, "step": 3738 }, { "epoch": 2.4091494845360826, "grad_norm": 0.6082509756088257, "learning_rate": 1.135988442082025e-06, "loss": 0.5174, "step": 3739 }, { "epoch": 2.4097938144329896, "grad_norm": 0.657049298286438, "learning_rate": 1.1336102900177903e-06, "loss": 0.5306, "step": 3740 }, { "epoch": 2.410438144329897, "grad_norm": 0.6661145687103271, "learning_rate": 1.1312343115419466e-06, "loss": 0.5092, "step": 3741 }, { "epoch": 2.411082474226804, "grad_norm": 0.6149865984916687, "learning_rate": 1.1288605079902053e-06, "loss": 0.5017, "step": 3742 }, { "epoch": 2.4117268041237114, "grad_norm": 0.568623960018158, "learning_rate": 1.126488880697063e-06, "loss": 0.5155, "step": 3743 }, { "epoch": 2.4123711340206184, "grad_norm": 0.654058039188385, "learning_rate": 1.1241194309957892e-06, "loss": 0.4748, "step": 3744 }, { "epoch": 2.413015463917526, "grad_norm": 0.6128855347633362, "learning_rate": 1.1217521602184296e-06, "loss": 0.5355, "step": 3745 }, { "epoch": 2.413659793814433, "grad_norm": 0.5998807549476624, "learning_rate": 1.1193870696958058e-06, "loss": 0.5083, "step": 3746 }, { "epoch": 2.4143041237113403, "grad_norm": 0.6120300889015198, "learning_rate": 1.1170241607575116e-06, "loss": 0.4964, "step": 3747 }, { "epoch": 2.4149484536082473, "grad_norm": 0.5867307186126709, "learning_rate": 1.1146634347319163e-06, "loss": 0.5478, "step": 3748 }, { "epoch": 2.4155927835051547, "grad_norm": 0.616269052028656, "learning_rate": 1.1123048929461628e-06, "loss": 0.4936, "step": 3749 }, { "epoch": 2.4162371134020617, "grad_norm": 0.6242615580558777, "learning_rate": 1.1099485367261625e-06, "loss": 0.5184, "step": 3750 }, { "epoch": 2.416881443298969, "grad_norm": 0.5674960613250732, "learning_rate": 1.1075943673966022e-06, "loss": 0.5254, "step": 3751 }, { "epoch": 2.417525773195876, "grad_norm": 0.6007330417633057, "learning_rate": 1.1052423862809376e-06, "loss": 0.5175, "step": 3752 }, { "epoch": 2.4181701030927836, "grad_norm": 0.6784449219703674, "learning_rate": 1.1028925947013925e-06, "loss": 0.5159, "step": 3753 }, { "epoch": 2.4188144329896906, "grad_norm": 0.6789349317550659, "learning_rate": 1.100544993978962e-06, "loss": 0.5162, "step": 3754 }, { "epoch": 2.419458762886598, "grad_norm": 0.5989450812339783, "learning_rate": 1.0981995854334109e-06, "loss": 0.5102, "step": 3755 }, { "epoch": 2.420103092783505, "grad_norm": 0.5569459199905396, "learning_rate": 1.0958563703832675e-06, "loss": 0.5148, "step": 3756 }, { "epoch": 2.4207474226804124, "grad_norm": 0.6118831634521484, "learning_rate": 1.0935153501458307e-06, "loss": 0.4999, "step": 3757 }, { "epoch": 2.4213917525773194, "grad_norm": 0.6238945722579956, "learning_rate": 1.091176526037162e-06, "loss": 0.4742, "step": 3758 }, { "epoch": 2.422036082474227, "grad_norm": 0.5490675568580627, "learning_rate": 1.088839899372091e-06, "loss": 0.5071, "step": 3759 }, { "epoch": 2.422680412371134, "grad_norm": 0.5706282258033752, "learning_rate": 1.0865054714642138e-06, "loss": 0.5214, "step": 3760 }, { "epoch": 2.4233247422680413, "grad_norm": 0.6553308963775635, "learning_rate": 1.084173243625884e-06, "loss": 0.5181, "step": 3761 }, { "epoch": 2.4239690721649483, "grad_norm": 0.48244157433509827, "learning_rate": 1.081843217168224e-06, "loss": 0.5272, "step": 3762 }, { "epoch": 2.4246134020618557, "grad_norm": 0.598346471786499, "learning_rate": 1.0795153934011176e-06, "loss": 0.5424, "step": 3763 }, { "epoch": 2.4252577319587627, "grad_norm": 0.5416222214698792, "learning_rate": 1.0771897736332083e-06, "loss": 0.5119, "step": 3764 }, { "epoch": 2.42590206185567, "grad_norm": 0.5278427004814148, "learning_rate": 1.0748663591719038e-06, "loss": 0.5168, "step": 3765 }, { "epoch": 2.426546391752577, "grad_norm": 0.6348315477371216, "learning_rate": 1.0725451513233676e-06, "loss": 0.4997, "step": 3766 }, { "epoch": 2.4271907216494846, "grad_norm": 0.5942865610122681, "learning_rate": 1.0702261513925265e-06, "loss": 0.4943, "step": 3767 }, { "epoch": 2.4278350515463916, "grad_norm": 0.5588842630386353, "learning_rate": 1.0679093606830658e-06, "loss": 0.5204, "step": 3768 }, { "epoch": 2.428479381443299, "grad_norm": 0.5648747682571411, "learning_rate": 1.0655947804974249e-06, "loss": 0.5053, "step": 3769 }, { "epoch": 2.429123711340206, "grad_norm": 0.5319509506225586, "learning_rate": 1.0632824121368051e-06, "loss": 0.5311, "step": 3770 }, { "epoch": 2.4297680412371134, "grad_norm": 0.5692919492721558, "learning_rate": 1.0609722569011639e-06, "loss": 0.5252, "step": 3771 }, { "epoch": 2.430412371134021, "grad_norm": 0.6064240336418152, "learning_rate": 1.0586643160892102e-06, "loss": 0.4867, "step": 3772 }, { "epoch": 2.431056701030928, "grad_norm": 0.559991717338562, "learning_rate": 1.056358590998413e-06, "loss": 0.4919, "step": 3773 }, { "epoch": 2.431701030927835, "grad_norm": 0.5816062092781067, "learning_rate": 1.0540550829249946e-06, "loss": 0.532, "step": 3774 }, { "epoch": 2.4323453608247423, "grad_norm": 0.5695400834083557, "learning_rate": 1.0517537931639276e-06, "loss": 0.5201, "step": 3775 }, { "epoch": 2.4329896907216497, "grad_norm": 0.5641035437583923, "learning_rate": 1.0494547230089413e-06, "loss": 0.5035, "step": 3776 }, { "epoch": 2.4336340206185567, "grad_norm": 0.574859082698822, "learning_rate": 1.0471578737525173e-06, "loss": 0.484, "step": 3777 }, { "epoch": 2.4342783505154637, "grad_norm": 0.5510883927345276, "learning_rate": 1.0448632466858843e-06, "loss": 0.4887, "step": 3778 }, { "epoch": 2.434922680412371, "grad_norm": 0.6045156717300415, "learning_rate": 1.0425708430990266e-06, "loss": 0.504, "step": 3779 }, { "epoch": 2.4355670103092786, "grad_norm": 0.5439822673797607, "learning_rate": 1.0402806642806779e-06, "loss": 0.5217, "step": 3780 }, { "epoch": 2.4362113402061856, "grad_norm": 0.5784657597541809, "learning_rate": 1.0379927115183153e-06, "loss": 0.5064, "step": 3781 }, { "epoch": 2.4368556701030926, "grad_norm": 0.5957522988319397, "learning_rate": 1.0357069860981756e-06, "loss": 0.5343, "step": 3782 }, { "epoch": 2.4375, "grad_norm": 0.5708243250846863, "learning_rate": 1.0334234893052314e-06, "loss": 0.5215, "step": 3783 }, { "epoch": 2.4381443298969074, "grad_norm": 0.5277881026268005, "learning_rate": 1.0311422224232103e-06, "loss": 0.5174, "step": 3784 }, { "epoch": 2.4387886597938144, "grad_norm": 0.6525755524635315, "learning_rate": 1.0288631867345856e-06, "loss": 0.5113, "step": 3785 }, { "epoch": 2.4394329896907214, "grad_norm": 0.5918504595756531, "learning_rate": 1.0265863835205709e-06, "loss": 0.4916, "step": 3786 }, { "epoch": 2.440077319587629, "grad_norm": 0.5804482698440552, "learning_rate": 1.0243118140611303e-06, "loss": 0.5157, "step": 3787 }, { "epoch": 2.4407216494845363, "grad_norm": 0.63120436668396, "learning_rate": 1.022039479634972e-06, "loss": 0.5253, "step": 3788 }, { "epoch": 2.4413659793814433, "grad_norm": 0.6293531060218811, "learning_rate": 1.0197693815195424e-06, "loss": 0.505, "step": 3789 }, { "epoch": 2.4420103092783503, "grad_norm": 0.5442696809768677, "learning_rate": 1.0175015209910372e-06, "loss": 0.5295, "step": 3790 }, { "epoch": 2.4426546391752577, "grad_norm": 0.5933384895324707, "learning_rate": 1.0152358993243915e-06, "loss": 0.4982, "step": 3791 }, { "epoch": 2.443298969072165, "grad_norm": 0.5718438625335693, "learning_rate": 1.0129725177932798e-06, "loss": 0.5009, "step": 3792 }, { "epoch": 2.443943298969072, "grad_norm": 0.5499704480171204, "learning_rate": 1.0107113776701206e-06, "loss": 0.4993, "step": 3793 }, { "epoch": 2.444587628865979, "grad_norm": 0.5409027934074402, "learning_rate": 1.0084524802260686e-06, "loss": 0.5174, "step": 3794 }, { "epoch": 2.4452319587628866, "grad_norm": 0.5673830509185791, "learning_rate": 1.0061958267310217e-06, "loss": 0.5221, "step": 3795 }, { "epoch": 2.445876288659794, "grad_norm": 0.5526725053787231, "learning_rate": 1.0039414184536161e-06, "loss": 0.5188, "step": 3796 }, { "epoch": 2.446520618556701, "grad_norm": 0.5670957565307617, "learning_rate": 1.0016892566612202e-06, "loss": 0.5265, "step": 3797 }, { "epoch": 2.4471649484536084, "grad_norm": 0.5748312473297119, "learning_rate": 9.994393426199456e-07, "loss": 0.5267, "step": 3798 }, { "epoch": 2.4478092783505154, "grad_norm": 0.6390751004219055, "learning_rate": 9.971916775946389e-07, "loss": 0.5077, "step": 3799 }, { "epoch": 2.448453608247423, "grad_norm": 0.6088677644729614, "learning_rate": 9.949462628488798e-07, "loss": 0.4794, "step": 3800 }, { "epoch": 2.44909793814433, "grad_norm": 0.6574788093566895, "learning_rate": 9.927030996449882e-07, "loss": 0.4744, "step": 3801 }, { "epoch": 2.4497422680412373, "grad_norm": 0.6105773448944092, "learning_rate": 9.904621892440108e-07, "loss": 0.5259, "step": 3802 }, { "epoch": 2.4503865979381443, "grad_norm": 0.5841469764709473, "learning_rate": 9.882235329057333e-07, "loss": 0.5184, "step": 3803 }, { "epoch": 2.4510309278350517, "grad_norm": 0.5694709420204163, "learning_rate": 9.859871318886743e-07, "loss": 0.5182, "step": 3804 }, { "epoch": 2.4516752577319587, "grad_norm": 0.6814290285110474, "learning_rate": 9.837529874500807e-07, "loss": 0.5246, "step": 3805 }, { "epoch": 2.452319587628866, "grad_norm": 0.6133627891540527, "learning_rate": 9.815211008459336e-07, "loss": 0.4945, "step": 3806 }, { "epoch": 2.452963917525773, "grad_norm": 0.6084442138671875, "learning_rate": 9.792914733309461e-07, "loss": 0.4961, "step": 3807 }, { "epoch": 2.4536082474226806, "grad_norm": 0.6770297288894653, "learning_rate": 9.770641061585573e-07, "loss": 0.5145, "step": 3808 }, { "epoch": 2.4542525773195876, "grad_norm": 0.6724525690078735, "learning_rate": 9.748390005809377e-07, "loss": 0.5009, "step": 3809 }, { "epoch": 2.454896907216495, "grad_norm": 0.5715776681900024, "learning_rate": 9.726161578489895e-07, "loss": 0.5058, "step": 3810 }, { "epoch": 2.455541237113402, "grad_norm": 0.5782635807991028, "learning_rate": 9.70395579212336e-07, "loss": 0.5088, "step": 3811 }, { "epoch": 2.4561855670103094, "grad_norm": 0.667339563369751, "learning_rate": 9.681772659193328e-07, "loss": 0.5157, "step": 3812 }, { "epoch": 2.4568298969072164, "grad_norm": 0.5397548079490662, "learning_rate": 9.659612192170625e-07, "loss": 0.5123, "step": 3813 }, { "epoch": 2.457474226804124, "grad_norm": 0.5965487957000732, "learning_rate": 9.63747440351328e-07, "loss": 0.5001, "step": 3814 }, { "epoch": 2.458118556701031, "grad_norm": 0.6401337385177612, "learning_rate": 9.615359305666633e-07, "loss": 0.494, "step": 3815 }, { "epoch": 2.4587628865979383, "grad_norm": 0.59635329246521, "learning_rate": 9.593266911063253e-07, "loss": 0.5119, "step": 3816 }, { "epoch": 2.4594072164948453, "grad_norm": 0.6034019589424133, "learning_rate": 9.57119723212289e-07, "loss": 0.5295, "step": 3817 }, { "epoch": 2.4600515463917527, "grad_norm": 0.6613209843635559, "learning_rate": 9.549150281252633e-07, "loss": 0.4976, "step": 3818 }, { "epoch": 2.4606958762886597, "grad_norm": 0.6171526312828064, "learning_rate": 9.527126070846681e-07, "loss": 0.526, "step": 3819 }, { "epoch": 2.461340206185567, "grad_norm": 0.6456401944160461, "learning_rate": 9.50512461328652e-07, "loss": 0.5163, "step": 3820 }, { "epoch": 2.461984536082474, "grad_norm": 0.6076686382293701, "learning_rate": 9.483145920940834e-07, "loss": 0.5115, "step": 3821 }, { "epoch": 2.4626288659793816, "grad_norm": 0.5861247777938843, "learning_rate": 9.461190006165472e-07, "loss": 0.5097, "step": 3822 }, { "epoch": 2.4632731958762886, "grad_norm": 0.631146252155304, "learning_rate": 9.439256881303521e-07, "loss": 0.502, "step": 3823 }, { "epoch": 2.463917525773196, "grad_norm": 0.581619143486023, "learning_rate": 9.417346558685248e-07, "loss": 0.5153, "step": 3824 }, { "epoch": 2.464561855670103, "grad_norm": 0.5456150770187378, "learning_rate": 9.39545905062807e-07, "loss": 0.5065, "step": 3825 }, { "epoch": 2.4652061855670104, "grad_norm": 0.5867785215377808, "learning_rate": 9.373594369436611e-07, "loss": 0.5266, "step": 3826 }, { "epoch": 2.4658505154639174, "grad_norm": 0.5940548181533813, "learning_rate": 9.351752527402663e-07, "loss": 0.5108, "step": 3827 }, { "epoch": 2.466494845360825, "grad_norm": 0.5774127840995789, "learning_rate": 9.329933536805152e-07, "loss": 0.4917, "step": 3828 }, { "epoch": 2.467139175257732, "grad_norm": 0.5585120320320129, "learning_rate": 9.308137409910195e-07, "loss": 0.5085, "step": 3829 }, { "epoch": 2.4677835051546393, "grad_norm": 0.6161280274391174, "learning_rate": 9.286364158971001e-07, "loss": 0.5241, "step": 3830 }, { "epoch": 2.4684278350515463, "grad_norm": 0.6318204998970032, "learning_rate": 9.264613796227973e-07, "loss": 0.522, "step": 3831 }, { "epoch": 2.4690721649484537, "grad_norm": 0.5889613032341003, "learning_rate": 9.242886333908635e-07, "loss": 0.5242, "step": 3832 }, { "epoch": 2.4697164948453607, "grad_norm": 0.5779937505722046, "learning_rate": 9.221181784227601e-07, "loss": 0.5139, "step": 3833 }, { "epoch": 2.470360824742268, "grad_norm": 0.6017043590545654, "learning_rate": 9.199500159386648e-07, "loss": 0.5029, "step": 3834 }, { "epoch": 2.471005154639175, "grad_norm": 0.6132926344871521, "learning_rate": 9.177841471574639e-07, "loss": 0.5053, "step": 3835 }, { "epoch": 2.4716494845360826, "grad_norm": 0.6293132305145264, "learning_rate": 9.15620573296756e-07, "loss": 0.5285, "step": 3836 }, { "epoch": 2.4722938144329896, "grad_norm": 1.8485051393508911, "learning_rate": 9.134592955728494e-07, "loss": 0.5475, "step": 3837 }, { "epoch": 2.472938144329897, "grad_norm": 0.549220621585846, "learning_rate": 9.113003152007582e-07, "loss": 0.5195, "step": 3838 }, { "epoch": 2.473582474226804, "grad_norm": 0.567238986492157, "learning_rate": 9.091436333942088e-07, "loss": 0.525, "step": 3839 }, { "epoch": 2.4742268041237114, "grad_norm": 0.5855289101600647, "learning_rate": 9.069892513656364e-07, "loss": 0.5232, "step": 3840 }, { "epoch": 2.4748711340206184, "grad_norm": 0.5672393441200256, "learning_rate": 9.048371703261777e-07, "loss": 0.4825, "step": 3841 }, { "epoch": 2.475515463917526, "grad_norm": 0.6231303811073303, "learning_rate": 9.026873914856809e-07, "loss": 0.521, "step": 3842 }, { "epoch": 2.476159793814433, "grad_norm": 0.5949978828430176, "learning_rate": 9.005399160526995e-07, "loss": 0.5, "step": 3843 }, { "epoch": 2.4768041237113403, "grad_norm": 0.5666894912719727, "learning_rate": 8.983947452344882e-07, "loss": 0.5112, "step": 3844 }, { "epoch": 2.4774484536082473, "grad_norm": 0.6304309964179993, "learning_rate": 8.962518802370101e-07, "loss": 0.4993, "step": 3845 }, { "epoch": 2.4780927835051547, "grad_norm": 0.5822027921676636, "learning_rate": 8.941113222649328e-07, "loss": 0.4938, "step": 3846 }, { "epoch": 2.4787371134020617, "grad_norm": 0.4892144501209259, "learning_rate": 8.919730725216214e-07, "loss": 0.5328, "step": 3847 }, { "epoch": 2.479381443298969, "grad_norm": 0.5810314416885376, "learning_rate": 8.898371322091487e-07, "loss": 0.5054, "step": 3848 }, { "epoch": 2.480025773195876, "grad_norm": 0.537018895149231, "learning_rate": 8.877035025282882e-07, "loss": 0.5255, "step": 3849 }, { "epoch": 2.4806701030927836, "grad_norm": 0.5051756501197815, "learning_rate": 8.855721846785114e-07, "loss": 0.5272, "step": 3850 }, { "epoch": 2.4813144329896906, "grad_norm": 0.5710557103157043, "learning_rate": 8.834431798579935e-07, "loss": 0.491, "step": 3851 }, { "epoch": 2.481958762886598, "grad_norm": 0.6382042169570923, "learning_rate": 8.813164892636095e-07, "loss": 0.4828, "step": 3852 }, { "epoch": 2.482603092783505, "grad_norm": 0.5676485300064087, "learning_rate": 8.791921140909282e-07, "loss": 0.5231, "step": 3853 }, { "epoch": 2.4832474226804124, "grad_norm": 0.61455899477005, "learning_rate": 8.770700555342254e-07, "loss": 0.545, "step": 3854 }, { "epoch": 2.4838917525773194, "grad_norm": 0.589311420917511, "learning_rate": 8.749503147864669e-07, "loss": 0.5211, "step": 3855 }, { "epoch": 2.484536082474227, "grad_norm": 0.5711119771003723, "learning_rate": 8.728328930393188e-07, "loss": 0.5035, "step": 3856 }, { "epoch": 2.485180412371134, "grad_norm": 0.6448460221290588, "learning_rate": 8.707177914831443e-07, "loss": 0.5246, "step": 3857 }, { "epoch": 2.4858247422680413, "grad_norm": 0.641692578792572, "learning_rate": 8.686050113069988e-07, "loss": 0.5021, "step": 3858 }, { "epoch": 2.4864690721649483, "grad_norm": 0.5367250442504883, "learning_rate": 8.664945536986369e-07, "loss": 0.5364, "step": 3859 }, { "epoch": 2.4871134020618557, "grad_norm": 0.6428788900375366, "learning_rate": 8.643864198445051e-07, "loss": 0.4949, "step": 3860 }, { "epoch": 2.4877577319587627, "grad_norm": 0.6085548996925354, "learning_rate": 8.622806109297432e-07, "loss": 0.5144, "step": 3861 }, { "epoch": 2.48840206185567, "grad_norm": 0.6069763898849487, "learning_rate": 8.601771281381849e-07, "loss": 0.5112, "step": 3862 }, { "epoch": 2.489046391752577, "grad_norm": 0.5878574252128601, "learning_rate": 8.580759726523574e-07, "loss": 0.5215, "step": 3863 }, { "epoch": 2.4896907216494846, "grad_norm": 0.523068368434906, "learning_rate": 8.55977145653476e-07, "loss": 0.4883, "step": 3864 }, { "epoch": 2.4903350515463916, "grad_norm": 0.5458766222000122, "learning_rate": 8.538806483214512e-07, "loss": 0.514, "step": 3865 }, { "epoch": 2.490979381443299, "grad_norm": 0.5282140970230103, "learning_rate": 8.517864818348804e-07, "loss": 0.5064, "step": 3866 }, { "epoch": 2.491623711340206, "grad_norm": 0.5148192048072815, "learning_rate": 8.496946473710515e-07, "loss": 0.5309, "step": 3867 }, { "epoch": 2.4922680412371134, "grad_norm": 0.6347101330757141, "learning_rate": 8.476051461059454e-07, "loss": 0.5214, "step": 3868 }, { "epoch": 2.492912371134021, "grad_norm": 0.5246029496192932, "learning_rate": 8.455179792142237e-07, "loss": 0.5002, "step": 3869 }, { "epoch": 2.493556701030928, "grad_norm": 0.5842447876930237, "learning_rate": 8.434331478692414e-07, "loss": 0.5239, "step": 3870 }, { "epoch": 2.494201030927835, "grad_norm": 0.5680087208747864, "learning_rate": 8.413506532430393e-07, "loss": 0.5054, "step": 3871 }, { "epoch": 2.4948453608247423, "grad_norm": 0.5465587973594666, "learning_rate": 8.392704965063442e-07, "loss": 0.5061, "step": 3872 }, { "epoch": 2.4954896907216497, "grad_norm": 0.5676049590110779, "learning_rate": 8.3719267882857e-07, "loss": 0.5019, "step": 3873 }, { "epoch": 2.4961340206185567, "grad_norm": 0.6996015906333923, "learning_rate": 8.351172013778114e-07, "loss": 0.5183, "step": 3874 }, { "epoch": 2.4967783505154637, "grad_norm": 0.6185671091079712, "learning_rate": 8.330440653208516e-07, "loss": 0.4919, "step": 3875 }, { "epoch": 2.497422680412371, "grad_norm": 0.6525441408157349, "learning_rate": 8.309732718231578e-07, "loss": 0.5151, "step": 3876 }, { "epoch": 2.4980670103092786, "grad_norm": 0.6142886281013489, "learning_rate": 8.289048220488766e-07, "loss": 0.5197, "step": 3877 }, { "epoch": 2.4987113402061856, "grad_norm": 0.5682188868522644, "learning_rate": 8.268387171608394e-07, "loss": 0.5155, "step": 3878 }, { "epoch": 2.4993556701030926, "grad_norm": 0.6187595725059509, "learning_rate": 8.247749583205605e-07, "loss": 0.5065, "step": 3879 }, { "epoch": 2.5, "grad_norm": 0.5691089630126953, "learning_rate": 8.227135466882325e-07, "loss": 0.5106, "step": 3880 }, { "epoch": 2.5006443298969074, "grad_norm": 0.5265897512435913, "learning_rate": 8.206544834227298e-07, "loss": 0.5213, "step": 3881 }, { "epoch": 2.5012886597938144, "grad_norm": 0.598703920841217, "learning_rate": 8.185977696816088e-07, "loss": 0.5234, "step": 3882 }, { "epoch": 2.5019329896907214, "grad_norm": 0.5860509872436523, "learning_rate": 8.165434066211003e-07, "loss": 0.499, "step": 3883 }, { "epoch": 2.502577319587629, "grad_norm": 0.5417135953903198, "learning_rate": 8.144913953961175e-07, "loss": 0.5117, "step": 3884 }, { "epoch": 2.5032216494845363, "grad_norm": 0.5758655071258545, "learning_rate": 8.12441737160251e-07, "loss": 0.4995, "step": 3885 }, { "epoch": 2.5038659793814433, "grad_norm": 0.575588583946228, "learning_rate": 8.103944330657665e-07, "loss": 0.5219, "step": 3886 }, { "epoch": 2.5045103092783503, "grad_norm": 0.6110833287239075, "learning_rate": 8.083494842636074e-07, "loss": 0.5097, "step": 3887 }, { "epoch": 2.5051546391752577, "grad_norm": 0.5464273691177368, "learning_rate": 8.063068919033951e-07, "loss": 0.509, "step": 3888 }, { "epoch": 2.505798969072165, "grad_norm": 0.5607302188873291, "learning_rate": 8.042666571334206e-07, "loss": 0.5023, "step": 3889 }, { "epoch": 2.506443298969072, "grad_norm": 0.6271824836730957, "learning_rate": 8.022287811006585e-07, "loss": 0.5326, "step": 3890 }, { "epoch": 2.507087628865979, "grad_norm": 0.5376031994819641, "learning_rate": 8.001932649507482e-07, "loss": 0.5077, "step": 3891 }, { "epoch": 2.5077319587628866, "grad_norm": 0.566096305847168, "learning_rate": 7.981601098280073e-07, "loss": 0.5287, "step": 3892 }, { "epoch": 2.508376288659794, "grad_norm": 0.5815293788909912, "learning_rate": 7.961293168754269e-07, "loss": 0.4827, "step": 3893 }, { "epoch": 2.509020618556701, "grad_norm": 0.5996228456497192, "learning_rate": 7.941008872346661e-07, "loss": 0.5071, "step": 3894 }, { "epoch": 2.509664948453608, "grad_norm": 0.5353671908378601, "learning_rate": 7.920748220460583e-07, "loss": 0.5109, "step": 3895 }, { "epoch": 2.5103092783505154, "grad_norm": 0.5735422968864441, "learning_rate": 7.900511224486085e-07, "loss": 0.5245, "step": 3896 }, { "epoch": 2.510953608247423, "grad_norm": 0.5781648755073547, "learning_rate": 7.880297895799882e-07, "loss": 0.507, "step": 3897 }, { "epoch": 2.51159793814433, "grad_norm": 0.5998399257659912, "learning_rate": 7.860108245765418e-07, "loss": 0.5079, "step": 3898 }, { "epoch": 2.512242268041237, "grad_norm": 0.5813242197036743, "learning_rate": 7.839942285732827e-07, "loss": 0.5005, "step": 3899 }, { "epoch": 2.5128865979381443, "grad_norm": 0.5850886106491089, "learning_rate": 7.819800027038882e-07, "loss": 0.5104, "step": 3900 }, { "epoch": 2.5135309278350517, "grad_norm": 0.6935377717018127, "learning_rate": 7.799681481007087e-07, "loss": 0.54, "step": 3901 }, { "epoch": 2.5141752577319587, "grad_norm": 0.6019220352172852, "learning_rate": 7.779586658947574e-07, "loss": 0.508, "step": 3902 }, { "epoch": 2.5148195876288657, "grad_norm": 0.5452452301979065, "learning_rate": 7.759515572157156e-07, "loss": 0.5253, "step": 3903 }, { "epoch": 2.515463917525773, "grad_norm": 0.6262795329093933, "learning_rate": 7.739468231919311e-07, "loss": 0.5067, "step": 3904 }, { "epoch": 2.5161082474226806, "grad_norm": 0.5724849700927734, "learning_rate": 7.719444649504138e-07, "loss": 0.5284, "step": 3905 }, { "epoch": 2.5167525773195876, "grad_norm": 0.5957221984863281, "learning_rate": 7.699444836168413e-07, "loss": 0.4808, "step": 3906 }, { "epoch": 2.517396907216495, "grad_norm": 0.5756139159202576, "learning_rate": 7.679468803155526e-07, "loss": 0.5173, "step": 3907 }, { "epoch": 2.518041237113402, "grad_norm": 0.6060143113136292, "learning_rate": 7.659516561695512e-07, "loss": 0.5236, "step": 3908 }, { "epoch": 2.5186855670103094, "grad_norm": 0.579490602016449, "learning_rate": 7.639588123005037e-07, "loss": 0.5159, "step": 3909 }, { "epoch": 2.5193298969072164, "grad_norm": 0.5895804166793823, "learning_rate": 7.619683498287356e-07, "loss": 0.5019, "step": 3910 }, { "epoch": 2.519974226804124, "grad_norm": 0.5554822683334351, "learning_rate": 7.599802698732356e-07, "loss": 0.492, "step": 3911 }, { "epoch": 2.520618556701031, "grad_norm": 0.6342623233795166, "learning_rate": 7.579945735516553e-07, "loss": 0.5121, "step": 3912 }, { "epoch": 2.5212628865979383, "grad_norm": 0.5962876081466675, "learning_rate": 7.560112619803012e-07, "loss": 0.5094, "step": 3913 }, { "epoch": 2.5219072164948453, "grad_norm": 0.5273809432983398, "learning_rate": 7.54030336274143e-07, "loss": 0.5188, "step": 3914 }, { "epoch": 2.5225515463917527, "grad_norm": 0.6082810163497925, "learning_rate": 7.520517975468089e-07, "loss": 0.4856, "step": 3915 }, { "epoch": 2.5231958762886597, "grad_norm": 0.5320258140563965, "learning_rate": 7.500756469105819e-07, "loss": 0.5163, "step": 3916 }, { "epoch": 2.523840206185567, "grad_norm": 0.5323888659477234, "learning_rate": 7.481018854764066e-07, "loss": 0.4859, "step": 3917 }, { "epoch": 2.524484536082474, "grad_norm": 0.5678150057792664, "learning_rate": 7.461305143538839e-07, "loss": 0.5007, "step": 3918 }, { "epoch": 2.5251288659793816, "grad_norm": 0.5835501551628113, "learning_rate": 7.441615346512665e-07, "loss": 0.5216, "step": 3919 }, { "epoch": 2.5257731958762886, "grad_norm": 0.5608299970626831, "learning_rate": 7.42194947475468e-07, "loss": 0.5339, "step": 3920 }, { "epoch": 2.526417525773196, "grad_norm": 0.5719382762908936, "learning_rate": 7.402307539320558e-07, "loss": 0.5017, "step": 3921 }, { "epoch": 2.527061855670103, "grad_norm": 0.5923765301704407, "learning_rate": 7.382689551252487e-07, "loss": 0.5097, "step": 3922 }, { "epoch": 2.5277061855670104, "grad_norm": 0.6837469339370728, "learning_rate": 7.363095521579216e-07, "loss": 0.5142, "step": 3923 }, { "epoch": 2.5283505154639174, "grad_norm": 0.5769017338752747, "learning_rate": 7.343525461316042e-07, "loss": 0.4971, "step": 3924 }, { "epoch": 2.528994845360825, "grad_norm": 0.5687041282653809, "learning_rate": 7.323979381464735e-07, "loss": 0.4931, "step": 3925 }, { "epoch": 2.529639175257732, "grad_norm": 0.5814142227172852, "learning_rate": 7.304457293013656e-07, "loss": 0.5152, "step": 3926 }, { "epoch": 2.5302835051546393, "grad_norm": 0.5711424946784973, "learning_rate": 7.284959206937609e-07, "loss": 0.487, "step": 3927 }, { "epoch": 2.5309278350515463, "grad_norm": 0.577282190322876, "learning_rate": 7.265485134197947e-07, "loss": 0.5114, "step": 3928 }, { "epoch": 2.5315721649484537, "grad_norm": 0.5640125274658203, "learning_rate": 7.246035085742514e-07, "loss": 0.5159, "step": 3929 }, { "epoch": 2.5322164948453607, "grad_norm": 0.5827906727790833, "learning_rate": 7.226609072505625e-07, "loss": 0.493, "step": 3930 }, { "epoch": 2.532860824742268, "grad_norm": 0.5741549134254456, "learning_rate": 7.20720710540812e-07, "loss": 0.5136, "step": 3931 }, { "epoch": 2.533505154639175, "grad_norm": 0.6196715831756592, "learning_rate": 7.18782919535731e-07, "loss": 0.5055, "step": 3932 }, { "epoch": 2.5341494845360826, "grad_norm": 0.6098926067352295, "learning_rate": 7.168475353246956e-07, "loss": 0.5081, "step": 3933 }, { "epoch": 2.5347938144329896, "grad_norm": 0.6347229480743408, "learning_rate": 7.149145589957313e-07, "loss": 0.52, "step": 3934 }, { "epoch": 2.535438144329897, "grad_norm": 0.5969951152801514, "learning_rate": 7.129839916355114e-07, "loss": 0.5171, "step": 3935 }, { "epoch": 2.536082474226804, "grad_norm": 0.6011204719543457, "learning_rate": 7.1105583432935e-07, "loss": 0.5013, "step": 3936 }, { "epoch": 2.5367268041237114, "grad_norm": 0.6164532899856567, "learning_rate": 7.091300881612118e-07, "loss": 0.5105, "step": 3937 }, { "epoch": 2.5373711340206184, "grad_norm": 0.533269464969635, "learning_rate": 7.072067542137012e-07, "loss": 0.5232, "step": 3938 }, { "epoch": 2.538015463917526, "grad_norm": 0.5845893025398254, "learning_rate": 7.052858335680707e-07, "loss": 0.4964, "step": 3939 }, { "epoch": 2.538659793814433, "grad_norm": 0.5504436492919922, "learning_rate": 7.033673273042146e-07, "loss": 0.506, "step": 3940 }, { "epoch": 2.5393041237113403, "grad_norm": 0.5651990175247192, "learning_rate": 7.01451236500667e-07, "loss": 0.5287, "step": 3941 }, { "epoch": 2.5399484536082473, "grad_norm": 0.5860133171081543, "learning_rate": 6.995375622346085e-07, "loss": 0.4872, "step": 3942 }, { "epoch": 2.5405927835051547, "grad_norm": 0.5476179122924805, "learning_rate": 6.976263055818588e-07, "loss": 0.5085, "step": 3943 }, { "epoch": 2.5412371134020617, "grad_norm": 0.5924586653709412, "learning_rate": 6.957174676168793e-07, "loss": 0.506, "step": 3944 }, { "epoch": 2.541881443298969, "grad_norm": 0.6359657645225525, "learning_rate": 6.938110494127726e-07, "loss": 0.5087, "step": 3945 }, { "epoch": 2.542525773195876, "grad_norm": 0.5506124496459961, "learning_rate": 6.919070520412768e-07, "loss": 0.5016, "step": 3946 }, { "epoch": 2.5431701030927836, "grad_norm": 0.6055700182914734, "learning_rate": 6.900054765727743e-07, "loss": 0.5159, "step": 3947 }, { "epoch": 2.5438144329896906, "grad_norm": 0.5909888744354248, "learning_rate": 6.88106324076283e-07, "loss": 0.5014, "step": 3948 }, { "epoch": 2.544458762886598, "grad_norm": 0.6639335751533508, "learning_rate": 6.862095956194587e-07, "loss": 0.5272, "step": 3949 }, { "epoch": 2.545103092783505, "grad_norm": 0.5884895920753479, "learning_rate": 6.843152922685958e-07, "loss": 0.5239, "step": 3950 }, { "epoch": 2.5457474226804124, "grad_norm": 0.6417267322540283, "learning_rate": 6.824234150886255e-07, "loss": 0.5297, "step": 3951 }, { "epoch": 2.5463917525773194, "grad_norm": 0.6389366984367371, "learning_rate": 6.80533965143112e-07, "loss": 0.508, "step": 3952 }, { "epoch": 2.547036082474227, "grad_norm": 0.6120231747627258, "learning_rate": 6.786469434942583e-07, "loss": 0.5029, "step": 3953 }, { "epoch": 2.5476804123711343, "grad_norm": 0.5520997643470764, "learning_rate": 6.76762351202902e-07, "loss": 0.5099, "step": 3954 }, { "epoch": 2.5483247422680413, "grad_norm": 0.667405366897583, "learning_rate": 6.748801893285129e-07, "loss": 0.5114, "step": 3955 }, { "epoch": 2.5489690721649483, "grad_norm": 0.5701883435249329, "learning_rate": 6.730004589291961e-07, "loss": 0.5256, "step": 3956 }, { "epoch": 2.5496134020618557, "grad_norm": 0.5024256706237793, "learning_rate": 6.711231610616903e-07, "loss": 0.5019, "step": 3957 }, { "epoch": 2.550257731958763, "grad_norm": 0.5786945223808289, "learning_rate": 6.692482967813647e-07, "loss": 0.5192, "step": 3958 }, { "epoch": 2.55090206185567, "grad_norm": 0.589684784412384, "learning_rate": 6.673758671422209e-07, "loss": 0.4977, "step": 3959 }, { "epoch": 2.551546391752577, "grad_norm": 0.5437315702438354, "learning_rate": 6.65505873196895e-07, "loss": 0.5084, "step": 3960 }, { "epoch": 2.5521907216494846, "grad_norm": 0.539569079875946, "learning_rate": 6.63638315996647e-07, "loss": 0.511, "step": 3961 }, { "epoch": 2.552835051546392, "grad_norm": 0.5429437756538391, "learning_rate": 6.61773196591376e-07, "loss": 0.5252, "step": 3962 }, { "epoch": 2.553479381443299, "grad_norm": 0.534574031829834, "learning_rate": 6.599105160296027e-07, "loss": 0.5183, "step": 3963 }, { "epoch": 2.554123711340206, "grad_norm": 0.6054503917694092, "learning_rate": 6.580502753584806e-07, "loss": 0.4979, "step": 3964 }, { "epoch": 2.5547680412371134, "grad_norm": 0.5683809518814087, "learning_rate": 6.56192475623792e-07, "loss": 0.4903, "step": 3965 }, { "epoch": 2.555412371134021, "grad_norm": 0.572766125202179, "learning_rate": 6.543371178699442e-07, "loss": 0.4782, "step": 3966 }, { "epoch": 2.556056701030928, "grad_norm": 0.5394784212112427, "learning_rate": 6.524842031399737e-07, "loss": 0.5226, "step": 3967 }, { "epoch": 2.556701030927835, "grad_norm": 0.5729382038116455, "learning_rate": 6.506337324755447e-07, "loss": 0.5065, "step": 3968 }, { "epoch": 2.5573453608247423, "grad_norm": 0.5911117196083069, "learning_rate": 6.487857069169439e-07, "loss": 0.5047, "step": 3969 }, { "epoch": 2.5579896907216497, "grad_norm": 0.5617167353630066, "learning_rate": 6.469401275030857e-07, "loss": 0.5359, "step": 3970 }, { "epoch": 2.5586340206185567, "grad_norm": 0.5835085511207581, "learning_rate": 6.450969952715114e-07, "loss": 0.5141, "step": 3971 }, { "epoch": 2.5592783505154637, "grad_norm": 0.5641992092132568, "learning_rate": 6.432563112583806e-07, "loss": 0.5065, "step": 3972 }, { "epoch": 2.559922680412371, "grad_norm": 0.6262556910514832, "learning_rate": 6.414180764984829e-07, "loss": 0.5056, "step": 3973 }, { "epoch": 2.5605670103092786, "grad_norm": 0.5916532278060913, "learning_rate": 6.395822920252276e-07, "loss": 0.516, "step": 3974 }, { "epoch": 2.5612113402061856, "grad_norm": 0.5823740363121033, "learning_rate": 6.377489588706465e-07, "loss": 0.4971, "step": 3975 }, { "epoch": 2.5618556701030926, "grad_norm": 0.6451955437660217, "learning_rate": 6.359180780653957e-07, "loss": 0.514, "step": 3976 }, { "epoch": 2.5625, "grad_norm": 0.530180037021637, "learning_rate": 6.340896506387484e-07, "loss": 0.5054, "step": 3977 }, { "epoch": 2.5631443298969074, "grad_norm": 0.6298691034317017, "learning_rate": 6.322636776186025e-07, "loss": 0.507, "step": 3978 }, { "epoch": 2.5637886597938144, "grad_norm": 0.5655044317245483, "learning_rate": 6.304401600314752e-07, "loss": 0.5254, "step": 3979 }, { "epoch": 2.5644329896907214, "grad_norm": 0.5448644757270813, "learning_rate": 6.286190989025021e-07, "loss": 0.5155, "step": 3980 }, { "epoch": 2.565077319587629, "grad_norm": 0.5736231803894043, "learning_rate": 6.268004952554402e-07, "loss": 0.4957, "step": 3981 }, { "epoch": 2.5657216494845363, "grad_norm": 0.5456445813179016, "learning_rate": 6.249843501126607e-07, "loss": 0.5063, "step": 3982 }, { "epoch": 2.5663659793814433, "grad_norm": 0.5547133088111877, "learning_rate": 6.231706644951574e-07, "loss": 0.5348, "step": 3983 }, { "epoch": 2.5670103092783503, "grad_norm": 0.5256608128547668, "learning_rate": 6.213594394225391e-07, "loss": 0.5143, "step": 3984 }, { "epoch": 2.5676546391752577, "grad_norm": 0.6046155691146851, "learning_rate": 6.195506759130304e-07, "loss": 0.5064, "step": 3985 }, { "epoch": 2.568298969072165, "grad_norm": 0.6391527652740479, "learning_rate": 6.177443749834743e-07, "loss": 0.5185, "step": 3986 }, { "epoch": 2.568943298969072, "grad_norm": 0.6193614602088928, "learning_rate": 6.159405376493288e-07, "loss": 0.4972, "step": 3987 }, { "epoch": 2.569587628865979, "grad_norm": 0.5542547106742859, "learning_rate": 6.141391649246653e-07, "loss": 0.51, "step": 3988 }, { "epoch": 2.5702319587628866, "grad_norm": 0.5944623351097107, "learning_rate": 6.12340257822171e-07, "loss": 0.4961, "step": 3989 }, { "epoch": 2.570876288659794, "grad_norm": 0.5785998106002808, "learning_rate": 6.105438173531486e-07, "loss": 0.5202, "step": 3990 }, { "epoch": 2.571520618556701, "grad_norm": 0.5534639954566956, "learning_rate": 6.087498445275103e-07, "loss": 0.5078, "step": 3991 }, { "epoch": 2.572164948453608, "grad_norm": 0.5983073115348816, "learning_rate": 6.069583403537838e-07, "loss": 0.4953, "step": 3992 }, { "epoch": 2.5728092783505154, "grad_norm": 0.5910734534263611, "learning_rate": 6.051693058391089e-07, "loss": 0.4944, "step": 3993 }, { "epoch": 2.573453608247423, "grad_norm": 0.5598301291465759, "learning_rate": 6.033827419892352e-07, "loss": 0.4957, "step": 3994 }, { "epoch": 2.57409793814433, "grad_norm": 0.5460503101348877, "learning_rate": 6.015986498085258e-07, "loss": 0.5109, "step": 3995 }, { "epoch": 2.574742268041237, "grad_norm": 0.5170513987541199, "learning_rate": 5.998170302999529e-07, "loss": 0.5142, "step": 3996 }, { "epoch": 2.5753865979381443, "grad_norm": 0.5417015552520752, "learning_rate": 5.980378844650964e-07, "loss": 0.5152, "step": 3997 }, { "epoch": 2.5760309278350517, "grad_norm": 0.5566166639328003, "learning_rate": 5.962612133041523e-07, "loss": 0.5011, "step": 3998 }, { "epoch": 2.5766752577319587, "grad_norm": 0.5355037450790405, "learning_rate": 5.944870178159179e-07, "loss": 0.4995, "step": 3999 }, { "epoch": 2.5773195876288657, "grad_norm": 0.5412173867225647, "learning_rate": 5.92715298997803e-07, "loss": 0.5107, "step": 4000 }, { "epoch": 2.577963917525773, "grad_norm": 0.5497490763664246, "learning_rate": 5.909460578458243e-07, "loss": 0.5149, "step": 4001 }, { "epoch": 2.5786082474226806, "grad_norm": 0.5581592917442322, "learning_rate": 5.891792953546039e-07, "loss": 0.5179, "step": 4002 }, { "epoch": 2.5792525773195876, "grad_norm": 0.5204585790634155, "learning_rate": 5.874150125173727e-07, "loss": 0.5326, "step": 4003 }, { "epoch": 2.579896907216495, "grad_norm": 0.49713626503944397, "learning_rate": 5.856532103259677e-07, "loss": 0.5084, "step": 4004 }, { "epoch": 2.580541237113402, "grad_norm": 0.5899934768676758, "learning_rate": 5.838938897708285e-07, "loss": 0.5335, "step": 4005 }, { "epoch": 2.5811855670103094, "grad_norm": 0.5983029007911682, "learning_rate": 5.821370518410019e-07, "loss": 0.503, "step": 4006 }, { "epoch": 2.5818298969072164, "grad_norm": 0.5667003989219666, "learning_rate": 5.803826975241406e-07, "loss": 0.4687, "step": 4007 }, { "epoch": 2.582474226804124, "grad_norm": 0.6477274894714355, "learning_rate": 5.786308278064967e-07, "loss": 0.506, "step": 4008 }, { "epoch": 2.583118556701031, "grad_norm": 0.576697051525116, "learning_rate": 5.768814436729287e-07, "loss": 0.5134, "step": 4009 }, { "epoch": 2.5837628865979383, "grad_norm": 0.5756586790084839, "learning_rate": 5.75134546106898e-07, "loss": 0.4968, "step": 4010 }, { "epoch": 2.5844072164948453, "grad_norm": 0.6409735083580017, "learning_rate": 5.733901360904648e-07, "loss": 0.4761, "step": 4011 }, { "epoch": 2.5850515463917527, "grad_norm": 0.5371703505516052, "learning_rate": 5.716482146042957e-07, "loss": 0.5204, "step": 4012 }, { "epoch": 2.5856958762886597, "grad_norm": 0.5879557132720947, "learning_rate": 5.699087826276539e-07, "loss": 0.5176, "step": 4013 }, { "epoch": 2.586340206185567, "grad_norm": 0.6314774751663208, "learning_rate": 5.681718411384052e-07, "loss": 0.501, "step": 4014 }, { "epoch": 2.586984536082474, "grad_norm": 0.5401092171669006, "learning_rate": 5.664373911130149e-07, "loss": 0.4972, "step": 4015 }, { "epoch": 2.5876288659793816, "grad_norm": 0.6852840781211853, "learning_rate": 5.647054335265489e-07, "loss": 0.4898, "step": 4016 }, { "epoch": 2.5882731958762886, "grad_norm": 0.5953755974769592, "learning_rate": 5.629759693526698e-07, "loss": 0.5161, "step": 4017 }, { "epoch": 2.588917525773196, "grad_norm": 0.5267965793609619, "learning_rate": 5.612489995636411e-07, "loss": 0.5252, "step": 4018 }, { "epoch": 2.589561855670103, "grad_norm": 0.6078029274940491, "learning_rate": 5.5952452513032e-07, "loss": 0.5167, "step": 4019 }, { "epoch": 2.5902061855670104, "grad_norm": 0.6043415665626526, "learning_rate": 5.578025470221648e-07, "loss": 0.5061, "step": 4020 }, { "epoch": 2.5908505154639174, "grad_norm": 0.5979164838790894, "learning_rate": 5.56083066207228e-07, "loss": 0.5135, "step": 4021 }, { "epoch": 2.591494845360825, "grad_norm": 0.5940758585929871, "learning_rate": 5.543660836521592e-07, "loss": 0.5104, "step": 4022 }, { "epoch": 2.592139175257732, "grad_norm": 0.5511759519577026, "learning_rate": 5.526516003222049e-07, "loss": 0.5306, "step": 4023 }, { "epoch": 2.5927835051546393, "grad_norm": 0.5754338502883911, "learning_rate": 5.509396171812026e-07, "loss": 0.5409, "step": 4024 }, { "epoch": 2.5934278350515463, "grad_norm": 0.5364115834236145, "learning_rate": 5.49230135191588e-07, "loss": 0.5208, "step": 4025 }, { "epoch": 2.5940721649484537, "grad_norm": 0.6179932355880737, "learning_rate": 5.475231553143906e-07, "loss": 0.5205, "step": 4026 }, { "epoch": 2.5947164948453607, "grad_norm": 0.5981595516204834, "learning_rate": 5.458186785092301e-07, "loss": 0.5232, "step": 4027 }, { "epoch": 2.595360824742268, "grad_norm": 0.525852382183075, "learning_rate": 5.441167057343216e-07, "loss": 0.536, "step": 4028 }, { "epoch": 2.596005154639175, "grad_norm": 0.5910958051681519, "learning_rate": 5.424172379464737e-07, "loss": 0.5114, "step": 4029 }, { "epoch": 2.5966494845360826, "grad_norm": 0.5978402495384216, "learning_rate": 5.407202761010827e-07, "loss": 0.5038, "step": 4030 }, { "epoch": 2.5972938144329896, "grad_norm": 0.5346774458885193, "learning_rate": 5.390258211521398e-07, "loss": 0.5206, "step": 4031 }, { "epoch": 2.597938144329897, "grad_norm": 0.6103032231330872, "learning_rate": 5.37333874052226e-07, "loss": 0.5018, "step": 4032 }, { "epoch": 2.598582474226804, "grad_norm": 0.5615730881690979, "learning_rate": 5.356444357525092e-07, "loss": 0.508, "step": 4033 }, { "epoch": 2.5992268041237114, "grad_norm": 0.5577130913734436, "learning_rate": 5.339575072027536e-07, "loss": 0.5174, "step": 4034 }, { "epoch": 2.5998711340206184, "grad_norm": 0.6179705858230591, "learning_rate": 5.322730893513056e-07, "loss": 0.4918, "step": 4035 }, { "epoch": 2.600515463917526, "grad_norm": 0.5816167593002319, "learning_rate": 5.305911831451044e-07, "loss": 0.4998, "step": 4036 }, { "epoch": 2.601159793814433, "grad_norm": 0.6231922507286072, "learning_rate": 5.289117895296769e-07, "loss": 0.5141, "step": 4037 }, { "epoch": 2.6018041237113403, "grad_norm": 0.6045860648155212, "learning_rate": 5.272349094491341e-07, "loss": 0.5129, "step": 4038 }, { "epoch": 2.6024484536082473, "grad_norm": 0.6097016334533691, "learning_rate": 5.255605438461781e-07, "loss": 0.5263, "step": 4039 }, { "epoch": 2.6030927835051547, "grad_norm": 0.5618062019348145, "learning_rate": 5.238886936620957e-07, "loss": 0.504, "step": 4040 }, { "epoch": 2.6037371134020617, "grad_norm": 0.5221797823905945, "learning_rate": 5.222193598367586e-07, "loss": 0.5156, "step": 4041 }, { "epoch": 2.604381443298969, "grad_norm": 0.5370350480079651, "learning_rate": 5.205525433086256e-07, "loss": 0.5212, "step": 4042 }, { "epoch": 2.605025773195876, "grad_norm": 0.5955010056495667, "learning_rate": 5.188882450147403e-07, "loss": 0.5038, "step": 4043 }, { "epoch": 2.6056701030927836, "grad_norm": 0.5505319237709045, "learning_rate": 5.172264658907284e-07, "loss": 0.4987, "step": 4044 }, { "epoch": 2.6063144329896906, "grad_norm": 0.5419166684150696, "learning_rate": 5.155672068708011e-07, "loss": 0.5108, "step": 4045 }, { "epoch": 2.606958762886598, "grad_norm": 0.596188485622406, "learning_rate": 5.139104688877549e-07, "loss": 0.5197, "step": 4046 }, { "epoch": 2.607603092783505, "grad_norm": 0.522609293460846, "learning_rate": 5.122562528729636e-07, "loss": 0.5307, "step": 4047 }, { "epoch": 2.6082474226804124, "grad_norm": 0.5784689784049988, "learning_rate": 5.106045597563885e-07, "loss": 0.5264, "step": 4048 }, { "epoch": 2.6088917525773194, "grad_norm": 0.5608997344970703, "learning_rate": 5.089553904665695e-07, "loss": 0.5172, "step": 4049 }, { "epoch": 2.609536082474227, "grad_norm": 0.5834921598434448, "learning_rate": 5.073087459306291e-07, "loss": 0.5235, "step": 4050 }, { "epoch": 2.6101804123711343, "grad_norm": 0.540431797504425, "learning_rate": 5.056646270742699e-07, "loss": 0.5148, "step": 4051 }, { "epoch": 2.6108247422680413, "grad_norm": 0.5432286858558655, "learning_rate": 5.040230348217745e-07, "loss": 0.5125, "step": 4052 }, { "epoch": 2.6114690721649483, "grad_norm": 0.5524317622184753, "learning_rate": 5.023839700960059e-07, "loss": 0.4955, "step": 4053 }, { "epoch": 2.6121134020618557, "grad_norm": 0.5855949521064758, "learning_rate": 5.007474338184059e-07, "loss": 0.5018, "step": 4054 }, { "epoch": 2.612757731958763, "grad_norm": 0.6361154913902283, "learning_rate": 4.991134269089936e-07, "loss": 0.4982, "step": 4055 }, { "epoch": 2.61340206185567, "grad_norm": 0.5513914823532104, "learning_rate": 4.974819502863687e-07, "loss": 0.5025, "step": 4056 }, { "epoch": 2.614046391752577, "grad_norm": 0.625952422618866, "learning_rate": 4.958530048677045e-07, "loss": 0.4965, "step": 4057 }, { "epoch": 2.6146907216494846, "grad_norm": 0.5822234153747559, "learning_rate": 4.94226591568755e-07, "loss": 0.5293, "step": 4058 }, { "epoch": 2.615335051546392, "grad_norm": 0.5558634400367737, "learning_rate": 4.926027113038507e-07, "loss": 0.4977, "step": 4059 }, { "epoch": 2.615979381443299, "grad_norm": 0.6505028009414673, "learning_rate": 4.909813649858941e-07, "loss": 0.5361, "step": 4060 }, { "epoch": 2.616623711340206, "grad_norm": 0.5984810590744019, "learning_rate": 4.89362553526368e-07, "loss": 0.5167, "step": 4061 }, { "epoch": 2.6172680412371134, "grad_norm": 0.6045730710029602, "learning_rate": 4.877462778353275e-07, "loss": 0.5091, "step": 4062 }, { "epoch": 2.617912371134021, "grad_norm": 0.5563434958457947, "learning_rate": 4.861325388214022e-07, "loss": 0.5246, "step": 4063 }, { "epoch": 2.618556701030928, "grad_norm": 0.6049947142601013, "learning_rate": 4.845213373917968e-07, "loss": 0.5101, "step": 4064 }, { "epoch": 2.619201030927835, "grad_norm": 0.5940573215484619, "learning_rate": 4.829126744522888e-07, "loss": 0.5059, "step": 4065 }, { "epoch": 2.6198453608247423, "grad_norm": 0.5261126756668091, "learning_rate": 4.813065509072279e-07, "loss": 0.521, "step": 4066 }, { "epoch": 2.6204896907216497, "grad_norm": 0.5959347486495972, "learning_rate": 4.797029676595366e-07, "loss": 0.5076, "step": 4067 }, { "epoch": 2.6211340206185567, "grad_norm": 0.6117706298828125, "learning_rate": 4.781019256107116e-07, "loss": 0.5115, "step": 4068 }, { "epoch": 2.6217783505154637, "grad_norm": 0.5512093305587769, "learning_rate": 4.76503425660817e-07, "loss": 0.5069, "step": 4069 }, { "epoch": 2.622422680412371, "grad_norm": 0.6334697008132935, "learning_rate": 4.749074687084909e-07, "loss": 0.4959, "step": 4070 }, { "epoch": 2.6230670103092786, "grad_norm": 0.6187028288841248, "learning_rate": 4.7331405565093967e-07, "loss": 0.494, "step": 4071 }, { "epoch": 2.6237113402061856, "grad_norm": 0.7392948865890503, "learning_rate": 4.7172318738394173e-07, "loss": 0.5258, "step": 4072 }, { "epoch": 2.6243556701030926, "grad_norm": 0.6911803483963013, "learning_rate": 4.7013486480184443e-07, "loss": 0.4999, "step": 4073 }, { "epoch": 2.625, "grad_norm": 0.6623152494430542, "learning_rate": 4.685490887975613e-07, "loss": 0.5174, "step": 4074 }, { "epoch": 2.6256443298969074, "grad_norm": 0.6914599537849426, "learning_rate": 4.6696586026257706e-07, "loss": 0.5031, "step": 4075 }, { "epoch": 2.6262886597938144, "grad_norm": 0.7492258548736572, "learning_rate": 4.6538518008694465e-07, "loss": 0.5258, "step": 4076 }, { "epoch": 2.6269329896907214, "grad_norm": 0.5765107274055481, "learning_rate": 4.638070491592811e-07, "loss": 0.5035, "step": 4077 }, { "epoch": 2.627577319587629, "grad_norm": 0.745932400226593, "learning_rate": 4.622314683667745e-07, "loss": 0.5181, "step": 4078 }, { "epoch": 2.6282216494845363, "grad_norm": 0.7643501162528992, "learning_rate": 4.606584385951768e-07, "loss": 0.5151, "step": 4079 }, { "epoch": 2.6288659793814433, "grad_norm": 0.7881836891174316, "learning_rate": 4.590879607288051e-07, "loss": 0.5034, "step": 4080 }, { "epoch": 2.6295103092783503, "grad_norm": 0.6642072796821594, "learning_rate": 4.575200356505438e-07, "loss": 0.4865, "step": 4081 }, { "epoch": 2.6301546391752577, "grad_norm": 0.8136774301528931, "learning_rate": 4.5595466424184276e-07, "loss": 0.5039, "step": 4082 }, { "epoch": 2.630798969072165, "grad_norm": 0.6542945504188538, "learning_rate": 4.54391847382713e-07, "loss": 0.5228, "step": 4083 }, { "epoch": 2.631443298969072, "grad_norm": 0.5564175844192505, "learning_rate": 4.5283158595173293e-07, "loss": 0.5069, "step": 4084 }, { "epoch": 2.632087628865979, "grad_norm": 0.7194393873214722, "learning_rate": 4.512738808260414e-07, "loss": 0.5031, "step": 4085 }, { "epoch": 2.6327319587628866, "grad_norm": 0.6088846325874329, "learning_rate": 4.4971873288134237e-07, "loss": 0.4995, "step": 4086 }, { "epoch": 2.633376288659794, "grad_norm": 0.6590967178344727, "learning_rate": 4.4816614299190264e-07, "loss": 0.5181, "step": 4087 }, { "epoch": 2.634020618556701, "grad_norm": 0.5358991622924805, "learning_rate": 4.466161120305462e-07, "loss": 0.5122, "step": 4088 }, { "epoch": 2.634664948453608, "grad_norm": 0.6606979370117188, "learning_rate": 4.450686408686655e-07, "loss": 0.5034, "step": 4089 }, { "epoch": 2.6353092783505154, "grad_norm": 0.654866099357605, "learning_rate": 4.4352373037621023e-07, "loss": 0.4943, "step": 4090 }, { "epoch": 2.635953608247423, "grad_norm": 0.6692166924476624, "learning_rate": 4.4198138142168834e-07, "loss": 0.5037, "step": 4091 }, { "epoch": 2.63659793814433, "grad_norm": 0.6145294308662415, "learning_rate": 4.40441594872173e-07, "loss": 0.5315, "step": 4092 }, { "epoch": 2.637242268041237, "grad_norm": 0.6467417478561401, "learning_rate": 4.389043715932917e-07, "loss": 0.5258, "step": 4093 }, { "epoch": 2.6378865979381443, "grad_norm": 0.6893128752708435, "learning_rate": 4.373697124492338e-07, "loss": 0.5228, "step": 4094 }, { "epoch": 2.6385309278350517, "grad_norm": 0.5774792432785034, "learning_rate": 4.3583761830274795e-07, "loss": 0.5139, "step": 4095 }, { "epoch": 2.6391752577319587, "grad_norm": 0.6026521325111389, "learning_rate": 4.343080900151375e-07, "loss": 0.5268, "step": 4096 }, { "epoch": 2.6398195876288657, "grad_norm": 0.6545516848564148, "learning_rate": 4.3278112844626683e-07, "loss": 0.5183, "step": 4097 }, { "epoch": 2.640463917525773, "grad_norm": 0.6393605470657349, "learning_rate": 4.312567344545554e-07, "loss": 0.5429, "step": 4098 }, { "epoch": 2.6411082474226806, "grad_norm": 0.5801621079444885, "learning_rate": 4.2973490889697934e-07, "loss": 0.5091, "step": 4099 }, { "epoch": 2.6417525773195876, "grad_norm": 0.6297934055328369, "learning_rate": 4.282156526290715e-07, "loss": 0.4933, "step": 4100 }, { "epoch": 2.642396907216495, "grad_norm": 0.685049831867218, "learning_rate": 4.266989665049204e-07, "loss": 0.5162, "step": 4101 }, { "epoch": 2.643041237113402, "grad_norm": 0.5356438755989075, "learning_rate": 4.251848513771684e-07, "loss": 0.5301, "step": 4102 }, { "epoch": 2.6436855670103094, "grad_norm": 0.6242610216140747, "learning_rate": 4.2367330809701404e-07, "loss": 0.5136, "step": 4103 }, { "epoch": 2.6443298969072164, "grad_norm": 0.60305255651474, "learning_rate": 4.221643375142109e-07, "loss": 0.508, "step": 4104 }, { "epoch": 2.644974226804124, "grad_norm": 0.5690869688987732, "learning_rate": 4.2065794047706166e-07, "loss": 0.5256, "step": 4105 }, { "epoch": 2.645618556701031, "grad_norm": 0.5664268136024475, "learning_rate": 4.1915411783242766e-07, "loss": 0.529, "step": 4106 }, { "epoch": 2.6462628865979383, "grad_norm": 0.5917067527770996, "learning_rate": 4.1765287042572047e-07, "loss": 0.5405, "step": 4107 }, { "epoch": 2.6469072164948453, "grad_norm": 0.5395490527153015, "learning_rate": 4.161541991009033e-07, "loss": 0.5003, "step": 4108 }, { "epoch": 2.6475515463917527, "grad_norm": 0.5781093835830688, "learning_rate": 4.1465810470049327e-07, "loss": 0.4936, "step": 4109 }, { "epoch": 2.6481958762886597, "grad_norm": 0.6104745268821716, "learning_rate": 4.1316458806555606e-07, "loss": 0.5044, "step": 4110 }, { "epoch": 2.648840206185567, "grad_norm": 0.5545309782028198, "learning_rate": 4.1167365003571047e-07, "loss": 0.5084, "step": 4111 }, { "epoch": 2.649484536082474, "grad_norm": 0.6034955978393555, "learning_rate": 4.101852914491261e-07, "loss": 0.5056, "step": 4112 }, { "epoch": 2.6501288659793816, "grad_norm": 0.517917811870575, "learning_rate": 4.086995131425187e-07, "loss": 0.5175, "step": 4113 }, { "epoch": 2.6507731958762886, "grad_norm": 0.5687193870544434, "learning_rate": 4.0721631595115716e-07, "loss": 0.5097, "step": 4114 }, { "epoch": 2.651417525773196, "grad_norm": 0.6941184997558594, "learning_rate": 4.0573570070885916e-07, "loss": 0.4934, "step": 4115 }, { "epoch": 2.652061855670103, "grad_norm": 0.544152557849884, "learning_rate": 4.0425766824798817e-07, "loss": 0.5188, "step": 4116 }, { "epoch": 2.6527061855670104, "grad_norm": 0.598566472530365, "learning_rate": 4.0278221939945795e-07, "loss": 0.5116, "step": 4117 }, { "epoch": 2.6533505154639174, "grad_norm": 0.6171483397483826, "learning_rate": 4.0130935499272994e-07, "loss": 0.4924, "step": 4118 }, { "epoch": 2.653994845360825, "grad_norm": 0.5857178568840027, "learning_rate": 3.9983907585581127e-07, "loss": 0.5115, "step": 4119 }, { "epoch": 2.654639175257732, "grad_norm": 0.6508338451385498, "learning_rate": 3.983713828152569e-07, "loss": 0.5027, "step": 4120 }, { "epoch": 2.6552835051546393, "grad_norm": 0.5620213150978088, "learning_rate": 3.9690627669616676e-07, "loss": 0.5199, "step": 4121 }, { "epoch": 2.6559278350515463, "grad_norm": 0.6147875785827637, "learning_rate": 3.9544375832218706e-07, "loss": 0.517, "step": 4122 }, { "epoch": 2.6565721649484537, "grad_norm": 0.6413827538490295, "learning_rate": 3.939838285155112e-07, "loss": 0.5176, "step": 4123 }, { "epoch": 2.6572164948453607, "grad_norm": 0.5277988314628601, "learning_rate": 3.925264880968721e-07, "loss": 0.5238, "step": 4124 }, { "epoch": 2.657860824742268, "grad_norm": 0.555444061756134, "learning_rate": 3.9107173788555333e-07, "loss": 0.5317, "step": 4125 }, { "epoch": 2.658505154639175, "grad_norm": 0.535132884979248, "learning_rate": 3.8961957869937893e-07, "loss": 0.5226, "step": 4126 }, { "epoch": 2.6591494845360826, "grad_norm": 0.545684814453125, "learning_rate": 3.88170011354716e-07, "loss": 0.5226, "step": 4127 }, { "epoch": 2.6597938144329896, "grad_norm": 0.5827793478965759, "learning_rate": 3.867230366664765e-07, "loss": 0.5254, "step": 4128 }, { "epoch": 2.660438144329897, "grad_norm": 0.6357749104499817, "learning_rate": 3.8527865544811194e-07, "loss": 0.5003, "step": 4129 }, { "epoch": 2.661082474226804, "grad_norm": 0.5090994238853455, "learning_rate": 3.838368685116184e-07, "loss": 0.5106, "step": 4130 }, { "epoch": 2.6617268041237114, "grad_norm": 0.5923934578895569, "learning_rate": 3.8239767666753356e-07, "loss": 0.5173, "step": 4131 }, { "epoch": 2.6623711340206184, "grad_norm": 0.5443612337112427, "learning_rate": 3.809610807249342e-07, "loss": 0.4826, "step": 4132 }, { "epoch": 2.663015463917526, "grad_norm": 0.5697106122970581, "learning_rate": 3.7952708149143926e-07, "loss": 0.4939, "step": 4133 }, { "epoch": 2.663659793814433, "grad_norm": 0.5876659750938416, "learning_rate": 3.780956797732088e-07, "loss": 0.5209, "step": 4134 }, { "epoch": 2.6643041237113403, "grad_norm": 0.5419542193412781, "learning_rate": 3.766668763749393e-07, "loss": 0.5077, "step": 4135 }, { "epoch": 2.6649484536082473, "grad_norm": 0.5780937671661377, "learning_rate": 3.752406720998691e-07, "loss": 0.5195, "step": 4136 }, { "epoch": 2.6655927835051547, "grad_norm": 0.593481183052063, "learning_rate": 3.7381706774977657e-07, "loss": 0.5235, "step": 4137 }, { "epoch": 2.6662371134020617, "grad_norm": 0.5956684947013855, "learning_rate": 3.723960641249746e-07, "loss": 0.4953, "step": 4138 }, { "epoch": 2.666881443298969, "grad_norm": 0.5656968355178833, "learning_rate": 3.709776620243166e-07, "loss": 0.5105, "step": 4139 }, { "epoch": 2.667525773195876, "grad_norm": 0.6292653679847717, "learning_rate": 3.695618622451952e-07, "loss": 0.5169, "step": 4140 }, { "epoch": 2.6681701030927836, "grad_norm": 0.5433037877082825, "learning_rate": 3.681486655835348e-07, "loss": 0.5047, "step": 4141 }, { "epoch": 2.6688144329896906, "grad_norm": 0.5461229085922241, "learning_rate": 3.6673807283380103e-07, "loss": 0.5197, "step": 4142 }, { "epoch": 2.669458762886598, "grad_norm": 0.5965960025787354, "learning_rate": 3.6533008478899456e-07, "loss": 0.5076, "step": 4143 }, { "epoch": 2.670103092783505, "grad_norm": 0.6078892946243286, "learning_rate": 3.6392470224065e-07, "loss": 0.5195, "step": 4144 }, { "epoch": 2.6707474226804124, "grad_norm": 0.6039332151412964, "learning_rate": 3.6252192597884095e-07, "loss": 0.4994, "step": 4145 }, { "epoch": 2.6713917525773194, "grad_norm": 0.6282094120979309, "learning_rate": 3.61121756792171e-07, "loss": 0.4951, "step": 4146 }, { "epoch": 2.672036082474227, "grad_norm": 0.6072065234184265, "learning_rate": 3.5972419546778106e-07, "loss": 0.5005, "step": 4147 }, { "epoch": 2.6726804123711343, "grad_norm": 0.5632229447364807, "learning_rate": 3.5832924279134705e-07, "loss": 0.4928, "step": 4148 }, { "epoch": 2.6733247422680413, "grad_norm": 0.6099588871002197, "learning_rate": 3.56936899547074e-07, "loss": 0.4926, "step": 4149 }, { "epoch": 2.6739690721649483, "grad_norm": 0.6520832777023315, "learning_rate": 3.555471665177046e-07, "loss": 0.4979, "step": 4150 }, { "epoch": 2.6746134020618557, "grad_norm": 0.608242392539978, "learning_rate": 3.5416004448451337e-07, "loss": 0.5033, "step": 4151 }, { "epoch": 2.675257731958763, "grad_norm": 0.556420087814331, "learning_rate": 3.5277553422730316e-07, "loss": 0.4727, "step": 4152 }, { "epoch": 2.67590206185567, "grad_norm": 0.5856521129608154, "learning_rate": 3.513936365244125e-07, "loss": 0.5333, "step": 4153 }, { "epoch": 2.676546391752577, "grad_norm": 0.5984749794006348, "learning_rate": 3.5001435215271105e-07, "loss": 0.5122, "step": 4154 }, { "epoch": 2.6771907216494846, "grad_norm": 0.5796511173248291, "learning_rate": 3.486376818875964e-07, "loss": 0.5046, "step": 4155 }, { "epoch": 2.677835051546392, "grad_norm": 0.5510055422782898, "learning_rate": 3.472636265030005e-07, "loss": 0.5122, "step": 4156 }, { "epoch": 2.678479381443299, "grad_norm": 0.5274403691291809, "learning_rate": 3.45892186771381e-07, "loss": 0.5064, "step": 4157 }, { "epoch": 2.679123711340206, "grad_norm": 0.5576527714729309, "learning_rate": 3.445233634637279e-07, "loss": 0.4984, "step": 4158 }, { "epoch": 2.6797680412371134, "grad_norm": 0.5783284306526184, "learning_rate": 3.431571573495618e-07, "loss": 0.5128, "step": 4159 }, { "epoch": 2.680412371134021, "grad_norm": 0.5730151534080505, "learning_rate": 3.417935691969265e-07, "loss": 0.5314, "step": 4160 }, { "epoch": 2.681056701030928, "grad_norm": 0.565512478351593, "learning_rate": 3.404325997723995e-07, "loss": 0.5131, "step": 4161 }, { "epoch": 2.681701030927835, "grad_norm": 0.5613424777984619, "learning_rate": 3.3907424984108486e-07, "loss": 0.5146, "step": 4162 }, { "epoch": 2.6823453608247423, "grad_norm": 0.6190486550331116, "learning_rate": 3.3771852016661166e-07, "loss": 0.5225, "step": 4163 }, { "epoch": 2.6829896907216497, "grad_norm": 0.5494011044502258, "learning_rate": 3.3636541151113813e-07, "loss": 0.5062, "step": 4164 }, { "epoch": 2.6836340206185567, "grad_norm": 0.5115302801132202, "learning_rate": 3.3501492463534793e-07, "loss": 0.5165, "step": 4165 }, { "epoch": 2.6842783505154637, "grad_norm": 0.5586063265800476, "learning_rate": 3.3366706029845097e-07, "loss": 0.5216, "step": 4166 }, { "epoch": 2.684922680412371, "grad_norm": 0.5477322936058044, "learning_rate": 3.323218192581851e-07, "loss": 0.5023, "step": 4167 }, { "epoch": 2.6855670103092786, "grad_norm": 0.5414209365844727, "learning_rate": 3.3097920227080917e-07, "loss": 0.5077, "step": 4168 }, { "epoch": 2.6862113402061856, "grad_norm": 0.5058279633522034, "learning_rate": 3.2963921009111e-07, "loss": 0.5204, "step": 4169 }, { "epoch": 2.6868556701030926, "grad_norm": 0.5271046757698059, "learning_rate": 3.283018434723989e-07, "loss": 0.511, "step": 4170 }, { "epoch": 2.6875, "grad_norm": 0.546602189540863, "learning_rate": 3.2696710316650836e-07, "loss": 0.5139, "step": 4171 }, { "epoch": 2.6881443298969074, "grad_norm": 0.5777022838592529, "learning_rate": 3.2563498992379703e-07, "loss": 0.5098, "step": 4172 }, { "epoch": 2.6887886597938144, "grad_norm": 0.5692847371101379, "learning_rate": 3.243055044931476e-07, "loss": 0.4981, "step": 4173 }, { "epoch": 2.6894329896907214, "grad_norm": 0.5605366826057434, "learning_rate": 3.2297864762196095e-07, "loss": 0.5302, "step": 4174 }, { "epoch": 2.690077319587629, "grad_norm": 0.6415053606033325, "learning_rate": 3.2165442005616497e-07, "loss": 0.5117, "step": 4175 }, { "epoch": 2.6907216494845363, "grad_norm": 0.5773971676826477, "learning_rate": 3.2033282254020747e-07, "loss": 0.5092, "step": 4176 }, { "epoch": 2.6913659793814433, "grad_norm": 0.5327409505844116, "learning_rate": 3.1901385581705693e-07, "loss": 0.5263, "step": 4177 }, { "epoch": 2.6920103092783503, "grad_norm": 0.6957939863204956, "learning_rate": 3.1769752062820405e-07, "loss": 0.516, "step": 4178 }, { "epoch": 2.6926546391752577, "grad_norm": 0.5728719234466553, "learning_rate": 3.163838177136597e-07, "loss": 0.5167, "step": 4179 }, { "epoch": 2.693298969072165, "grad_norm": 0.5573298335075378, "learning_rate": 3.1507274781195587e-07, "loss": 0.5082, "step": 4180 }, { "epoch": 2.693943298969072, "grad_norm": 0.6466101408004761, "learning_rate": 3.1376431166014354e-07, "loss": 0.4972, "step": 4181 }, { "epoch": 2.694587628865979, "grad_norm": 0.628017008304596, "learning_rate": 3.124585099937916e-07, "loss": 0.5223, "step": 4182 }, { "epoch": 2.6952319587628866, "grad_norm": 0.5699337124824524, "learning_rate": 3.111553435469905e-07, "loss": 0.4964, "step": 4183 }, { "epoch": 2.695876288659794, "grad_norm": 0.5913154482841492, "learning_rate": 3.098548130523482e-07, "loss": 0.5202, "step": 4184 }, { "epoch": 2.696520618556701, "grad_norm": 0.6492338180541992, "learning_rate": 3.085569192409904e-07, "loss": 0.5116, "step": 4185 }, { "epoch": 2.697164948453608, "grad_norm": 0.556464672088623, "learning_rate": 3.072616628425601e-07, "loss": 0.5179, "step": 4186 }, { "epoch": 2.6978092783505154, "grad_norm": 0.5164768695831299, "learning_rate": 3.0596904458521983e-07, "loss": 0.5181, "step": 4187 }, { "epoch": 2.698453608247423, "grad_norm": 0.5604942440986633, "learning_rate": 3.04679065195646e-07, "loss": 0.5162, "step": 4188 }, { "epoch": 2.69909793814433, "grad_norm": 0.6408377289772034, "learning_rate": 3.033917253990337e-07, "loss": 0.5226, "step": 4189 }, { "epoch": 2.699742268041237, "grad_norm": 0.6063274145126343, "learning_rate": 3.021070259190945e-07, "loss": 0.5232, "step": 4190 }, { "epoch": 2.7003865979381443, "grad_norm": 0.5733911991119385, "learning_rate": 3.0082496747805225e-07, "loss": 0.5132, "step": 4191 }, { "epoch": 2.7010309278350517, "grad_norm": 0.5322358012199402, "learning_rate": 2.9954555079664996e-07, "loss": 0.5319, "step": 4192 }, { "epoch": 2.7016752577319587, "grad_norm": 0.6530497670173645, "learning_rate": 2.982687765941444e-07, "loss": 0.538, "step": 4193 }, { "epoch": 2.7023195876288657, "grad_norm": 0.5688532590866089, "learning_rate": 2.969946455883044e-07, "loss": 0.4988, "step": 4194 }, { "epoch": 2.702963917525773, "grad_norm": 0.5575458407402039, "learning_rate": 2.957231584954173e-07, "loss": 0.5303, "step": 4195 }, { "epoch": 2.7036082474226806, "grad_norm": 0.6717913150787354, "learning_rate": 2.9445431603027876e-07, "loss": 0.5035, "step": 4196 }, { "epoch": 2.7042525773195876, "grad_norm": 0.6185145974159241, "learning_rate": 2.931881189062025e-07, "loss": 0.5274, "step": 4197 }, { "epoch": 2.704896907216495, "grad_norm": 0.5410360097885132, "learning_rate": 2.9192456783501353e-07, "loss": 0.4976, "step": 4198 }, { "epoch": 2.705541237113402, "grad_norm": 0.6465786099433899, "learning_rate": 2.9066366352704765e-07, "loss": 0.5261, "step": 4199 }, { "epoch": 2.7061855670103094, "grad_norm": 0.6569363474845886, "learning_rate": 2.89405406691155e-07, "loss": 0.5052, "step": 4200 }, { "epoch": 2.7068298969072164, "grad_norm": 0.5724846720695496, "learning_rate": 2.881497980346953e-07, "loss": 0.497, "step": 4201 }, { "epoch": 2.707474226804124, "grad_norm": 0.6895914077758789, "learning_rate": 2.8689683826354067e-07, "loss": 0.5163, "step": 4202 }, { "epoch": 2.708118556701031, "grad_norm": 0.6269958019256592, "learning_rate": 2.856465280820753e-07, "loss": 0.5256, "step": 4203 }, { "epoch": 2.7087628865979383, "grad_norm": 0.5459779500961304, "learning_rate": 2.843988681931914e-07, "loss": 0.515, "step": 4204 }, { "epoch": 2.7094072164948453, "grad_norm": 0.6091318130493164, "learning_rate": 2.831538592982924e-07, "loss": 0.5035, "step": 4205 }, { "epoch": 2.7100515463917527, "grad_norm": 0.626583993434906, "learning_rate": 2.8191150209729233e-07, "loss": 0.5166, "step": 4206 }, { "epoch": 2.7106958762886597, "grad_norm": 0.5687763094902039, "learning_rate": 2.806717972886125e-07, "loss": 0.5154, "step": 4207 }, { "epoch": 2.711340206185567, "grad_norm": 0.6023850440979004, "learning_rate": 2.794347455691848e-07, "loss": 0.5265, "step": 4208 }, { "epoch": 2.711984536082474, "grad_norm": 0.666304886341095, "learning_rate": 2.7820034763444915e-07, "loss": 0.5056, "step": 4209 }, { "epoch": 2.7126288659793816, "grad_norm": 0.5212836861610413, "learning_rate": 2.7696860417835315e-07, "loss": 0.5158, "step": 4210 }, { "epoch": 2.7132731958762886, "grad_norm": 0.5762307047843933, "learning_rate": 2.7573951589335177e-07, "loss": 0.5021, "step": 4211 }, { "epoch": 2.713917525773196, "grad_norm": 0.5702105760574341, "learning_rate": 2.745130834704102e-07, "loss": 0.4979, "step": 4212 }, { "epoch": 2.714561855670103, "grad_norm": 0.5330665111541748, "learning_rate": 2.732893075989956e-07, "loss": 0.5124, "step": 4213 }, { "epoch": 2.7152061855670104, "grad_norm": 0.5111385583877563, "learning_rate": 2.7206818896708565e-07, "loss": 0.5216, "step": 4214 }, { "epoch": 2.7158505154639174, "grad_norm": 0.5063999891281128, "learning_rate": 2.7084972826116274e-07, "loss": 0.5128, "step": 4215 }, { "epoch": 2.716494845360825, "grad_norm": 0.560964047908783, "learning_rate": 2.696339261662156e-07, "loss": 0.4812, "step": 4216 }, { "epoch": 2.717139175257732, "grad_norm": 0.5570883750915527, "learning_rate": 2.684207833657387e-07, "loss": 0.5408, "step": 4217 }, { "epoch": 2.7177835051546393, "grad_norm": 0.5385285019874573, "learning_rate": 2.6721030054172924e-07, "loss": 0.525, "step": 4218 }, { "epoch": 2.7184278350515463, "grad_norm": 0.5453154444694519, "learning_rate": 2.660024783746912e-07, "loss": 0.5333, "step": 4219 }, { "epoch": 2.7190721649484537, "grad_norm": 0.5754627585411072, "learning_rate": 2.647973175436336e-07, "loss": 0.5301, "step": 4220 }, { "epoch": 2.7197164948453607, "grad_norm": 0.6050049662590027, "learning_rate": 2.6359481872606563e-07, "loss": 0.5336, "step": 4221 }, { "epoch": 2.720360824742268, "grad_norm": 0.549436628818512, "learning_rate": 2.623949825980032e-07, "loss": 0.5168, "step": 4222 }, { "epoch": 2.721005154639175, "grad_norm": 0.6213060021400452, "learning_rate": 2.611978098339657e-07, "loss": 0.5403, "step": 4223 }, { "epoch": 2.7216494845360826, "grad_norm": 0.6363527178764343, "learning_rate": 2.600033011069725e-07, "loss": 0.5089, "step": 4224 }, { "epoch": 2.7222938144329896, "grad_norm": 0.5096006393432617, "learning_rate": 2.588114570885464e-07, "loss": 0.5178, "step": 4225 }, { "epoch": 2.722938144329897, "grad_norm": 0.5904222130775452, "learning_rate": 2.576222784487148e-07, "loss": 0.5103, "step": 4226 }, { "epoch": 2.723582474226804, "grad_norm": 0.5988823771476746, "learning_rate": 2.5643576585600184e-07, "loss": 0.5177, "step": 4227 }, { "epoch": 2.7242268041237114, "grad_norm": 0.6290488839149475, "learning_rate": 2.552519199774367e-07, "loss": 0.5324, "step": 4228 }, { "epoch": 2.7248711340206184, "grad_norm": 0.5418846607208252, "learning_rate": 2.5407074147854925e-07, "loss": 0.5042, "step": 4229 }, { "epoch": 2.725515463917526, "grad_norm": 0.5807674527168274, "learning_rate": 2.528922310233661e-07, "loss": 0.515, "step": 4230 }, { "epoch": 2.726159793814433, "grad_norm": 0.5170537829399109, "learning_rate": 2.51716389274419e-07, "loss": 0.5206, "step": 4231 }, { "epoch": 2.7268041237113403, "grad_norm": 0.5198007822036743, "learning_rate": 2.505432168927352e-07, "loss": 0.5228, "step": 4232 }, { "epoch": 2.7274484536082473, "grad_norm": 0.5495055317878723, "learning_rate": 2.4937271453784386e-07, "loss": 0.5098, "step": 4233 }, { "epoch": 2.7280927835051547, "grad_norm": 0.5654629468917847, "learning_rate": 2.482048828677741e-07, "loss": 0.4998, "step": 4234 }, { "epoch": 2.7287371134020617, "grad_norm": 0.591261088848114, "learning_rate": 2.470397225390486e-07, "loss": 0.4919, "step": 4235 }, { "epoch": 2.729381443298969, "grad_norm": 0.5608907341957092, "learning_rate": 2.45877234206694e-07, "loss": 0.4987, "step": 4236 }, { "epoch": 2.730025773195876, "grad_norm": 0.5499082207679749, "learning_rate": 2.447174185242324e-07, "loss": 0.5097, "step": 4237 }, { "epoch": 2.7306701030927836, "grad_norm": 0.6130073666572571, "learning_rate": 2.435602761436817e-07, "loss": 0.4835, "step": 4238 }, { "epoch": 2.7313144329896906, "grad_norm": 0.5205788612365723, "learning_rate": 2.4240580771556056e-07, "loss": 0.5019, "step": 4239 }, { "epoch": 2.731958762886598, "grad_norm": 0.5331275463104248, "learning_rate": 2.412540138888808e-07, "loss": 0.5021, "step": 4240 }, { "epoch": 2.732603092783505, "grad_norm": 0.5069374442100525, "learning_rate": 2.4010489531115343e-07, "loss": 0.5375, "step": 4241 }, { "epoch": 2.7332474226804124, "grad_norm": 0.5657426118850708, "learning_rate": 2.389584526283845e-07, "loss": 0.5258, "step": 4242 }, { "epoch": 2.7338917525773194, "grad_norm": 0.5670508146286011, "learning_rate": 2.3781468648507445e-07, "loss": 0.5163, "step": 4243 }, { "epoch": 2.734536082474227, "grad_norm": 0.5703803896903992, "learning_rate": 2.3667359752422127e-07, "loss": 0.4955, "step": 4244 }, { "epoch": 2.7351804123711343, "grad_norm": 0.5760935544967651, "learning_rate": 2.3553518638731731e-07, "loss": 0.5013, "step": 4245 }, { "epoch": 2.7358247422680413, "grad_norm": 0.5616676211357117, "learning_rate": 2.3439945371434792e-07, "loss": 0.5219, "step": 4246 }, { "epoch": 2.7364690721649483, "grad_norm": 0.5546470284461975, "learning_rate": 2.3326640014379397e-07, "loss": 0.5023, "step": 4247 }, { "epoch": 2.7371134020618557, "grad_norm": 0.5426527261734009, "learning_rate": 2.3213602631263155e-07, "loss": 0.5047, "step": 4248 }, { "epoch": 2.737757731958763, "grad_norm": 0.5248534083366394, "learning_rate": 2.3100833285632674e-07, "loss": 0.5201, "step": 4249 }, { "epoch": 2.73840206185567, "grad_norm": 0.6272618174552917, "learning_rate": 2.2988332040884133e-07, "loss": 0.5114, "step": 4250 }, { "epoch": 2.739046391752577, "grad_norm": 0.5712012052536011, "learning_rate": 2.287609896026305e-07, "loss": 0.5295, "step": 4251 }, { "epoch": 2.7396907216494846, "grad_norm": 0.5949087142944336, "learning_rate": 2.2764134106863957e-07, "loss": 0.5384, "step": 4252 }, { "epoch": 2.740335051546392, "grad_norm": 0.6023072600364685, "learning_rate": 2.265243754363089e-07, "loss": 0.5122, "step": 4253 }, { "epoch": 2.740979381443299, "grad_norm": 0.5270093083381653, "learning_rate": 2.254100933335668e-07, "loss": 0.5215, "step": 4254 }, { "epoch": 2.741623711340206, "grad_norm": 0.5473041534423828, "learning_rate": 2.24298495386836e-07, "loss": 0.5074, "step": 4255 }, { "epoch": 2.7422680412371134, "grad_norm": 0.560833752155304, "learning_rate": 2.2318958222103004e-07, "loss": 0.5306, "step": 4256 }, { "epoch": 2.742912371134021, "grad_norm": 0.5918184518814087, "learning_rate": 2.2208335445955077e-07, "loss": 0.52, "step": 4257 }, { "epoch": 2.743556701030928, "grad_norm": 0.5597960352897644, "learning_rate": 2.2097981272429236e-07, "loss": 0.5201, "step": 4258 }, { "epoch": 2.744201030927835, "grad_norm": 0.5259117484092712, "learning_rate": 2.1987895763564027e-07, "loss": 0.4873, "step": 4259 }, { "epoch": 2.7448453608247423, "grad_norm": 0.6005602478981018, "learning_rate": 2.1878078981246554e-07, "loss": 0.5349, "step": 4260 }, { "epoch": 2.7454896907216497, "grad_norm": 0.5658758878707886, "learning_rate": 2.1768530987213267e-07, "loss": 0.524, "step": 4261 }, { "epoch": 2.7461340206185567, "grad_norm": 0.5817517638206482, "learning_rate": 2.1659251843049289e-07, "loss": 0.5151, "step": 4262 }, { "epoch": 2.7467783505154637, "grad_norm": 0.5356471538543701, "learning_rate": 2.1550241610188538e-07, "loss": 0.5121, "step": 4263 }, { "epoch": 2.747422680412371, "grad_norm": 0.5293049216270447, "learning_rate": 2.144150034991399e-07, "loss": 0.5107, "step": 4264 }, { "epoch": 2.7480670103092786, "grad_norm": 0.5856543779373169, "learning_rate": 2.1333028123357302e-07, "loss": 0.5228, "step": 4265 }, { "epoch": 2.7487113402061856, "grad_norm": 0.561481237411499, "learning_rate": 2.1224824991498695e-07, "loss": 0.5232, "step": 4266 }, { "epoch": 2.7493556701030926, "grad_norm": 0.5529563426971436, "learning_rate": 2.1116891015167506e-07, "loss": 0.5375, "step": 4267 }, { "epoch": 2.75, "grad_norm": 0.5893868207931519, "learning_rate": 2.1009226255041315e-07, "loss": 0.494, "step": 4268 }, { "epoch": 2.7506443298969074, "grad_norm": 0.6161097288131714, "learning_rate": 2.0901830771646703e-07, "loss": 0.4966, "step": 4269 }, { "epoch": 2.7512886597938144, "grad_norm": 0.6089330911636353, "learning_rate": 2.0794704625358876e-07, "loss": 0.4921, "step": 4270 }, { "epoch": 2.7519329896907214, "grad_norm": 0.5943873524665833, "learning_rate": 2.0687847876401268e-07, "loss": 0.5247, "step": 4271 }, { "epoch": 2.752577319587629, "grad_norm": 0.5154080390930176, "learning_rate": 2.0581260584846164e-07, "loss": 0.4994, "step": 4272 }, { "epoch": 2.7532216494845363, "grad_norm": 0.628688395023346, "learning_rate": 2.0474942810614407e-07, "loss": 0.515, "step": 4273 }, { "epoch": 2.7538659793814433, "grad_norm": 0.565757155418396, "learning_rate": 2.0368894613475022e-07, "loss": 0.5281, "step": 4274 }, { "epoch": 2.7545103092783503, "grad_norm": 0.5139604806900024, "learning_rate": 2.0263116053045873e-07, "loss": 0.5128, "step": 4275 }, { "epoch": 2.7551546391752577, "grad_norm": 0.5616028308868408, "learning_rate": 2.0157607188792894e-07, "loss": 0.5268, "step": 4276 }, { "epoch": 2.755798969072165, "grad_norm": 0.5538349747657776, "learning_rate": 2.0052368080030583e-07, "loss": 0.5157, "step": 4277 }, { "epoch": 2.756443298969072, "grad_norm": 0.5440319180488586, "learning_rate": 1.9947398785921835e-07, "loss": 0.5021, "step": 4278 }, { "epoch": 2.757087628865979, "grad_norm": 0.5634934902191162, "learning_rate": 1.984269936547767e-07, "loss": 0.5058, "step": 4279 }, { "epoch": 2.7577319587628866, "grad_norm": 0.5898926854133606, "learning_rate": 1.9738269877557558e-07, "loss": 0.4938, "step": 4280 }, { "epoch": 2.758376288659794, "grad_norm": 0.5110098123550415, "learning_rate": 1.9634110380869208e-07, "loss": 0.52, "step": 4281 }, { "epoch": 2.759020618556701, "grad_norm": 0.5123064517974854, "learning_rate": 1.9530220933968446e-07, "loss": 0.5044, "step": 4282 }, { "epoch": 2.759664948453608, "grad_norm": 0.5694167017936707, "learning_rate": 1.9426601595259332e-07, "loss": 0.5009, "step": 4283 }, { "epoch": 2.7603092783505154, "grad_norm": 0.5981396436691284, "learning_rate": 1.9323252422994153e-07, "loss": 0.5253, "step": 4284 }, { "epoch": 2.760953608247423, "grad_norm": 0.5547893047332764, "learning_rate": 1.922017347527322e-07, "loss": 0.5185, "step": 4285 }, { "epoch": 2.76159793814433, "grad_norm": 0.5989051461219788, "learning_rate": 1.911736481004489e-07, "loss": 0.522, "step": 4286 }, { "epoch": 2.762242268041237, "grad_norm": 0.6066784858703613, "learning_rate": 1.9014826485105775e-07, "loss": 0.5318, "step": 4287 }, { "epoch": 2.7628865979381443, "grad_norm": 0.5775738954544067, "learning_rate": 1.8912558558100258e-07, "loss": 0.5245, "step": 4288 }, { "epoch": 2.7635309278350517, "grad_norm": 0.5357692241668701, "learning_rate": 1.8810561086520963e-07, "loss": 0.5143, "step": 4289 }, { "epoch": 2.7641752577319587, "grad_norm": 0.6609684228897095, "learning_rate": 1.8708834127708187e-07, "loss": 0.5176, "step": 4290 }, { "epoch": 2.7648195876288657, "grad_norm": 0.5958391427993774, "learning_rate": 1.8607377738850406e-07, "loss": 0.5184, "step": 4291 }, { "epoch": 2.765463917525773, "grad_norm": 0.6030781865119934, "learning_rate": 1.8506191976983888e-07, "loss": 0.5096, "step": 4292 }, { "epoch": 2.7661082474226806, "grad_norm": 0.5351424813270569, "learning_rate": 1.8405276898992687e-07, "loss": 0.5137, "step": 4293 }, { "epoch": 2.7667525773195876, "grad_norm": 0.5913203954696655, "learning_rate": 1.8304632561608748e-07, "loss": 0.4956, "step": 4294 }, { "epoch": 2.767396907216495, "grad_norm": 0.6259756088256836, "learning_rate": 1.8204259021411985e-07, "loss": 0.5109, "step": 4295 }, { "epoch": 2.768041237113402, "grad_norm": 0.541279137134552, "learning_rate": 1.8104156334829703e-07, "loss": 0.487, "step": 4296 }, { "epoch": 2.7686855670103094, "grad_norm": 0.5891938805580139, "learning_rate": 1.8004324558137277e-07, "loss": 0.5085, "step": 4297 }, { "epoch": 2.7693298969072164, "grad_norm": 0.6051477193832397, "learning_rate": 1.7904763747457642e-07, "loss": 0.499, "step": 4298 }, { "epoch": 2.769974226804124, "grad_norm": 0.528426468372345, "learning_rate": 1.7805473958761364e-07, "loss": 0.5246, "step": 4299 }, { "epoch": 2.770618556701031, "grad_norm": 0.593986988067627, "learning_rate": 1.7706455247866684e-07, "loss": 0.5139, "step": 4300 }, { "epoch": 2.7712628865979383, "grad_norm": 0.6265544295310974, "learning_rate": 1.7607707670439623e-07, "loss": 0.5091, "step": 4301 }, { "epoch": 2.7719072164948453, "grad_norm": 0.6986631751060486, "learning_rate": 1.7509231281993444e-07, "loss": 0.4994, "step": 4302 }, { "epoch": 2.7725515463917527, "grad_norm": 0.7363611459732056, "learning_rate": 1.7411026137889198e-07, "loss": 0.4978, "step": 4303 }, { "epoch": 2.7731958762886597, "grad_norm": 0.5467268824577332, "learning_rate": 1.731309229333533e-07, "loss": 0.517, "step": 4304 }, { "epoch": 2.773840206185567, "grad_norm": 0.6987946629524231, "learning_rate": 1.721542980338786e-07, "loss": 0.5058, "step": 4305 }, { "epoch": 2.774484536082474, "grad_norm": 0.5773074626922607, "learning_rate": 1.7118038722950313e-07, "loss": 0.5359, "step": 4306 }, { "epoch": 2.7751288659793816, "grad_norm": 0.5445587038993835, "learning_rate": 1.7020919106773393e-07, "loss": 0.5233, "step": 4307 }, { "epoch": 2.7757731958762886, "grad_norm": 0.5521110892295837, "learning_rate": 1.6924071009455422e-07, "loss": 0.4911, "step": 4308 }, { "epoch": 2.776417525773196, "grad_norm": 0.5909976959228516, "learning_rate": 1.6827494485441963e-07, "loss": 0.5085, "step": 4309 }, { "epoch": 2.777061855670103, "grad_norm": 0.6049740314483643, "learning_rate": 1.673118958902592e-07, "loss": 0.5044, "step": 4310 }, { "epoch": 2.7777061855670104, "grad_norm": 0.5391202569007874, "learning_rate": 1.663515637434765e-07, "loss": 0.4985, "step": 4311 }, { "epoch": 2.7783505154639174, "grad_norm": 0.5202606916427612, "learning_rate": 1.6539394895394412e-07, "loss": 0.5195, "step": 4312 }, { "epoch": 2.778994845360825, "grad_norm": 0.5606990456581116, "learning_rate": 1.6443905206001087e-07, "loss": 0.4978, "step": 4313 }, { "epoch": 2.779639175257732, "grad_norm": 0.5703455209732056, "learning_rate": 1.634868735984968e-07, "loss": 0.5205, "step": 4314 }, { "epoch": 2.7802835051546393, "grad_norm": 0.5061217546463013, "learning_rate": 1.6253741410469093e-07, "loss": 0.5084, "step": 4315 }, { "epoch": 2.7809278350515463, "grad_norm": 0.547258734703064, "learning_rate": 1.615906741123574e-07, "loss": 0.5274, "step": 4316 }, { "epoch": 2.7815721649484537, "grad_norm": 0.598960280418396, "learning_rate": 1.6064665415372993e-07, "loss": 0.5075, "step": 4317 }, { "epoch": 2.7822164948453607, "grad_norm": 0.5583512783050537, "learning_rate": 1.5970535475951176e-07, "loss": 0.5394, "step": 4318 }, { "epoch": 2.782860824742268, "grad_norm": 0.5747784972190857, "learning_rate": 1.587667764588796e-07, "loss": 0.5013, "step": 4319 }, { "epoch": 2.783505154639175, "grad_norm": 0.6001756191253662, "learning_rate": 1.5783091977947863e-07, "loss": 0.5332, "step": 4320 }, { "epoch": 2.7841494845360826, "grad_norm": 0.5069338083267212, "learning_rate": 1.568977852474235e-07, "loss": 0.5056, "step": 4321 }, { "epoch": 2.7847938144329896, "grad_norm": 0.5174503326416016, "learning_rate": 1.559673733872996e-07, "loss": 0.5433, "step": 4322 }, { "epoch": 2.785438144329897, "grad_norm": 0.5694887042045593, "learning_rate": 1.550396847221619e-07, "loss": 0.5209, "step": 4323 }, { "epoch": 2.786082474226804, "grad_norm": 0.5289561748504639, "learning_rate": 1.541147197735332e-07, "loss": 0.5174, "step": 4324 }, { "epoch": 2.7867268041237114, "grad_norm": 0.5297551155090332, "learning_rate": 1.53192479061407e-07, "loss": 0.4954, "step": 4325 }, { "epoch": 2.7873711340206184, "grad_norm": 0.5471329092979431, "learning_rate": 1.5227296310424244e-07, "loss": 0.5121, "step": 4326 }, { "epoch": 2.788015463917526, "grad_norm": 0.5727365612983704, "learning_rate": 1.5135617241896993e-07, "loss": 0.5047, "step": 4327 }, { "epoch": 2.788659793814433, "grad_norm": 0.5978706479072571, "learning_rate": 1.5044210752098666e-07, "loss": 0.513, "step": 4328 }, { "epoch": 2.7893041237113403, "grad_norm": 0.5036062598228455, "learning_rate": 1.4953076892415596e-07, "loss": 0.5314, "step": 4329 }, { "epoch": 2.7899484536082473, "grad_norm": 0.5435452461242676, "learning_rate": 1.4862215714081029e-07, "loss": 0.5127, "step": 4330 }, { "epoch": 2.7905927835051547, "grad_norm": 0.51711106300354, "learning_rate": 1.4771627268174993e-07, "loss": 0.5065, "step": 4331 }, { "epoch": 2.7912371134020617, "grad_norm": 0.571220874786377, "learning_rate": 1.468131160562386e-07, "loss": 0.5062, "step": 4332 }, { "epoch": 2.791881443298969, "grad_norm": 0.5573796033859253, "learning_rate": 1.4591268777200962e-07, "loss": 0.5318, "step": 4333 }, { "epoch": 2.792525773195876, "grad_norm": 0.5328905582427979, "learning_rate": 1.450149883352614e-07, "loss": 0.5267, "step": 4334 }, { "epoch": 2.7931701030927836, "grad_norm": 0.5131634473800659, "learning_rate": 1.4412001825065858e-07, "loss": 0.5223, "step": 4335 }, { "epoch": 2.7938144329896906, "grad_norm": 0.5895485281944275, "learning_rate": 1.432277780213298e-07, "loss": 0.5023, "step": 4336 }, { "epoch": 2.794458762886598, "grad_norm": 0.5799592733383179, "learning_rate": 1.4233826814887274e-07, "loss": 0.5019, "step": 4337 }, { "epoch": 2.795103092783505, "grad_norm": 0.5811841487884521, "learning_rate": 1.4145148913334516e-07, "loss": 0.4852, "step": 4338 }, { "epoch": 2.7957474226804124, "grad_norm": 0.5649160146713257, "learning_rate": 1.405674414732744e-07, "loss": 0.524, "step": 4339 }, { "epoch": 2.7963917525773194, "grad_norm": 0.5534244179725647, "learning_rate": 1.3968612566564844e-07, "loss": 0.4947, "step": 4340 }, { "epoch": 2.797036082474227, "grad_norm": 0.6158068180084229, "learning_rate": 1.3880754220592206e-07, "loss": 0.5145, "step": 4341 }, { "epoch": 2.7976804123711343, "grad_norm": 0.6022154688835144, "learning_rate": 1.3793169158801357e-07, "loss": 0.507, "step": 4342 }, { "epoch": 2.7983247422680413, "grad_norm": 0.5960054397583008, "learning_rate": 1.3705857430430347e-07, "loss": 0.5219, "step": 4343 }, { "epoch": 2.7989690721649483, "grad_norm": 0.5750449895858765, "learning_rate": 1.3618819084563696e-07, "loss": 0.5037, "step": 4344 }, { "epoch": 2.7996134020618557, "grad_norm": 0.5125272274017334, "learning_rate": 1.3532054170132258e-07, "loss": 0.5307, "step": 4345 }, { "epoch": 2.800257731958763, "grad_norm": 0.6043824553489685, "learning_rate": 1.3445562735912965e-07, "loss": 0.4988, "step": 4346 }, { "epoch": 2.80090206185567, "grad_norm": 0.5813114643096924, "learning_rate": 1.3359344830529363e-07, "loss": 0.5105, "step": 4347 }, { "epoch": 2.801546391752577, "grad_norm": 0.5704647898674011, "learning_rate": 1.327340050245085e-07, "loss": 0.5282, "step": 4348 }, { "epoch": 2.8021907216494846, "grad_norm": 0.5395419597625732, "learning_rate": 1.3187729799993277e-07, "loss": 0.509, "step": 4349 }, { "epoch": 2.802835051546392, "grad_norm": 0.556004524230957, "learning_rate": 1.3102332771318617e-07, "loss": 0.5086, "step": 4350 }, { "epoch": 2.803479381443299, "grad_norm": 0.6070494055747986, "learning_rate": 1.3017209464434856e-07, "loss": 0.4892, "step": 4351 }, { "epoch": 2.804123711340206, "grad_norm": 0.5785179734230042, "learning_rate": 1.293235992719627e-07, "loss": 0.5095, "step": 4352 }, { "epoch": 2.8047680412371134, "grad_norm": 0.6135639548301697, "learning_rate": 1.2847784207303203e-07, "loss": 0.4967, "step": 4353 }, { "epoch": 2.805412371134021, "grad_norm": 0.574277937412262, "learning_rate": 1.2763482352301903e-07, "loss": 0.5039, "step": 4354 }, { "epoch": 2.806056701030928, "grad_norm": 0.5602821707725525, "learning_rate": 1.26794544095849e-07, "loss": 0.5099, "step": 4355 }, { "epoch": 2.806701030927835, "grad_norm": 0.6018095016479492, "learning_rate": 1.2595700426390633e-07, "loss": 0.5027, "step": 4356 }, { "epoch": 2.8073453608247423, "grad_norm": 0.6344881653785706, "learning_rate": 1.251222044980338e-07, "loss": 0.5022, "step": 4357 }, { "epoch": 2.8079896907216497, "grad_norm": 0.5572983622550964, "learning_rate": 1.2429014526753602e-07, "loss": 0.5099, "step": 4358 }, { "epoch": 2.8086340206185567, "grad_norm": 0.6031149625778198, "learning_rate": 1.2346082704017605e-07, "loss": 0.5113, "step": 4359 }, { "epoch": 2.8092783505154637, "grad_norm": 0.6128017902374268, "learning_rate": 1.2263425028217592e-07, "loss": 0.499, "step": 4360 }, { "epoch": 2.809922680412371, "grad_norm": 0.5952181220054626, "learning_rate": 1.2181041545821727e-07, "loss": 0.51, "step": 4361 }, { "epoch": 2.8105670103092786, "grad_norm": 0.5662224292755127, "learning_rate": 1.209893230314385e-07, "loss": 0.5216, "step": 4362 }, { "epoch": 2.8112113402061856, "grad_norm": 0.53944331407547, "learning_rate": 1.2017097346343765e-07, "loss": 0.5018, "step": 4363 }, { "epoch": 2.8118556701030926, "grad_norm": 0.5653399229049683, "learning_rate": 1.193553672142711e-07, "loss": 0.5187, "step": 4364 }, { "epoch": 2.8125, "grad_norm": 0.6157789826393127, "learning_rate": 1.1854250474245154e-07, "loss": 0.5223, "step": 4365 }, { "epoch": 2.8131443298969074, "grad_norm": 0.5666000843048096, "learning_rate": 1.1773238650495122e-07, "loss": 0.5143, "step": 4366 }, { "epoch": 2.8137886597938144, "grad_norm": 0.626249372959137, "learning_rate": 1.1692501295719805e-07, "loss": 0.5198, "step": 4367 }, { "epoch": 2.8144329896907214, "grad_norm": 0.6133968830108643, "learning_rate": 1.1612038455307673e-07, "loss": 0.4935, "step": 4368 }, { "epoch": 2.815077319587629, "grad_norm": 0.5347908139228821, "learning_rate": 1.1531850174492987e-07, "loss": 0.4997, "step": 4369 }, { "epoch": 2.8157216494845363, "grad_norm": 0.6222226023674011, "learning_rate": 1.1451936498355686e-07, "loss": 0.5352, "step": 4370 }, { "epoch": 2.8163659793814433, "grad_norm": 0.5949161648750305, "learning_rate": 1.137229747182117e-07, "loss": 0.5442, "step": 4371 }, { "epoch": 2.8170103092783503, "grad_norm": 0.5981602668762207, "learning_rate": 1.1292933139660511e-07, "loss": 0.5257, "step": 4372 }, { "epoch": 2.8176546391752577, "grad_norm": 0.5937691926956177, "learning_rate": 1.1213843546490466e-07, "loss": 0.5058, "step": 4373 }, { "epoch": 2.818298969072165, "grad_norm": 0.5937234163284302, "learning_rate": 1.1135028736773079e-07, "loss": 0.5005, "step": 4374 }, { "epoch": 2.818943298969072, "grad_norm": 0.6363233327865601, "learning_rate": 1.1056488754816297e-07, "loss": 0.4928, "step": 4375 }, { "epoch": 2.819587628865979, "grad_norm": 0.5968869924545288, "learning_rate": 1.0978223644773134e-07, "loss": 0.5178, "step": 4376 }, { "epoch": 2.8202319587628866, "grad_norm": 0.5897549390792847, "learning_rate": 1.0900233450642339e-07, "loss": 0.5163, "step": 4377 }, { "epoch": 2.820876288659794, "grad_norm": 0.5364176630973816, "learning_rate": 1.082251821626823e-07, "loss": 0.5375, "step": 4378 }, { "epoch": 2.821520618556701, "grad_norm": 0.5732208490371704, "learning_rate": 1.0745077985340247e-07, "loss": 0.4913, "step": 4379 }, { "epoch": 2.822164948453608, "grad_norm": 0.49765506386756897, "learning_rate": 1.0667912801393343e-07, "loss": 0.4984, "step": 4380 }, { "epoch": 2.8228092783505154, "grad_norm": 0.5728128552436829, "learning_rate": 1.0591022707807986e-07, "loss": 0.5041, "step": 4381 }, { "epoch": 2.823453608247423, "grad_norm": 0.5397997498512268, "learning_rate": 1.0514407747809763e-07, "loss": 0.5129, "step": 4382 }, { "epoch": 2.82409793814433, "grad_norm": 0.5344615578651428, "learning_rate": 1.0438067964469834e-07, "loss": 0.5019, "step": 4383 }, { "epoch": 2.824742268041237, "grad_norm": 0.5343238115310669, "learning_rate": 1.0362003400704368e-07, "loss": 0.5292, "step": 4384 }, { "epoch": 2.8253865979381443, "grad_norm": 0.48034438490867615, "learning_rate": 1.0286214099275049e-07, "loss": 0.5077, "step": 4385 }, { "epoch": 2.8260309278350517, "grad_norm": 0.5519316792488098, "learning_rate": 1.0210700102788796e-07, "loss": 0.4899, "step": 4386 }, { "epoch": 2.8266752577319587, "grad_norm": 0.5851917862892151, "learning_rate": 1.0135461453697648e-07, "loss": 0.4954, "step": 4387 }, { "epoch": 2.8273195876288657, "grad_norm": 0.5450327396392822, "learning_rate": 1.0060498194298884e-07, "loss": 0.5101, "step": 4388 }, { "epoch": 2.827963917525773, "grad_norm": 0.5157790184020996, "learning_rate": 9.985810366735128e-08, "loss": 0.5266, "step": 4389 }, { "epoch": 2.8286082474226806, "grad_norm": 0.5851464867591858, "learning_rate": 9.911398012993789e-08, "loss": 0.5187, "step": 4390 }, { "epoch": 2.8292525773195876, "grad_norm": 0.5901756286621094, "learning_rate": 9.837261174907853e-08, "loss": 0.5062, "step": 4391 }, { "epoch": 2.829896907216495, "grad_norm": 0.5450929403305054, "learning_rate": 9.763399894155145e-08, "loss": 0.5011, "step": 4392 }, { "epoch": 2.830541237113402, "grad_norm": 0.5705446004867554, "learning_rate": 9.689814212258564e-08, "loss": 0.5139, "step": 4393 }, { "epoch": 2.8311855670103094, "grad_norm": 0.5780133605003357, "learning_rate": 9.616504170586238e-08, "loss": 0.4943, "step": 4394 }, { "epoch": 2.8318298969072164, "grad_norm": 0.5834952592849731, "learning_rate": 9.543469810351314e-08, "loss": 0.5027, "step": 4395 }, { "epoch": 2.832474226804124, "grad_norm": 0.5778325200080872, "learning_rate": 9.470711172611724e-08, "loss": 0.5038, "step": 4396 }, { "epoch": 2.833118556701031, "grad_norm": 0.5412313938140869, "learning_rate": 9.398228298270806e-08, "loss": 0.5016, "step": 4397 }, { "epoch": 2.8337628865979383, "grad_norm": 0.5719434022903442, "learning_rate": 9.32602122807641e-08, "loss": 0.5185, "step": 4398 }, { "epoch": 2.8344072164948453, "grad_norm": 0.6064339280128479, "learning_rate": 9.254090002621619e-08, "loss": 0.5019, "step": 4399 }, { "epoch": 2.8350515463917527, "grad_norm": 0.6230506896972656, "learning_rate": 9.18243466234453e-08, "loss": 0.5073, "step": 4400 }, { "epoch": 2.8356958762886597, "grad_norm": 0.5691801309585571, "learning_rate": 9.111055247527756e-08, "loss": 0.5175, "step": 4401 }, { "epoch": 2.836340206185567, "grad_norm": 0.6593831181526184, "learning_rate": 9.039951798299252e-08, "loss": 0.5184, "step": 4402 }, { "epoch": 2.836984536082474, "grad_norm": 0.5759062767028809, "learning_rate": 8.969124354631486e-08, "loss": 0.5257, "step": 4403 }, { "epoch": 2.8376288659793816, "grad_norm": 0.5517955422401428, "learning_rate": 8.898572956341945e-08, "loss": 0.4977, "step": 4404 }, { "epoch": 2.8382731958762886, "grad_norm": 0.6135358810424805, "learning_rate": 8.828297643092787e-08, "loss": 0.5071, "step": 4405 }, { "epoch": 2.838917525773196, "grad_norm": 0.517795979976654, "learning_rate": 8.75829845439119e-08, "loss": 0.5081, "step": 4406 }, { "epoch": 2.839561855670103, "grad_norm": 0.568693220615387, "learning_rate": 8.688575429588842e-08, "loss": 0.5272, "step": 4407 }, { "epoch": 2.8402061855670104, "grad_norm": 0.5973080992698669, "learning_rate": 8.619128607882388e-08, "loss": 0.5217, "step": 4408 }, { "epoch": 2.8408505154639174, "grad_norm": 0.5453476309776306, "learning_rate": 8.549958028313044e-08, "loss": 0.5163, "step": 4409 }, { "epoch": 2.841494845360825, "grad_norm": 0.6462000608444214, "learning_rate": 8.481063729766814e-08, "loss": 0.5169, "step": 4410 }, { "epoch": 2.842139175257732, "grad_norm": 0.5793086290359497, "learning_rate": 8.412445750974385e-08, "loss": 0.5216, "step": 4411 }, { "epoch": 2.8427835051546393, "grad_norm": 0.5988099575042725, "learning_rate": 8.344104130511065e-08, "loss": 0.5199, "step": 4412 }, { "epoch": 2.8434278350515463, "grad_norm": 0.5780421495437622, "learning_rate": 8.276038906796736e-08, "loss": 0.521, "step": 4413 }, { "epoch": 2.8440721649484537, "grad_norm": 0.59730064868927, "learning_rate": 8.208250118096173e-08, "loss": 0.5036, "step": 4414 }, { "epoch": 2.8447164948453607, "grad_norm": 0.6931182146072388, "learning_rate": 8.140737802518339e-08, "loss": 0.4984, "step": 4415 }, { "epoch": 2.845360824742268, "grad_norm": 0.65669846534729, "learning_rate": 8.073501998017152e-08, "loss": 0.5036, "step": 4416 }, { "epoch": 2.846005154639175, "grad_norm": 0.6250931620597839, "learning_rate": 8.006542742390821e-08, "loss": 0.5545, "step": 4417 }, { "epoch": 2.8466494845360826, "grad_norm": 0.6471143364906311, "learning_rate": 7.939860073282179e-08, "loss": 0.4922, "step": 4418 }, { "epoch": 2.8472938144329896, "grad_norm": 0.6173228025436401, "learning_rate": 7.873454028178684e-08, "loss": 0.5261, "step": 4419 }, { "epoch": 2.847938144329897, "grad_norm": 0.5641868114471436, "learning_rate": 7.807324644411974e-08, "loss": 0.4672, "step": 4420 }, { "epoch": 2.848582474226804, "grad_norm": 0.5527377128601074, "learning_rate": 7.74147195915842e-08, "loss": 0.5004, "step": 4421 }, { "epoch": 2.8492268041237114, "grad_norm": 0.5616501569747925, "learning_rate": 7.67589600943891e-08, "loss": 0.5378, "step": 4422 }, { "epoch": 2.8498711340206184, "grad_norm": 0.5275139212608337, "learning_rate": 7.6105968321184e-08, "loss": 0.5184, "step": 4423 }, { "epoch": 2.850515463917526, "grad_norm": 0.5833497643470764, "learning_rate": 7.545574463906636e-08, "loss": 0.513, "step": 4424 }, { "epoch": 2.851159793814433, "grad_norm": 0.5489292144775391, "learning_rate": 7.480828941357488e-08, "loss": 0.5283, "step": 4425 }, { "epoch": 2.8518041237113403, "grad_norm": 0.5633611679077148, "learning_rate": 7.416360300869285e-08, "loss": 0.5225, "step": 4426 }, { "epoch": 2.8524484536082473, "grad_norm": 0.5951183438301086, "learning_rate": 7.35216857868476e-08, "loss": 0.5089, "step": 4427 }, { "epoch": 2.8530927835051547, "grad_norm": 0.5807955265045166, "learning_rate": 7.288253810890878e-08, "loss": 0.5102, "step": 4428 }, { "epoch": 2.8537371134020617, "grad_norm": 0.5498061180114746, "learning_rate": 7.224616033418897e-08, "loss": 0.4978, "step": 4429 }, { "epoch": 2.854381443298969, "grad_norm": 0.545661211013794, "learning_rate": 7.161255282044421e-08, "loss": 0.5088, "step": 4430 }, { "epoch": 2.855025773195876, "grad_norm": 0.5724973082542419, "learning_rate": 7.098171592387294e-08, "loss": 0.5179, "step": 4431 }, { "epoch": 2.8556701030927836, "grad_norm": 0.6176350116729736, "learning_rate": 7.035364999911587e-08, "loss": 0.5051, "step": 4432 }, { "epoch": 2.8563144329896906, "grad_norm": 0.4873586893081665, "learning_rate": 6.972835539925671e-08, "loss": 0.5186, "step": 4433 }, { "epoch": 2.856958762886598, "grad_norm": 0.530602753162384, "learning_rate": 6.910583247581981e-08, "loss": 0.5043, "step": 4434 }, { "epoch": 2.857603092783505, "grad_norm": 0.5947563648223877, "learning_rate": 6.848608157877246e-08, "loss": 0.4858, "step": 4435 }, { "epoch": 2.8582474226804124, "grad_norm": 0.5409435629844666, "learning_rate": 6.786910305652373e-08, "loss": 0.4867, "step": 4436 }, { "epoch": 2.8588917525773194, "grad_norm": 0.5359827280044556, "learning_rate": 6.725489725592283e-08, "loss": 0.5163, "step": 4437 }, { "epoch": 2.859536082474227, "grad_norm": 0.5443687438964844, "learning_rate": 6.664346452226134e-08, "loss": 0.501, "step": 4438 }, { "epoch": 2.8601804123711343, "grad_norm": 0.5284249782562256, "learning_rate": 6.603480519927152e-08, "loss": 0.5008, "step": 4439 }, { "epoch": 2.8608247422680413, "grad_norm": 0.5713971257209778, "learning_rate": 6.542891962912634e-08, "loss": 0.5306, "step": 4440 }, { "epoch": 2.8614690721649483, "grad_norm": 0.4896419644355774, "learning_rate": 6.482580815244e-08, "loss": 0.5196, "step": 4441 }, { "epoch": 2.8621134020618557, "grad_norm": 0.5094815492630005, "learning_rate": 6.422547110826738e-08, "loss": 0.5177, "step": 4442 }, { "epoch": 2.862757731958763, "grad_norm": 0.5857886075973511, "learning_rate": 6.362790883410185e-08, "loss": 0.5211, "step": 4443 }, { "epoch": 2.86340206185567, "grad_norm": 0.625239908695221, "learning_rate": 6.30331216658786e-08, "loss": 0.5008, "step": 4444 }, { "epoch": 2.864046391752577, "grad_norm": 0.4989881217479706, "learning_rate": 6.244110993797237e-08, "loss": 0.5187, "step": 4445 }, { "epoch": 2.8646907216494846, "grad_norm": 0.6284793019294739, "learning_rate": 6.185187398319691e-08, "loss": 0.5101, "step": 4446 }, { "epoch": 2.865335051546392, "grad_norm": 0.5820673704147339, "learning_rate": 6.126541413280674e-08, "loss": 0.5116, "step": 4447 }, { "epoch": 2.865979381443299, "grad_norm": 0.5715214610099792, "learning_rate": 6.06817307164953e-08, "loss": 0.5188, "step": 4448 }, { "epoch": 2.866623711340206, "grad_norm": 0.6267213225364685, "learning_rate": 6.010082406239348e-08, "loss": 0.5201, "step": 4449 }, { "epoch": 2.8672680412371134, "grad_norm": 0.5689154267311096, "learning_rate": 5.952269449707449e-08, "loss": 0.5097, "step": 4450 }, { "epoch": 2.867912371134021, "grad_norm": 0.5231937766075134, "learning_rate": 5.894734234554722e-08, "loss": 0.5272, "step": 4451 }, { "epoch": 2.868556701030928, "grad_norm": 0.5608552694320679, "learning_rate": 5.8374767931260176e-08, "loss": 0.4972, "step": 4452 }, { "epoch": 2.869201030927835, "grad_norm": 0.540209174156189, "learning_rate": 5.780497157610199e-08, "loss": 0.5254, "step": 4453 }, { "epoch": 2.8698453608247423, "grad_norm": 0.5470871925354004, "learning_rate": 5.723795360039586e-08, "loss": 0.5133, "step": 4454 }, { "epoch": 2.8704896907216497, "grad_norm": 0.5794188976287842, "learning_rate": 5.6673714322906803e-08, "loss": 0.5127, "step": 4455 }, { "epoch": 2.8711340206185567, "grad_norm": 0.5217306613922119, "learning_rate": 5.611225406083609e-08, "loss": 0.5188, "step": 4456 }, { "epoch": 2.8717783505154637, "grad_norm": 0.4977891445159912, "learning_rate": 5.5553573129822346e-08, "loss": 0.521, "step": 4457 }, { "epoch": 2.872422680412371, "grad_norm": 0.6094259023666382, "learning_rate": 5.4997671843942116e-08, "loss": 0.5165, "step": 4458 }, { "epoch": 2.8730670103092786, "grad_norm": 0.5227946639060974, "learning_rate": 5.444455051570874e-08, "loss": 0.5156, "step": 4459 }, { "epoch": 2.8737113402061856, "grad_norm": 0.5061483383178711, "learning_rate": 5.38942094560746e-08, "loss": 0.5321, "step": 4460 }, { "epoch": 2.8743556701030926, "grad_norm": 0.5547683835029602, "learning_rate": 5.334664897442721e-08, "loss": 0.5347, "step": 4461 }, { "epoch": 2.875, "grad_norm": 0.5102273225784302, "learning_rate": 5.2801869378590885e-08, "loss": 0.5253, "step": 4462 }, { "epoch": 2.8756443298969074, "grad_norm": 0.5313528776168823, "learning_rate": 5.2259870974827876e-08, "loss": 0.5109, "step": 4463 }, { "epoch": 2.8762886597938144, "grad_norm": 0.5447465777397156, "learning_rate": 5.172065406783611e-08, "loss": 0.5383, "step": 4464 }, { "epoch": 2.8769329896907214, "grad_norm": 0.5100845098495483, "learning_rate": 5.118421896074977e-08, "loss": 0.5282, "step": 4465 }, { "epoch": 2.877577319587629, "grad_norm": 0.5709891319274902, "learning_rate": 5.065056595513984e-08, "loss": 0.5212, "step": 4466 }, { "epoch": 2.8782216494845363, "grad_norm": 0.5353963971138, "learning_rate": 5.0119695351012465e-08, "loss": 0.5252, "step": 4467 }, { "epoch": 2.8788659793814433, "grad_norm": 0.5120770931243896, "learning_rate": 4.9591607446809465e-08, "loss": 0.5177, "step": 4468 }, { "epoch": 2.8795103092783503, "grad_norm": 0.583452582359314, "learning_rate": 4.9066302539410024e-08, "loss": 0.5099, "step": 4469 }, { "epoch": 2.8801546391752577, "grad_norm": 0.5723310708999634, "learning_rate": 4.8543780924126264e-08, "loss": 0.509, "step": 4470 }, { "epoch": 2.880798969072165, "grad_norm": 0.5604143738746643, "learning_rate": 4.8024042894707635e-08, "loss": 0.5026, "step": 4471 }, { "epoch": 2.881443298969072, "grad_norm": 0.5489606857299805, "learning_rate": 4.750708874333765e-08, "loss": 0.5006, "step": 4472 }, { "epoch": 2.882087628865979, "grad_norm": 0.5552851557731628, "learning_rate": 4.699291876063494e-08, "loss": 0.5251, "step": 4473 }, { "epoch": 2.8827319587628866, "grad_norm": 0.5560833215713501, "learning_rate": 4.64815332356533e-08, "loss": 0.5161, "step": 4474 }, { "epoch": 2.883376288659794, "grad_norm": 0.5203348398208618, "learning_rate": 4.597293245588164e-08, "loss": 0.5334, "step": 4475 }, { "epoch": 2.884020618556701, "grad_norm": 0.5791516304016113, "learning_rate": 4.546711670724124e-08, "loss": 0.5097, "step": 4476 }, { "epoch": 2.884664948453608, "grad_norm": 0.670473039150238, "learning_rate": 4.496408627409021e-08, "loss": 0.5138, "step": 4477 }, { "epoch": 2.8853092783505154, "grad_norm": 0.5669862031936646, "learning_rate": 4.446384143922011e-08, "loss": 0.5223, "step": 4478 }, { "epoch": 2.885953608247423, "grad_norm": 0.5646092891693115, "learning_rate": 4.396638248385543e-08, "loss": 0.5083, "step": 4479 }, { "epoch": 2.88659793814433, "grad_norm": 0.6049929261207581, "learning_rate": 4.347170968765524e-08, "loss": 0.5189, "step": 4480 }, { "epoch": 2.887242268041237, "grad_norm": 0.5372269749641418, "learning_rate": 4.297982332871265e-08, "loss": 0.5224, "step": 4481 }, { "epoch": 2.8878865979381443, "grad_norm": 0.5672149062156677, "learning_rate": 4.249072368355367e-08, "loss": 0.4937, "step": 4482 }, { "epoch": 2.8885309278350517, "grad_norm": 0.4967897832393646, "learning_rate": 4.200441102713837e-08, "loss": 0.5303, "step": 4483 }, { "epoch": 2.8891752577319587, "grad_norm": 0.5954452157020569, "learning_rate": 4.152088563285972e-08, "loss": 0.5143, "step": 4484 }, { "epoch": 2.8898195876288657, "grad_norm": 0.55016028881073, "learning_rate": 4.104014777254306e-08, "loss": 0.5252, "step": 4485 }, { "epoch": 2.890463917525773, "grad_norm": 0.6163472533226013, "learning_rate": 4.0562197716448316e-08, "loss": 0.5152, "step": 4486 }, { "epoch": 2.8911082474226806, "grad_norm": 0.5955190658569336, "learning_rate": 4.0087035733266114e-08, "loss": 0.5036, "step": 4487 }, { "epoch": 2.8917525773195876, "grad_norm": 0.665378987789154, "learning_rate": 3.96146620901211e-08, "loss": 0.505, "step": 4488 }, { "epoch": 2.892396907216495, "grad_norm": 0.5669302344322205, "learning_rate": 3.914507705257031e-08, "loss": 0.511, "step": 4489 }, { "epoch": 2.893041237113402, "grad_norm": 0.5279627442359924, "learning_rate": 3.8678280884602016e-08, "loss": 0.5006, "step": 4490 }, { "epoch": 2.8936855670103094, "grad_norm": 0.5831915140151978, "learning_rate": 3.8214273848638516e-08, "loss": 0.5088, "step": 4491 }, { "epoch": 2.8943298969072164, "grad_norm": 0.5703362226486206, "learning_rate": 3.7753056205532265e-08, "loss": 0.5148, "step": 4492 }, { "epoch": 2.894974226804124, "grad_norm": 0.6483855247497559, "learning_rate": 3.729462821456919e-08, "loss": 0.502, "step": 4493 }, { "epoch": 2.895618556701031, "grad_norm": 0.5673989057540894, "learning_rate": 3.683899013346537e-08, "loss": 0.4985, "step": 4494 }, { "epoch": 2.8962628865979383, "grad_norm": 0.6420431733131409, "learning_rate": 3.638614221836923e-08, "loss": 0.5203, "step": 4495 }, { "epoch": 2.8969072164948453, "grad_norm": 0.6556150913238525, "learning_rate": 3.593608472386045e-08, "loss": 0.518, "step": 4496 }, { "epoch": 2.8975515463917527, "grad_norm": 0.528160810470581, "learning_rate": 3.5488817902951666e-08, "loss": 0.5174, "step": 4497 }, { "epoch": 2.8981958762886597, "grad_norm": 0.6643354296684265, "learning_rate": 3.50443420070834e-08, "loss": 0.5176, "step": 4498 }, { "epoch": 2.898840206185567, "grad_norm": 0.5191599726676941, "learning_rate": 3.460265728612966e-08, "loss": 0.5285, "step": 4499 }, { "epoch": 2.899484536082474, "grad_norm": 0.6007695198059082, "learning_rate": 3.416376398839516e-08, "loss": 0.5096, "step": 4500 }, { "epoch": 2.9001288659793816, "grad_norm": 0.6690667271614075, "learning_rate": 3.3727662360613644e-08, "loss": 0.519, "step": 4501 }, { "epoch": 2.9007731958762886, "grad_norm": 0.6102307438850403, "learning_rate": 3.3294352647951776e-08, "loss": 0.5287, "step": 4502 }, { "epoch": 2.901417525773196, "grad_norm": 0.5614368915557861, "learning_rate": 3.286383509400526e-08, "loss": 0.521, "step": 4503 }, { "epoch": 2.902061855670103, "grad_norm": 0.5899942517280579, "learning_rate": 3.243610994079993e-08, "loss": 0.4758, "step": 4504 }, { "epoch": 2.9027061855670104, "grad_norm": 0.6953026652336121, "learning_rate": 3.201117742879234e-08, "loss": 0.4967, "step": 4505 }, { "epoch": 2.9033505154639174, "grad_norm": 0.6203656196594238, "learning_rate": 3.1589037796869725e-08, "loss": 0.4971, "step": 4506 }, { "epoch": 2.903994845360825, "grad_norm": 0.5700042247772217, "learning_rate": 3.1169691282347817e-08, "loss": 0.5325, "step": 4507 }, { "epoch": 2.904639175257732, "grad_norm": 0.7028594017028809, "learning_rate": 3.075313812097302e-08, "loss": 0.507, "step": 4508 }, { "epoch": 2.9052835051546393, "grad_norm": 0.5949023962020874, "learning_rate": 3.0339378546920815e-08, "loss": 0.4906, "step": 4509 }, { "epoch": 2.9059278350515463, "grad_norm": 0.6281190514564514, "learning_rate": 2.992841279279734e-08, "loss": 0.5214, "step": 4510 }, { "epoch": 2.9065721649484537, "grad_norm": 0.6024733781814575, "learning_rate": 2.9520241089636692e-08, "loss": 0.5074, "step": 4511 }, { "epoch": 2.9072164948453607, "grad_norm": 0.575535774230957, "learning_rate": 2.9114863666903638e-08, "loss": 0.5082, "step": 4512 }, { "epoch": 2.907860824742268, "grad_norm": 0.5833802819252014, "learning_rate": 2.871228075249033e-08, "loss": 0.5318, "step": 4513 }, { "epoch": 2.908505154639175, "grad_norm": 0.6711902022361755, "learning_rate": 2.831249257271962e-08, "loss": 0.5157, "step": 4514 }, { "epoch": 2.9091494845360826, "grad_norm": 0.6628119945526123, "learning_rate": 2.7915499352341724e-08, "loss": 0.5237, "step": 4515 }, { "epoch": 2.9097938144329896, "grad_norm": 0.5524478554725647, "learning_rate": 2.7521301314537564e-08, "loss": 0.5295, "step": 4516 }, { "epoch": 2.910438144329897, "grad_norm": 0.5848913788795471, "learning_rate": 2.7129898680914314e-08, "loss": 0.5146, "step": 4517 }, { "epoch": 2.911082474226804, "grad_norm": 0.740699291229248, "learning_rate": 2.674129167150985e-08, "loss": 0.5053, "step": 4518 }, { "epoch": 2.9117268041237114, "grad_norm": 0.5190373659133911, "learning_rate": 2.635548050478831e-08, "loss": 0.5122, "step": 4519 }, { "epoch": 2.9123711340206184, "grad_norm": 0.6682837605476379, "learning_rate": 2.5972465397643977e-08, "loss": 0.5475, "step": 4520 }, { "epoch": 2.913015463917526, "grad_norm": 0.5929891467094421, "learning_rate": 2.5592246565398493e-08, "loss": 0.5133, "step": 4521 }, { "epoch": 2.913659793814433, "grad_norm": 0.6085596680641174, "learning_rate": 2.5214824221800882e-08, "loss": 0.5145, "step": 4522 }, { "epoch": 2.9143041237113403, "grad_norm": 0.6106369495391846, "learning_rate": 2.4840198579029195e-08, "loss": 0.5048, "step": 4523 }, { "epoch": 2.9149484536082473, "grad_norm": 0.6128644347190857, "learning_rate": 2.4468369847687746e-08, "loss": 0.5273, "step": 4524 }, { "epoch": 2.9155927835051547, "grad_norm": 0.7422507405281067, "learning_rate": 2.4099338236810988e-08, "loss": 0.5085, "step": 4525 }, { "epoch": 2.9162371134020617, "grad_norm": 0.5965543985366821, "learning_rate": 2.373310395385797e-08, "loss": 0.5155, "step": 4526 }, { "epoch": 2.916881443298969, "grad_norm": 0.573313295841217, "learning_rate": 2.3369667204716782e-08, "loss": 0.5066, "step": 4527 }, { "epoch": 2.917525773195876, "grad_norm": 0.6361870765686035, "learning_rate": 2.3009028193703432e-08, "loss": 0.5034, "step": 4528 }, { "epoch": 2.9181701030927836, "grad_norm": 0.5652021765708923, "learning_rate": 2.2651187123558515e-08, "loss": 0.52, "step": 4529 }, { "epoch": 2.9188144329896906, "grad_norm": 0.5658990740776062, "learning_rate": 2.229614419545334e-08, "loss": 0.506, "step": 4530 }, { "epoch": 2.919458762886598, "grad_norm": 0.5576101541519165, "learning_rate": 2.1943899608982687e-08, "loss": 0.4928, "step": 4531 }, { "epoch": 2.920103092783505, "grad_norm": 0.5416404008865356, "learning_rate": 2.1594453562170937e-08, "loss": 0.4865, "step": 4532 }, { "epoch": 2.9207474226804124, "grad_norm": 0.5848390460014343, "learning_rate": 2.1247806251467052e-08, "loss": 0.4956, "step": 4533 }, { "epoch": 2.9213917525773194, "grad_norm": 0.5917619466781616, "learning_rate": 2.0903957871747925e-08, "loss": 0.5224, "step": 4534 }, { "epoch": 2.922036082474227, "grad_norm": 0.54554283618927, "learning_rate": 2.056290861631671e-08, "loss": 0.5046, "step": 4535 }, { "epoch": 2.9226804123711343, "grad_norm": 0.5656002759933472, "learning_rate": 2.022465867690282e-08, "loss": 0.5109, "step": 4536 }, { "epoch": 2.9233247422680413, "grad_norm": 0.551986575126648, "learning_rate": 1.9889208243661363e-08, "loss": 0.5153, "step": 4537 }, { "epoch": 2.9239690721649483, "grad_norm": 0.5308968424797058, "learning_rate": 1.955655750517538e-08, "loss": 0.5162, "step": 4538 }, { "epoch": 2.9246134020618557, "grad_norm": 0.5325641632080078, "learning_rate": 1.9226706648451942e-08, "loss": 0.5281, "step": 4539 }, { "epoch": 2.925257731958763, "grad_norm": 0.5803263187408447, "learning_rate": 1.889965585892495e-08, "loss": 0.4965, "step": 4540 }, { "epoch": 2.92590206185567, "grad_norm": 0.6447356343269348, "learning_rate": 1.85754053204551e-08, "loss": 0.4899, "step": 4541 }, { "epoch": 2.926546391752577, "grad_norm": 0.5650786757469177, "learning_rate": 1.8253955215327135e-08, "loss": 0.5144, "step": 4542 }, { "epoch": 2.9271907216494846, "grad_norm": 0.5477502942085266, "learning_rate": 1.7935305724253172e-08, "loss": 0.5021, "step": 4543 }, { "epoch": 2.927835051546392, "grad_norm": 0.6359169483184814, "learning_rate": 1.7619457026369357e-08, "loss": 0.5219, "step": 4544 }, { "epoch": 2.928479381443299, "grad_norm": 0.6863119602203369, "learning_rate": 1.7306409299238103e-08, "loss": 0.5375, "step": 4545 }, { "epoch": 2.929123711340206, "grad_norm": 0.5342296361923218, "learning_rate": 1.699616271884752e-08, "loss": 0.5252, "step": 4546 }, { "epoch": 2.9297680412371134, "grad_norm": 0.6508448719978333, "learning_rate": 1.6688717459610315e-08, "loss": 0.5091, "step": 4547 }, { "epoch": 2.930412371134021, "grad_norm": 0.7356709837913513, "learning_rate": 1.6384073694363788e-08, "loss": 0.5069, "step": 4548 }, { "epoch": 2.931056701030928, "grad_norm": 0.6557682752609253, "learning_rate": 1.6082231594372054e-08, "loss": 0.5254, "step": 4549 }, { "epoch": 2.931701030927835, "grad_norm": 0.656166136264801, "learning_rate": 1.5783191329323268e-08, "loss": 0.5057, "step": 4550 }, { "epoch": 2.9323453608247423, "grad_norm": 0.7429404854774475, "learning_rate": 1.548695306733017e-08, "loss": 0.5216, "step": 4551 }, { "epoch": 2.9329896907216497, "grad_norm": 0.7679611444473267, "learning_rate": 1.5193516974930656e-08, "loss": 0.4991, "step": 4552 }, { "epoch": 2.9336340206185567, "grad_norm": 0.6786885857582092, "learning_rate": 1.4902883217087216e-08, "loss": 0.5228, "step": 4553 }, { "epoch": 2.9342783505154637, "grad_norm": 0.7371875047683716, "learning_rate": 1.4615051957186376e-08, "loss": 0.5231, "step": 4554 }, { "epoch": 2.934922680412371, "grad_norm": 0.7439994215965271, "learning_rate": 1.4330023357040367e-08, "loss": 0.4956, "step": 4555 }, { "epoch": 2.9355670103092786, "grad_norm": 0.5911663770675659, "learning_rate": 1.4047797576885458e-08, "loss": 0.5193, "step": 4556 }, { "epoch": 2.9362113402061856, "grad_norm": 0.6872801184654236, "learning_rate": 1.3768374775380843e-08, "loss": 0.4989, "step": 4557 }, { "epoch": 2.9368556701030926, "grad_norm": 0.6739519834518433, "learning_rate": 1.3491755109612536e-08, "loss": 0.5276, "step": 4558 }, { "epoch": 2.9375, "grad_norm": 0.5708818435668945, "learning_rate": 1.3217938735088365e-08, "loss": 0.4974, "step": 4559 }, { "epoch": 2.9381443298969074, "grad_norm": 0.5928760766983032, "learning_rate": 1.294692580574075e-08, "loss": 0.5213, "step": 4560 }, { "epoch": 2.9387886597938144, "grad_norm": 0.532185435295105, "learning_rate": 1.2678716473927821e-08, "loss": 0.5057, "step": 4561 }, { "epoch": 2.9394329896907214, "grad_norm": 0.5794960856437683, "learning_rate": 1.241331089042841e-08, "loss": 0.5217, "step": 4562 }, { "epoch": 2.940077319587629, "grad_norm": 0.5419527888298035, "learning_rate": 1.2150709204448719e-08, "loss": 0.5061, "step": 4563 }, { "epoch": 2.9407216494845363, "grad_norm": 0.5797025561332703, "learning_rate": 1.1890911563615659e-08, "loss": 0.5297, "step": 4564 }, { "epoch": 2.9413659793814433, "grad_norm": 0.5388844609260559, "learning_rate": 1.1633918113981846e-08, "loss": 0.5091, "step": 4565 }, { "epoch": 2.9420103092783503, "grad_norm": 0.5408979058265686, "learning_rate": 1.1379729000021711e-08, "loss": 0.5292, "step": 4566 }, { "epoch": 2.9426546391752577, "grad_norm": 0.5361945033073425, "learning_rate": 1.1128344364634835e-08, "loss": 0.4984, "step": 4567 }, { "epoch": 2.943298969072165, "grad_norm": 0.5420482158660889, "learning_rate": 1.0879764349143174e-08, "loss": 0.5193, "step": 4568 }, { "epoch": 2.943943298969072, "grad_norm": 0.49375295639038086, "learning_rate": 1.0633989093292163e-08, "loss": 0.518, "step": 4569 }, { "epoch": 2.944587628865979, "grad_norm": 0.5284269452095032, "learning_rate": 1.0391018735250725e-08, "loss": 0.5157, "step": 4570 }, { "epoch": 2.9452319587628866, "grad_norm": 0.6008866429328918, "learning_rate": 1.0150853411610706e-08, "loss": 0.5102, "step": 4571 }, { "epoch": 2.945876288659794, "grad_norm": 0.5242319107055664, "learning_rate": 9.913493257386885e-09, "loss": 0.5127, "step": 4572 }, { "epoch": 2.946520618556701, "grad_norm": 0.5353492498397827, "learning_rate": 9.678938406017524e-09, "loss": 0.523, "step": 4573 }, { "epoch": 2.947164948453608, "grad_norm": 0.5592984557151794, "learning_rate": 9.447188989363254e-09, "loss": 0.506, "step": 4574 }, { "epoch": 2.9478092783505154, "grad_norm": 0.5437702536582947, "learning_rate": 9.218245137708193e-09, "loss": 0.5096, "step": 4575 }, { "epoch": 2.948453608247423, "grad_norm": 0.5115686058998108, "learning_rate": 8.99210697975883e-09, "loss": 0.5198, "step": 4576 }, { "epoch": 2.94909793814433, "grad_norm": 0.5548731684684753, "learning_rate": 8.768774642644029e-09, "loss": 0.4923, "step": 4577 }, { "epoch": 2.949742268041237, "grad_norm": 0.5852733254432678, "learning_rate": 8.548248251916136e-09, "loss": 0.5041, "step": 4578 }, { "epoch": 2.9503865979381443, "grad_norm": 0.5282109379768372, "learning_rate": 8.33052793154987e-09, "loss": 0.4961, "step": 4579 }, { "epoch": 2.9510309278350517, "grad_norm": 0.5460972189903259, "learning_rate": 8.115613803941214e-09, "loss": 0.4923, "step": 4580 }, { "epoch": 2.9516752577319587, "grad_norm": 0.5341330766677856, "learning_rate": 7.903505989910743e-09, "loss": 0.5216, "step": 4581 }, { "epoch": 2.9523195876288657, "grad_norm": 0.5368437170982361, "learning_rate": 7.694204608699186e-09, "loss": 0.5141, "step": 4582 }, { "epoch": 2.952963917525773, "grad_norm": 0.5518361926078796, "learning_rate": 7.487709777971308e-09, "loss": 0.5067, "step": 4583 }, { "epoch": 2.9536082474226806, "grad_norm": 0.5749544501304626, "learning_rate": 7.2840216138131415e-09, "loss": 0.4898, "step": 4584 }, { "epoch": 2.9542525773195876, "grad_norm": 0.5400862693786621, "learning_rate": 7.083140230732533e-09, "loss": 0.5077, "step": 4585 }, { "epoch": 2.954896907216495, "grad_norm": 0.506264865398407, "learning_rate": 6.885065741661367e-09, "loss": 0.524, "step": 4586 }, { "epoch": 2.955541237113402, "grad_norm": 0.5480736494064331, "learning_rate": 6.689798257950575e-09, "loss": 0.5142, "step": 4587 }, { "epoch": 2.9561855670103094, "grad_norm": 0.560556173324585, "learning_rate": 6.497337889375677e-09, "loss": 0.4998, "step": 4588 }, { "epoch": 2.9568298969072164, "grad_norm": 0.5194936394691467, "learning_rate": 6.307684744133458e-09, "loss": 0.5033, "step": 4589 }, { "epoch": 2.957474226804124, "grad_norm": 0.5529318451881409, "learning_rate": 6.1208389288414104e-09, "loss": 0.5282, "step": 4590 }, { "epoch": 2.958118556701031, "grad_norm": 0.5132611989974976, "learning_rate": 5.936800548539956e-09, "loss": 0.5235, "step": 4591 }, { "epoch": 2.9587628865979383, "grad_norm": 0.5519582629203796, "learning_rate": 5.7555697066907775e-09, "loss": 0.5259, "step": 4592 }, { "epoch": 2.9594072164948453, "grad_norm": 0.5337942838668823, "learning_rate": 5.577146505177378e-09, "loss": 0.5129, "step": 4593 }, { "epoch": 2.9600515463917527, "grad_norm": 0.5767365097999573, "learning_rate": 5.401531044305075e-09, "loss": 0.4927, "step": 4594 }, { "epoch": 2.9606958762886597, "grad_norm": 0.5442022681236267, "learning_rate": 5.228723422800452e-09, "loss": 0.5093, "step": 4595 }, { "epoch": 2.961340206185567, "grad_norm": 0.5668591856956482, "learning_rate": 5.058723737811355e-09, "loss": 0.5303, "step": 4596 }, { "epoch": 2.961984536082474, "grad_norm": 0.5837090611457825, "learning_rate": 4.891532084907447e-09, "loss": 0.5318, "step": 4597 }, { "epoch": 2.9626288659793816, "grad_norm": 0.5434551239013672, "learning_rate": 4.727148558080208e-09, "loss": 0.506, "step": 4598 }, { "epoch": 2.9632731958762886, "grad_norm": 0.5230947732925415, "learning_rate": 4.5655732497418284e-09, "loss": 0.5057, "step": 4599 }, { "epoch": 2.963917525773196, "grad_norm": 0.5531985759735107, "learning_rate": 4.406806250725759e-09, "loss": 0.5221, "step": 4600 }, { "epoch": 2.964561855670103, "grad_norm": 0.5208873152732849, "learning_rate": 4.2508476502867154e-09, "loss": 0.5099, "step": 4601 }, { "epoch": 2.9652061855670104, "grad_norm": 0.517869770526886, "learning_rate": 4.097697536100676e-09, "loss": 0.5277, "step": 4602 }, { "epoch": 2.9658505154639174, "grad_norm": 0.6051976084709167, "learning_rate": 3.947355994265434e-09, "loss": 0.5003, "step": 4603 }, { "epoch": 2.966494845360825, "grad_norm": 0.5635430812835693, "learning_rate": 3.7998231092989394e-09, "loss": 0.5115, "step": 4604 }, { "epoch": 2.967139175257732, "grad_norm": 0.5854930281639099, "learning_rate": 3.6550989641404023e-09, "loss": 0.4885, "step": 4605 }, { "epoch": 2.9677835051546393, "grad_norm": 0.5694024562835693, "learning_rate": 3.5131836401502972e-09, "loss": 0.4767, "step": 4606 }, { "epoch": 2.9684278350515463, "grad_norm": 0.7533397078514099, "learning_rate": 3.3740772171098057e-09, "loss": 0.483, "step": 4607 }, { "epoch": 2.9690721649484537, "grad_norm": 0.5806443095207214, "learning_rate": 3.237779773221372e-09, "loss": 0.5323, "step": 4608 }, { "epoch": 2.9697164948453607, "grad_norm": 0.658989429473877, "learning_rate": 3.104291385107594e-09, "loss": 0.5085, "step": 4609 }, { "epoch": 2.970360824742268, "grad_norm": 0.6446241140365601, "learning_rate": 2.973612127812886e-09, "loss": 0.5034, "step": 4610 }, { "epoch": 2.971005154639175, "grad_norm": 0.5872840285301208, "learning_rate": 2.845742074801261e-09, "loss": 0.5132, "step": 4611 }, { "epoch": 2.9716494845360826, "grad_norm": 0.6283121109008789, "learning_rate": 2.7206812979579944e-09, "loss": 0.5082, "step": 4612 }, { "epoch": 2.9722938144329896, "grad_norm": 0.5762442350387573, "learning_rate": 2.59842986759018e-09, "loss": 0.4958, "step": 4613 }, { "epoch": 2.972938144329897, "grad_norm": 0.5810965895652771, "learning_rate": 2.478987852423953e-09, "loss": 0.5068, "step": 4614 }, { "epoch": 2.973582474226804, "grad_norm": 0.6231669187545776, "learning_rate": 2.3623553196061577e-09, "loss": 0.5026, "step": 4615 }, { "epoch": 2.9742268041237114, "grad_norm": 0.552555501461029, "learning_rate": 2.2485323347054555e-09, "loss": 0.5163, "step": 4616 }, { "epoch": 2.9748711340206184, "grad_norm": 0.5805788636207581, "learning_rate": 2.1375189617101055e-09, "loss": 0.5201, "step": 4617 }, { "epoch": 2.975515463917526, "grad_norm": 0.5655517578125, "learning_rate": 2.0293152630290748e-09, "loss": 0.5201, "step": 4618 }, { "epoch": 2.976159793814433, "grad_norm": 0.5693396329879761, "learning_rate": 1.9239212994914826e-09, "loss": 0.5087, "step": 4619 }, { "epoch": 2.9768041237113403, "grad_norm": 0.5741912126541138, "learning_rate": 1.8213371303477112e-09, "loss": 0.51, "step": 4620 }, { "epoch": 2.9774484536082473, "grad_norm": 0.590345025062561, "learning_rate": 1.7215628132671857e-09, "loss": 0.5103, "step": 4621 }, { "epoch": 2.9780927835051547, "grad_norm": 0.5741516351699829, "learning_rate": 1.624598404341704e-09, "loss": 0.5204, "step": 4622 }, { "epoch": 2.9787371134020617, "grad_norm": 0.5560887455940247, "learning_rate": 1.530443958080996e-09, "loss": 0.5314, "step": 4623 }, { "epoch": 2.979381443298969, "grad_norm": 0.5414069294929504, "learning_rate": 1.4390995274177199e-09, "loss": 0.5122, "step": 4624 }, { "epoch": 2.980025773195876, "grad_norm": 0.5503866076469421, "learning_rate": 1.3505651637024664e-09, "loss": 0.5197, "step": 4625 }, { "epoch": 2.9806701030927836, "grad_norm": 0.5408263206481934, "learning_rate": 1.2648409167070886e-09, "loss": 0.5283, "step": 4626 }, { "epoch": 2.9813144329896906, "grad_norm": 0.5945572853088379, "learning_rate": 1.1819268346241475e-09, "loss": 0.5183, "step": 4627 }, { "epoch": 2.981958762886598, "grad_norm": 0.5758721232414246, "learning_rate": 1.1018229640658018e-09, "loss": 0.5285, "step": 4628 }, { "epoch": 2.982603092783505, "grad_norm": 0.6566312909126282, "learning_rate": 1.0245293500638077e-09, "loss": 0.4978, "step": 4629 }, { "epoch": 2.9832474226804124, "grad_norm": 0.556899905204773, "learning_rate": 9.50046036071739e-10, "loss": 0.513, "step": 4630 }, { "epoch": 2.9838917525773194, "grad_norm": 0.6153331995010376, "learning_rate": 8.783730639616572e-10, "loss": 0.5162, "step": 4631 }, { "epoch": 2.984536082474227, "grad_norm": 0.5659255385398865, "learning_rate": 8.095104740257764e-10, "loss": 0.5138, "step": 4632 }, { "epoch": 2.9851804123711343, "grad_norm": 0.6614088416099548, "learning_rate": 7.434583049781286e-10, "loss": 0.5163, "step": 4633 }, { "epoch": 2.9858247422680413, "grad_norm": 0.6274524927139282, "learning_rate": 6.802165939512329e-10, "loss": 0.5031, "step": 4634 }, { "epoch": 2.9864690721649483, "grad_norm": 0.5860931277275085, "learning_rate": 6.197853764972062e-10, "loss": 0.511, "step": 4635 }, { "epoch": 2.9871134020618557, "grad_norm": 0.5965206027030945, "learning_rate": 5.621646865899832e-10, "loss": 0.5154, "step": 4636 }, { "epoch": 2.987757731958763, "grad_norm": 0.6151943206787109, "learning_rate": 5.073545566219861e-10, "loss": 0.5199, "step": 4637 }, { "epoch": 2.98840206185567, "grad_norm": 0.6209184527397156, "learning_rate": 4.553550174063448e-10, "loss": 0.5164, "step": 4638 }, { "epoch": 2.989046391752577, "grad_norm": 0.6127544045448303, "learning_rate": 4.0616609817523135e-10, "loss": 0.5151, "step": 4639 }, { "epoch": 2.9896907216494846, "grad_norm": 0.6568924784660339, "learning_rate": 3.59787826582636e-10, "loss": 0.5085, "step": 4640 }, { "epoch": 2.990335051546392, "grad_norm": 0.7383096814155579, "learning_rate": 3.162202286999261e-10, "loss": 0.5357, "step": 4641 }, { "epoch": 2.990979381443299, "grad_norm": 0.5669957399368286, "learning_rate": 2.7546332902084195e-10, "loss": 0.5046, "step": 4642 }, { "epoch": 2.991623711340206, "grad_norm": 0.6778392195701599, "learning_rate": 2.3751715045761125e-10, "loss": 0.4956, "step": 4643 }, { "epoch": 2.9922680412371134, "grad_norm": 0.6945351362228394, "learning_rate": 2.0238171434261433e-10, "loss": 0.5119, "step": 4644 }, { "epoch": 2.992912371134021, "grad_norm": 0.7070682644844055, "learning_rate": 1.700570404272739e-10, "loss": 0.5133, "step": 4645 }, { "epoch": 2.993556701030928, "grad_norm": 0.598598837852478, "learning_rate": 1.405431468848306e-10, "loss": 0.5191, "step": 4646 }, { "epoch": 2.994201030927835, "grad_norm": 0.6289284229278564, "learning_rate": 1.1384005030701251e-10, "loss": 0.5029, "step": 4647 }, { "epoch": 2.9948453608247423, "grad_norm": 0.6145480275154114, "learning_rate": 8.994776570570019e-11, "loss": 0.516, "step": 4648 }, { "epoch": 2.9954896907216497, "grad_norm": 0.5965689420700073, "learning_rate": 6.886630651181669e-11, "loss": 0.526, "step": 4649 }, { "epoch": 2.9961340206185567, "grad_norm": 0.5733819603919983, "learning_rate": 5.0595684578103045e-11, "loss": 0.4993, "step": 4650 }, { "epoch": 2.9967783505154637, "grad_norm": 0.6139348745346069, "learning_rate": 3.5135910174677325e-11, "loss": 0.506, "step": 4651 }, { "epoch": 2.997422680412371, "grad_norm": 0.6007868647575378, "learning_rate": 2.248699199292048e-11, "loss": 0.5297, "step": 4652 }, { "epoch": 2.9980670103092786, "grad_norm": 0.5594995021820068, "learning_rate": 1.2648937143810991e-11, "loss": 0.52, "step": 4653 }, { "epoch": 2.9987113402061856, "grad_norm": 0.6329032182693481, "learning_rate": 5.62175115847996e-12, "loss": 0.5306, "step": 4654 }, { "epoch": 2.9993556701030926, "grad_norm": 0.5982125401496887, "learning_rate": 1.4054379871009106e-12, "loss": 0.4984, "step": 4655 }, { "epoch": 3.0, "grad_norm": 0.5997171401977539, "learning_rate": 0.0, "loss": 0.5069, "step": 4656 }, { "epoch": 3.0, "step": 4656, "total_flos": 5.799223180316901e+18, "train_loss": 0.17011004269660748, "train_runtime": 48515.4596, "train_samples_per_second": 9.212, "train_steps_per_second": 0.096 } ], "logging_steps": 1, "max_steps": 4656, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.799223180316901e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }