{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3157721616139466, "eval_steps": 500, "global_step": 720, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00043857244668603693, "grad_norm": 1.5650805234909058, "learning_rate": 1.4492753623188406e-08, "loss": 0.4933, "step": 1 }, { "epoch": 0.0008771448933720739, "grad_norm": 1.2313175201416016, "learning_rate": 2.898550724637681e-08, "loss": 0.4744, "step": 2 }, { "epoch": 0.001315717340058111, "grad_norm": 1.4933998584747314, "learning_rate": 4.347826086956521e-08, "loss": 0.4723, "step": 3 }, { "epoch": 0.0017542897867441477, "grad_norm": 1.4024336338043213, "learning_rate": 5.797101449275362e-08, "loss": 0.4569, "step": 4 }, { "epoch": 0.0021928622334301848, "grad_norm": 1.2763197422027588, "learning_rate": 7.246376811594203e-08, "loss": 0.4702, "step": 5 }, { "epoch": 0.002631434680116222, "grad_norm": 1.3957324028015137, "learning_rate": 8.695652173913042e-08, "loss": 0.4788, "step": 6 }, { "epoch": 0.003070007126802259, "grad_norm": 1.3984495401382446, "learning_rate": 1.0144927536231885e-07, "loss": 0.4864, "step": 7 }, { "epoch": 0.0035085795734882954, "grad_norm": 1.321620225906372, "learning_rate": 1.1594202898550725e-07, "loss": 0.4611, "step": 8 }, { "epoch": 0.0039471520201743325, "grad_norm": 1.2592936754226685, "learning_rate": 1.3043478260869563e-07, "loss": 0.4508, "step": 9 }, { "epoch": 0.0043857244668603695, "grad_norm": 1.4119383096694946, "learning_rate": 1.4492753623188405e-07, "loss": 0.5231, "step": 10 }, { "epoch": 0.0048242969135464065, "grad_norm": 1.343310832977295, "learning_rate": 1.5942028985507245e-07, "loss": 0.4483, "step": 11 }, { "epoch": 0.005262869360232444, "grad_norm": 1.522481918334961, "learning_rate": 1.7391304347826085e-07, "loss": 0.4727, "step": 12 }, { "epoch": 0.005701441806918481, "grad_norm": 1.2751290798187256, "learning_rate": 1.8840579710144927e-07, "loss": 0.4973, "step": 13 }, { "epoch": 0.006140014253604518, "grad_norm": 1.29275643825531, "learning_rate": 2.028985507246377e-07, "loss": 0.4627, "step": 14 }, { "epoch": 0.006578586700290554, "grad_norm": 1.2619212865829468, "learning_rate": 2.1739130434782607e-07, "loss": 0.4771, "step": 15 }, { "epoch": 0.007017159146976591, "grad_norm": 1.4083727598190308, "learning_rate": 2.318840579710145e-07, "loss": 0.4892, "step": 16 }, { "epoch": 0.007455731593662628, "grad_norm": 1.468392252922058, "learning_rate": 2.463768115942029e-07, "loss": 0.4481, "step": 17 }, { "epoch": 0.007894304040348665, "grad_norm": 1.4284064769744873, "learning_rate": 2.6086956521739126e-07, "loss": 0.454, "step": 18 }, { "epoch": 0.008332876487034702, "grad_norm": 1.2147692441940308, "learning_rate": 2.753623188405797e-07, "loss": 0.4357, "step": 19 }, { "epoch": 0.008771448933720739, "grad_norm": 1.2484781742095947, "learning_rate": 2.898550724637681e-07, "loss": 0.47, "step": 20 }, { "epoch": 0.009210021380406776, "grad_norm": 1.3065662384033203, "learning_rate": 3.043478260869565e-07, "loss": 0.515, "step": 21 }, { "epoch": 0.009648593827092813, "grad_norm": 1.2731220722198486, "learning_rate": 3.188405797101449e-07, "loss": 0.4567, "step": 22 }, { "epoch": 0.01008716627377885, "grad_norm": 1.2372796535491943, "learning_rate": 3.333333333333333e-07, "loss": 0.4886, "step": 23 }, { "epoch": 0.010525738720464887, "grad_norm": 1.2341394424438477, "learning_rate": 3.478260869565217e-07, "loss": 0.4618, "step": 24 }, { "epoch": 0.010964311167150924, "grad_norm": 1.2708252668380737, "learning_rate": 3.6231884057971015e-07, "loss": 0.4695, "step": 25 }, { "epoch": 0.011402883613836961, "grad_norm": 1.326284646987915, "learning_rate": 3.7681159420289855e-07, "loss": 0.4807, "step": 26 }, { "epoch": 0.011841456060522998, "grad_norm": 1.267345666885376, "learning_rate": 3.9130434782608694e-07, "loss": 0.4729, "step": 27 }, { "epoch": 0.012280028507209035, "grad_norm": 1.302316665649414, "learning_rate": 4.057971014492754e-07, "loss": 0.4427, "step": 28 }, { "epoch": 0.012718600953895072, "grad_norm": 1.2135436534881592, "learning_rate": 4.2028985507246374e-07, "loss": 0.4371, "step": 29 }, { "epoch": 0.013157173400581108, "grad_norm": 1.162211298942566, "learning_rate": 4.3478260869565214e-07, "loss": 0.4611, "step": 30 }, { "epoch": 0.013595745847267145, "grad_norm": 1.305415391921997, "learning_rate": 4.4927536231884053e-07, "loss": 0.4789, "step": 31 }, { "epoch": 0.014034318293953182, "grad_norm": 1.545601725578308, "learning_rate": 4.63768115942029e-07, "loss": 0.5328, "step": 32 }, { "epoch": 0.014472890740639219, "grad_norm": 1.4781523942947388, "learning_rate": 4.782608695652174e-07, "loss": 0.4827, "step": 33 }, { "epoch": 0.014911463187325256, "grad_norm": 1.2530491352081299, "learning_rate": 4.927536231884058e-07, "loss": 0.4836, "step": 34 }, { "epoch": 0.015350035634011293, "grad_norm": 1.2580320835113525, "learning_rate": 5.072463768115942e-07, "loss": 0.4536, "step": 35 }, { "epoch": 0.01578860808069733, "grad_norm": 1.2908357381820679, "learning_rate": 5.217391304347825e-07, "loss": 0.4578, "step": 36 }, { "epoch": 0.016227180527383367, "grad_norm": 1.2363778352737427, "learning_rate": 5.36231884057971e-07, "loss": 0.4768, "step": 37 }, { "epoch": 0.016665752974069404, "grad_norm": 1.395221471786499, "learning_rate": 5.507246376811594e-07, "loss": 0.4261, "step": 38 }, { "epoch": 0.01710432542075544, "grad_norm": 1.5333712100982666, "learning_rate": 5.652173913043477e-07, "loss": 0.5309, "step": 39 }, { "epoch": 0.017542897867441478, "grad_norm": 1.4437744617462158, "learning_rate": 5.797101449275362e-07, "loss": 0.5116, "step": 40 }, { "epoch": 0.017981470314127515, "grad_norm": 1.3201457262039185, "learning_rate": 5.942028985507246e-07, "loss": 0.4506, "step": 41 }, { "epoch": 0.018420042760813552, "grad_norm": 1.2283574342727661, "learning_rate": 6.08695652173913e-07, "loss": 0.5068, "step": 42 }, { "epoch": 0.01885861520749959, "grad_norm": 1.355634093284607, "learning_rate": 6.231884057971014e-07, "loss": 0.5099, "step": 43 }, { "epoch": 0.019297187654185626, "grad_norm": 1.284976601600647, "learning_rate": 6.376811594202898e-07, "loss": 0.4596, "step": 44 }, { "epoch": 0.019735760100871663, "grad_norm": 1.3104331493377686, "learning_rate": 6.521739130434782e-07, "loss": 0.4222, "step": 45 }, { "epoch": 0.0201743325475577, "grad_norm": 1.2699097394943237, "learning_rate": 6.666666666666666e-07, "loss": 0.4565, "step": 46 }, { "epoch": 0.020612904994243737, "grad_norm": 1.401367425918579, "learning_rate": 6.811594202898551e-07, "loss": 0.4207, "step": 47 }, { "epoch": 0.021051477440929774, "grad_norm": 1.2467551231384277, "learning_rate": 6.956521739130434e-07, "loss": 0.448, "step": 48 }, { "epoch": 0.02149004988761581, "grad_norm": 1.304906964302063, "learning_rate": 7.101449275362319e-07, "loss": 0.451, "step": 49 }, { "epoch": 0.02192862233430185, "grad_norm": 1.21262526512146, "learning_rate": 7.246376811594203e-07, "loss": 0.4607, "step": 50 }, { "epoch": 0.022367194780987885, "grad_norm": 1.1871509552001953, "learning_rate": 7.391304347826086e-07, "loss": 0.4355, "step": 51 }, { "epoch": 0.022805767227673922, "grad_norm": 1.2784358263015747, "learning_rate": 7.536231884057971e-07, "loss": 0.4316, "step": 52 }, { "epoch": 0.02324433967435996, "grad_norm": 1.2877992391586304, "learning_rate": 7.681159420289855e-07, "loss": 0.4737, "step": 53 }, { "epoch": 0.023682912121045997, "grad_norm": 1.315425157546997, "learning_rate": 7.826086956521739e-07, "loss": 0.4885, "step": 54 }, { "epoch": 0.024121484567732034, "grad_norm": 1.2742252349853516, "learning_rate": 7.971014492753623e-07, "loss": 0.4956, "step": 55 }, { "epoch": 0.02456005701441807, "grad_norm": 1.2137497663497925, "learning_rate": 8.115942028985508e-07, "loss": 0.3931, "step": 56 }, { "epoch": 0.024998629461104108, "grad_norm": 1.2331868410110474, "learning_rate": 8.260869565217391e-07, "loss": 0.463, "step": 57 }, { "epoch": 0.025437201907790145, "grad_norm": 1.3358384370803833, "learning_rate": 8.405797101449275e-07, "loss": 0.4744, "step": 58 }, { "epoch": 0.025875774354476182, "grad_norm": 1.3751511573791504, "learning_rate": 8.550724637681159e-07, "loss": 0.4179, "step": 59 }, { "epoch": 0.026314346801162215, "grad_norm": 1.3184670209884644, "learning_rate": 8.695652173913043e-07, "loss": 0.4813, "step": 60 }, { "epoch": 0.026752919247848252, "grad_norm": 1.2712794542312622, "learning_rate": 8.840579710144928e-07, "loss": 0.4931, "step": 61 }, { "epoch": 0.02719149169453429, "grad_norm": 1.3398113250732422, "learning_rate": 8.985507246376811e-07, "loss": 0.54, "step": 62 }, { "epoch": 0.027630064141220326, "grad_norm": 1.2505016326904297, "learning_rate": 9.130434782608695e-07, "loss": 0.4464, "step": 63 }, { "epoch": 0.028068636587906363, "grad_norm": 1.1980383396148682, "learning_rate": 9.27536231884058e-07, "loss": 0.4118, "step": 64 }, { "epoch": 0.0285072090345924, "grad_norm": 1.27497136592865, "learning_rate": 9.420289855072463e-07, "loss": 0.4491, "step": 65 }, { "epoch": 0.028945781481278438, "grad_norm": 1.2081972360610962, "learning_rate": 9.565217391304349e-07, "loss": 0.3582, "step": 66 }, { "epoch": 0.029384353927964475, "grad_norm": 1.2842012643814087, "learning_rate": 9.710144927536232e-07, "loss": 0.4584, "step": 67 }, { "epoch": 0.02982292637465051, "grad_norm": 1.2653955221176147, "learning_rate": 9.855072463768117e-07, "loss": 0.463, "step": 68 }, { "epoch": 0.03026149882133655, "grad_norm": 1.190528392791748, "learning_rate": 1e-06, "loss": 0.4628, "step": 69 }, { "epoch": 0.030700071268022586, "grad_norm": 1.1633963584899902, "learning_rate": 9.999994952664242e-07, "loss": 0.4204, "step": 70 }, { "epoch": 0.031138643714708623, "grad_norm": 1.4251805543899536, "learning_rate": 9.999979810667154e-07, "loss": 0.492, "step": 71 }, { "epoch": 0.03157721616139466, "grad_norm": 1.4169254302978516, "learning_rate": 9.99995457403931e-07, "loss": 0.4918, "step": 72 }, { "epoch": 0.0320157886080807, "grad_norm": 1.2996279001235962, "learning_rate": 9.999919242831662e-07, "loss": 0.4389, "step": 73 }, { "epoch": 0.032454361054766734, "grad_norm": 1.1896483898162842, "learning_rate": 9.999873817115539e-07, "loss": 0.4169, "step": 74 }, { "epoch": 0.032892933501452774, "grad_norm": 1.3035953044891357, "learning_rate": 9.999818296982652e-07, "loss": 0.4625, "step": 75 }, { "epoch": 0.03333150594813881, "grad_norm": 1.2109572887420654, "learning_rate": 9.999752682545095e-07, "loss": 0.4225, "step": 76 }, { "epoch": 0.03377007839482485, "grad_norm": 1.1753894090652466, "learning_rate": 9.999676973935336e-07, "loss": 0.4265, "step": 77 }, { "epoch": 0.03420865084151088, "grad_norm": 1.2878810167312622, "learning_rate": 9.99959117130623e-07, "loss": 0.4736, "step": 78 }, { "epoch": 0.03464722328819692, "grad_norm": 1.29270601272583, "learning_rate": 9.999495274831003e-07, "loss": 0.4767, "step": 79 }, { "epoch": 0.035085795734882956, "grad_norm": 1.1873911619186401, "learning_rate": 9.999389284703264e-07, "loss": 0.4282, "step": 80 }, { "epoch": 0.03552436818156899, "grad_norm": 1.4078459739685059, "learning_rate": 9.999273201137004e-07, "loss": 0.4702, "step": 81 }, { "epoch": 0.03596294062825503, "grad_norm": 1.209952473640442, "learning_rate": 9.999147024366583e-07, "loss": 0.4174, "step": 82 }, { "epoch": 0.036401513074941064, "grad_norm": 1.1953450441360474, "learning_rate": 9.999010754646748e-07, "loss": 0.4217, "step": 83 }, { "epoch": 0.036840085521627104, "grad_norm": 1.3063303232192993, "learning_rate": 9.998864392252614e-07, "loss": 0.4839, "step": 84 }, { "epoch": 0.03727865796831314, "grad_norm": 1.1942335367202759, "learning_rate": 9.998707937479682e-07, "loss": 0.4443, "step": 85 }, { "epoch": 0.03771723041499918, "grad_norm": 1.6526542901992798, "learning_rate": 9.99854139064382e-07, "loss": 0.4641, "step": 86 }, { "epoch": 0.03815580286168521, "grad_norm": 1.4341564178466797, "learning_rate": 9.998364752081277e-07, "loss": 0.4928, "step": 87 }, { "epoch": 0.03859437530837125, "grad_norm": 1.1596306562423706, "learning_rate": 9.998178022148676e-07, "loss": 0.4209, "step": 88 }, { "epoch": 0.039032947755057286, "grad_norm": 1.2552282810211182, "learning_rate": 9.997981201223009e-07, "loss": 0.5077, "step": 89 }, { "epoch": 0.039471520201743326, "grad_norm": 1.4964027404785156, "learning_rate": 9.997774289701647e-07, "loss": 0.4983, "step": 90 }, { "epoch": 0.03991009264842936, "grad_norm": 1.221358060836792, "learning_rate": 9.99755728800233e-07, "loss": 0.4285, "step": 91 }, { "epoch": 0.0403486650951154, "grad_norm": 1.2171251773834229, "learning_rate": 9.997330196563169e-07, "loss": 0.4797, "step": 92 }, { "epoch": 0.040787237541801434, "grad_norm": 1.2266311645507812, "learning_rate": 9.99709301584265e-07, "loss": 0.4371, "step": 93 }, { "epoch": 0.041225809988487475, "grad_norm": 1.4037154912948608, "learning_rate": 9.99684574631962e-07, "loss": 0.5075, "step": 94 }, { "epoch": 0.04166438243517351, "grad_norm": 1.2918537855148315, "learning_rate": 9.996588388493306e-07, "loss": 0.4561, "step": 95 }, { "epoch": 0.04210295488185955, "grad_norm": 1.422583818435669, "learning_rate": 9.996320942883295e-07, "loss": 0.4975, "step": 96 }, { "epoch": 0.04254152732854558, "grad_norm": 1.3405108451843262, "learning_rate": 9.996043410029537e-07, "loss": 0.4725, "step": 97 }, { "epoch": 0.04298009977523162, "grad_norm": 1.2558434009552002, "learning_rate": 9.995755790492359e-07, "loss": 0.4548, "step": 98 }, { "epoch": 0.043418672221917656, "grad_norm": 1.3062516450881958, "learning_rate": 9.99545808485244e-07, "loss": 0.4432, "step": 99 }, { "epoch": 0.0438572446686037, "grad_norm": 1.277416467666626, "learning_rate": 9.995150293710838e-07, "loss": 0.4464, "step": 100 }, { "epoch": 0.04429581711528973, "grad_norm": 1.2594455480575562, "learning_rate": 9.99483241768895e-07, "loss": 0.4271, "step": 101 }, { "epoch": 0.04473438956197577, "grad_norm": 1.4411756992340088, "learning_rate": 9.994504457428556e-07, "loss": 0.4759, "step": 102 }, { "epoch": 0.045172962008661804, "grad_norm": 1.3192076683044434, "learning_rate": 9.994166413591784e-07, "loss": 0.4448, "step": 103 }, { "epoch": 0.045611534455347845, "grad_norm": 1.247989296913147, "learning_rate": 9.993818286861122e-07, "loss": 0.491, "step": 104 }, { "epoch": 0.04605010690203388, "grad_norm": 1.2985870838165283, "learning_rate": 9.993460077939414e-07, "loss": 0.4688, "step": 105 }, { "epoch": 0.04648867934871992, "grad_norm": 1.3416420221328735, "learning_rate": 9.993091787549862e-07, "loss": 0.4958, "step": 106 }, { "epoch": 0.04692725179540595, "grad_norm": 1.4006417989730835, "learning_rate": 9.99271341643602e-07, "loss": 0.4829, "step": 107 }, { "epoch": 0.04736582424209199, "grad_norm": 1.422197699546814, "learning_rate": 9.99232496536179e-07, "loss": 0.4628, "step": 108 }, { "epoch": 0.04780439668877803, "grad_norm": 1.2422194480895996, "learning_rate": 9.991926435111437e-07, "loss": 0.4086, "step": 109 }, { "epoch": 0.04824296913546407, "grad_norm": 1.185133695602417, "learning_rate": 9.99151782648956e-07, "loss": 0.4374, "step": 110 }, { "epoch": 0.0486815415821501, "grad_norm": 1.2822550535202026, "learning_rate": 9.99109914032112e-07, "loss": 0.457, "step": 111 }, { "epoch": 0.04912011402883614, "grad_norm": 1.3902647495269775, "learning_rate": 9.99067037745141e-07, "loss": 0.4281, "step": 112 }, { "epoch": 0.049558686475522175, "grad_norm": 1.244568943977356, "learning_rate": 9.990231538746078e-07, "loss": 0.443, "step": 113 }, { "epoch": 0.049997258922208215, "grad_norm": 1.1944634914398193, "learning_rate": 9.989782625091113e-07, "loss": 0.3956, "step": 114 }, { "epoch": 0.05043583136889425, "grad_norm": 1.330433964729309, "learning_rate": 9.989323637392834e-07, "loss": 0.5167, "step": 115 }, { "epoch": 0.05087440381558029, "grad_norm": 1.2958663702011108, "learning_rate": 9.988854576577913e-07, "loss": 0.5158, "step": 116 }, { "epoch": 0.05131297626226632, "grad_norm": 1.2222250699996948, "learning_rate": 9.988375443593354e-07, "loss": 0.4502, "step": 117 }, { "epoch": 0.051751548708952363, "grad_norm": 1.3030247688293457, "learning_rate": 9.987886239406491e-07, "loss": 0.387, "step": 118 }, { "epoch": 0.0521901211556384, "grad_norm": 1.281866431236267, "learning_rate": 9.987386965004997e-07, "loss": 0.4687, "step": 119 }, { "epoch": 0.05262869360232443, "grad_norm": 1.273772120475769, "learning_rate": 9.986877621396877e-07, "loss": 0.4064, "step": 120 }, { "epoch": 0.05306726604901047, "grad_norm": 1.3751946687698364, "learning_rate": 9.986358209610457e-07, "loss": 0.4919, "step": 121 }, { "epoch": 0.053505838495696505, "grad_norm": 1.3724075555801392, "learning_rate": 9.985828730694396e-07, "loss": 0.4633, "step": 122 }, { "epoch": 0.053944410942382545, "grad_norm": 1.2193125486373901, "learning_rate": 9.985289185717683e-07, "loss": 0.4356, "step": 123 }, { "epoch": 0.05438298338906858, "grad_norm": 1.424628734588623, "learning_rate": 9.984739575769617e-07, "loss": 0.4395, "step": 124 }, { "epoch": 0.05482155583575462, "grad_norm": 1.2202560901641846, "learning_rate": 9.984179901959828e-07, "loss": 0.4765, "step": 125 }, { "epoch": 0.05526012828244065, "grad_norm": 1.3114961385726929, "learning_rate": 9.983610165418259e-07, "loss": 0.4197, "step": 126 }, { "epoch": 0.05569870072912669, "grad_norm": 1.2029987573623657, "learning_rate": 9.983030367295173e-07, "loss": 0.429, "step": 127 }, { "epoch": 0.05613727317581273, "grad_norm": 1.4868053197860718, "learning_rate": 9.982440508761143e-07, "loss": 0.4736, "step": 128 }, { "epoch": 0.05657584562249877, "grad_norm": 1.212609887123108, "learning_rate": 9.981840591007051e-07, "loss": 0.4235, "step": 129 }, { "epoch": 0.0570144180691848, "grad_norm": 1.3265501260757446, "learning_rate": 9.981230615244099e-07, "loss": 0.459, "step": 130 }, { "epoch": 0.05745299051587084, "grad_norm": 1.2474690675735474, "learning_rate": 9.980610582703782e-07, "loss": 0.4498, "step": 131 }, { "epoch": 0.057891562962556875, "grad_norm": 1.1922861337661743, "learning_rate": 9.979980494637908e-07, "loss": 0.4227, "step": 132 }, { "epoch": 0.058330135409242916, "grad_norm": 1.351237416267395, "learning_rate": 9.979340352318582e-07, "loss": 0.441, "step": 133 }, { "epoch": 0.05876870785592895, "grad_norm": 1.1860822439193726, "learning_rate": 9.978690157038208e-07, "loss": 0.412, "step": 134 }, { "epoch": 0.05920728030261499, "grad_norm": 1.3116803169250488, "learning_rate": 9.97802991010949e-07, "loss": 0.5344, "step": 135 }, { "epoch": 0.05964585274930102, "grad_norm": 1.3794771432876587, "learning_rate": 9.977359612865422e-07, "loss": 0.4982, "step": 136 }, { "epoch": 0.060084425195987064, "grad_norm": 1.197117567062378, "learning_rate": 9.976679266659292e-07, "loss": 0.4201, "step": 137 }, { "epoch": 0.0605229976426731, "grad_norm": 1.2389888763427734, "learning_rate": 9.97598887286467e-07, "loss": 0.4692, "step": 138 }, { "epoch": 0.06096157008935914, "grad_norm": 1.3116774559020996, "learning_rate": 9.975288432875422e-07, "loss": 0.4854, "step": 139 }, { "epoch": 0.06140014253604517, "grad_norm": 1.2350678443908691, "learning_rate": 9.974577948105684e-07, "loss": 0.4724, "step": 140 }, { "epoch": 0.06183871498273121, "grad_norm": 1.3790651559829712, "learning_rate": 9.973857419989881e-07, "loss": 0.5353, "step": 141 }, { "epoch": 0.062277287429417245, "grad_norm": 1.3146638870239258, "learning_rate": 9.973126849982713e-07, "loss": 0.4545, "step": 142 }, { "epoch": 0.06271585987610329, "grad_norm": 1.4369558095932007, "learning_rate": 9.972386239559152e-07, "loss": 0.4897, "step": 143 }, { "epoch": 0.06315443232278932, "grad_norm": 1.3575363159179688, "learning_rate": 9.97163559021444e-07, "loss": 0.4034, "step": 144 }, { "epoch": 0.06359300476947535, "grad_norm": 1.2711018323898315, "learning_rate": 9.97087490346409e-07, "loss": 0.4584, "step": 145 }, { "epoch": 0.0640315772161614, "grad_norm": 1.2842679023742676, "learning_rate": 9.970104180843878e-07, "loss": 0.4689, "step": 146 }, { "epoch": 0.06447014966284743, "grad_norm": 1.2036633491516113, "learning_rate": 9.969323423909846e-07, "loss": 0.417, "step": 147 }, { "epoch": 0.06490872210953347, "grad_norm": 1.199987769126892, "learning_rate": 9.968532634238287e-07, "loss": 0.4346, "step": 148 }, { "epoch": 0.0653472945562195, "grad_norm": 1.280448317527771, "learning_rate": 9.967731813425752e-07, "loss": 0.4319, "step": 149 }, { "epoch": 0.06578586700290555, "grad_norm": 1.2478801012039185, "learning_rate": 9.966920963089051e-07, "loss": 0.4424, "step": 150 }, { "epoch": 0.06622443944959158, "grad_norm": 1.2198948860168457, "learning_rate": 9.966100084865232e-07, "loss": 0.4236, "step": 151 }, { "epoch": 0.06666301189627762, "grad_norm": 1.28062903881073, "learning_rate": 9.965269180411598e-07, "loss": 0.4262, "step": 152 }, { "epoch": 0.06710158434296365, "grad_norm": 1.137345790863037, "learning_rate": 9.96442825140569e-07, "loss": 0.4692, "step": 153 }, { "epoch": 0.0675401567896497, "grad_norm": 1.2888954877853394, "learning_rate": 9.963577299545286e-07, "loss": 0.5076, "step": 154 }, { "epoch": 0.06797872923633573, "grad_norm": 1.2218375205993652, "learning_rate": 9.962716326548404e-07, "loss": 0.4651, "step": 155 }, { "epoch": 0.06841730168302176, "grad_norm": 1.3181045055389404, "learning_rate": 9.961845334153293e-07, "loss": 0.4151, "step": 156 }, { "epoch": 0.0688558741297078, "grad_norm": 1.2577183246612549, "learning_rate": 9.960964324118425e-07, "loss": 0.4713, "step": 157 }, { "epoch": 0.06929444657639384, "grad_norm": 1.2748234272003174, "learning_rate": 9.960073298222508e-07, "loss": 0.4551, "step": 158 }, { "epoch": 0.06973301902307988, "grad_norm": 1.2437751293182373, "learning_rate": 9.959172258264458e-07, "loss": 0.4388, "step": 159 }, { "epoch": 0.07017159146976591, "grad_norm": 1.3476532697677612, "learning_rate": 9.95826120606342e-07, "loss": 0.5177, "step": 160 }, { "epoch": 0.07061016391645195, "grad_norm": 1.3056862354278564, "learning_rate": 9.957340143458747e-07, "loss": 0.4781, "step": 161 }, { "epoch": 0.07104873636313798, "grad_norm": 1.4270120859146118, "learning_rate": 9.956409072310004e-07, "loss": 0.4588, "step": 162 }, { "epoch": 0.07148730880982403, "grad_norm": 1.3080512285232544, "learning_rate": 9.95546799449696e-07, "loss": 0.4265, "step": 163 }, { "epoch": 0.07192588125651006, "grad_norm": 1.4175631999969482, "learning_rate": 9.954516911919595e-07, "loss": 0.4487, "step": 164 }, { "epoch": 0.0723644537031961, "grad_norm": 1.2619612216949463, "learning_rate": 9.953555826498077e-07, "loss": 0.4375, "step": 165 }, { "epoch": 0.07280302614988213, "grad_norm": 1.301414966583252, "learning_rate": 9.952584740172777e-07, "loss": 0.4425, "step": 166 }, { "epoch": 0.07324159859656817, "grad_norm": 1.2964798212051392, "learning_rate": 9.951603654904254e-07, "loss": 0.4687, "step": 167 }, { "epoch": 0.07368017104325421, "grad_norm": 1.2225522994995117, "learning_rate": 9.950612572673255e-07, "loss": 0.4119, "step": 168 }, { "epoch": 0.07411874348994024, "grad_norm": 1.2875328063964844, "learning_rate": 9.949611495480708e-07, "loss": 0.4658, "step": 169 }, { "epoch": 0.07455731593662628, "grad_norm": 1.168454885482788, "learning_rate": 9.948600425347724e-07, "loss": 0.409, "step": 170 }, { "epoch": 0.07499588838331232, "grad_norm": 1.3310086727142334, "learning_rate": 9.947579364315587e-07, "loss": 0.4657, "step": 171 }, { "epoch": 0.07543446082999836, "grad_norm": 1.254629135131836, "learning_rate": 9.946548314445751e-07, "loss": 0.4406, "step": 172 }, { "epoch": 0.07587303327668439, "grad_norm": 1.2079718112945557, "learning_rate": 9.94550727781984e-07, "loss": 0.4325, "step": 173 }, { "epoch": 0.07631160572337042, "grad_norm": 1.2881447076797485, "learning_rate": 9.944456256539636e-07, "loss": 0.4459, "step": 174 }, { "epoch": 0.07675017817005647, "grad_norm": 1.4596824645996094, "learning_rate": 9.943395252727085e-07, "loss": 0.4727, "step": 175 }, { "epoch": 0.0771887506167425, "grad_norm": 1.2925268411636353, "learning_rate": 9.94232426852428e-07, "loss": 0.482, "step": 176 }, { "epoch": 0.07762732306342854, "grad_norm": 1.2629894018173218, "learning_rate": 9.94124330609347e-07, "loss": 0.45, "step": 177 }, { "epoch": 0.07806589551011457, "grad_norm": 1.238906741142273, "learning_rate": 9.940152367617049e-07, "loss": 0.4358, "step": 178 }, { "epoch": 0.07850446795680062, "grad_norm": 1.33396315574646, "learning_rate": 9.939051455297548e-07, "loss": 0.4683, "step": 179 }, { "epoch": 0.07894304040348665, "grad_norm": 1.4251528978347778, "learning_rate": 9.937940571357636e-07, "loss": 0.4883, "step": 180 }, { "epoch": 0.07938161285017269, "grad_norm": 1.2432212829589844, "learning_rate": 9.936819718040116e-07, "loss": 0.4295, "step": 181 }, { "epoch": 0.07982018529685872, "grad_norm": 1.2691503763198853, "learning_rate": 9.935688897607915e-07, "loss": 0.49, "step": 182 }, { "epoch": 0.08025875774354477, "grad_norm": 1.3803023099899292, "learning_rate": 9.934548112344087e-07, "loss": 0.4467, "step": 183 }, { "epoch": 0.0806973301902308, "grad_norm": 1.3414316177368164, "learning_rate": 9.933397364551805e-07, "loss": 0.4608, "step": 184 }, { "epoch": 0.08113590263691683, "grad_norm": 1.3661813735961914, "learning_rate": 9.93223665655435e-07, "loss": 0.4901, "step": 185 }, { "epoch": 0.08157447508360287, "grad_norm": 1.2298403978347778, "learning_rate": 9.931065990695113e-07, "loss": 0.4408, "step": 186 }, { "epoch": 0.08201304753028892, "grad_norm": 1.1648515462875366, "learning_rate": 9.929885369337596e-07, "loss": 0.4097, "step": 187 }, { "epoch": 0.08245161997697495, "grad_norm": 1.1954864263534546, "learning_rate": 9.928694794865395e-07, "loss": 0.4162, "step": 188 }, { "epoch": 0.08289019242366098, "grad_norm": 1.2637189626693726, "learning_rate": 9.9274942696822e-07, "loss": 0.4361, "step": 189 }, { "epoch": 0.08332876487034702, "grad_norm": 1.2444093227386475, "learning_rate": 9.926283796211794e-07, "loss": 0.4692, "step": 190 }, { "epoch": 0.08376733731703306, "grad_norm": 1.282979965209961, "learning_rate": 9.925063376898044e-07, "loss": 0.3999, "step": 191 }, { "epoch": 0.0842059097637191, "grad_norm": 1.2734895944595337, "learning_rate": 9.923833014204893e-07, "loss": 0.4343, "step": 192 }, { "epoch": 0.08464448221040513, "grad_norm": 1.208871603012085, "learning_rate": 9.922592710616364e-07, "loss": 0.4414, "step": 193 }, { "epoch": 0.08508305465709116, "grad_norm": 1.2236772775650024, "learning_rate": 9.92134246863655e-07, "loss": 0.4345, "step": 194 }, { "epoch": 0.08552162710377721, "grad_norm": 1.2935453653335571, "learning_rate": 9.920082290789607e-07, "loss": 0.4637, "step": 195 }, { "epoch": 0.08596019955046325, "grad_norm": 1.307611107826233, "learning_rate": 9.91881217961975e-07, "loss": 0.4666, "step": 196 }, { "epoch": 0.08639877199714928, "grad_norm": 1.2390419244766235, "learning_rate": 9.917532137691252e-07, "loss": 0.4072, "step": 197 }, { "epoch": 0.08683734444383531, "grad_norm": 1.3468120098114014, "learning_rate": 9.916242167588432e-07, "loss": 0.4486, "step": 198 }, { "epoch": 0.08727591689052135, "grad_norm": 1.270461916923523, "learning_rate": 9.914942271915655e-07, "loss": 0.4438, "step": 199 }, { "epoch": 0.0877144893372074, "grad_norm": 1.2275724411010742, "learning_rate": 9.913632453297325e-07, "loss": 0.4503, "step": 200 }, { "epoch": 0.08815306178389343, "grad_norm": 1.2674055099487305, "learning_rate": 9.912312714377879e-07, "loss": 0.425, "step": 201 }, { "epoch": 0.08859163423057946, "grad_norm": 1.2971616983413696, "learning_rate": 9.910983057821786e-07, "loss": 0.4633, "step": 202 }, { "epoch": 0.0890302066772655, "grad_norm": 1.3597564697265625, "learning_rate": 9.909643486313533e-07, "loss": 0.4709, "step": 203 }, { "epoch": 0.08946877912395154, "grad_norm": 1.403801441192627, "learning_rate": 9.908294002557627e-07, "loss": 0.4508, "step": 204 }, { "epoch": 0.08990735157063758, "grad_norm": 1.1720523834228516, "learning_rate": 9.906934609278588e-07, "loss": 0.4182, "step": 205 }, { "epoch": 0.09034592401732361, "grad_norm": 1.4372029304504395, "learning_rate": 9.90556530922094e-07, "loss": 0.4495, "step": 206 }, { "epoch": 0.09078449646400964, "grad_norm": 1.2117644548416138, "learning_rate": 9.904186105149211e-07, "loss": 0.4486, "step": 207 }, { "epoch": 0.09122306891069569, "grad_norm": 1.253118872642517, "learning_rate": 9.902796999847923e-07, "loss": 0.4683, "step": 208 }, { "epoch": 0.09166164135738172, "grad_norm": 1.1815754175186157, "learning_rate": 9.901397996121587e-07, "loss": 0.4208, "step": 209 }, { "epoch": 0.09210021380406776, "grad_norm": 1.1744418144226074, "learning_rate": 9.899989096794704e-07, "loss": 0.4086, "step": 210 }, { "epoch": 0.09253878625075379, "grad_norm": 1.194718599319458, "learning_rate": 9.898570304711746e-07, "loss": 0.4343, "step": 211 }, { "epoch": 0.09297735869743984, "grad_norm": 1.3020581007003784, "learning_rate": 9.897141622737159e-07, "loss": 0.4369, "step": 212 }, { "epoch": 0.09341593114412587, "grad_norm": 1.3153276443481445, "learning_rate": 9.895703053755363e-07, "loss": 0.4609, "step": 213 }, { "epoch": 0.0938545035908119, "grad_norm": 1.3371548652648926, "learning_rate": 9.89425460067073e-07, "loss": 0.459, "step": 214 }, { "epoch": 0.09429307603749794, "grad_norm": 1.2718877792358398, "learning_rate": 9.892796266407595e-07, "loss": 0.4669, "step": 215 }, { "epoch": 0.09473164848418399, "grad_norm": 1.2996882200241089, "learning_rate": 9.891328053910237e-07, "loss": 0.4598, "step": 216 }, { "epoch": 0.09517022093087002, "grad_norm": 1.3045529127120972, "learning_rate": 9.88984996614288e-07, "loss": 0.4628, "step": 217 }, { "epoch": 0.09560879337755605, "grad_norm": 1.2552134990692139, "learning_rate": 9.888362006089688e-07, "loss": 0.423, "step": 218 }, { "epoch": 0.09604736582424209, "grad_norm": 1.3412526845932007, "learning_rate": 9.886864176754754e-07, "loss": 0.4708, "step": 219 }, { "epoch": 0.09648593827092813, "grad_norm": 1.2526757717132568, "learning_rate": 9.885356481162096e-07, "loss": 0.4416, "step": 220 }, { "epoch": 0.09692451071761417, "grad_norm": 1.2230411767959595, "learning_rate": 9.883838922355653e-07, "loss": 0.4032, "step": 221 }, { "epoch": 0.0973630831643002, "grad_norm": 1.3482338190078735, "learning_rate": 9.882311503399277e-07, "loss": 0.445, "step": 222 }, { "epoch": 0.09780165561098624, "grad_norm": 1.3570636510849, "learning_rate": 9.880774227376725e-07, "loss": 0.519, "step": 223 }, { "epoch": 0.09824022805767228, "grad_norm": 1.2754287719726562, "learning_rate": 9.879227097391658e-07, "loss": 0.4919, "step": 224 }, { "epoch": 0.09867880050435832, "grad_norm": 1.3529372215270996, "learning_rate": 9.87767011656763e-07, "loss": 0.4193, "step": 225 }, { "epoch": 0.09911737295104435, "grad_norm": 1.3907129764556885, "learning_rate": 9.876103288048084e-07, "loss": 0.4202, "step": 226 }, { "epoch": 0.09955594539773038, "grad_norm": 1.2041593790054321, "learning_rate": 9.87452661499634e-07, "loss": 0.4326, "step": 227 }, { "epoch": 0.09999451784441643, "grad_norm": 1.3893024921417236, "learning_rate": 9.872940100595597e-07, "loss": 0.4933, "step": 228 }, { "epoch": 0.10043309029110246, "grad_norm": 1.1358956098556519, "learning_rate": 9.871343748048929e-07, "loss": 0.3958, "step": 229 }, { "epoch": 0.1008716627377885, "grad_norm": 1.258034586906433, "learning_rate": 9.869737560579262e-07, "loss": 0.4579, "step": 230 }, { "epoch": 0.10131023518447453, "grad_norm": 1.3062182664871216, "learning_rate": 9.868121541429386e-07, "loss": 0.4671, "step": 231 }, { "epoch": 0.10174880763116058, "grad_norm": 1.2269024848937988, "learning_rate": 9.866495693861934e-07, "loss": 0.4541, "step": 232 }, { "epoch": 0.10218738007784661, "grad_norm": 1.3577250242233276, "learning_rate": 9.86486002115939e-07, "loss": 0.5296, "step": 233 }, { "epoch": 0.10262595252453265, "grad_norm": 1.330257534980774, "learning_rate": 9.863214526624063e-07, "loss": 0.4678, "step": 234 }, { "epoch": 0.10306452497121868, "grad_norm": 1.1935516595840454, "learning_rate": 9.861559213578107e-07, "loss": 0.4826, "step": 235 }, { "epoch": 0.10350309741790473, "grad_norm": 1.4533356428146362, "learning_rate": 9.859894085363485e-07, "loss": 0.5029, "step": 236 }, { "epoch": 0.10394166986459076, "grad_norm": 1.274011254310608, "learning_rate": 9.85821914534198e-07, "loss": 0.4471, "step": 237 }, { "epoch": 0.1043802423112768, "grad_norm": 1.3055455684661865, "learning_rate": 9.856534396895193e-07, "loss": 0.4521, "step": 238 }, { "epoch": 0.10481881475796283, "grad_norm": 1.290225625038147, "learning_rate": 9.854839843424512e-07, "loss": 0.4444, "step": 239 }, { "epoch": 0.10525738720464886, "grad_norm": 1.256276249885559, "learning_rate": 9.853135488351132e-07, "loss": 0.4397, "step": 240 }, { "epoch": 0.10569595965133491, "grad_norm": 1.2009029388427734, "learning_rate": 9.851421335116036e-07, "loss": 0.464, "step": 241 }, { "epoch": 0.10613453209802094, "grad_norm": 1.3538483381271362, "learning_rate": 9.849697387179987e-07, "loss": 0.4543, "step": 242 }, { "epoch": 0.10657310454470698, "grad_norm": 1.314831018447876, "learning_rate": 9.84796364802352e-07, "loss": 0.4675, "step": 243 }, { "epoch": 0.10701167699139301, "grad_norm": 1.334395408630371, "learning_rate": 9.846220121146943e-07, "loss": 0.4541, "step": 244 }, { "epoch": 0.10745024943807906, "grad_norm": 1.2025914192199707, "learning_rate": 9.844466810070317e-07, "loss": 0.3998, "step": 245 }, { "epoch": 0.10788882188476509, "grad_norm": 1.438745141029358, "learning_rate": 9.842703718333468e-07, "loss": 0.4303, "step": 246 }, { "epoch": 0.10832739433145112, "grad_norm": 1.2495355606079102, "learning_rate": 9.84093084949596e-07, "loss": 0.4415, "step": 247 }, { "epoch": 0.10876596677813716, "grad_norm": 1.3335902690887451, "learning_rate": 9.8391482071371e-07, "loss": 0.4923, "step": 248 }, { "epoch": 0.1092045392248232, "grad_norm": 1.359126091003418, "learning_rate": 9.837355794855923e-07, "loss": 0.4598, "step": 249 }, { "epoch": 0.10964311167150924, "grad_norm": 1.1839964389801025, "learning_rate": 9.835553616271194e-07, "loss": 0.4053, "step": 250 }, { "epoch": 0.11008168411819527, "grad_norm": 1.2885066270828247, "learning_rate": 9.83374167502139e-07, "loss": 0.4521, "step": 251 }, { "epoch": 0.1105202565648813, "grad_norm": 1.381494402885437, "learning_rate": 9.83191997476471e-07, "loss": 0.5172, "step": 252 }, { "epoch": 0.11095882901156735, "grad_norm": 1.3523175716400146, "learning_rate": 9.830088519179035e-07, "loss": 0.4813, "step": 253 }, { "epoch": 0.11139740145825339, "grad_norm": 1.1988216638565063, "learning_rate": 9.82824731196196e-07, "loss": 0.4377, "step": 254 }, { "epoch": 0.11183597390493942, "grad_norm": 1.1344822645187378, "learning_rate": 9.826396356830764e-07, "loss": 0.4638, "step": 255 }, { "epoch": 0.11227454635162545, "grad_norm": 1.2201296091079712, "learning_rate": 9.824535657522397e-07, "loss": 0.4378, "step": 256 }, { "epoch": 0.1127131187983115, "grad_norm": 1.4678555727005005, "learning_rate": 9.822665217793496e-07, "loss": 0.4644, "step": 257 }, { "epoch": 0.11315169124499753, "grad_norm": 1.251297950744629, "learning_rate": 9.820785041420348e-07, "loss": 0.4329, "step": 258 }, { "epoch": 0.11359026369168357, "grad_norm": 1.2199487686157227, "learning_rate": 9.818895132198913e-07, "loss": 0.4474, "step": 259 }, { "epoch": 0.1140288361383696, "grad_norm": 1.311081051826477, "learning_rate": 9.81699549394479e-07, "loss": 0.4531, "step": 260 }, { "epoch": 0.11446740858505565, "grad_norm": 1.3128409385681152, "learning_rate": 9.815086130493221e-07, "loss": 0.4166, "step": 261 }, { "epoch": 0.11490598103174168, "grad_norm": 1.300299882888794, "learning_rate": 9.81316704569909e-07, "loss": 0.4667, "step": 262 }, { "epoch": 0.11534455347842772, "grad_norm": 1.1987988948822021, "learning_rate": 9.811238243436904e-07, "loss": 0.4499, "step": 263 }, { "epoch": 0.11578312592511375, "grad_norm": 1.2745237350463867, "learning_rate": 9.809299727600783e-07, "loss": 0.4198, "step": 264 }, { "epoch": 0.1162216983717998, "grad_norm": 1.3099387884140015, "learning_rate": 9.807351502104468e-07, "loss": 0.4695, "step": 265 }, { "epoch": 0.11666027081848583, "grad_norm": 1.3721712827682495, "learning_rate": 9.805393570881295e-07, "loss": 0.4978, "step": 266 }, { "epoch": 0.11709884326517186, "grad_norm": 1.3487995862960815, "learning_rate": 9.8034259378842e-07, "loss": 0.4425, "step": 267 }, { "epoch": 0.1175374157118579, "grad_norm": 1.3157742023468018, "learning_rate": 9.801448607085704e-07, "loss": 0.4446, "step": 268 }, { "epoch": 0.11797598815854395, "grad_norm": 1.368003487586975, "learning_rate": 9.799461582477909e-07, "loss": 0.462, "step": 269 }, { "epoch": 0.11841456060522998, "grad_norm": 1.254192590713501, "learning_rate": 9.797464868072486e-07, "loss": 0.4711, "step": 270 }, { "epoch": 0.11885313305191601, "grad_norm": 1.3297892808914185, "learning_rate": 9.795458467900672e-07, "loss": 0.4489, "step": 271 }, { "epoch": 0.11929170549860205, "grad_norm": 1.3227829933166504, "learning_rate": 9.793442386013255e-07, "loss": 0.4494, "step": 272 }, { "epoch": 0.1197302779452881, "grad_norm": 1.3050018548965454, "learning_rate": 9.79141662648057e-07, "loss": 0.4533, "step": 273 }, { "epoch": 0.12016885039197413, "grad_norm": 1.2946937084197998, "learning_rate": 9.789381193392498e-07, "loss": 0.4566, "step": 274 }, { "epoch": 0.12060742283866016, "grad_norm": 1.1722913980484009, "learning_rate": 9.787336090858441e-07, "loss": 0.4402, "step": 275 }, { "epoch": 0.1210459952853462, "grad_norm": 1.3350423574447632, "learning_rate": 9.78528132300733e-07, "loss": 0.4497, "step": 276 }, { "epoch": 0.12148456773203223, "grad_norm": 1.2697076797485352, "learning_rate": 9.7832168939876e-07, "loss": 0.4525, "step": 277 }, { "epoch": 0.12192314017871828, "grad_norm": 1.3002448081970215, "learning_rate": 9.781142807967204e-07, "loss": 0.4955, "step": 278 }, { "epoch": 0.12236171262540431, "grad_norm": 1.319963812828064, "learning_rate": 9.779059069133582e-07, "loss": 0.4588, "step": 279 }, { "epoch": 0.12280028507209034, "grad_norm": 1.4620566368103027, "learning_rate": 9.776965681693666e-07, "loss": 0.4813, "step": 280 }, { "epoch": 0.12323885751877638, "grad_norm": 1.4124107360839844, "learning_rate": 9.774862649873868e-07, "loss": 0.4535, "step": 281 }, { "epoch": 0.12367742996546242, "grad_norm": 1.3538119792938232, "learning_rate": 9.772749977920071e-07, "loss": 0.4501, "step": 282 }, { "epoch": 0.12411600241214846, "grad_norm": 1.2902806997299194, "learning_rate": 9.770627670097623e-07, "loss": 0.4242, "step": 283 }, { "epoch": 0.12455457485883449, "grad_norm": 1.2163552045822144, "learning_rate": 9.768495730691321e-07, "loss": 0.4376, "step": 284 }, { "epoch": 0.12499314730552052, "grad_norm": 1.2694215774536133, "learning_rate": 9.766354164005414e-07, "loss": 0.4893, "step": 285 }, { "epoch": 0.12543171975220657, "grad_norm": 1.3819620609283447, "learning_rate": 9.76420297436358e-07, "loss": 0.4442, "step": 286 }, { "epoch": 0.1258702921988926, "grad_norm": 1.2240216732025146, "learning_rate": 9.762042166108932e-07, "loss": 0.4411, "step": 287 }, { "epoch": 0.12630886464557864, "grad_norm": 1.3175026178359985, "learning_rate": 9.759871743604001e-07, "loss": 0.3919, "step": 288 }, { "epoch": 0.1267474370922647, "grad_norm": 1.219504714012146, "learning_rate": 9.757691711230727e-07, "loss": 0.4241, "step": 289 }, { "epoch": 0.1271860095389507, "grad_norm": 1.389089584350586, "learning_rate": 9.75550207339045e-07, "loss": 0.4885, "step": 290 }, { "epoch": 0.12762458198563675, "grad_norm": 1.3088550567626953, "learning_rate": 9.753302834503908e-07, "loss": 0.4106, "step": 291 }, { "epoch": 0.1280631544323228, "grad_norm": 1.2590484619140625, "learning_rate": 9.751093999011216e-07, "loss": 0.472, "step": 292 }, { "epoch": 0.12850172687900882, "grad_norm": 1.2284749746322632, "learning_rate": 9.74887557137187e-07, "loss": 0.4192, "step": 293 }, { "epoch": 0.12894029932569487, "grad_norm": 1.2713497877120972, "learning_rate": 9.746647556064732e-07, "loss": 0.4552, "step": 294 }, { "epoch": 0.1293788717723809, "grad_norm": 1.276537537574768, "learning_rate": 9.744409957588014e-07, "loss": 0.4545, "step": 295 }, { "epoch": 0.12981744421906694, "grad_norm": 1.3731132745742798, "learning_rate": 9.742162780459281e-07, "loss": 0.4143, "step": 296 }, { "epoch": 0.13025601666575298, "grad_norm": 1.246085524559021, "learning_rate": 9.73990602921544e-07, "loss": 0.4145, "step": 297 }, { "epoch": 0.130694589112439, "grad_norm": 1.2259469032287598, "learning_rate": 9.737639708412721e-07, "loss": 0.4474, "step": 298 }, { "epoch": 0.13113316155912505, "grad_norm": 1.3381538391113281, "learning_rate": 9.735363822626676e-07, "loss": 0.4735, "step": 299 }, { "epoch": 0.1315717340058111, "grad_norm": 1.2572318315505981, "learning_rate": 9.73307837645217e-07, "loss": 0.406, "step": 300 }, { "epoch": 0.13201030645249712, "grad_norm": 1.3827505111694336, "learning_rate": 9.730783374503369e-07, "loss": 0.405, "step": 301 }, { "epoch": 0.13244887889918316, "grad_norm": 1.2806872129440308, "learning_rate": 9.728478821413728e-07, "loss": 0.439, "step": 302 }, { "epoch": 0.13288745134586918, "grad_norm": 1.2340537309646606, "learning_rate": 9.726164721835995e-07, "loss": 0.4199, "step": 303 }, { "epoch": 0.13332602379255523, "grad_norm": 1.3651092052459717, "learning_rate": 9.723841080442176e-07, "loss": 0.4593, "step": 304 }, { "epoch": 0.13376459623924128, "grad_norm": 1.4623818397521973, "learning_rate": 9.721507901923559e-07, "loss": 0.4769, "step": 305 }, { "epoch": 0.1342031686859273, "grad_norm": 1.2197152376174927, "learning_rate": 9.719165190990673e-07, "loss": 0.4333, "step": 306 }, { "epoch": 0.13464174113261335, "grad_norm": 1.3367750644683838, "learning_rate": 9.716812952373297e-07, "loss": 0.4313, "step": 307 }, { "epoch": 0.1350803135792994, "grad_norm": 1.3246334791183472, "learning_rate": 9.714451190820449e-07, "loss": 0.4228, "step": 308 }, { "epoch": 0.1355188860259854, "grad_norm": 1.2900984287261963, "learning_rate": 9.71207991110037e-07, "loss": 0.414, "step": 309 }, { "epoch": 0.13595745847267146, "grad_norm": 1.2274436950683594, "learning_rate": 9.709699118000517e-07, "loss": 0.4576, "step": 310 }, { "epoch": 0.13639603091935748, "grad_norm": 1.288940191268921, "learning_rate": 9.707308816327556e-07, "loss": 0.4692, "step": 311 }, { "epoch": 0.13683460336604353, "grad_norm": 1.3872264623641968, "learning_rate": 9.704909010907348e-07, "loss": 0.4913, "step": 312 }, { "epoch": 0.13727317581272958, "grad_norm": 1.2140966653823853, "learning_rate": 9.702499706584943e-07, "loss": 0.4335, "step": 313 }, { "epoch": 0.1377117482594156, "grad_norm": 1.399438738822937, "learning_rate": 9.700080908224567e-07, "loss": 0.4552, "step": 314 }, { "epoch": 0.13815032070610164, "grad_norm": 1.334566354751587, "learning_rate": 9.697652620709615e-07, "loss": 0.4204, "step": 315 }, { "epoch": 0.1385888931527877, "grad_norm": 1.1500535011291504, "learning_rate": 9.695214848942641e-07, "loss": 0.419, "step": 316 }, { "epoch": 0.1390274655994737, "grad_norm": 1.257188081741333, "learning_rate": 9.692767597845349e-07, "loss": 0.4364, "step": 317 }, { "epoch": 0.13946603804615976, "grad_norm": 1.3126678466796875, "learning_rate": 9.690310872358571e-07, "loss": 0.4729, "step": 318 }, { "epoch": 0.13990461049284578, "grad_norm": 1.2673767805099487, "learning_rate": 9.687844677442282e-07, "loss": 0.4365, "step": 319 }, { "epoch": 0.14034318293953182, "grad_norm": 1.3040989637374878, "learning_rate": 9.685369018075562e-07, "loss": 0.4715, "step": 320 }, { "epoch": 0.14078175538621787, "grad_norm": 1.3866188526153564, "learning_rate": 9.682883899256607e-07, "loss": 0.438, "step": 321 }, { "epoch": 0.1412203278329039, "grad_norm": 1.2850979566574097, "learning_rate": 9.680389326002707e-07, "loss": 0.4325, "step": 322 }, { "epoch": 0.14165890027958994, "grad_norm": 1.244288682937622, "learning_rate": 9.677885303350244e-07, "loss": 0.447, "step": 323 }, { "epoch": 0.14209747272627596, "grad_norm": 1.327620029449463, "learning_rate": 9.675371836354673e-07, "loss": 0.4371, "step": 324 }, { "epoch": 0.142536045172962, "grad_norm": 1.311292290687561, "learning_rate": 9.672848930090522e-07, "loss": 0.4506, "step": 325 }, { "epoch": 0.14297461761964805, "grad_norm": 1.2936956882476807, "learning_rate": 9.670316589651367e-07, "loss": 0.4778, "step": 326 }, { "epoch": 0.14341319006633407, "grad_norm": 1.2305073738098145, "learning_rate": 9.667774820149843e-07, "loss": 0.4253, "step": 327 }, { "epoch": 0.14385176251302012, "grad_norm": 1.366324782371521, "learning_rate": 9.665223626717613e-07, "loss": 0.463, "step": 328 }, { "epoch": 0.14429033495970617, "grad_norm": 1.3423768281936646, "learning_rate": 9.66266301450537e-07, "loss": 0.4535, "step": 329 }, { "epoch": 0.1447289074063922, "grad_norm": 1.348443627357483, "learning_rate": 9.66009298868282e-07, "loss": 0.4935, "step": 330 }, { "epoch": 0.14516747985307824, "grad_norm": 1.1361678838729858, "learning_rate": 9.657513554438677e-07, "loss": 0.3826, "step": 331 }, { "epoch": 0.14560605229976425, "grad_norm": 1.1119284629821777, "learning_rate": 9.65492471698065e-07, "loss": 0.4397, "step": 332 }, { "epoch": 0.1460446247464503, "grad_norm": 1.3069874048233032, "learning_rate": 9.652326481535433e-07, "loss": 0.4402, "step": 333 }, { "epoch": 0.14648319719313635, "grad_norm": 1.3529688119888306, "learning_rate": 9.64971885334869e-07, "loss": 0.4139, "step": 334 }, { "epoch": 0.14692176963982237, "grad_norm": 1.2486492395401, "learning_rate": 9.647101837685052e-07, "loss": 0.3979, "step": 335 }, { "epoch": 0.14736034208650842, "grad_norm": 1.3289217948913574, "learning_rate": 9.644475439828102e-07, "loss": 0.451, "step": 336 }, { "epoch": 0.14779891453319446, "grad_norm": 1.3076781034469604, "learning_rate": 9.641839665080363e-07, "loss": 0.482, "step": 337 }, { "epoch": 0.14823748697988048, "grad_norm": 1.285366415977478, "learning_rate": 9.639194518763294e-07, "loss": 0.4418, "step": 338 }, { "epoch": 0.14867605942656653, "grad_norm": 1.2070653438568115, "learning_rate": 9.636540006217268e-07, "loss": 0.439, "step": 339 }, { "epoch": 0.14911463187325255, "grad_norm": 1.2946048974990845, "learning_rate": 9.633876132801577e-07, "loss": 0.4141, "step": 340 }, { "epoch": 0.1495532043199386, "grad_norm": 1.31601881980896, "learning_rate": 9.6312029038944e-07, "loss": 0.4488, "step": 341 }, { "epoch": 0.14999177676662465, "grad_norm": 1.2885645627975464, "learning_rate": 9.628520324892816e-07, "loss": 0.4398, "step": 342 }, { "epoch": 0.15043034921331067, "grad_norm": 1.262040138244629, "learning_rate": 9.625828401212772e-07, "loss": 0.4839, "step": 343 }, { "epoch": 0.1508689216599967, "grad_norm": 1.273505687713623, "learning_rate": 9.623127138289087e-07, "loss": 0.419, "step": 344 }, { "epoch": 0.15130749410668276, "grad_norm": 1.374241590499878, "learning_rate": 9.620416541575432e-07, "loss": 0.4567, "step": 345 }, { "epoch": 0.15174606655336878, "grad_norm": 1.118963360786438, "learning_rate": 9.617696616544325e-07, "loss": 0.3974, "step": 346 }, { "epoch": 0.15218463900005483, "grad_norm": 1.2339282035827637, "learning_rate": 9.614967368687115e-07, "loss": 0.4428, "step": 347 }, { "epoch": 0.15262321144674085, "grad_norm": 1.2473440170288086, "learning_rate": 9.612228803513975e-07, "loss": 0.4359, "step": 348 }, { "epoch": 0.1530617838934269, "grad_norm": 1.1782147884368896, "learning_rate": 9.609480926553887e-07, "loss": 0.4493, "step": 349 }, { "epoch": 0.15350035634011294, "grad_norm": 1.233314037322998, "learning_rate": 9.606723743354637e-07, "loss": 0.4602, "step": 350 }, { "epoch": 0.15393892878679896, "grad_norm": 1.267967700958252, "learning_rate": 9.60395725948279e-07, "loss": 0.4777, "step": 351 }, { "epoch": 0.154377501233485, "grad_norm": 1.3795247077941895, "learning_rate": 9.601181480523702e-07, "loss": 0.425, "step": 352 }, { "epoch": 0.15481607368017106, "grad_norm": 1.411054253578186, "learning_rate": 9.598396412081488e-07, "loss": 0.479, "step": 353 }, { "epoch": 0.15525464612685708, "grad_norm": 1.2811428308486938, "learning_rate": 9.595602059779015e-07, "loss": 0.4635, "step": 354 }, { "epoch": 0.15569321857354312, "grad_norm": 1.3159114122390747, "learning_rate": 9.592798429257899e-07, "loss": 0.444, "step": 355 }, { "epoch": 0.15613179102022914, "grad_norm": 1.1596899032592773, "learning_rate": 9.589985526178484e-07, "loss": 0.3732, "step": 356 }, { "epoch": 0.1565703634669152, "grad_norm": 1.2319432497024536, "learning_rate": 9.587163356219836e-07, "loss": 0.426, "step": 357 }, { "epoch": 0.15700893591360124, "grad_norm": 1.502126932144165, "learning_rate": 9.584331925079734e-07, "loss": 0.4757, "step": 358 }, { "epoch": 0.15744750836028726, "grad_norm": 1.2502065896987915, "learning_rate": 9.58149123847465e-07, "loss": 0.4382, "step": 359 }, { "epoch": 0.1578860808069733, "grad_norm": 1.2441209554672241, "learning_rate": 9.578641302139742e-07, "loss": 0.431, "step": 360 }, { "epoch": 0.15832465325365933, "grad_norm": 1.2779874801635742, "learning_rate": 9.575782121828845e-07, "loss": 0.4354, "step": 361 }, { "epoch": 0.15876322570034537, "grad_norm": 1.3990627527236938, "learning_rate": 9.572913703314454e-07, "loss": 0.4955, "step": 362 }, { "epoch": 0.15920179814703142, "grad_norm": 1.2426034212112427, "learning_rate": 9.570036052387724e-07, "loss": 0.4749, "step": 363 }, { "epoch": 0.15964037059371744, "grad_norm": 1.1871353387832642, "learning_rate": 9.567149174858438e-07, "loss": 0.4638, "step": 364 }, { "epoch": 0.1600789430404035, "grad_norm": 1.3922412395477295, "learning_rate": 9.564253076555013e-07, "loss": 0.4672, "step": 365 }, { "epoch": 0.16051751548708953, "grad_norm": 1.1922954320907593, "learning_rate": 9.561347763324483e-07, "loss": 0.4251, "step": 366 }, { "epoch": 0.16095608793377555, "grad_norm": 1.1256937980651855, "learning_rate": 9.558433241032483e-07, "loss": 0.4816, "step": 367 }, { "epoch": 0.1613946603804616, "grad_norm": 1.2206835746765137, "learning_rate": 9.55550951556324e-07, "loss": 0.442, "step": 368 }, { "epoch": 0.16183323282714762, "grad_norm": 1.230305790901184, "learning_rate": 9.55257659281957e-07, "loss": 0.4258, "step": 369 }, { "epoch": 0.16227180527383367, "grad_norm": 1.1930097341537476, "learning_rate": 9.549634478722843e-07, "loss": 0.4303, "step": 370 }, { "epoch": 0.16271037772051972, "grad_norm": 1.1612834930419922, "learning_rate": 9.546683179213e-07, "loss": 0.4636, "step": 371 }, { "epoch": 0.16314895016720574, "grad_norm": 1.4160550832748413, "learning_rate": 9.54372270024852e-07, "loss": 0.4658, "step": 372 }, { "epoch": 0.16358752261389178, "grad_norm": 1.2340151071548462, "learning_rate": 9.540753047806413e-07, "loss": 0.4572, "step": 373 }, { "epoch": 0.16402609506057783, "grad_norm": 1.3071835041046143, "learning_rate": 9.537774227882215e-07, "loss": 0.4409, "step": 374 }, { "epoch": 0.16446466750726385, "grad_norm": 1.381076455116272, "learning_rate": 9.534786246489966e-07, "loss": 0.466, "step": 375 }, { "epoch": 0.1649032399539499, "grad_norm": 1.2368775606155396, "learning_rate": 9.531789109662204e-07, "loss": 0.4359, "step": 376 }, { "epoch": 0.16534181240063592, "grad_norm": 1.3651288747787476, "learning_rate": 9.528782823449953e-07, "loss": 0.4409, "step": 377 }, { "epoch": 0.16578038484732197, "grad_norm": 1.466973900794983, "learning_rate": 9.525767393922706e-07, "loss": 0.4804, "step": 378 }, { "epoch": 0.166218957294008, "grad_norm": 1.3845632076263428, "learning_rate": 9.522742827168416e-07, "loss": 0.4723, "step": 379 }, { "epoch": 0.16665752974069403, "grad_norm": 1.242058277130127, "learning_rate": 9.519709129293488e-07, "loss": 0.4361, "step": 380 }, { "epoch": 0.16709610218738008, "grad_norm": 1.2678145170211792, "learning_rate": 9.516666306422755e-07, "loss": 0.4524, "step": 381 }, { "epoch": 0.16753467463406613, "grad_norm": 1.2828023433685303, "learning_rate": 9.51361436469948e-07, "loss": 0.4527, "step": 382 }, { "epoch": 0.16797324708075215, "grad_norm": 1.3856935501098633, "learning_rate": 9.510553310285331e-07, "loss": 0.4676, "step": 383 }, { "epoch": 0.1684118195274382, "grad_norm": 1.3793553113937378, "learning_rate": 9.507483149360375e-07, "loss": 0.4231, "step": 384 }, { "epoch": 0.16885039197412421, "grad_norm": 1.2514710426330566, "learning_rate": 9.504403888123066e-07, "loss": 0.4119, "step": 385 }, { "epoch": 0.16928896442081026, "grad_norm": 1.2418358325958252, "learning_rate": 9.50131553279023e-07, "loss": 0.4563, "step": 386 }, { "epoch": 0.1697275368674963, "grad_norm": 1.291680097579956, "learning_rate": 9.498218089597054e-07, "loss": 0.4411, "step": 387 }, { "epoch": 0.17016610931418233, "grad_norm": 1.2920475006103516, "learning_rate": 9.495111564797073e-07, "loss": 0.4543, "step": 388 }, { "epoch": 0.17060468176086838, "grad_norm": 1.341176152229309, "learning_rate": 9.491995964662154e-07, "loss": 0.4597, "step": 389 }, { "epoch": 0.17104325420755442, "grad_norm": 1.2180360555648804, "learning_rate": 9.488871295482491e-07, "loss": 0.4578, "step": 390 }, { "epoch": 0.17148182665424044, "grad_norm": 1.3022912740707397, "learning_rate": 9.485737563566585e-07, "loss": 0.4191, "step": 391 }, { "epoch": 0.1719203991009265, "grad_norm": 1.261922001838684, "learning_rate": 9.482594775241236e-07, "loss": 0.4556, "step": 392 }, { "epoch": 0.1723589715476125, "grad_norm": 1.392162799835205, "learning_rate": 9.479442936851526e-07, "loss": 0.456, "step": 393 }, { "epoch": 0.17279754399429856, "grad_norm": 1.3650363683700562, "learning_rate": 9.476282054760809e-07, "loss": 0.4934, "step": 394 }, { "epoch": 0.1732361164409846, "grad_norm": 1.3042758703231812, "learning_rate": 9.4731121353507e-07, "loss": 0.4519, "step": 395 }, { "epoch": 0.17367468888767063, "grad_norm": 1.3327823877334595, "learning_rate": 9.469933185021058e-07, "loss": 0.4212, "step": 396 }, { "epoch": 0.17411326133435667, "grad_norm": 1.2559716701507568, "learning_rate": 9.466745210189972e-07, "loss": 0.4643, "step": 397 }, { "epoch": 0.1745518337810427, "grad_norm": 1.188941478729248, "learning_rate": 9.463548217293759e-07, "loss": 0.4303, "step": 398 }, { "epoch": 0.17499040622772874, "grad_norm": 1.3942642211914062, "learning_rate": 9.460342212786932e-07, "loss": 0.4702, "step": 399 }, { "epoch": 0.1754289786744148, "grad_norm": 1.2623578310012817, "learning_rate": 9.457127203142206e-07, "loss": 0.4399, "step": 400 }, { "epoch": 0.1758675511211008, "grad_norm": 1.3886511325836182, "learning_rate": 9.453903194850475e-07, "loss": 0.4714, "step": 401 }, { "epoch": 0.17630612356778685, "grad_norm": 1.1634944677352905, "learning_rate": 9.4506701944208e-07, "loss": 0.4144, "step": 402 }, { "epoch": 0.1767446960144729, "grad_norm": 1.3607139587402344, "learning_rate": 9.447428208380395e-07, "loss": 0.4516, "step": 403 }, { "epoch": 0.17718326846115892, "grad_norm": 1.351024866104126, "learning_rate": 9.444177243274617e-07, "loss": 0.4194, "step": 404 }, { "epoch": 0.17762184090784497, "grad_norm": 1.2279891967773438, "learning_rate": 9.440917305666951e-07, "loss": 0.482, "step": 405 }, { "epoch": 0.178060413354531, "grad_norm": 1.2650517225265503, "learning_rate": 9.437648402138998e-07, "loss": 0.414, "step": 406 }, { "epoch": 0.17849898580121704, "grad_norm": 1.3450144529342651, "learning_rate": 9.434370539290459e-07, "loss": 0.4864, "step": 407 }, { "epoch": 0.17893755824790308, "grad_norm": 1.2379168272018433, "learning_rate": 9.431083723739124e-07, "loss": 0.4029, "step": 408 }, { "epoch": 0.1793761306945891, "grad_norm": 1.3272401094436646, "learning_rate": 9.427787962120857e-07, "loss": 0.4524, "step": 409 }, { "epoch": 0.17981470314127515, "grad_norm": 1.3062082529067993, "learning_rate": 9.424483261089583e-07, "loss": 0.4385, "step": 410 }, { "epoch": 0.1802532755879612, "grad_norm": 1.2583765983581543, "learning_rate": 9.421169627317278e-07, "loss": 0.4247, "step": 411 }, { "epoch": 0.18069184803464722, "grad_norm": 1.2366396188735962, "learning_rate": 9.417847067493952e-07, "loss": 0.4399, "step": 412 }, { "epoch": 0.18113042048133327, "grad_norm": 1.334314227104187, "learning_rate": 9.414515588327631e-07, "loss": 0.438, "step": 413 }, { "epoch": 0.18156899292801928, "grad_norm": 1.3514381647109985, "learning_rate": 9.411175196544358e-07, "loss": 0.5151, "step": 414 }, { "epoch": 0.18200756537470533, "grad_norm": 1.2838636636734009, "learning_rate": 9.40782589888816e-07, "loss": 0.4465, "step": 415 }, { "epoch": 0.18244613782139138, "grad_norm": 1.25214684009552, "learning_rate": 9.404467702121051e-07, "loss": 0.3996, "step": 416 }, { "epoch": 0.1828847102680774, "grad_norm": 1.323198676109314, "learning_rate": 9.40110061302301e-07, "loss": 0.4093, "step": 417 }, { "epoch": 0.18332328271476345, "grad_norm": 1.3139373064041138, "learning_rate": 9.397724638391967e-07, "loss": 0.4789, "step": 418 }, { "epoch": 0.1837618551614495, "grad_norm": 1.1997904777526855, "learning_rate": 9.394339785043794e-07, "loss": 0.4365, "step": 419 }, { "epoch": 0.18420042760813551, "grad_norm": 1.297034502029419, "learning_rate": 9.390946059812289e-07, "loss": 0.4518, "step": 420 }, { "epoch": 0.18463900005482156, "grad_norm": 1.1115834712982178, "learning_rate": 9.387543469549155e-07, "loss": 0.4254, "step": 421 }, { "epoch": 0.18507757250150758, "grad_norm": 1.141788363456726, "learning_rate": 9.384132021124004e-07, "loss": 0.3842, "step": 422 }, { "epoch": 0.18551614494819363, "grad_norm": 1.3769128322601318, "learning_rate": 9.380711721424326e-07, "loss": 0.4629, "step": 423 }, { "epoch": 0.18595471739487968, "grad_norm": 1.3783560991287231, "learning_rate": 9.377282577355478e-07, "loss": 0.4906, "step": 424 }, { "epoch": 0.1863932898415657, "grad_norm": 1.255490779876709, "learning_rate": 9.373844595840678e-07, "loss": 0.4537, "step": 425 }, { "epoch": 0.18683186228825174, "grad_norm": 1.354235053062439, "learning_rate": 9.370397783820984e-07, "loss": 0.4497, "step": 426 }, { "epoch": 0.1872704347349378, "grad_norm": 1.347273826599121, "learning_rate": 9.366942148255285e-07, "loss": 0.4267, "step": 427 }, { "epoch": 0.1877090071816238, "grad_norm": 1.2848392724990845, "learning_rate": 9.363477696120283e-07, "loss": 0.4738, "step": 428 }, { "epoch": 0.18814757962830986, "grad_norm": 1.2392444610595703, "learning_rate": 9.360004434410476e-07, "loss": 0.4834, "step": 429 }, { "epoch": 0.18858615207499588, "grad_norm": 1.195491909980774, "learning_rate": 9.356522370138154e-07, "loss": 0.3701, "step": 430 }, { "epoch": 0.18902472452168192, "grad_norm": 1.2303284406661987, "learning_rate": 9.353031510333373e-07, "loss": 0.4281, "step": 431 }, { "epoch": 0.18946329696836797, "grad_norm": 1.2569408416748047, "learning_rate": 9.349531862043951e-07, "loss": 0.4603, "step": 432 }, { "epoch": 0.189901869415054, "grad_norm": 1.3959214687347412, "learning_rate": 9.346023432335449e-07, "loss": 0.4663, "step": 433 }, { "epoch": 0.19034044186174004, "grad_norm": 1.31195068359375, "learning_rate": 9.342506228291156e-07, "loss": 0.4729, "step": 434 }, { "epoch": 0.1907790143084261, "grad_norm": 1.33975088596344, "learning_rate": 9.338980257012074e-07, "loss": 0.4647, "step": 435 }, { "epoch": 0.1912175867551121, "grad_norm": 1.283033013343811, "learning_rate": 9.335445525616909e-07, "loss": 0.4047, "step": 436 }, { "epoch": 0.19165615920179815, "grad_norm": 1.341052770614624, "learning_rate": 9.331902041242053e-07, "loss": 0.4434, "step": 437 }, { "epoch": 0.19209473164848417, "grad_norm": 1.3923901319503784, "learning_rate": 9.328349811041564e-07, "loss": 0.4544, "step": 438 }, { "epoch": 0.19253330409517022, "grad_norm": 1.2661969661712646, "learning_rate": 9.324788842187163e-07, "loss": 0.4461, "step": 439 }, { "epoch": 0.19297187654185627, "grad_norm": 1.2110265493392944, "learning_rate": 9.321219141868215e-07, "loss": 0.4707, "step": 440 }, { "epoch": 0.1934104489885423, "grad_norm": 1.4487905502319336, "learning_rate": 9.317640717291708e-07, "loss": 0.4779, "step": 441 }, { "epoch": 0.19384902143522834, "grad_norm": 1.3091298341751099, "learning_rate": 9.314053575682246e-07, "loss": 0.4674, "step": 442 }, { "epoch": 0.19428759388191436, "grad_norm": 1.2794967889785767, "learning_rate": 9.310457724282033e-07, "loss": 0.3946, "step": 443 }, { "epoch": 0.1947261663286004, "grad_norm": 1.2558822631835938, "learning_rate": 9.306853170350854e-07, "loss": 0.4358, "step": 444 }, { "epoch": 0.19516473877528645, "grad_norm": 1.1746054887771606, "learning_rate": 9.303239921166071e-07, "loss": 0.434, "step": 445 }, { "epoch": 0.19560331122197247, "grad_norm": 1.2854063510894775, "learning_rate": 9.299617984022597e-07, "loss": 0.4739, "step": 446 }, { "epoch": 0.19604188366865852, "grad_norm": 1.214192509651184, "learning_rate": 9.29598736623288e-07, "loss": 0.4272, "step": 447 }, { "epoch": 0.19648045611534457, "grad_norm": 1.248357892036438, "learning_rate": 9.292348075126901e-07, "loss": 0.431, "step": 448 }, { "epoch": 0.19691902856203058, "grad_norm": 1.2610982656478882, "learning_rate": 9.288700118052151e-07, "loss": 0.4582, "step": 449 }, { "epoch": 0.19735760100871663, "grad_norm": 1.3101803064346313, "learning_rate": 9.285043502373615e-07, "loss": 0.4896, "step": 450 }, { "epoch": 0.19779617345540265, "grad_norm": 1.4143675565719604, "learning_rate": 9.281378235473761e-07, "loss": 0.4681, "step": 451 }, { "epoch": 0.1982347459020887, "grad_norm": 1.2853251695632935, "learning_rate": 9.27770432475252e-07, "loss": 0.4826, "step": 452 }, { "epoch": 0.19867331834877475, "grad_norm": 1.3457460403442383, "learning_rate": 9.274021777627276e-07, "loss": 0.4796, "step": 453 }, { "epoch": 0.19911189079546077, "grad_norm": 1.3269720077514648, "learning_rate": 9.270330601532854e-07, "loss": 0.4544, "step": 454 }, { "epoch": 0.1995504632421468, "grad_norm": 1.2965871095657349, "learning_rate": 9.266630803921491e-07, "loss": 0.4134, "step": 455 }, { "epoch": 0.19998903568883286, "grad_norm": 1.250986933708191, "learning_rate": 9.262922392262837e-07, "loss": 0.4538, "step": 456 }, { "epoch": 0.20042760813551888, "grad_norm": 1.3280874490737915, "learning_rate": 9.259205374043933e-07, "loss": 0.457, "step": 457 }, { "epoch": 0.20086618058220493, "grad_norm": 1.3324693441390991, "learning_rate": 9.255479756769193e-07, "loss": 0.4568, "step": 458 }, { "epoch": 0.20130475302889095, "grad_norm": 1.4459820985794067, "learning_rate": 9.251745547960393e-07, "loss": 0.5189, "step": 459 }, { "epoch": 0.201743325475577, "grad_norm": 1.2799947261810303, "learning_rate": 9.248002755156659e-07, "loss": 0.4259, "step": 460 }, { "epoch": 0.20218189792226304, "grad_norm": 1.3187311887741089, "learning_rate": 9.244251385914437e-07, "loss": 0.4845, "step": 461 }, { "epoch": 0.20262047036894906, "grad_norm": 1.3244918584823608, "learning_rate": 9.240491447807501e-07, "loss": 0.4272, "step": 462 }, { "epoch": 0.2030590428156351, "grad_norm": 1.355553150177002, "learning_rate": 9.236722948426918e-07, "loss": 0.4544, "step": 463 }, { "epoch": 0.20349761526232116, "grad_norm": 1.241315245628357, "learning_rate": 9.232945895381039e-07, "loss": 0.419, "step": 464 }, { "epoch": 0.20393618770900718, "grad_norm": 1.3401538133621216, "learning_rate": 9.229160296295487e-07, "loss": 0.3983, "step": 465 }, { "epoch": 0.20437476015569322, "grad_norm": 1.4097830057144165, "learning_rate": 9.225366158813138e-07, "loss": 0.4733, "step": 466 }, { "epoch": 0.20481333260237924, "grad_norm": 1.2525343894958496, "learning_rate": 9.221563490594103e-07, "loss": 0.4384, "step": 467 }, { "epoch": 0.2052519050490653, "grad_norm": 1.1235498189926147, "learning_rate": 9.217752299315724e-07, "loss": 0.4004, "step": 468 }, { "epoch": 0.20569047749575134, "grad_norm": 1.2831324338912964, "learning_rate": 9.213932592672544e-07, "loss": 0.4974, "step": 469 }, { "epoch": 0.20612904994243736, "grad_norm": 1.2940030097961426, "learning_rate": 9.2101043783763e-07, "loss": 0.4909, "step": 470 }, { "epoch": 0.2065676223891234, "grad_norm": 1.3095605373382568, "learning_rate": 9.206267664155906e-07, "loss": 0.4604, "step": 471 }, { "epoch": 0.20700619483580945, "grad_norm": 1.2282830476760864, "learning_rate": 9.202422457757433e-07, "loss": 0.4773, "step": 472 }, { "epoch": 0.20744476728249547, "grad_norm": 1.3521174192428589, "learning_rate": 9.198568766944102e-07, "loss": 0.4535, "step": 473 }, { "epoch": 0.20788333972918152, "grad_norm": 1.260043978691101, "learning_rate": 9.194706599496262e-07, "loss": 0.4328, "step": 474 }, { "epoch": 0.20832191217586754, "grad_norm": 1.3410409688949585, "learning_rate": 9.190835963211376e-07, "loss": 0.4879, "step": 475 }, { "epoch": 0.2087604846225536, "grad_norm": 1.3885812759399414, "learning_rate": 9.186956865904003e-07, "loss": 0.4449, "step": 476 }, { "epoch": 0.20919905706923964, "grad_norm": 1.3543665409088135, "learning_rate": 9.183069315405784e-07, "loss": 0.5023, "step": 477 }, { "epoch": 0.20963762951592566, "grad_norm": 1.3189713954925537, "learning_rate": 9.179173319565432e-07, "loss": 0.4523, "step": 478 }, { "epoch": 0.2100762019626117, "grad_norm": 1.262413501739502, "learning_rate": 9.175268886248704e-07, "loss": 0.4632, "step": 479 }, { "epoch": 0.21051477440929772, "grad_norm": 1.3755730390548706, "learning_rate": 9.171356023338395e-07, "loss": 0.4475, "step": 480 }, { "epoch": 0.21095334685598377, "grad_norm": 1.2970229387283325, "learning_rate": 9.167434738734318e-07, "loss": 0.4604, "step": 481 }, { "epoch": 0.21139191930266982, "grad_norm": 1.2604976892471313, "learning_rate": 9.163505040353287e-07, "loss": 0.5005, "step": 482 }, { "epoch": 0.21183049174935584, "grad_norm": 1.4260404109954834, "learning_rate": 9.159566936129111e-07, "loss": 0.4655, "step": 483 }, { "epoch": 0.21226906419604188, "grad_norm": 1.4161478281021118, "learning_rate": 9.155620434012556e-07, "loss": 0.4763, "step": 484 }, { "epoch": 0.21270763664272793, "grad_norm": 1.3012006282806396, "learning_rate": 9.151665541971356e-07, "loss": 0.4265, "step": 485 }, { "epoch": 0.21314620908941395, "grad_norm": 1.3242242336273193, "learning_rate": 9.147702267990177e-07, "loss": 0.4468, "step": 486 }, { "epoch": 0.2135847815361, "grad_norm": 1.350150465965271, "learning_rate": 9.143730620070608e-07, "loss": 0.4252, "step": 487 }, { "epoch": 0.21402335398278602, "grad_norm": 1.3396698236465454, "learning_rate": 9.139750606231145e-07, "loss": 0.4238, "step": 488 }, { "epoch": 0.21446192642947207, "grad_norm": 1.2412779331207275, "learning_rate": 9.135762234507174e-07, "loss": 0.4363, "step": 489 }, { "epoch": 0.2149004988761581, "grad_norm": 1.2845489978790283, "learning_rate": 9.131765512950958e-07, "loss": 0.447, "step": 490 }, { "epoch": 0.21533907132284413, "grad_norm": 1.3290574550628662, "learning_rate": 9.127760449631612e-07, "loss": 0.5195, "step": 491 }, { "epoch": 0.21577764376953018, "grad_norm": 1.325376033782959, "learning_rate": 9.123747052635098e-07, "loss": 0.4565, "step": 492 }, { "epoch": 0.21621621621621623, "grad_norm": 1.39933180809021, "learning_rate": 9.119725330064201e-07, "loss": 0.4568, "step": 493 }, { "epoch": 0.21665478866290225, "grad_norm": 1.3452569246292114, "learning_rate": 9.115695290038513e-07, "loss": 0.4797, "step": 494 }, { "epoch": 0.2170933611095883, "grad_norm": 1.330816388130188, "learning_rate": 9.111656940694421e-07, "loss": 0.4302, "step": 495 }, { "epoch": 0.21753193355627432, "grad_norm": 1.2303009033203125, "learning_rate": 9.107610290185087e-07, "loss": 0.4297, "step": 496 }, { "epoch": 0.21797050600296036, "grad_norm": 1.2003905773162842, "learning_rate": 9.103555346680433e-07, "loss": 0.4407, "step": 497 }, { "epoch": 0.2184090784496464, "grad_norm": 1.2907445430755615, "learning_rate": 9.099492118367122e-07, "loss": 0.4384, "step": 498 }, { "epoch": 0.21884765089633243, "grad_norm": 1.2537270784378052, "learning_rate": 9.095420613448548e-07, "loss": 0.4237, "step": 499 }, { "epoch": 0.21928622334301848, "grad_norm": 1.3271981477737427, "learning_rate": 9.091340840144807e-07, "loss": 0.4436, "step": 500 }, { "epoch": 0.21972479578970452, "grad_norm": 1.2225565910339355, "learning_rate": 9.087252806692699e-07, "loss": 0.4514, "step": 501 }, { "epoch": 0.22016336823639054, "grad_norm": 1.273189663887024, "learning_rate": 9.083156521345692e-07, "loss": 0.4481, "step": 502 }, { "epoch": 0.2206019406830766, "grad_norm": 1.3194257020950317, "learning_rate": 9.079051992373916e-07, "loss": 0.4672, "step": 503 }, { "epoch": 0.2210405131297626, "grad_norm": 1.24362051486969, "learning_rate": 9.074939228064147e-07, "loss": 0.4002, "step": 504 }, { "epoch": 0.22147908557644866, "grad_norm": 1.4185460805892944, "learning_rate": 9.070818236719785e-07, "loss": 0.4555, "step": 505 }, { "epoch": 0.2219176580231347, "grad_norm": 1.2558276653289795, "learning_rate": 9.066689026660842e-07, "loss": 0.4683, "step": 506 }, { "epoch": 0.22235623046982073, "grad_norm": 1.3859413862228394, "learning_rate": 9.062551606223921e-07, "loss": 0.4426, "step": 507 }, { "epoch": 0.22279480291650677, "grad_norm": 1.3655338287353516, "learning_rate": 9.058405983762201e-07, "loss": 0.405, "step": 508 }, { "epoch": 0.22323337536319282, "grad_norm": 1.2751038074493408, "learning_rate": 9.054252167645425e-07, "loss": 0.4462, "step": 509 }, { "epoch": 0.22367194780987884, "grad_norm": 1.3899219036102295, "learning_rate": 9.050090166259872e-07, "loss": 0.412, "step": 510 }, { "epoch": 0.2241105202565649, "grad_norm": 1.3597509860992432, "learning_rate": 9.045919988008348e-07, "loss": 0.4391, "step": 511 }, { "epoch": 0.2245490927032509, "grad_norm": 1.4710191488265991, "learning_rate": 9.041741641310172e-07, "loss": 0.44, "step": 512 }, { "epoch": 0.22498766514993696, "grad_norm": 1.2938218116760254, "learning_rate": 9.037555134601149e-07, "loss": 0.433, "step": 513 }, { "epoch": 0.225426237596623, "grad_norm": 1.3852367401123047, "learning_rate": 9.033360476333565e-07, "loss": 0.4891, "step": 514 }, { "epoch": 0.22586481004330902, "grad_norm": 1.2105050086975098, "learning_rate": 9.029157674976154e-07, "loss": 0.4549, "step": 515 }, { "epoch": 0.22630338248999507, "grad_norm": 1.290735125541687, "learning_rate": 9.0249467390141e-07, "loss": 0.436, "step": 516 }, { "epoch": 0.2267419549366811, "grad_norm": 1.3402304649353027, "learning_rate": 9.020727676949004e-07, "loss": 0.4032, "step": 517 }, { "epoch": 0.22718052738336714, "grad_norm": 1.257806420326233, "learning_rate": 9.016500497298876e-07, "loss": 0.4829, "step": 518 }, { "epoch": 0.22761909983005318, "grad_norm": 1.203140139579773, "learning_rate": 9.012265208598113e-07, "loss": 0.4132, "step": 519 }, { "epoch": 0.2280576722767392, "grad_norm": 1.2450907230377197, "learning_rate": 9.008021819397486e-07, "loss": 0.4349, "step": 520 }, { "epoch": 0.22849624472342525, "grad_norm": 1.3185664415359497, "learning_rate": 9.003770338264118e-07, "loss": 0.4742, "step": 521 }, { "epoch": 0.2289348171701113, "grad_norm": 1.2235329151153564, "learning_rate": 8.999510773781471e-07, "loss": 0.4181, "step": 522 }, { "epoch": 0.22937338961679732, "grad_norm": 1.2654236555099487, "learning_rate": 8.995243134549326e-07, "loss": 0.4767, "step": 523 }, { "epoch": 0.22981196206348337, "grad_norm": 1.2756022214889526, "learning_rate": 8.990967429183765e-07, "loss": 0.4335, "step": 524 }, { "epoch": 0.23025053451016939, "grad_norm": 1.337327241897583, "learning_rate": 8.986683666317157e-07, "loss": 0.4182, "step": 525 }, { "epoch": 0.23068910695685543, "grad_norm": 1.3000355958938599, "learning_rate": 8.982391854598137e-07, "loss": 0.4844, "step": 526 }, { "epoch": 0.23112767940354148, "grad_norm": 1.393452525138855, "learning_rate": 8.97809200269159e-07, "loss": 0.4648, "step": 527 }, { "epoch": 0.2315662518502275, "grad_norm": 1.2905426025390625, "learning_rate": 8.973784119278639e-07, "loss": 0.4342, "step": 528 }, { "epoch": 0.23200482429691355, "grad_norm": 1.1917750835418701, "learning_rate": 8.969468213056613e-07, "loss": 0.4271, "step": 529 }, { "epoch": 0.2324433967435996, "grad_norm": 1.3469759225845337, "learning_rate": 8.965144292739046e-07, "loss": 0.4533, "step": 530 }, { "epoch": 0.23288196919028561, "grad_norm": 1.2097269296646118, "learning_rate": 8.960812367055646e-07, "loss": 0.4292, "step": 531 }, { "epoch": 0.23332054163697166, "grad_norm": 1.1486494541168213, "learning_rate": 8.95647244475229e-07, "loss": 0.4925, "step": 532 }, { "epoch": 0.23375911408365768, "grad_norm": 1.2630372047424316, "learning_rate": 8.952124534590993e-07, "loss": 0.4106, "step": 533 }, { "epoch": 0.23419768653034373, "grad_norm": 1.3153948783874512, "learning_rate": 8.947768645349903e-07, "loss": 0.4591, "step": 534 }, { "epoch": 0.23463625897702978, "grad_norm": 1.280995488166809, "learning_rate": 8.943404785823269e-07, "loss": 0.4071, "step": 535 }, { "epoch": 0.2350748314237158, "grad_norm": 1.2669576406478882, "learning_rate": 8.939032964821442e-07, "loss": 0.4185, "step": 536 }, { "epoch": 0.23551340387040184, "grad_norm": 1.197329044342041, "learning_rate": 8.93465319117084e-07, "loss": 0.4476, "step": 537 }, { "epoch": 0.2359519763170879, "grad_norm": 1.1159907579421997, "learning_rate": 8.930265473713937e-07, "loss": 0.429, "step": 538 }, { "epoch": 0.2363905487637739, "grad_norm": 1.3145620822906494, "learning_rate": 8.925869821309247e-07, "loss": 0.4711, "step": 539 }, { "epoch": 0.23682912121045996, "grad_norm": 1.306536078453064, "learning_rate": 8.921466242831303e-07, "loss": 0.4491, "step": 540 }, { "epoch": 0.23726769365714598, "grad_norm": 1.2363736629486084, "learning_rate": 8.917054747170642e-07, "loss": 0.44, "step": 541 }, { "epoch": 0.23770626610383203, "grad_norm": 1.340304970741272, "learning_rate": 8.912635343233783e-07, "loss": 0.4919, "step": 542 }, { "epoch": 0.23814483855051807, "grad_norm": 1.2518609762191772, "learning_rate": 8.908208039943213e-07, "loss": 0.4362, "step": 543 }, { "epoch": 0.2385834109972041, "grad_norm": 1.2911936044692993, "learning_rate": 8.903772846237364e-07, "loss": 0.4416, "step": 544 }, { "epoch": 0.23902198344389014, "grad_norm": 1.2846801280975342, "learning_rate": 8.899329771070602e-07, "loss": 0.4885, "step": 545 }, { "epoch": 0.2394605558905762, "grad_norm": 1.3153619766235352, "learning_rate": 8.894878823413207e-07, "loss": 0.4922, "step": 546 }, { "epoch": 0.2398991283372622, "grad_norm": 1.4053038358688354, "learning_rate": 8.890420012251346e-07, "loss": 0.4912, "step": 547 }, { "epoch": 0.24033770078394825, "grad_norm": 1.302511215209961, "learning_rate": 8.885953346587065e-07, "loss": 0.4019, "step": 548 }, { "epoch": 0.24077627323063427, "grad_norm": 1.3971174955368042, "learning_rate": 8.881478835438272e-07, "loss": 0.4438, "step": 549 }, { "epoch": 0.24121484567732032, "grad_norm": 1.2062628269195557, "learning_rate": 8.876996487838711e-07, "loss": 0.383, "step": 550 }, { "epoch": 0.24165341812400637, "grad_norm": 1.3044462203979492, "learning_rate": 8.872506312837944e-07, "loss": 0.451, "step": 551 }, { "epoch": 0.2420919905706924, "grad_norm": 1.3488870859146118, "learning_rate": 8.868008319501341e-07, "loss": 0.4249, "step": 552 }, { "epoch": 0.24253056301737844, "grad_norm": 1.1751151084899902, "learning_rate": 8.863502516910057e-07, "loss": 0.4581, "step": 553 }, { "epoch": 0.24296913546406446, "grad_norm": 1.3605632781982422, "learning_rate": 8.858988914161009e-07, "loss": 0.4447, "step": 554 }, { "epoch": 0.2434077079107505, "grad_norm": 1.2312195301055908, "learning_rate": 8.854467520366864e-07, "loss": 0.4164, "step": 555 }, { "epoch": 0.24384628035743655, "grad_norm": 1.3179501295089722, "learning_rate": 8.849938344656021e-07, "loss": 0.4674, "step": 556 }, { "epoch": 0.24428485280412257, "grad_norm": 1.2114101648330688, "learning_rate": 8.845401396172588e-07, "loss": 0.4529, "step": 557 }, { "epoch": 0.24472342525080862, "grad_norm": 1.238842487335205, "learning_rate": 8.840856684076365e-07, "loss": 0.4685, "step": 558 }, { "epoch": 0.24516199769749467, "grad_norm": 1.1789189577102661, "learning_rate": 8.836304217542828e-07, "loss": 0.4282, "step": 559 }, { "epoch": 0.24560057014418069, "grad_norm": 1.304688811302185, "learning_rate": 8.831744005763107e-07, "loss": 0.4424, "step": 560 }, { "epoch": 0.24603914259086673, "grad_norm": 1.3729431629180908, "learning_rate": 8.827176057943969e-07, "loss": 0.4683, "step": 561 }, { "epoch": 0.24647771503755275, "grad_norm": 1.3452732563018799, "learning_rate": 8.822600383307802e-07, "loss": 0.4498, "step": 562 }, { "epoch": 0.2469162874842388, "grad_norm": 1.27022385597229, "learning_rate": 8.818016991092594e-07, "loss": 0.4617, "step": 563 }, { "epoch": 0.24735485993092485, "grad_norm": 1.2468392848968506, "learning_rate": 8.813425890551909e-07, "loss": 0.4356, "step": 564 }, { "epoch": 0.24779343237761087, "grad_norm": 1.2570581436157227, "learning_rate": 8.808827090954881e-07, "loss": 0.3973, "step": 565 }, { "epoch": 0.24823200482429691, "grad_norm": 1.2952882051467896, "learning_rate": 8.804220601586183e-07, "loss": 0.4624, "step": 566 }, { "epoch": 0.24867057727098296, "grad_norm": 1.2446449995040894, "learning_rate": 8.799606431746013e-07, "loss": 0.3957, "step": 567 }, { "epoch": 0.24910914971766898, "grad_norm": 1.4092421531677246, "learning_rate": 8.794984590750079e-07, "loss": 0.4968, "step": 568 }, { "epoch": 0.24954772216435503, "grad_norm": 1.3203057050704956, "learning_rate": 8.790355087929573e-07, "loss": 0.4874, "step": 569 }, { "epoch": 0.24998629461104105, "grad_norm": 1.468639850616455, "learning_rate": 8.785717932631155e-07, "loss": 0.4666, "step": 570 }, { "epoch": 0.2504248670577271, "grad_norm": 1.3520934581756592, "learning_rate": 8.781073134216943e-07, "loss": 0.4469, "step": 571 }, { "epoch": 0.25086343950441314, "grad_norm": 1.266510248184204, "learning_rate": 8.776420702064473e-07, "loss": 0.3974, "step": 572 }, { "epoch": 0.2513020119510992, "grad_norm": 1.2830793857574463, "learning_rate": 8.771760645566705e-07, "loss": 0.4821, "step": 573 }, { "epoch": 0.2517405843977852, "grad_norm": 1.3247817754745483, "learning_rate": 8.767092974131984e-07, "loss": 0.4426, "step": 574 }, { "epoch": 0.25217915684447123, "grad_norm": 1.340187907218933, "learning_rate": 8.762417697184032e-07, "loss": 0.4512, "step": 575 }, { "epoch": 0.2526177292911573, "grad_norm": 1.2662535905838013, "learning_rate": 8.757734824161929e-07, "loss": 0.4661, "step": 576 }, { "epoch": 0.2530563017378433, "grad_norm": 1.2929002046585083, "learning_rate": 8.753044364520083e-07, "loss": 0.4576, "step": 577 }, { "epoch": 0.2534948741845294, "grad_norm": 1.3643664121627808, "learning_rate": 8.748346327728228e-07, "loss": 0.4642, "step": 578 }, { "epoch": 0.2539334466312154, "grad_norm": 1.4776318073272705, "learning_rate": 8.74364072327139e-07, "loss": 0.4554, "step": 579 }, { "epoch": 0.2543720190779014, "grad_norm": 1.1963412761688232, "learning_rate": 8.738927560649876e-07, "loss": 0.4026, "step": 580 }, { "epoch": 0.25481059152458746, "grad_norm": 1.3636819124221802, "learning_rate": 8.734206849379253e-07, "loss": 0.4415, "step": 581 }, { "epoch": 0.2552491639712735, "grad_norm": 1.3011325597763062, "learning_rate": 8.729478598990323e-07, "loss": 0.4289, "step": 582 }, { "epoch": 0.25568773641795955, "grad_norm": 1.2742562294006348, "learning_rate": 8.724742819029116e-07, "loss": 0.4317, "step": 583 }, { "epoch": 0.2561263088646456, "grad_norm": 1.4198004007339478, "learning_rate": 8.719999519056859e-07, "loss": 0.4656, "step": 584 }, { "epoch": 0.2565648813113316, "grad_norm": 1.376230001449585, "learning_rate": 8.715248708649963e-07, "loss": 0.459, "step": 585 }, { "epoch": 0.25700345375801764, "grad_norm": 1.1574681997299194, "learning_rate": 8.710490397400005e-07, "loss": 0.4235, "step": 586 }, { "epoch": 0.2574420262047037, "grad_norm": 1.2351993322372437, "learning_rate": 8.7057245949137e-07, "loss": 0.4214, "step": 587 }, { "epoch": 0.25788059865138974, "grad_norm": 1.2393819093704224, "learning_rate": 8.70095131081289e-07, "loss": 0.415, "step": 588 }, { "epoch": 0.2583191710980758, "grad_norm": 1.267637014389038, "learning_rate": 8.696170554734523e-07, "loss": 0.4274, "step": 589 }, { "epoch": 0.2587577435447618, "grad_norm": 1.3031575679779053, "learning_rate": 8.691382336330631e-07, "loss": 0.4582, "step": 590 }, { "epoch": 0.2591963159914478, "grad_norm": 1.3642998933792114, "learning_rate": 8.686586665268313e-07, "loss": 0.465, "step": 591 }, { "epoch": 0.25963488843813387, "grad_norm": 1.3312978744506836, "learning_rate": 8.681783551229713e-07, "loss": 0.4679, "step": 592 }, { "epoch": 0.2600734608848199, "grad_norm": 1.3136823177337646, "learning_rate": 8.676973003912004e-07, "loss": 0.485, "step": 593 }, { "epoch": 0.26051203333150597, "grad_norm": 1.2545801401138306, "learning_rate": 8.672155033027364e-07, "loss": 0.4402, "step": 594 }, { "epoch": 0.26095060577819196, "grad_norm": 1.227077603340149, "learning_rate": 8.667329648302959e-07, "loss": 0.4597, "step": 595 }, { "epoch": 0.261389178224878, "grad_norm": 1.3195209503173828, "learning_rate": 8.662496859480925e-07, "loss": 0.4567, "step": 596 }, { "epoch": 0.26182775067156405, "grad_norm": 1.235021710395813, "learning_rate": 8.657656676318345e-07, "loss": 0.4215, "step": 597 }, { "epoch": 0.2622663231182501, "grad_norm": 1.2290730476379395, "learning_rate": 8.652809108587231e-07, "loss": 0.4139, "step": 598 }, { "epoch": 0.26270489556493615, "grad_norm": 1.2141824960708618, "learning_rate": 8.647954166074503e-07, "loss": 0.4234, "step": 599 }, { "epoch": 0.2631434680116222, "grad_norm": 1.3154791593551636, "learning_rate": 8.64309185858197e-07, "loss": 0.4791, "step": 600 }, { "epoch": 0.2635820404583082, "grad_norm": 1.2648977041244507, "learning_rate": 8.638222195926313e-07, "loss": 0.4227, "step": 601 }, { "epoch": 0.26402061290499423, "grad_norm": 1.231683373451233, "learning_rate": 8.633345187939061e-07, "loss": 0.4806, "step": 602 }, { "epoch": 0.2644591853516803, "grad_norm": 1.3356053829193115, "learning_rate": 8.628460844466572e-07, "loss": 0.4733, "step": 603 }, { "epoch": 0.26489775779836633, "grad_norm": 1.3214858770370483, "learning_rate": 8.623569175370016e-07, "loss": 0.4439, "step": 604 }, { "epoch": 0.2653363302450524, "grad_norm": 1.3323317766189575, "learning_rate": 8.61867019052535e-07, "loss": 0.4439, "step": 605 }, { "epoch": 0.26577490269173837, "grad_norm": 1.388413906097412, "learning_rate": 8.613763899823303e-07, "loss": 0.46, "step": 606 }, { "epoch": 0.2662134751384244, "grad_norm": 1.3175569772720337, "learning_rate": 8.608850313169355e-07, "loss": 0.3937, "step": 607 }, { "epoch": 0.26665204758511046, "grad_norm": 1.3802878856658936, "learning_rate": 8.603929440483713e-07, "loss": 0.4245, "step": 608 }, { "epoch": 0.2670906200317965, "grad_norm": 1.294541835784912, "learning_rate": 8.599001291701294e-07, "loss": 0.4257, "step": 609 }, { "epoch": 0.26752919247848256, "grad_norm": 1.3904340267181396, "learning_rate": 8.59406587677171e-07, "loss": 0.47, "step": 610 }, { "epoch": 0.26796776492516855, "grad_norm": 1.310609221458435, "learning_rate": 8.589123205659237e-07, "loss": 0.4693, "step": 611 }, { "epoch": 0.2684063373718546, "grad_norm": 1.508705496788025, "learning_rate": 8.584173288342806e-07, "loss": 0.4818, "step": 612 }, { "epoch": 0.26884490981854064, "grad_norm": 1.258613109588623, "learning_rate": 8.579216134815972e-07, "loss": 0.4391, "step": 613 }, { "epoch": 0.2692834822652267, "grad_norm": 1.309943437576294, "learning_rate": 8.574251755086905e-07, "loss": 0.4345, "step": 614 }, { "epoch": 0.26972205471191274, "grad_norm": 1.2155766487121582, "learning_rate": 8.569280159178358e-07, "loss": 0.4451, "step": 615 }, { "epoch": 0.2701606271585988, "grad_norm": 1.3451566696166992, "learning_rate": 8.564301357127662e-07, "loss": 0.446, "step": 616 }, { "epoch": 0.2705991996052848, "grad_norm": 1.581193447113037, "learning_rate": 8.559315358986684e-07, "loss": 0.455, "step": 617 }, { "epoch": 0.2710377720519708, "grad_norm": 1.3008910417556763, "learning_rate": 8.554322174821833e-07, "loss": 0.4209, "step": 618 }, { "epoch": 0.2714763444986569, "grad_norm": 1.2218880653381348, "learning_rate": 8.549321814714017e-07, "loss": 0.4627, "step": 619 }, { "epoch": 0.2719149169453429, "grad_norm": 1.2023916244506836, "learning_rate": 8.544314288758634e-07, "loss": 0.4502, "step": 620 }, { "epoch": 0.27235348939202897, "grad_norm": 1.3902664184570312, "learning_rate": 8.539299607065551e-07, "loss": 0.4435, "step": 621 }, { "epoch": 0.27279206183871496, "grad_norm": 1.35061776638031, "learning_rate": 8.534277779759081e-07, "loss": 0.4476, "step": 622 }, { "epoch": 0.273230634285401, "grad_norm": 1.2908838987350464, "learning_rate": 8.529248816977963e-07, "loss": 0.4829, "step": 623 }, { "epoch": 0.27366920673208706, "grad_norm": 1.2177608013153076, "learning_rate": 8.524212728875342e-07, "loss": 0.4346, "step": 624 }, { "epoch": 0.2741077791787731, "grad_norm": 1.3742519617080688, "learning_rate": 8.51916952561875e-07, "loss": 0.463, "step": 625 }, { "epoch": 0.27454635162545915, "grad_norm": 1.1771491765975952, "learning_rate": 8.514119217390084e-07, "loss": 0.4665, "step": 626 }, { "epoch": 0.27498492407214514, "grad_norm": 1.4156662225723267, "learning_rate": 8.509061814385581e-07, "loss": 0.5086, "step": 627 }, { "epoch": 0.2754234965188312, "grad_norm": 1.4472264051437378, "learning_rate": 8.503997326815811e-07, "loss": 0.4758, "step": 628 }, { "epoch": 0.27586206896551724, "grad_norm": 1.3086676597595215, "learning_rate": 8.498925764905635e-07, "loss": 0.4232, "step": 629 }, { "epoch": 0.2763006414122033, "grad_norm": 1.2598360776901245, "learning_rate": 8.493847138894208e-07, "loss": 0.4545, "step": 630 }, { "epoch": 0.27673921385888933, "grad_norm": 1.3427696228027344, "learning_rate": 8.488761459034941e-07, "loss": 0.4266, "step": 631 }, { "epoch": 0.2771777863055754, "grad_norm": 1.257368803024292, "learning_rate": 8.483668735595486e-07, "loss": 0.4527, "step": 632 }, { "epoch": 0.27761635875226137, "grad_norm": 1.3200676441192627, "learning_rate": 8.478568978857721e-07, "loss": 0.5008, "step": 633 }, { "epoch": 0.2780549311989474, "grad_norm": 1.3147386312484741, "learning_rate": 8.473462199117715e-07, "loss": 0.4622, "step": 634 }, { "epoch": 0.27849350364563347, "grad_norm": 1.241818904876709, "learning_rate": 8.468348406685724e-07, "loss": 0.4707, "step": 635 }, { "epoch": 0.2789320760923195, "grad_norm": 1.1343870162963867, "learning_rate": 8.463227611886157e-07, "loss": 0.4625, "step": 636 }, { "epoch": 0.27937064853900556, "grad_norm": 1.3791996240615845, "learning_rate": 8.458099825057565e-07, "loss": 0.4212, "step": 637 }, { "epoch": 0.27980922098569155, "grad_norm": 1.2124180793762207, "learning_rate": 8.45296505655261e-07, "loss": 0.4738, "step": 638 }, { "epoch": 0.2802477934323776, "grad_norm": 1.332551121711731, "learning_rate": 8.447823316738054e-07, "loss": 0.445, "step": 639 }, { "epoch": 0.28068636587906365, "grad_norm": 1.3134421110153198, "learning_rate": 8.442674615994731e-07, "loss": 0.4764, "step": 640 }, { "epoch": 0.2811249383257497, "grad_norm": 1.2080031633377075, "learning_rate": 8.43751896471753e-07, "loss": 0.4158, "step": 641 }, { "epoch": 0.28156351077243574, "grad_norm": 1.3157594203948975, "learning_rate": 8.432356373315371e-07, "loss": 0.4535, "step": 642 }, { "epoch": 0.28200208321912174, "grad_norm": 1.3137787580490112, "learning_rate": 8.427186852211188e-07, "loss": 0.4377, "step": 643 }, { "epoch": 0.2824406556658078, "grad_norm": 1.255462408065796, "learning_rate": 8.422010411841905e-07, "loss": 0.3872, "step": 644 }, { "epoch": 0.28287922811249383, "grad_norm": 1.292777180671692, "learning_rate": 8.416827062658415e-07, "loss": 0.4944, "step": 645 }, { "epoch": 0.2833178005591799, "grad_norm": 1.4436018466949463, "learning_rate": 8.411636815125558e-07, "loss": 0.4772, "step": 646 }, { "epoch": 0.2837563730058659, "grad_norm": 1.2209681272506714, "learning_rate": 8.406439679722104e-07, "loss": 0.4437, "step": 647 }, { "epoch": 0.2841949454525519, "grad_norm": 1.3344581127166748, "learning_rate": 8.401235666940727e-07, "loss": 0.4715, "step": 648 }, { "epoch": 0.28463351789923796, "grad_norm": 1.1497730016708374, "learning_rate": 8.396024787287988e-07, "loss": 0.4298, "step": 649 }, { "epoch": 0.285072090345924, "grad_norm": 1.341781735420227, "learning_rate": 8.390807051284309e-07, "loss": 0.4257, "step": 650 }, { "epoch": 0.28551066279261006, "grad_norm": 1.280103087425232, "learning_rate": 8.385582469463959e-07, "loss": 0.4368, "step": 651 }, { "epoch": 0.2859492352392961, "grad_norm": 1.2544479370117188, "learning_rate": 8.380351052375023e-07, "loss": 0.4723, "step": 652 }, { "epoch": 0.28638780768598215, "grad_norm": 1.1628601551055908, "learning_rate": 8.375112810579389e-07, "loss": 0.4515, "step": 653 }, { "epoch": 0.28682638013266815, "grad_norm": 1.3643077611923218, "learning_rate": 8.369867754652724e-07, "loss": 0.4235, "step": 654 }, { "epoch": 0.2872649525793542, "grad_norm": 1.205430030822754, "learning_rate": 8.36461589518445e-07, "loss": 0.4375, "step": 655 }, { "epoch": 0.28770352502604024, "grad_norm": 1.2593382596969604, "learning_rate": 8.359357242777728e-07, "loss": 0.4494, "step": 656 }, { "epoch": 0.2881420974727263, "grad_norm": 1.226905107498169, "learning_rate": 8.354091808049431e-07, "loss": 0.4476, "step": 657 }, { "epoch": 0.28858066991941234, "grad_norm": 1.2723194360733032, "learning_rate": 8.348819601630124e-07, "loss": 0.4706, "step": 658 }, { "epoch": 0.28901924236609833, "grad_norm": 1.2584813833236694, "learning_rate": 8.343540634164047e-07, "loss": 0.4518, "step": 659 }, { "epoch": 0.2894578148127844, "grad_norm": 1.19351327419281, "learning_rate": 8.338254916309089e-07, "loss": 0.3974, "step": 660 }, { "epoch": 0.2898963872594704, "grad_norm": 1.3015458583831787, "learning_rate": 8.332962458736765e-07, "loss": 0.451, "step": 661 }, { "epoch": 0.29033495970615647, "grad_norm": 1.206496000289917, "learning_rate": 8.327663272132202e-07, "loss": 0.4202, "step": 662 }, { "epoch": 0.2907735321528425, "grad_norm": 1.1564853191375732, "learning_rate": 8.322357367194108e-07, "loss": 0.4161, "step": 663 }, { "epoch": 0.2912121045995285, "grad_norm": 1.2696150541305542, "learning_rate": 8.317044754634756e-07, "loss": 0.4461, "step": 664 }, { "epoch": 0.29165067704621456, "grad_norm": 1.227945327758789, "learning_rate": 8.311725445179964e-07, "loss": 0.4303, "step": 665 }, { "epoch": 0.2920892494929006, "grad_norm": 1.2026177644729614, "learning_rate": 8.306399449569066e-07, "loss": 0.4329, "step": 666 }, { "epoch": 0.29252782193958665, "grad_norm": 1.330443263053894, "learning_rate": 8.301066778554897e-07, "loss": 0.4263, "step": 667 }, { "epoch": 0.2929663943862727, "grad_norm": 1.2969058752059937, "learning_rate": 8.295727442903772e-07, "loss": 0.4558, "step": 668 }, { "epoch": 0.29340496683295875, "grad_norm": 1.3471171855926514, "learning_rate": 8.290381453395457e-07, "loss": 0.4873, "step": 669 }, { "epoch": 0.29384353927964474, "grad_norm": 1.3648743629455566, "learning_rate": 8.285028820823153e-07, "loss": 0.4626, "step": 670 }, { "epoch": 0.2942821117263308, "grad_norm": 1.481045126914978, "learning_rate": 8.279669555993475e-07, "loss": 0.4924, "step": 671 }, { "epoch": 0.29472068417301683, "grad_norm": 1.1982755661010742, "learning_rate": 8.274303669726426e-07, "loss": 0.4365, "step": 672 }, { "epoch": 0.2951592566197029, "grad_norm": 1.2949172258377075, "learning_rate": 8.268931172855378e-07, "loss": 0.4929, "step": 673 }, { "epoch": 0.29559782906638893, "grad_norm": 1.2132493257522583, "learning_rate": 8.263552076227047e-07, "loss": 0.413, "step": 674 }, { "epoch": 0.2960364015130749, "grad_norm": 1.3627359867095947, "learning_rate": 8.258166390701481e-07, "loss": 0.486, "step": 675 }, { "epoch": 0.29647497395976097, "grad_norm": 1.3661818504333496, "learning_rate": 8.25277412715202e-07, "loss": 0.4704, "step": 676 }, { "epoch": 0.296913546406447, "grad_norm": 1.2108001708984375, "learning_rate": 8.247375296465293e-07, "loss": 0.402, "step": 677 }, { "epoch": 0.29735211885313306, "grad_norm": 1.4531514644622803, "learning_rate": 8.241969909541183e-07, "loss": 0.4896, "step": 678 }, { "epoch": 0.2977906912998191, "grad_norm": 1.3670005798339844, "learning_rate": 8.236557977292813e-07, "loss": 0.4511, "step": 679 }, { "epoch": 0.2982292637465051, "grad_norm": 1.4220718145370483, "learning_rate": 8.231139510646515e-07, "loss": 0.4765, "step": 680 }, { "epoch": 0.29866783619319115, "grad_norm": 1.2339059114456177, "learning_rate": 8.22571452054182e-07, "loss": 0.4748, "step": 681 }, { "epoch": 0.2991064086398772, "grad_norm": 1.2986462116241455, "learning_rate": 8.220283017931427e-07, "loss": 0.4738, "step": 682 }, { "epoch": 0.29954498108656324, "grad_norm": 1.2862370014190674, "learning_rate": 8.214845013781183e-07, "loss": 0.4658, "step": 683 }, { "epoch": 0.2999835535332493, "grad_norm": 1.3527040481567383, "learning_rate": 8.209400519070057e-07, "loss": 0.4905, "step": 684 }, { "epoch": 0.3004221259799353, "grad_norm": 1.2541968822479248, "learning_rate": 8.203949544790129e-07, "loss": 0.4487, "step": 685 }, { "epoch": 0.30086069842662133, "grad_norm": 1.3078125715255737, "learning_rate": 8.198492101946562e-07, "loss": 0.4253, "step": 686 }, { "epoch": 0.3012992708733074, "grad_norm": 1.210807204246521, "learning_rate": 8.193028201557567e-07, "loss": 0.414, "step": 687 }, { "epoch": 0.3017378433199934, "grad_norm": 1.3149670362472534, "learning_rate": 8.187557854654406e-07, "loss": 0.4341, "step": 688 }, { "epoch": 0.3021764157666795, "grad_norm": 1.2983028888702393, "learning_rate": 8.182081072281346e-07, "loss": 0.471, "step": 689 }, { "epoch": 0.3026149882133655, "grad_norm": 1.1206141710281372, "learning_rate": 8.176597865495653e-07, "loss": 0.4074, "step": 690 }, { "epoch": 0.3030535606600515, "grad_norm": 1.276363492012024, "learning_rate": 8.171108245367561e-07, "loss": 0.4624, "step": 691 }, { "epoch": 0.30349213310673756, "grad_norm": 1.3328988552093506, "learning_rate": 8.165612222980251e-07, "loss": 0.4394, "step": 692 }, { "epoch": 0.3039307055534236, "grad_norm": 1.4199296236038208, "learning_rate": 8.160109809429834e-07, "loss": 0.4394, "step": 693 }, { "epoch": 0.30436927800010966, "grad_norm": 1.363855004310608, "learning_rate": 8.154601015825318e-07, "loss": 0.4267, "step": 694 }, { "epoch": 0.3048078504467957, "grad_norm": 1.2294096946716309, "learning_rate": 8.149085853288597e-07, "loss": 0.4361, "step": 695 }, { "epoch": 0.3052464228934817, "grad_norm": 1.3705590963363647, "learning_rate": 8.143564332954425e-07, "loss": 0.465, "step": 696 }, { "epoch": 0.30568499534016774, "grad_norm": 1.2903246879577637, "learning_rate": 8.138036465970384e-07, "loss": 0.4469, "step": 697 }, { "epoch": 0.3061235677868538, "grad_norm": 1.2863341569900513, "learning_rate": 8.132502263496875e-07, "loss": 0.4226, "step": 698 }, { "epoch": 0.30656214023353984, "grad_norm": 1.333102822303772, "learning_rate": 8.126961736707091e-07, "loss": 0.4326, "step": 699 }, { "epoch": 0.3070007126802259, "grad_norm": 1.2394230365753174, "learning_rate": 8.121414896786991e-07, "loss": 0.4383, "step": 700 }, { "epoch": 0.3074392851269119, "grad_norm": 1.4252898693084717, "learning_rate": 8.115861754935279e-07, "loss": 0.4935, "step": 701 }, { "epoch": 0.3078778575735979, "grad_norm": 1.2032008171081543, "learning_rate": 8.110302322363387e-07, "loss": 0.4592, "step": 702 }, { "epoch": 0.30831643002028397, "grad_norm": 1.212705135345459, "learning_rate": 8.10473661029544e-07, "loss": 0.4354, "step": 703 }, { "epoch": 0.30875500246697, "grad_norm": 1.3495712280273438, "learning_rate": 8.099164629968247e-07, "loss": 0.4419, "step": 704 }, { "epoch": 0.30919357491365607, "grad_norm": 1.325252652168274, "learning_rate": 8.093586392631271e-07, "loss": 0.4354, "step": 705 }, { "epoch": 0.3096321473603421, "grad_norm": 1.2952181100845337, "learning_rate": 8.088001909546606e-07, "loss": 0.4495, "step": 706 }, { "epoch": 0.3100707198070281, "grad_norm": 1.4086304903030396, "learning_rate": 8.082411191988956e-07, "loss": 0.4763, "step": 707 }, { "epoch": 0.31050929225371415, "grad_norm": 1.5200958251953125, "learning_rate": 8.076814251245612e-07, "loss": 0.4554, "step": 708 }, { "epoch": 0.3109478647004002, "grad_norm": 1.2769299745559692, "learning_rate": 8.071211098616433e-07, "loss": 0.4089, "step": 709 }, { "epoch": 0.31138643714708625, "grad_norm": 1.294661283493042, "learning_rate": 8.06560174541381e-07, "loss": 0.5231, "step": 710 }, { "epoch": 0.3118250095937723, "grad_norm": 1.3333851099014282, "learning_rate": 8.059986202962666e-07, "loss": 0.4307, "step": 711 }, { "epoch": 0.3122635820404583, "grad_norm": 1.2841682434082031, "learning_rate": 8.054364482600405e-07, "loss": 0.4259, "step": 712 }, { "epoch": 0.31270215448714433, "grad_norm": 1.257988452911377, "learning_rate": 8.048736595676916e-07, "loss": 0.4293, "step": 713 }, { "epoch": 0.3131407269338304, "grad_norm": 1.2623865604400635, "learning_rate": 8.043102553554531e-07, "loss": 0.4311, "step": 714 }, { "epoch": 0.31357929938051643, "grad_norm": 1.398133635520935, "learning_rate": 8.037462367608012e-07, "loss": 0.4415, "step": 715 }, { "epoch": 0.3140178718272025, "grad_norm": 1.3084006309509277, "learning_rate": 8.031816049224523e-07, "loss": 0.4699, "step": 716 }, { "epoch": 0.31445644427388847, "grad_norm": 1.1205732822418213, "learning_rate": 8.026163609803611e-07, "loss": 0.4297, "step": 717 }, { "epoch": 0.3148950167205745, "grad_norm": 1.3080092668533325, "learning_rate": 8.020505060757178e-07, "loss": 0.3942, "step": 718 }, { "epoch": 0.31533358916726056, "grad_norm": 1.2918123006820679, "learning_rate": 8.014840413509464e-07, "loss": 0.4946, "step": 719 }, { "epoch": 0.3157721616139466, "grad_norm": 1.257453441619873, "learning_rate": 8.009169679497019e-07, "loss": 0.4411, "step": 720 } ], "logging_steps": 1.0, "max_steps": 2280, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 40, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.522795283589628e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }