{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 295, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003395585738539898, "grad_norm": 6.320343624704885, "learning_rate": 0.0, "loss": 0.9899, "step": 1 }, { "epoch": 0.006791171477079796, "grad_norm": 16.545756923443665, "learning_rate": 3.3333333333333335e-07, "loss": 0.9683, "step": 2 }, { "epoch": 0.010186757215619695, "grad_norm": 6.124626290117518, "learning_rate": 6.666666666666667e-07, "loss": 0.9722, "step": 3 }, { "epoch": 0.013582342954159592, "grad_norm": 6.907471369837716, "learning_rate": 1.0000000000000002e-06, "loss": 0.9667, "step": 4 }, { "epoch": 0.01697792869269949, "grad_norm": 6.275431523928528, "learning_rate": 1.3333333333333334e-06, "loss": 0.9814, "step": 5 }, { "epoch": 0.02037351443123939, "grad_norm": 6.074623009583982, "learning_rate": 1.6666666666666667e-06, "loss": 0.9502, "step": 6 }, { "epoch": 0.023769100169779286, "grad_norm": 89.99670872123082, "learning_rate": 2.0000000000000003e-06, "loss": 0.9411, "step": 7 }, { "epoch": 0.027164685908319185, "grad_norm": 6.224698090768632, "learning_rate": 2.3333333333333336e-06, "loss": 0.9075, "step": 8 }, { "epoch": 0.030560271646859084, "grad_norm": 5.913600661145363, "learning_rate": 2.666666666666667e-06, "loss": 0.8979, "step": 9 }, { "epoch": 0.03395585738539898, "grad_norm": 7.267004601688997, "learning_rate": 3e-06, "loss": 0.8589, "step": 10 }, { "epoch": 0.03735144312393888, "grad_norm": 4.920293108168373, "learning_rate": 3.3333333333333333e-06, "loss": 0.8797, "step": 11 }, { "epoch": 0.04074702886247878, "grad_norm": 7.0251120254603325, "learning_rate": 3.6666666666666666e-06, "loss": 0.785, "step": 12 }, { "epoch": 0.044142614601018676, "grad_norm": 4.620531235688851, "learning_rate": 4.000000000000001e-06, "loss": 0.7932, "step": 13 }, { "epoch": 0.04753820033955857, "grad_norm": 8.017042756220283, "learning_rate": 4.333333333333334e-06, "loss": 0.7878, "step": 14 }, { "epoch": 0.050933786078098474, "grad_norm": 7.168335735527789, "learning_rate": 4.666666666666667e-06, "loss": 0.7854, "step": 15 }, { "epoch": 0.05432937181663837, "grad_norm": 2.6102783506270137, "learning_rate": 5e-06, "loss": 0.7194, "step": 16 }, { "epoch": 0.057724957555178265, "grad_norm": 2.8281275726026074, "learning_rate": 5.333333333333334e-06, "loss": 0.7059, "step": 17 }, { "epoch": 0.06112054329371817, "grad_norm": 2.1330350735122505, "learning_rate": 5.666666666666667e-06, "loss": 0.7236, "step": 18 }, { "epoch": 0.06451612903225806, "grad_norm": 3.3828007198200694, "learning_rate": 6e-06, "loss": 0.7133, "step": 19 }, { "epoch": 0.06791171477079797, "grad_norm": 1.733407506222324, "learning_rate": 6.333333333333333e-06, "loss": 0.6807, "step": 20 }, { "epoch": 0.07130730050933787, "grad_norm": 1.5732888077219966, "learning_rate": 6.666666666666667e-06, "loss": 0.6717, "step": 21 }, { "epoch": 0.07470288624787776, "grad_norm": 2.869294784618442, "learning_rate": 7e-06, "loss": 0.6836, "step": 22 }, { "epoch": 0.07809847198641766, "grad_norm": 1.539403446338188, "learning_rate": 7.333333333333333e-06, "loss": 0.656, "step": 23 }, { "epoch": 0.08149405772495756, "grad_norm": 1.390711875015812, "learning_rate": 7.666666666666667e-06, "loss": 0.652, "step": 24 }, { "epoch": 0.08488964346349745, "grad_norm": 1.4133171062619616, "learning_rate": 8.000000000000001e-06, "loss": 0.6237, "step": 25 }, { "epoch": 0.08828522920203735, "grad_norm": 1.2675100082185478, "learning_rate": 8.333333333333334e-06, "loss": 0.5999, "step": 26 }, { "epoch": 0.09168081494057725, "grad_norm": 1.4010718655700412, "learning_rate": 8.666666666666668e-06, "loss": 0.6229, "step": 27 }, { "epoch": 0.09507640067911714, "grad_norm": 1.4302934299507057, "learning_rate": 9e-06, "loss": 0.5909, "step": 28 }, { "epoch": 0.09847198641765705, "grad_norm": 1.3417528783040296, "learning_rate": 9.333333333333334e-06, "loss": 0.6236, "step": 29 }, { "epoch": 0.10186757215619695, "grad_norm": 1.2550326498175997, "learning_rate": 9.666666666666667e-06, "loss": 0.5978, "step": 30 }, { "epoch": 0.10526315789473684, "grad_norm": 1.2949203936642646, "learning_rate": 1e-05, "loss": 0.6206, "step": 31 }, { "epoch": 0.10865874363327674, "grad_norm": 1.3404316606873072, "learning_rate": 9.999648647603774e-06, "loss": 0.6103, "step": 32 }, { "epoch": 0.11205432937181664, "grad_norm": 1.2464543142979883, "learning_rate": 9.998594639794502e-06, "loss": 0.5823, "step": 33 }, { "epoch": 0.11544991511035653, "grad_norm": 1.28881560865773, "learning_rate": 9.996838124703448e-06, "loss": 0.5724, "step": 34 }, { "epoch": 0.11884550084889643, "grad_norm": 1.236715524722473, "learning_rate": 9.994379349192927e-06, "loss": 0.6006, "step": 35 }, { "epoch": 0.12224108658743633, "grad_norm": 1.1748122882399419, "learning_rate": 9.991218658821609e-06, "loss": 0.5834, "step": 36 }, { "epoch": 0.12563667232597622, "grad_norm": 1.3335055567165142, "learning_rate": 9.987356497795944e-06, "loss": 0.5781, "step": 37 }, { "epoch": 0.12903225806451613, "grad_norm": 1.2573177388462913, "learning_rate": 9.982793408907747e-06, "loss": 0.5613, "step": 38 }, { "epoch": 0.13242784380305603, "grad_norm": 614.0974853168025, "learning_rate": 9.977530033457906e-06, "loss": 0.5518, "step": 39 }, { "epoch": 0.13582342954159593, "grad_norm": 1.8130801394796245, "learning_rate": 9.971567111166246e-06, "loss": 0.5875, "step": 40 }, { "epoch": 0.13921901528013583, "grad_norm": 3.021540974319598, "learning_rate": 9.964905480067585e-06, "loss": 0.5666, "step": 41 }, { "epoch": 0.14261460101867574, "grad_norm": 1.4307732415515269, "learning_rate": 9.957546076393944e-06, "loss": 0.5829, "step": 42 }, { "epoch": 0.1460101867572156, "grad_norm": 1.2948953962351142, "learning_rate": 9.949489934442966e-06, "loss": 0.592, "step": 43 }, { "epoch": 0.1494057724957555, "grad_norm": 1.28013431192408, "learning_rate": 9.940738186432565e-06, "loss": 0.554, "step": 44 }, { "epoch": 0.15280135823429541, "grad_norm": 1.364283730780204, "learning_rate": 9.931292062341793e-06, "loss": 0.5563, "step": 45 }, { "epoch": 0.15619694397283532, "grad_norm": 1.1789812327801241, "learning_rate": 9.921152889737985e-06, "loss": 0.5349, "step": 46 }, { "epoch": 0.15959252971137522, "grad_norm": 1.272995264216437, "learning_rate": 9.910322093590177e-06, "loss": 0.5885, "step": 47 }, { "epoch": 0.16298811544991512, "grad_norm": 1.2071797315958241, "learning_rate": 9.898801196068839e-06, "loss": 0.555, "step": 48 }, { "epoch": 0.166383701188455, "grad_norm": 1.2466300266575705, "learning_rate": 9.886591816331953e-06, "loss": 0.5412, "step": 49 }, { "epoch": 0.1697792869269949, "grad_norm": 1.1951025261533827, "learning_rate": 9.87369567029745e-06, "loss": 0.5266, "step": 50 }, { "epoch": 0.1731748726655348, "grad_norm": 1.1409578758956818, "learning_rate": 9.860114570402055e-06, "loss": 0.5364, "step": 51 }, { "epoch": 0.1765704584040747, "grad_norm": 1.194968225105999, "learning_rate": 9.845850425346563e-06, "loss": 0.544, "step": 52 }, { "epoch": 0.1799660441426146, "grad_norm": 1.105032206844106, "learning_rate": 9.830905239827592e-06, "loss": 0.526, "step": 53 }, { "epoch": 0.1833616298811545, "grad_norm": 1.0951143452700236, "learning_rate": 9.815281114255841e-06, "loss": 0.5384, "step": 54 }, { "epoch": 0.1867572156196944, "grad_norm": 1.237628011550819, "learning_rate": 9.798980244460892e-06, "loss": 0.5216, "step": 55 }, { "epoch": 0.19015280135823429, "grad_norm": 1.2049941993061364, "learning_rate": 9.782004921382612e-06, "loss": 0.5602, "step": 56 }, { "epoch": 0.1935483870967742, "grad_norm": 1.0963401467603866, "learning_rate": 9.764357530749178e-06, "loss": 0.5211, "step": 57 }, { "epoch": 0.1969439728353141, "grad_norm": 1.1153693044651702, "learning_rate": 9.74604055274178e-06, "loss": 0.5408, "step": 58 }, { "epoch": 0.200339558573854, "grad_norm": 1.1446372149192978, "learning_rate": 9.727056561646067e-06, "loss": 0.5297, "step": 59 }, { "epoch": 0.2037351443123939, "grad_norm": 1.1379654178567067, "learning_rate": 9.707408225490343e-06, "loss": 0.5241, "step": 60 }, { "epoch": 0.2071307300509338, "grad_norm": 1.1133344090884694, "learning_rate": 9.687098305670606e-06, "loss": 0.5373, "step": 61 }, { "epoch": 0.21052631578947367, "grad_norm": 1.14642700050245, "learning_rate": 9.66612965656245e-06, "loss": 0.5397, "step": 62 }, { "epoch": 0.21392190152801357, "grad_norm": 1.0251599768158084, "learning_rate": 9.644505225119922e-06, "loss": 0.5444, "step": 63 }, { "epoch": 0.21731748726655348, "grad_norm": 0.9890854380699226, "learning_rate": 9.622228050461345e-06, "loss": 0.4984, "step": 64 }, { "epoch": 0.22071307300509338, "grad_norm": 1.1117068387014335, "learning_rate": 9.599301263442194e-06, "loss": 0.5214, "step": 65 }, { "epoch": 0.22410865874363328, "grad_norm": 1.1785090789686485, "learning_rate": 9.575728086215093e-06, "loss": 0.5265, "step": 66 }, { "epoch": 0.22750424448217318, "grad_norm": 1.0565597242070186, "learning_rate": 9.551511831776966e-06, "loss": 0.5155, "step": 67 }, { "epoch": 0.23089983022071306, "grad_norm": 1.2278748412937952, "learning_rate": 9.526655903503423e-06, "loss": 0.5277, "step": 68 }, { "epoch": 0.23429541595925296, "grad_norm": 1.1213279864335683, "learning_rate": 9.501163794670445e-06, "loss": 0.5307, "step": 69 }, { "epoch": 0.23769100169779286, "grad_norm": 1.1237007845490015, "learning_rate": 9.475039087963443e-06, "loss": 0.51, "step": 70 }, { "epoch": 0.24108658743633277, "grad_norm": 1.0615774210564253, "learning_rate": 9.448285454973739e-06, "loss": 0.517, "step": 71 }, { "epoch": 0.24448217317487267, "grad_norm": 1.2104405507373044, "learning_rate": 9.420906655682553e-06, "loss": 0.5223, "step": 72 }, { "epoch": 0.24787775891341257, "grad_norm": 1.0985265075768098, "learning_rate": 9.392906537932582e-06, "loss": 0.5257, "step": 73 }, { "epoch": 0.25127334465195245, "grad_norm": 1.0626836434651132, "learning_rate": 9.364289036887214e-06, "loss": 0.5187, "step": 74 }, { "epoch": 0.2546689303904924, "grad_norm": 1.0234412989238388, "learning_rate": 9.335058174477472e-06, "loss": 0.5127, "step": 75 }, { "epoch": 0.25806451612903225, "grad_norm": 1.0702433549745685, "learning_rate": 9.305218058836778e-06, "loss": 0.4966, "step": 76 }, { "epoch": 0.2614601018675722, "grad_norm": 1.1228773632449267, "learning_rate": 9.274772883723587e-06, "loss": 0.5265, "step": 77 }, { "epoch": 0.26485568760611206, "grad_norm": 1.0076068781933833, "learning_rate": 9.24372692793199e-06, "loss": 0.5003, "step": 78 }, { "epoch": 0.26825127334465193, "grad_norm": 1.0987680518642855, "learning_rate": 9.21208455469037e-06, "loss": 0.4971, "step": 79 }, { "epoch": 0.27164685908319186, "grad_norm": 1.1051172285171345, "learning_rate": 9.179850211048193e-06, "loss": 0.491, "step": 80 }, { "epoch": 0.27504244482173174, "grad_norm": 1.0886761114445709, "learning_rate": 9.14702842725101e-06, "loss": 0.5033, "step": 81 }, { "epoch": 0.27843803056027167, "grad_norm": 1.101989710769037, "learning_rate": 9.113623816103775e-06, "loss": 0.5157, "step": 82 }, { "epoch": 0.28183361629881154, "grad_norm": 1.0846044794179348, "learning_rate": 9.079641072322555e-06, "loss": 0.5103, "step": 83 }, { "epoch": 0.28522920203735147, "grad_norm": 1.0994243701550208, "learning_rate": 9.045084971874738e-06, "loss": 0.5118, "step": 84 }, { "epoch": 0.28862478777589134, "grad_norm": 0.982679989364588, "learning_rate": 9.009960371307798e-06, "loss": 0.4753, "step": 85 }, { "epoch": 0.2920203735144312, "grad_norm": 1.0974557383394132, "learning_rate": 8.974272207066767e-06, "loss": 0.5065, "step": 86 }, { "epoch": 0.29541595925297115, "grad_norm": 1.191767026119928, "learning_rate": 8.938025494800454e-06, "loss": 0.496, "step": 87 }, { "epoch": 0.298811544991511, "grad_norm": 1.0863558052080193, "learning_rate": 8.901225328656543e-06, "loss": 0.4768, "step": 88 }, { "epoch": 0.30220713073005095, "grad_norm": 1.0292339490206497, "learning_rate": 8.863876880565656e-06, "loss": 0.4982, "step": 89 }, { "epoch": 0.30560271646859083, "grad_norm": 1.0263488228624327, "learning_rate": 8.825985399514488e-06, "loss": 0.49, "step": 90 }, { "epoch": 0.3089983022071307, "grad_norm": 1.1083934275101859, "learning_rate": 8.787556210808101e-06, "loss": 0.508, "step": 91 }, { "epoch": 0.31239388794567063, "grad_norm": 1.0759718773481644, "learning_rate": 8.748594715321512e-06, "loss": 0.493, "step": 92 }, { "epoch": 0.3157894736842105, "grad_norm": 1.1275176872041444, "learning_rate": 8.70910638874064e-06, "loss": 0.5063, "step": 93 }, { "epoch": 0.31918505942275044, "grad_norm": 1.0268180450930244, "learning_rate": 8.669096780792754e-06, "loss": 0.4953, "step": 94 }, { "epoch": 0.3225806451612903, "grad_norm": 1.0957503547012442, "learning_rate": 8.628571514466502e-06, "loss": 0.5062, "step": 95 }, { "epoch": 0.32597623089983024, "grad_norm": 1.0476160480506713, "learning_rate": 8.587536285221656e-06, "loss": 0.5254, "step": 96 }, { "epoch": 0.3293718166383701, "grad_norm": 0.973467408933069, "learning_rate": 8.545996860188668e-06, "loss": 0.5107, "step": 97 }, { "epoch": 0.33276740237691, "grad_norm": 0.9690640110532758, "learning_rate": 8.503959077358143e-06, "loss": 0.4767, "step": 98 }, { "epoch": 0.3361629881154499, "grad_norm": 1.1272513638936046, "learning_rate": 8.46142884476038e-06, "loss": 0.4991, "step": 99 }, { "epoch": 0.3395585738539898, "grad_norm": 1.089949180068211, "learning_rate": 8.418412139635026e-06, "loss": 0.5202, "step": 100 }, { "epoch": 0.34295415959252973, "grad_norm": 1.0904976650398113, "learning_rate": 8.374915007591053e-06, "loss": 0.5019, "step": 101 }, { "epoch": 0.3463497453310696, "grad_norm": 1.0132973869482662, "learning_rate": 8.330943561757092e-06, "loss": 0.4981, "step": 102 }, { "epoch": 0.34974533106960953, "grad_norm": 1.0144345934272716, "learning_rate": 8.286503981922284e-06, "loss": 0.4828, "step": 103 }, { "epoch": 0.3531409168081494, "grad_norm": 1.0708364693456955, "learning_rate": 8.241602513667775e-06, "loss": 0.4677, "step": 104 }, { "epoch": 0.3565365025466893, "grad_norm": 1.0734641865863697, "learning_rate": 8.19624546748895e-06, "loss": 0.4991, "step": 105 }, { "epoch": 0.3599320882852292, "grad_norm": 1.206730960165882, "learning_rate": 8.150439217908557e-06, "loss": 0.4874, "step": 106 }, { "epoch": 0.3633276740237691, "grad_norm": 0.990112721589732, "learning_rate": 8.104190202580811e-06, "loss": 0.4613, "step": 107 }, { "epoch": 0.366723259762309, "grad_norm": 1.135017537331381, "learning_rate": 8.057504921386661e-06, "loss": 0.4981, "step": 108 }, { "epoch": 0.3701188455008489, "grad_norm": 1.1168906124405507, "learning_rate": 8.010389935520269e-06, "loss": 0.4812, "step": 109 }, { "epoch": 0.3735144312393888, "grad_norm": 1.109365518761395, "learning_rate": 7.962851866566912e-06, "loss": 0.4901, "step": 110 }, { "epoch": 0.3769100169779287, "grad_norm": 1.1121750297406146, "learning_rate": 7.914897395572362e-06, "loss": 0.4841, "step": 111 }, { "epoch": 0.38030560271646857, "grad_norm": 1.10407244769018, "learning_rate": 7.866533262103937e-06, "loss": 0.4813, "step": 112 }, { "epoch": 0.3837011884550085, "grad_norm": 1.109502100982312, "learning_rate": 7.817766263303312e-06, "loss": 0.4731, "step": 113 }, { "epoch": 0.3870967741935484, "grad_norm": 1.0709239172717544, "learning_rate": 7.768603252931243e-06, "loss": 0.4988, "step": 114 }, { "epoch": 0.3904923599320883, "grad_norm": 1.0915174076138725, "learning_rate": 7.719051140404327e-06, "loss": 0.4825, "step": 115 }, { "epoch": 0.3938879456706282, "grad_norm": 1.0965896575304275, "learning_rate": 7.669116889823955e-06, "loss": 0.4742, "step": 116 }, { "epoch": 0.39728353140916806, "grad_norm": 1.0241307032424805, "learning_rate": 7.6188075189975644e-06, "loss": 0.4743, "step": 117 }, { "epoch": 0.400679117147708, "grad_norm": 0.9435780818732129, "learning_rate": 7.568130098452352e-06, "loss": 0.4739, "step": 118 }, { "epoch": 0.40407470288624786, "grad_norm": 1.0248298007190784, "learning_rate": 7.517091750441576e-06, "loss": 0.4814, "step": 119 }, { "epoch": 0.4074702886247878, "grad_norm": 1.0315243401737155, "learning_rate": 7.465699647943586e-06, "loss": 0.4805, "step": 120 }, { "epoch": 0.41086587436332767, "grad_norm": 1.091443614674752, "learning_rate": 7.413961013653725e-06, "loss": 0.4953, "step": 121 }, { "epoch": 0.4142614601018676, "grad_norm": 1.0680561484041602, "learning_rate": 7.361883118969248e-06, "loss": 0.4588, "step": 122 }, { "epoch": 0.41765704584040747, "grad_norm": 0.9761759048656751, "learning_rate": 7.309473282967387e-06, "loss": 0.469, "step": 123 }, { "epoch": 0.42105263157894735, "grad_norm": 1.0119186737160737, "learning_rate": 7.256738871376733e-06, "loss": 0.4812, "step": 124 }, { "epoch": 0.4244482173174873, "grad_norm": 0.9670302415957511, "learning_rate": 7.203687295542032e-06, "loss": 0.456, "step": 125 }, { "epoch": 0.42784380305602715, "grad_norm": 1.001107021910696, "learning_rate": 7.1503260113826035e-06, "loss": 0.4597, "step": 126 }, { "epoch": 0.4312393887945671, "grad_norm": 1.024857257795839, "learning_rate": 7.09666251834447e-06, "loss": 0.4884, "step": 127 }, { "epoch": 0.43463497453310695, "grad_norm": 1.0331435407774852, "learning_rate": 7.042704358346375e-06, "loss": 0.4581, "step": 128 }, { "epoch": 0.4380305602716469, "grad_norm": 1.00793619623988, "learning_rate": 6.988459114719849e-06, "loss": 0.4672, "step": 129 }, { "epoch": 0.44142614601018676, "grad_norm": 0.9145035341259203, "learning_rate": 6.933934411143419e-06, "loss": 0.4524, "step": 130 }, { "epoch": 0.44482173174872663, "grad_norm": 1.1118101624400085, "learning_rate": 6.879137910571191e-06, "loss": 0.4738, "step": 131 }, { "epoch": 0.44821731748726656, "grad_norm": 1.0044078413498474, "learning_rate": 6.824077314155877e-06, "loss": 0.4679, "step": 132 }, { "epoch": 0.45161290322580644, "grad_norm": 1.0568591065810937, "learning_rate": 6.768760360166471e-06, "loss": 0.4909, "step": 133 }, { "epoch": 0.45500848896434637, "grad_norm": 0.9937372079053678, "learning_rate": 6.713194822900707e-06, "loss": 0.4734, "step": 134 }, { "epoch": 0.45840407470288624, "grad_norm": 1.0380172124065203, "learning_rate": 6.657388511592453e-06, "loss": 0.4979, "step": 135 }, { "epoch": 0.4617996604414261, "grad_norm": 0.9710945454163669, "learning_rate": 6.601349269314188e-06, "loss": 0.4826, "step": 136 }, { "epoch": 0.46519524617996605, "grad_norm": 1.1195229655523649, "learning_rate": 6.545084971874738e-06, "loss": 0.4685, "step": 137 }, { "epoch": 0.4685908319185059, "grad_norm": 1.079871197310449, "learning_rate": 6.488603526712391e-06, "loss": 0.483, "step": 138 }, { "epoch": 0.47198641765704585, "grad_norm": 1.0089317127122743, "learning_rate": 6.431912871783587e-06, "loss": 0.5028, "step": 139 }, { "epoch": 0.47538200339558573, "grad_norm": 0.9671431137430636, "learning_rate": 6.3750209744473105e-06, "loss": 0.4669, "step": 140 }, { "epoch": 0.47877758913412566, "grad_norm": 1.0520557175009426, "learning_rate": 6.3179358303453386e-06, "loss": 0.4735, "step": 141 }, { "epoch": 0.48217317487266553, "grad_norm": 1.0538462837406155, "learning_rate": 6.260665462278544e-06, "loss": 0.5051, "step": 142 }, { "epoch": 0.4855687606112054, "grad_norm": 1.037965055623141, "learning_rate": 6.203217919079343e-06, "loss": 0.4762, "step": 143 }, { "epoch": 0.48896434634974534, "grad_norm": 1.0176275927771237, "learning_rate": 6.145601274480521e-06, "loss": 0.4869, "step": 144 }, { "epoch": 0.4923599320882852, "grad_norm": 1.0431340329887582, "learning_rate": 6.08782362598054e-06, "loss": 0.4611, "step": 145 }, { "epoch": 0.49575551782682514, "grad_norm": 1.0308571103490511, "learning_rate": 6.029893093705492e-06, "loss": 0.4763, "step": 146 }, { "epoch": 0.499151103565365, "grad_norm": 1.0355702334516321, "learning_rate": 5.971817819267914e-06, "loss": 0.4654, "step": 147 }, { "epoch": 0.5025466893039049, "grad_norm": 0.9893529623332218, "learning_rate": 5.9136059646225375e-06, "loss": 0.5044, "step": 148 }, { "epoch": 0.5059422750424448, "grad_norm": 0.960195438000855, "learning_rate": 5.855265710919211e-06, "loss": 0.4435, "step": 149 }, { "epoch": 0.5093378607809848, "grad_norm": 0.9646163806867624, "learning_rate": 5.796805257353109e-06, "loss": 0.4834, "step": 150 }, { "epoch": 0.5127334465195246, "grad_norm": 0.9931874900764169, "learning_rate": 5.738232820012407e-06, "loss": 0.461, "step": 151 }, { "epoch": 0.5161290322580645, "grad_norm": 1.0149629817842267, "learning_rate": 5.679556630723592e-06, "loss": 0.4611, "step": 152 }, { "epoch": 0.5195246179966044, "grad_norm": 1.0022805808479243, "learning_rate": 5.620784935894548e-06, "loss": 0.4712, "step": 153 }, { "epoch": 0.5229202037351444, "grad_norm": 0.9091219788824919, "learning_rate": 5.561925995355595e-06, "loss": 0.4906, "step": 154 }, { "epoch": 0.5263157894736842, "grad_norm": 0.9508691466403413, "learning_rate": 5.5029880811986546e-06, "loss": 0.4836, "step": 155 }, { "epoch": 0.5297113752122241, "grad_norm": 0.9816091460679524, "learning_rate": 5.443979476614674e-06, "loss": 0.4568, "step": 156 }, { "epoch": 0.533106960950764, "grad_norm": 0.9890714751673821, "learning_rate": 5.384908474729501e-06, "loss": 0.4669, "step": 157 }, { "epoch": 0.5365025466893039, "grad_norm": 0.9852345937629808, "learning_rate": 5.325783377438357e-06, "loss": 0.4631, "step": 158 }, { "epoch": 0.5398981324278438, "grad_norm": 0.9992258942689997, "learning_rate": 5.266612494239088e-06, "loss": 0.4824, "step": 159 }, { "epoch": 0.5432937181663837, "grad_norm": 1.0010167382867965, "learning_rate": 5.207404141064334e-06, "loss": 0.4619, "step": 160 }, { "epoch": 0.5466893039049237, "grad_norm": 1.117194827109753, "learning_rate": 5.148166639112799e-06, "loss": 0.518, "step": 161 }, { "epoch": 0.5500848896434635, "grad_norm": 0.9423156223035444, "learning_rate": 5.088908313679788e-06, "loss": 0.4567, "step": 162 }, { "epoch": 0.5534804753820034, "grad_norm": 0.919109798014551, "learning_rate": 5.029637492987153e-06, "loss": 0.4649, "step": 163 }, { "epoch": 0.5568760611205433, "grad_norm": 1.1232411363329386, "learning_rate": 4.970362507012848e-06, "loss": 0.4682, "step": 164 }, { "epoch": 0.5602716468590832, "grad_norm": 0.9560703825281516, "learning_rate": 4.911091686320213e-06, "loss": 0.4594, "step": 165 }, { "epoch": 0.5636672325976231, "grad_norm": 0.9927617945956294, "learning_rate": 4.8518333608872015e-06, "loss": 0.4266, "step": 166 }, { "epoch": 0.567062818336163, "grad_norm": 1.0053279255521879, "learning_rate": 4.792595858935668e-06, "loss": 0.4745, "step": 167 }, { "epoch": 0.5704584040747029, "grad_norm": 1.1247627665383018, "learning_rate": 4.733387505760913e-06, "loss": 0.4863, "step": 168 }, { "epoch": 0.5738539898132428, "grad_norm": 0.9506838521386066, "learning_rate": 4.674216622561645e-06, "loss": 0.4459, "step": 169 }, { "epoch": 0.5772495755517827, "grad_norm": 0.8918545360346245, "learning_rate": 4.6150915252705005e-06, "loss": 0.4571, "step": 170 }, { "epoch": 0.5806451612903226, "grad_norm": 0.9704497436094051, "learning_rate": 4.556020523385326e-06, "loss": 0.4701, "step": 171 }, { "epoch": 0.5840407470288624, "grad_norm": 0.9830084007827714, "learning_rate": 4.497011918801347e-06, "loss": 0.477, "step": 172 }, { "epoch": 0.5874363327674024, "grad_norm": 1.0446285595393363, "learning_rate": 4.438074004644407e-06, "loss": 0.4963, "step": 173 }, { "epoch": 0.5908319185059423, "grad_norm": 0.9430261659822053, "learning_rate": 4.379215064105454e-06, "loss": 0.4502, "step": 174 }, { "epoch": 0.5942275042444821, "grad_norm": 0.9528761706191039, "learning_rate": 4.32044336927641e-06, "loss": 0.4667, "step": 175 }, { "epoch": 0.597623089983022, "grad_norm": 0.9712945630496029, "learning_rate": 4.261767179987595e-06, "loss": 0.469, "step": 176 }, { "epoch": 0.601018675721562, "grad_norm": 0.9203712027671123, "learning_rate": 4.203194742646893e-06, "loss": 0.4549, "step": 177 }, { "epoch": 0.6044142614601019, "grad_norm": 0.9335750991253596, "learning_rate": 4.1447342890807905e-06, "loss": 0.4589, "step": 178 }, { "epoch": 0.6078098471986417, "grad_norm": 0.9302840956981335, "learning_rate": 4.086394035377463e-06, "loss": 0.4525, "step": 179 }, { "epoch": 0.6112054329371817, "grad_norm": 0.9205763667360544, "learning_rate": 4.028182180732088e-06, "loss": 0.4652, "step": 180 }, { "epoch": 0.6146010186757216, "grad_norm": 0.9053159533350853, "learning_rate": 3.970106906294509e-06, "loss": 0.4887, "step": 181 }, { "epoch": 0.6179966044142614, "grad_norm": 0.9639072424421392, "learning_rate": 3.912176374019462e-06, "loss": 0.4928, "step": 182 }, { "epoch": 0.6213921901528013, "grad_norm": 0.8934882136410526, "learning_rate": 3.85439872551948e-06, "loss": 0.4613, "step": 183 }, { "epoch": 0.6247877758913413, "grad_norm": 0.9142153731395757, "learning_rate": 3.796782080920659e-06, "loss": 0.4489, "step": 184 }, { "epoch": 0.6281833616298812, "grad_norm": 0.9197655133200423, "learning_rate": 3.7393345377214584e-06, "loss": 0.4433, "step": 185 }, { "epoch": 0.631578947368421, "grad_norm": 0.9569197839619695, "learning_rate": 3.682064169654663e-06, "loss": 0.4665, "step": 186 }, { "epoch": 0.634974533106961, "grad_norm": 0.9362494462879061, "learning_rate": 3.6249790255526916e-06, "loss": 0.4363, "step": 187 }, { "epoch": 0.6383701188455009, "grad_norm": 0.8671267136185908, "learning_rate": 3.568087128216414e-06, "loss": 0.4646, "step": 188 }, { "epoch": 0.6417657045840407, "grad_norm": 0.9203019838984255, "learning_rate": 3.511396473287611e-06, "loss": 0.4571, "step": 189 }, { "epoch": 0.6451612903225806, "grad_norm": 0.9768723974984408, "learning_rate": 3.4549150281252635e-06, "loss": 0.4484, "step": 190 }, { "epoch": 0.6485568760611206, "grad_norm": 0.9396504959111149, "learning_rate": 3.398650730685813e-06, "loss": 0.457, "step": 191 }, { "epoch": 0.6519524617996605, "grad_norm": 0.9336335579059931, "learning_rate": 3.3426114884075488e-06, "loss": 0.4467, "step": 192 }, { "epoch": 0.6553480475382003, "grad_norm": 0.8976643217426382, "learning_rate": 3.2868051770992935e-06, "loss": 0.4572, "step": 193 }, { "epoch": 0.6587436332767402, "grad_norm": 0.9373033603378723, "learning_rate": 3.2312396398335312e-06, "loss": 0.4687, "step": 194 }, { "epoch": 0.6621392190152802, "grad_norm": 0.9707294952319975, "learning_rate": 3.175922685844125e-06, "loss": 0.4336, "step": 195 }, { "epoch": 0.66553480475382, "grad_norm": 0.9583493306758707, "learning_rate": 3.1208620894288105e-06, "loss": 0.4486, "step": 196 }, { "epoch": 0.6689303904923599, "grad_norm": 0.9820021819293268, "learning_rate": 3.0660655888565827e-06, "loss": 0.4588, "step": 197 }, { "epoch": 0.6723259762308998, "grad_norm": 0.9400174461623624, "learning_rate": 3.0115408852801535e-06, "loss": 0.4423, "step": 198 }, { "epoch": 0.6757215619694398, "grad_norm": 0.9603934801371697, "learning_rate": 2.9572956416536267e-06, "loss": 0.444, "step": 199 }, { "epoch": 0.6791171477079796, "grad_norm": 0.9537870436181752, "learning_rate": 2.9033374816555338e-06, "loss": 0.4556, "step": 200 }, { "epoch": 0.6825127334465195, "grad_norm": 0.9149135094841083, "learning_rate": 2.8496739886173994e-06, "loss": 0.4727, "step": 201 }, { "epoch": 0.6859083191850595, "grad_norm": 0.9486316397589806, "learning_rate": 2.7963127044579697e-06, "loss": 0.432, "step": 202 }, { "epoch": 0.6893039049235993, "grad_norm": 0.9240748880967119, "learning_rate": 2.743261128623269e-06, "loss": 0.4596, "step": 203 }, { "epoch": 0.6926994906621392, "grad_norm": 0.9573847569611018, "learning_rate": 2.6905267170326143e-06, "loss": 0.4679, "step": 204 }, { "epoch": 0.6960950764006791, "grad_norm": 0.8845992056166951, "learning_rate": 2.6381168810307536e-06, "loss": 0.4631, "step": 205 }, { "epoch": 0.6994906621392191, "grad_norm": 1.004797015683845, "learning_rate": 2.5860389863462765e-06, "loss": 0.4475, "step": 206 }, { "epoch": 0.7028862478777589, "grad_norm": 0.9408081397458115, "learning_rate": 2.534300352056416e-06, "loss": 0.4529, "step": 207 }, { "epoch": 0.7062818336162988, "grad_norm": 0.8830371296945094, "learning_rate": 2.4829082495584244e-06, "loss": 0.4431, "step": 208 }, { "epoch": 0.7096774193548387, "grad_norm": 0.918591969840702, "learning_rate": 2.4318699015476495e-06, "loss": 0.4295, "step": 209 }, { "epoch": 0.7130730050933786, "grad_norm": 0.9126233293580099, "learning_rate": 2.3811924810024385e-06, "loss": 0.4462, "step": 210 }, { "epoch": 0.7164685908319185, "grad_norm": 0.944049132172904, "learning_rate": 2.330883110176049e-06, "loss": 0.4764, "step": 211 }, { "epoch": 0.7198641765704584, "grad_norm": 0.8594916774059891, "learning_rate": 2.2809488595956746e-06, "loss": 0.4245, "step": 212 }, { "epoch": 0.7232597623089984, "grad_norm": 0.9004237430035651, "learning_rate": 2.2313967470687593e-06, "loss": 0.4618, "step": 213 }, { "epoch": 0.7266553480475382, "grad_norm": 0.9087973453481795, "learning_rate": 2.18223373669669e-06, "loss": 0.4614, "step": 214 }, { "epoch": 0.7300509337860781, "grad_norm": 0.8919978483526617, "learning_rate": 2.1334667378960642e-06, "loss": 0.462, "step": 215 }, { "epoch": 0.733446519524618, "grad_norm": 0.9099211098465206, "learning_rate": 2.0851026044276405e-06, "loss": 0.4634, "step": 216 }, { "epoch": 0.7368421052631579, "grad_norm": 0.8896204484265512, "learning_rate": 2.0371481334330913e-06, "loss": 0.463, "step": 217 }, { "epoch": 0.7402376910016978, "grad_norm": 0.9519257493559129, "learning_rate": 1.9896100644797316e-06, "loss": 0.4788, "step": 218 }, { "epoch": 0.7436332767402377, "grad_norm": 0.9116718849117885, "learning_rate": 1.9424950786133414e-06, "loss": 0.4634, "step": 219 }, { "epoch": 0.7470288624787776, "grad_norm": 0.9263410885293661, "learning_rate": 1.8958097974191909e-06, "loss": 0.484, "step": 220 }, { "epoch": 0.7504244482173175, "grad_norm": 0.9023645875611066, "learning_rate": 1.8495607820914451e-06, "loss": 0.4442, "step": 221 }, { "epoch": 0.7538200339558574, "grad_norm": 0.8875288134755263, "learning_rate": 1.8037545325110506e-06, "loss": 0.4241, "step": 222 }, { "epoch": 0.7572156196943973, "grad_norm": 0.9432820849930534, "learning_rate": 1.7583974863322272e-06, "loss": 0.4837, "step": 223 }, { "epoch": 0.7606112054329371, "grad_norm": 0.9658199086993665, "learning_rate": 1.7134960180777171e-06, "loss": 0.4728, "step": 224 }, { "epoch": 0.7640067911714771, "grad_norm": 0.8457866400541412, "learning_rate": 1.6690564382429104e-06, "loss": 0.4301, "step": 225 }, { "epoch": 0.767402376910017, "grad_norm": 0.8989697214841547, "learning_rate": 1.6250849924089485e-06, "loss": 0.4527, "step": 226 }, { "epoch": 0.7707979626485568, "grad_norm": 0.9022700419018949, "learning_rate": 1.581587860364977e-06, "loss": 0.4543, "step": 227 }, { "epoch": 0.7741935483870968, "grad_norm": 0.9173241233268217, "learning_rate": 1.5385711552396227e-06, "loss": 0.4466, "step": 228 }, { "epoch": 0.7775891341256367, "grad_norm": 0.8617337826425147, "learning_rate": 1.4960409226418576e-06, "loss": 0.4386, "step": 229 }, { "epoch": 0.7809847198641766, "grad_norm": 0.8625616538339923, "learning_rate": 1.4540031398113335e-06, "loss": 0.4533, "step": 230 }, { "epoch": 0.7843803056027164, "grad_norm": 0.894048701617508, "learning_rate": 1.4124637147783431e-06, "loss": 0.4447, "step": 231 }, { "epoch": 0.7877758913412564, "grad_norm": 0.9208844935282854, "learning_rate": 1.371428485533498e-06, "loss": 0.4826, "step": 232 }, { "epoch": 0.7911714770797963, "grad_norm": 0.8883039726822113, "learning_rate": 1.3309032192072463e-06, "loss": 0.4486, "step": 233 }, { "epoch": 0.7945670628183361, "grad_norm": 0.899703690304073, "learning_rate": 1.29089361125936e-06, "loss": 0.4396, "step": 234 }, { "epoch": 0.797962648556876, "grad_norm": 0.942126628969688, "learning_rate": 1.251405284678488e-06, "loss": 0.4666, "step": 235 }, { "epoch": 0.801358234295416, "grad_norm": 0.876574882692073, "learning_rate": 1.2124437891918995e-06, "loss": 0.4338, "step": 236 }, { "epoch": 0.8047538200339559, "grad_norm": 0.8738541998828565, "learning_rate": 1.1740146004855141e-06, "loss": 0.4609, "step": 237 }, { "epoch": 0.8081494057724957, "grad_norm": 0.8732256896169818, "learning_rate": 1.1361231194343436e-06, "loss": 0.4505, "step": 238 }, { "epoch": 0.8115449915110357, "grad_norm": 0.8634374293174691, "learning_rate": 1.0987746713434578e-06, "loss": 0.4626, "step": 239 }, { "epoch": 0.8149405772495756, "grad_norm": 0.8709360064531477, "learning_rate": 1.0619745051995473e-06, "loss": 0.465, "step": 240 }, { "epoch": 0.8183361629881154, "grad_norm": 0.8612129831209381, "learning_rate": 1.0257277929332332e-06, "loss": 0.4543, "step": 241 }, { "epoch": 0.8217317487266553, "grad_norm": 0.9089176413771565, "learning_rate": 9.900396286922025e-07, "loss": 0.4676, "step": 242 }, { "epoch": 0.8251273344651953, "grad_norm": 0.8856710278778406, "learning_rate": 9.549150281252633e-07, "loss": 0.4366, "step": 243 }, { "epoch": 0.8285229202037352, "grad_norm": 0.8188623584178454, "learning_rate": 9.203589276774438e-07, "loss": 0.4404, "step": 244 }, { "epoch": 0.831918505942275, "grad_norm": 0.9252917982729971, "learning_rate": 8.86376183896226e-07, "loss": 0.4447, "step": 245 }, { "epoch": 0.8353140916808149, "grad_norm": 0.8784423429363009, "learning_rate": 8.529715727489912e-07, "loss": 0.4488, "step": 246 }, { "epoch": 0.8387096774193549, "grad_norm": 0.9207047553546756, "learning_rate": 8.201497889518073e-07, "loss": 0.4561, "step": 247 }, { "epoch": 0.8421052631578947, "grad_norm": 0.886908726068121, "learning_rate": 7.879154453096305e-07, "loss": 0.4556, "step": 248 }, { "epoch": 0.8455008488964346, "grad_norm": 0.8585705366949494, "learning_rate": 7.562730720680111e-07, "loss": 0.4333, "step": 249 }, { "epoch": 0.8488964346349746, "grad_norm": 0.8504815527006262, "learning_rate": 7.25227116276413e-07, "loss": 0.4512, "step": 250 }, { "epoch": 0.8522920203735145, "grad_norm": 0.8201572218259242, "learning_rate": 6.947819411632223e-07, "loss": 0.4492, "step": 251 }, { "epoch": 0.8556876061120543, "grad_norm": 0.8579475897548395, "learning_rate": 6.649418255225298e-07, "loss": 0.4633, "step": 252 }, { "epoch": 0.8590831918505942, "grad_norm": 0.8140416489339838, "learning_rate": 6.357109631127889e-07, "loss": 0.4418, "step": 253 }, { "epoch": 0.8624787775891342, "grad_norm": 0.8552881051968678, "learning_rate": 6.07093462067419e-07, "loss": 0.4572, "step": 254 }, { "epoch": 0.865874363327674, "grad_norm": 0.8971540169404577, "learning_rate": 5.79093344317449e-07, "loss": 0.4775, "step": 255 }, { "epoch": 0.8692699490662139, "grad_norm": 0.8058254068183305, "learning_rate": 5.517145450262639e-07, "loss": 0.4313, "step": 256 }, { "epoch": 0.8726655348047538, "grad_norm": 0.8530487589073922, "learning_rate": 5.249609120365579e-07, "loss": 0.4397, "step": 257 }, { "epoch": 0.8760611205432938, "grad_norm": 0.9429999718413201, "learning_rate": 4.988362053295564e-07, "loss": 0.4496, "step": 258 }, { "epoch": 0.8794567062818336, "grad_norm": 0.8375356078246441, "learning_rate": 4.733440964965791e-07, "loss": 0.4303, "step": 259 }, { "epoch": 0.8828522920203735, "grad_norm": 0.8256703834683652, "learning_rate": 4.484881682230341e-07, "loss": 0.4268, "step": 260 }, { "epoch": 0.8862478777589134, "grad_norm": 0.8543078561007811, "learning_rate": 4.242719137849077e-07, "loss": 0.4514, "step": 261 }, { "epoch": 0.8896434634974533, "grad_norm": 0.905414358204414, "learning_rate": 4.00698736557808e-07, "loss": 0.4475, "step": 262 }, { "epoch": 0.8930390492359932, "grad_norm": 0.8479836834072867, "learning_rate": 3.777719495386567e-07, "loss": 0.4479, "step": 263 }, { "epoch": 0.8964346349745331, "grad_norm": 0.8646327130622585, "learning_rate": 3.5549477488007853e-07, "loss": 0.4563, "step": 264 }, { "epoch": 0.8998302207130731, "grad_norm": 0.8363483102479007, "learning_rate": 3.3387034343755063e-07, "loss": 0.4447, "step": 265 }, { "epoch": 0.9032258064516129, "grad_norm": 0.8772923486186023, "learning_rate": 3.1290169432939556e-07, "loss": 0.4416, "step": 266 }, { "epoch": 0.9066213921901528, "grad_norm": 0.8320996733730202, "learning_rate": 2.925917745096568e-07, "loss": 0.4263, "step": 267 }, { "epoch": 0.9100169779286927, "grad_norm": 0.8735391402072182, "learning_rate": 2.7294343835393366e-07, "loss": 0.4635, "step": 268 }, { "epoch": 0.9134125636672326, "grad_norm": 0.8476560390812548, "learning_rate": 2.539594472582213e-07, "loss": 0.4387, "step": 269 }, { "epoch": 0.9168081494057725, "grad_norm": 0.8265472830061668, "learning_rate": 2.3564246925082358e-07, "loss": 0.4409, "step": 270 }, { "epoch": 0.9202037351443124, "grad_norm": 0.8275085794146296, "learning_rate": 2.179950786173879e-07, "loss": 0.422, "step": 271 }, { "epoch": 0.9235993208828522, "grad_norm": 0.8378368477541204, "learning_rate": 2.01019755539108e-07, "loss": 0.4533, "step": 272 }, { "epoch": 0.9269949066213922, "grad_norm": 0.8225280576542648, "learning_rate": 1.8471888574415953e-07, "loss": 0.4504, "step": 273 }, { "epoch": 0.9303904923599321, "grad_norm": 0.8689138287376853, "learning_rate": 1.690947601724091e-07, "loss": 0.4616, "step": 274 }, { "epoch": 0.933786078098472, "grad_norm": 0.8153963652578823, "learning_rate": 1.5414957465343883e-07, "loss": 0.4362, "step": 275 }, { "epoch": 0.9371816638370118, "grad_norm": 0.7970051219455544, "learning_rate": 1.3988542959794627e-07, "loss": 0.4476, "step": 276 }, { "epoch": 0.9405772495755518, "grad_norm": 0.8286263852975556, "learning_rate": 1.2630432970255014e-07, "loss": 0.4317, "step": 277 }, { "epoch": 0.9439728353140917, "grad_norm": 0.819703548732724, "learning_rate": 1.1340818366804728e-07, "loss": 0.4374, "step": 278 }, { "epoch": 0.9473684210526315, "grad_norm": 0.8630335285550543, "learning_rate": 1.0119880393116177e-07, "loss": 0.4463, "step": 279 }, { "epoch": 0.9507640067911715, "grad_norm": 0.8905416016998572, "learning_rate": 8.967790640982466e-08, "loss": 0.4534, "step": 280 }, { "epoch": 0.9541595925297114, "grad_norm": 0.869553726844365, "learning_rate": 7.884711026201586e-08, "loss": 0.4474, "step": 281 }, { "epoch": 0.9575551782682513, "grad_norm": 0.8516101851144577, "learning_rate": 6.870793765820783e-08, "loss": 0.4735, "step": 282 }, { "epoch": 0.9609507640067911, "grad_norm": 0.834166916764426, "learning_rate": 5.92618135674361e-08, "loss": 0.4506, "step": 283 }, { "epoch": 0.9643463497453311, "grad_norm": 0.8500281237762543, "learning_rate": 5.0510065557034526e-08, "loss": 0.4677, "step": 284 }, { "epoch": 0.967741935483871, "grad_norm": 0.833367943844974, "learning_rate": 4.245392360605727e-08, "loss": 0.4582, "step": 285 }, { "epoch": 0.9711375212224108, "grad_norm": 0.8458107328647104, "learning_rate": 3.5094519932415417e-08, "loss": 0.4505, "step": 286 }, { "epoch": 0.9745331069609507, "grad_norm": 0.8443895579435003, "learning_rate": 2.843288883375539e-08, "loss": 0.4538, "step": 287 }, { "epoch": 0.9779286926994907, "grad_norm": 0.8631958241337926, "learning_rate": 2.2469966542096323e-08, "loss": 0.4432, "step": 288 }, { "epoch": 0.9813242784380306, "grad_norm": 0.886260366562854, "learning_rate": 1.7206591092253642e-08, "loss": 0.4567, "step": 289 }, { "epoch": 0.9847198641765704, "grad_norm": 0.7994144470255627, "learning_rate": 1.264350220405719e-08, "loss": 0.4317, "step": 290 }, { "epoch": 0.9881154499151104, "grad_norm": 0.8700681537853684, "learning_rate": 8.781341178393244e-09, "loss": 0.4555, "step": 291 }, { "epoch": 0.9915110356536503, "grad_norm": 0.8575898812488901, "learning_rate": 5.620650807073857e-09, "loss": 0.4606, "step": 292 }, { "epoch": 0.9949066213921901, "grad_norm": 0.8616135380601608, "learning_rate": 3.1618752965534295e-09, "loss": 0.4408, "step": 293 }, { "epoch": 0.99830220713073, "grad_norm": 0.7784658740487929, "learning_rate": 1.4053602054991954e-09, "loss": 0.4157, "step": 294 }, { "epoch": 1.0, "grad_norm": 1.2063313976605035, "learning_rate": 3.513523962256349e-10, "loss": 0.4174, "step": 295 }, { "epoch": 1.0, "step": 295, "total_flos": 44242544295936.0, "train_loss": 0.5097209928399425, "train_runtime": 1256.4644, "train_samples_per_second": 14.995, "train_steps_per_second": 0.235 } ], "logging_steps": 1, "max_steps": 295, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 2500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 44242544295936.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }