{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 11035, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00045310376076121433, "grad_norm": 2.7540205907021864, "learning_rate": 7.246376811594204e-08, "loss": 1.4967, "step": 1 }, { "epoch": 0.0009062075215224287, "grad_norm": 2.759771915863116, "learning_rate": 1.4492753623188408e-07, "loss": 1.5085, "step": 2 }, { "epoch": 0.001359311282283643, "grad_norm": 2.791345441088159, "learning_rate": 2.173913043478261e-07, "loss": 1.5106, "step": 3 }, { "epoch": 0.0018124150430448573, "grad_norm": 2.7274320302763213, "learning_rate": 2.8985507246376816e-07, "loss": 1.4995, "step": 4 }, { "epoch": 0.0022655188038060714, "grad_norm": 2.762087809898038, "learning_rate": 3.623188405797102e-07, "loss": 1.483, "step": 5 }, { "epoch": 0.002718622564567286, "grad_norm": 2.7526691372288643, "learning_rate": 4.347826086956522e-07, "loss": 1.4829, "step": 6 }, { "epoch": 0.0031717263253285004, "grad_norm": 2.72824006419183, "learning_rate": 5.072463768115942e-07, "loss": 1.479, "step": 7 }, { "epoch": 0.0036248300860897147, "grad_norm": 2.7345760870851663, "learning_rate": 5.797101449275363e-07, "loss": 1.488, "step": 8 }, { "epoch": 0.0040779338468509285, "grad_norm": 2.703812574753596, "learning_rate": 6.521739130434783e-07, "loss": 1.4798, "step": 9 }, { "epoch": 0.004531037607612143, "grad_norm": 2.688806841956623, "learning_rate": 7.246376811594204e-07, "loss": 1.5099, "step": 10 }, { "epoch": 0.004984141368373357, "grad_norm": 2.6708612481715344, "learning_rate": 7.971014492753623e-07, "loss": 1.509, "step": 11 }, { "epoch": 0.005437245129134572, "grad_norm": 2.4815070820742644, "learning_rate": 8.695652173913044e-07, "loss": 1.4925, "step": 12 }, { "epoch": 0.0058903488898957865, "grad_norm": 2.4653158595323883, "learning_rate": 9.420289855072465e-07, "loss": 1.4958, "step": 13 }, { "epoch": 0.006343452650657001, "grad_norm": 2.448798219876715, "learning_rate": 1.0144927536231885e-06, "loss": 1.4903, "step": 14 }, { "epoch": 0.006796556411418215, "grad_norm": 2.420101451262591, "learning_rate": 1.0869565217391306e-06, "loss": 1.5174, "step": 15 }, { "epoch": 0.007249660172179429, "grad_norm": 2.350280309507433, "learning_rate": 1.1594202898550726e-06, "loss": 1.5039, "step": 16 }, { "epoch": 0.007702763932940644, "grad_norm": 1.903784507978735, "learning_rate": 1.2318840579710147e-06, "loss": 1.4663, "step": 17 }, { "epoch": 0.008155867693701857, "grad_norm": 1.8706106514184024, "learning_rate": 1.3043478260869566e-06, "loss": 1.466, "step": 18 }, { "epoch": 0.008608971454463071, "grad_norm": 1.880738364157002, "learning_rate": 1.3768115942028987e-06, "loss": 1.4796, "step": 19 }, { "epoch": 0.009062075215224286, "grad_norm": 1.7651640288977044, "learning_rate": 1.4492753623188408e-06, "loss": 1.4649, "step": 20 }, { "epoch": 0.0095151789759855, "grad_norm": 1.7860064661443993, "learning_rate": 1.521739130434783e-06, "loss": 1.4804, "step": 21 }, { "epoch": 0.009968282736746714, "grad_norm": 1.4361778423456308, "learning_rate": 1.5942028985507246e-06, "loss": 1.4456, "step": 22 }, { "epoch": 0.01042138649750793, "grad_norm": 1.3957509384183253, "learning_rate": 1.6666666666666667e-06, "loss": 1.4468, "step": 23 }, { "epoch": 0.010874490258269144, "grad_norm": 1.5249634592562644, "learning_rate": 1.7391304347826088e-06, "loss": 1.4575, "step": 24 }, { "epoch": 0.011327594019030359, "grad_norm": 1.540130373062518, "learning_rate": 1.8115942028985508e-06, "loss": 1.4395, "step": 25 }, { "epoch": 0.011780697779791573, "grad_norm": 1.4497783004478555, "learning_rate": 1.884057971014493e-06, "loss": 1.4399, "step": 26 }, { "epoch": 0.012233801540552787, "grad_norm": 1.323097240615225, "learning_rate": 1.956521739130435e-06, "loss": 1.444, "step": 27 }, { "epoch": 0.012686905301314002, "grad_norm": 1.2145025717830857, "learning_rate": 2.028985507246377e-06, "loss": 1.4206, "step": 28 }, { "epoch": 0.013140009062075216, "grad_norm": 1.0467110920881306, "learning_rate": 2.101449275362319e-06, "loss": 1.4292, "step": 29 }, { "epoch": 0.01359311282283643, "grad_norm": 1.219831363334628, "learning_rate": 2.173913043478261e-06, "loss": 1.3991, "step": 30 }, { "epoch": 0.014046216583597644, "grad_norm": 1.2255311619381022, "learning_rate": 2.246376811594203e-06, "loss": 1.4075, "step": 31 }, { "epoch": 0.014499320344358859, "grad_norm": 1.1427785073594934, "learning_rate": 2.3188405797101453e-06, "loss": 1.3818, "step": 32 }, { "epoch": 0.014952424105120073, "grad_norm": 1.0434587069188892, "learning_rate": 2.391304347826087e-06, "loss": 1.3966, "step": 33 }, { "epoch": 0.015405527865881287, "grad_norm": 1.0133669866162653, "learning_rate": 2.4637681159420295e-06, "loss": 1.3805, "step": 34 }, { "epoch": 0.0158586316266425, "grad_norm": 0.9784498150712031, "learning_rate": 2.5362318840579714e-06, "loss": 1.3963, "step": 35 }, { "epoch": 0.016311735387403714, "grad_norm": 0.9145284627121728, "learning_rate": 2.6086956521739132e-06, "loss": 1.3977, "step": 36 }, { "epoch": 0.01676483914816493, "grad_norm": 0.8201512022927249, "learning_rate": 2.6811594202898555e-06, "loss": 1.3573, "step": 37 }, { "epoch": 0.017217942908926143, "grad_norm": 0.738233456808168, "learning_rate": 2.7536231884057974e-06, "loss": 1.3878, "step": 38 }, { "epoch": 0.017671046669687357, "grad_norm": 0.6544541118238848, "learning_rate": 2.8260869565217393e-06, "loss": 1.3924, "step": 39 }, { "epoch": 0.01812415043044857, "grad_norm": 0.6022226727466975, "learning_rate": 2.8985507246376816e-06, "loss": 1.3553, "step": 40 }, { "epoch": 0.018577254191209785, "grad_norm": 0.7219617579121242, "learning_rate": 2.9710144927536235e-06, "loss": 1.3491, "step": 41 }, { "epoch": 0.019030357951971, "grad_norm": 0.8146989935871364, "learning_rate": 3.043478260869566e-06, "loss": 1.3367, "step": 42 }, { "epoch": 0.019483461712732214, "grad_norm": 0.7772284964843019, "learning_rate": 3.1159420289855073e-06, "loss": 1.3338, "step": 43 }, { "epoch": 0.01993656547349343, "grad_norm": 0.6794981730842932, "learning_rate": 3.188405797101449e-06, "loss": 1.3437, "step": 44 }, { "epoch": 0.020389669234254643, "grad_norm": 0.5632646719048948, "learning_rate": 3.2608695652173914e-06, "loss": 1.3291, "step": 45 }, { "epoch": 0.02084277299501586, "grad_norm": 0.49851233310861054, "learning_rate": 3.3333333333333333e-06, "loss": 1.3259, "step": 46 }, { "epoch": 0.021295876755777075, "grad_norm": 0.49908934546191075, "learning_rate": 3.4057971014492756e-06, "loss": 1.3237, "step": 47 }, { "epoch": 0.02174898051653829, "grad_norm": 0.5703544008655084, "learning_rate": 3.4782608695652175e-06, "loss": 1.3318, "step": 48 }, { "epoch": 0.022202084277299503, "grad_norm": 0.5771950937765048, "learning_rate": 3.55072463768116e-06, "loss": 1.3063, "step": 49 }, { "epoch": 0.022655188038060717, "grad_norm": 0.556006083426687, "learning_rate": 3.6231884057971017e-06, "loss": 1.3108, "step": 50 }, { "epoch": 0.023108291798821932, "grad_norm": 0.5060360265978124, "learning_rate": 3.6956521739130436e-06, "loss": 1.3098, "step": 51 }, { "epoch": 0.023561395559583146, "grad_norm": 0.45293715817843766, "learning_rate": 3.768115942028986e-06, "loss": 1.3227, "step": 52 }, { "epoch": 0.02401449932034436, "grad_norm": 0.4016194444846861, "learning_rate": 3.840579710144928e-06, "loss": 1.3321, "step": 53 }, { "epoch": 0.024467603081105575, "grad_norm": 0.3475388838071568, "learning_rate": 3.91304347826087e-06, "loss": 1.3026, "step": 54 }, { "epoch": 0.02492070684186679, "grad_norm": 0.3475690603016617, "learning_rate": 3.9855072463768115e-06, "loss": 1.294, "step": 55 }, { "epoch": 0.025373810602628003, "grad_norm": 0.36117019352138463, "learning_rate": 4.057971014492754e-06, "loss": 1.299, "step": 56 }, { "epoch": 0.025826914363389217, "grad_norm": 0.38113709429149534, "learning_rate": 4.130434782608696e-06, "loss": 1.3133, "step": 57 }, { "epoch": 0.02628001812415043, "grad_norm": 0.38760031717008975, "learning_rate": 4.202898550724638e-06, "loss": 1.2899, "step": 58 }, { "epoch": 0.026733121884911646, "grad_norm": 0.3624568286521374, "learning_rate": 4.27536231884058e-06, "loss": 1.2955, "step": 59 }, { "epoch": 0.02718622564567286, "grad_norm": 0.29653785730100624, "learning_rate": 4.347826086956522e-06, "loss": 1.3141, "step": 60 }, { "epoch": 0.027639329406434075, "grad_norm": 0.286304561008758, "learning_rate": 4.4202898550724645e-06, "loss": 1.2769, "step": 61 }, { "epoch": 0.02809243316719529, "grad_norm": 0.29924558928870865, "learning_rate": 4.492753623188406e-06, "loss": 1.2909, "step": 62 }, { "epoch": 0.028545536927956503, "grad_norm": 0.30831976265759875, "learning_rate": 4.565217391304348e-06, "loss": 1.314, "step": 63 }, { "epoch": 0.028998640688717717, "grad_norm": 0.28290196230704007, "learning_rate": 4.637681159420291e-06, "loss": 1.302, "step": 64 }, { "epoch": 0.02945174444947893, "grad_norm": 0.26105127121291155, "learning_rate": 4.710144927536232e-06, "loss": 1.2945, "step": 65 }, { "epoch": 0.029904848210240146, "grad_norm": 0.2427615554827512, "learning_rate": 4.782608695652174e-06, "loss": 1.3198, "step": 66 }, { "epoch": 0.03035795197100136, "grad_norm": 0.23004745852888506, "learning_rate": 4.855072463768117e-06, "loss": 1.2748, "step": 67 }, { "epoch": 0.030811055731762575, "grad_norm": 0.20566297710607717, "learning_rate": 4.927536231884059e-06, "loss": 1.2582, "step": 68 }, { "epoch": 0.031264159492523785, "grad_norm": 0.21692893542678926, "learning_rate": 5e-06, "loss": 1.2504, "step": 69 }, { "epoch": 0.031717263253285, "grad_norm": 0.22697604738660843, "learning_rate": 5.072463768115943e-06, "loss": 1.2813, "step": 70 }, { "epoch": 0.032170367014046214, "grad_norm": 0.20765832361745915, "learning_rate": 5.144927536231884e-06, "loss": 1.245, "step": 71 }, { "epoch": 0.03262347077480743, "grad_norm": 0.19079707717290276, "learning_rate": 5.2173913043478265e-06, "loss": 1.2559, "step": 72 }, { "epoch": 0.03307657453556864, "grad_norm": 0.19991003455950862, "learning_rate": 5.289855072463769e-06, "loss": 1.2427, "step": 73 }, { "epoch": 0.03352967829632986, "grad_norm": 0.20925895476283646, "learning_rate": 5.362318840579711e-06, "loss": 1.2508, "step": 74 }, { "epoch": 0.03398278205709107, "grad_norm": 0.21100175214238714, "learning_rate": 5.4347826086956525e-06, "loss": 1.2562, "step": 75 }, { "epoch": 0.034435885817852285, "grad_norm": 0.18205269181611544, "learning_rate": 5.507246376811595e-06, "loss": 1.2457, "step": 76 }, { "epoch": 0.0348889895786135, "grad_norm": 0.1789724748650181, "learning_rate": 5.579710144927537e-06, "loss": 1.2627, "step": 77 }, { "epoch": 0.035342093339374714, "grad_norm": 0.18501437834983528, "learning_rate": 5.652173913043479e-06, "loss": 1.2577, "step": 78 }, { "epoch": 0.03579519710013593, "grad_norm": 0.1585022725852172, "learning_rate": 5.724637681159421e-06, "loss": 1.2402, "step": 79 }, { "epoch": 0.03624830086089714, "grad_norm": 0.1588188148616759, "learning_rate": 5.797101449275363e-06, "loss": 1.2536, "step": 80 }, { "epoch": 0.03670140462165836, "grad_norm": 0.16385352844592405, "learning_rate": 5.8695652173913055e-06, "loss": 1.2617, "step": 81 }, { "epoch": 0.03715450838241957, "grad_norm": 0.14753077526654598, "learning_rate": 5.942028985507247e-06, "loss": 1.2407, "step": 82 }, { "epoch": 0.037607612143180785, "grad_norm": 0.1610557449308786, "learning_rate": 6.014492753623189e-06, "loss": 1.2481, "step": 83 }, { "epoch": 0.038060715903942, "grad_norm": 0.1556496331346778, "learning_rate": 6.086956521739132e-06, "loss": 1.2238, "step": 84 }, { "epoch": 0.038513819664703214, "grad_norm": 0.14170722555372517, "learning_rate": 6.159420289855072e-06, "loss": 1.2474, "step": 85 }, { "epoch": 0.03896692342546443, "grad_norm": 0.159819642522384, "learning_rate": 6.2318840579710145e-06, "loss": 1.2527, "step": 86 }, { "epoch": 0.03942002718622564, "grad_norm": 0.15063392870507789, "learning_rate": 6.304347826086958e-06, "loss": 1.2388, "step": 87 }, { "epoch": 0.03987313094698686, "grad_norm": 0.13788500087974867, "learning_rate": 6.376811594202898e-06, "loss": 1.2369, "step": 88 }, { "epoch": 0.04032623470774807, "grad_norm": 0.1478889073899375, "learning_rate": 6.449275362318841e-06, "loss": 1.2353, "step": 89 }, { "epoch": 0.040779338468509285, "grad_norm": 0.1505537201653033, "learning_rate": 6.521739130434783e-06, "loss": 1.2334, "step": 90 }, { "epoch": 0.0412324422292705, "grad_norm": 0.1335623695138423, "learning_rate": 6.594202898550725e-06, "loss": 1.2208, "step": 91 }, { "epoch": 0.04168554599003172, "grad_norm": 0.1445351214619499, "learning_rate": 6.666666666666667e-06, "loss": 1.2424, "step": 92 }, { "epoch": 0.042138649750792935, "grad_norm": 0.1496066231569064, "learning_rate": 6.739130434782609e-06, "loss": 1.2308, "step": 93 }, { "epoch": 0.04259175351155415, "grad_norm": 0.13056868308919578, "learning_rate": 6.811594202898551e-06, "loss": 1.2282, "step": 94 }, { "epoch": 0.043044857272315364, "grad_norm": 0.1423198307952714, "learning_rate": 6.884057971014493e-06, "loss": 1.2391, "step": 95 }, { "epoch": 0.04349796103307658, "grad_norm": 0.1484022593442689, "learning_rate": 6.956521739130435e-06, "loss": 1.2247, "step": 96 }, { "epoch": 0.04395106479383779, "grad_norm": 0.14206969616812365, "learning_rate": 7.028985507246377e-06, "loss": 1.2233, "step": 97 }, { "epoch": 0.044404168554599006, "grad_norm": 0.1403400138431183, "learning_rate": 7.10144927536232e-06, "loss": 1.2114, "step": 98 }, { "epoch": 0.04485727231536022, "grad_norm": 0.1403679275216342, "learning_rate": 7.173913043478261e-06, "loss": 1.2093, "step": 99 }, { "epoch": 0.045310376076121435, "grad_norm": 0.1382477795708966, "learning_rate": 7.246376811594203e-06, "loss": 1.2389, "step": 100 }, { "epoch": 0.04576347983688265, "grad_norm": 0.14455057308599017, "learning_rate": 7.318840579710146e-06, "loss": 1.2199, "step": 101 }, { "epoch": 0.046216583597643864, "grad_norm": 0.13267216089847295, "learning_rate": 7.391304347826087e-06, "loss": 1.2247, "step": 102 }, { "epoch": 0.04666968735840508, "grad_norm": 0.1260519289561795, "learning_rate": 7.4637681159420295e-06, "loss": 1.2273, "step": 103 }, { "epoch": 0.04712279111916629, "grad_norm": 0.14030652726959875, "learning_rate": 7.536231884057972e-06, "loss": 1.2145, "step": 104 }, { "epoch": 0.047575894879927506, "grad_norm": 0.14506109914519766, "learning_rate": 7.608695652173914e-06, "loss": 1.2165, "step": 105 }, { "epoch": 0.04802899864068872, "grad_norm": 0.15616407661914336, "learning_rate": 7.681159420289856e-06, "loss": 1.2162, "step": 106 }, { "epoch": 0.048482102401449935, "grad_norm": 0.13558942314029152, "learning_rate": 7.753623188405797e-06, "loss": 1.227, "step": 107 }, { "epoch": 0.04893520616221115, "grad_norm": 0.13566211355156743, "learning_rate": 7.82608695652174e-06, "loss": 1.2103, "step": 108 }, { "epoch": 0.049388309922972363, "grad_norm": 0.15917410507290075, "learning_rate": 7.898550724637682e-06, "loss": 1.2185, "step": 109 }, { "epoch": 0.04984141368373358, "grad_norm": 0.13195619282589566, "learning_rate": 7.971014492753623e-06, "loss": 1.2209, "step": 110 }, { "epoch": 0.05029451744449479, "grad_norm": 0.13378781830184408, "learning_rate": 8.043478260869566e-06, "loss": 1.2156, "step": 111 }, { "epoch": 0.050747621205256006, "grad_norm": 0.1297819249807937, "learning_rate": 8.115942028985508e-06, "loss": 1.2334, "step": 112 }, { "epoch": 0.05120072496601722, "grad_norm": 0.13650402005691822, "learning_rate": 8.188405797101449e-06, "loss": 1.1866, "step": 113 }, { "epoch": 0.051653828726778435, "grad_norm": 0.13366697643960854, "learning_rate": 8.260869565217392e-06, "loss": 1.208, "step": 114 }, { "epoch": 0.05210693248753965, "grad_norm": 0.14212385384329335, "learning_rate": 8.333333333333334e-06, "loss": 1.2219, "step": 115 }, { "epoch": 0.05256003624830086, "grad_norm": 0.135037906560412, "learning_rate": 8.405797101449275e-06, "loss": 1.1845, "step": 116 }, { "epoch": 0.05301314000906208, "grad_norm": 0.12460786736879492, "learning_rate": 8.478260869565218e-06, "loss": 1.2247, "step": 117 }, { "epoch": 0.05346624376982329, "grad_norm": 0.1298318472557901, "learning_rate": 8.55072463768116e-06, "loss": 1.1962, "step": 118 }, { "epoch": 0.053919347530584506, "grad_norm": 0.1311559932895696, "learning_rate": 8.623188405797103e-06, "loss": 1.2175, "step": 119 }, { "epoch": 0.05437245129134572, "grad_norm": 0.13796686560477314, "learning_rate": 8.695652173913044e-06, "loss": 1.2106, "step": 120 }, { "epoch": 0.054825555052106935, "grad_norm": 0.12686970488068378, "learning_rate": 8.768115942028986e-06, "loss": 1.1998, "step": 121 }, { "epoch": 0.05527865881286815, "grad_norm": 0.13029398233340597, "learning_rate": 8.840579710144929e-06, "loss": 1.1922, "step": 122 }, { "epoch": 0.05573176257362936, "grad_norm": 0.12154343160902772, "learning_rate": 8.91304347826087e-06, "loss": 1.2084, "step": 123 }, { "epoch": 0.05618486633439058, "grad_norm": 0.1365554482694075, "learning_rate": 8.985507246376812e-06, "loss": 1.1845, "step": 124 }, { "epoch": 0.05663797009515179, "grad_norm": 0.136349393555408, "learning_rate": 9.057971014492755e-06, "loss": 1.2142, "step": 125 }, { "epoch": 0.057091073855913006, "grad_norm": 0.14640567283060618, "learning_rate": 9.130434782608697e-06, "loss": 1.1821, "step": 126 }, { "epoch": 0.05754417761667422, "grad_norm": 0.162704740534985, "learning_rate": 9.202898550724638e-06, "loss": 1.2033, "step": 127 }, { "epoch": 0.057997281377435435, "grad_norm": 0.14620890194009586, "learning_rate": 9.275362318840581e-06, "loss": 1.1971, "step": 128 }, { "epoch": 0.05845038513819665, "grad_norm": 0.1511526117984611, "learning_rate": 9.347826086956523e-06, "loss": 1.1976, "step": 129 }, { "epoch": 0.05890348889895786, "grad_norm": 0.14153883351180233, "learning_rate": 9.420289855072464e-06, "loss": 1.1739, "step": 130 }, { "epoch": 0.05935659265971908, "grad_norm": 0.15814275919694315, "learning_rate": 9.492753623188407e-06, "loss": 1.1856, "step": 131 }, { "epoch": 0.05980969642048029, "grad_norm": 0.13186823973363634, "learning_rate": 9.565217391304349e-06, "loss": 1.1928, "step": 132 }, { "epoch": 0.060262800181241506, "grad_norm": 0.15741057654836269, "learning_rate": 9.63768115942029e-06, "loss": 1.2077, "step": 133 }, { "epoch": 0.06071590394200272, "grad_norm": 0.14864467163625872, "learning_rate": 9.710144927536233e-06, "loss": 1.2211, "step": 134 }, { "epoch": 0.061169007702763935, "grad_norm": 0.15138938681274097, "learning_rate": 9.782608695652175e-06, "loss": 1.2061, "step": 135 }, { "epoch": 0.06162211146352515, "grad_norm": 0.17987475484976362, "learning_rate": 9.855072463768118e-06, "loss": 1.2007, "step": 136 }, { "epoch": 0.06207521522428636, "grad_norm": 0.13296780040987535, "learning_rate": 9.927536231884058e-06, "loss": 1.1716, "step": 137 }, { "epoch": 0.06252831898504757, "grad_norm": 0.18666319447079466, "learning_rate": 1e-05, "loss": 1.1912, "step": 138 }, { "epoch": 0.06298142274580878, "grad_norm": 0.16669086671364958, "learning_rate": 1.0072463768115944e-05, "loss": 1.2018, "step": 139 }, { "epoch": 0.06343452650657, "grad_norm": 0.16750473224559076, "learning_rate": 1.0144927536231885e-05, "loss": 1.2185, "step": 140 }, { "epoch": 0.06388763026733121, "grad_norm": 0.13184183854275117, "learning_rate": 1.0217391304347829e-05, "loss": 1.1973, "step": 141 }, { "epoch": 0.06434073402809243, "grad_norm": 0.14725951911066765, "learning_rate": 1.0289855072463768e-05, "loss": 1.1901, "step": 142 }, { "epoch": 0.06479383778885364, "grad_norm": 0.13471002568690935, "learning_rate": 1.036231884057971e-05, "loss": 1.201, "step": 143 }, { "epoch": 0.06524694154961486, "grad_norm": 0.15650732752332805, "learning_rate": 1.0434782608695653e-05, "loss": 1.2188, "step": 144 }, { "epoch": 0.06570004531037607, "grad_norm": 0.1433833759368619, "learning_rate": 1.0507246376811594e-05, "loss": 1.1656, "step": 145 }, { "epoch": 0.06615314907113728, "grad_norm": 0.1426785943071348, "learning_rate": 1.0579710144927538e-05, "loss": 1.1807, "step": 146 }, { "epoch": 0.0666062528318985, "grad_norm": 0.14503219112661309, "learning_rate": 1.0652173913043479e-05, "loss": 1.2017, "step": 147 }, { "epoch": 0.06705935659265971, "grad_norm": 0.1492894899225257, "learning_rate": 1.0724637681159422e-05, "loss": 1.196, "step": 148 }, { "epoch": 0.06751246035342093, "grad_norm": 0.22273022277682494, "learning_rate": 1.0797101449275362e-05, "loss": 1.1905, "step": 149 }, { "epoch": 0.06796556411418214, "grad_norm": 0.17624192460215024, "learning_rate": 1.0869565217391305e-05, "loss": 1.1662, "step": 150 }, { "epoch": 0.06841866787494336, "grad_norm": 0.17152041660806583, "learning_rate": 1.0942028985507247e-05, "loss": 1.1931, "step": 151 }, { "epoch": 0.06887177163570457, "grad_norm": 0.14532442048767774, "learning_rate": 1.101449275362319e-05, "loss": 1.1851, "step": 152 }, { "epoch": 0.06932487539646578, "grad_norm": 0.16118481913636407, "learning_rate": 1.1086956521739131e-05, "loss": 1.1921, "step": 153 }, { "epoch": 0.069777979157227, "grad_norm": 0.15569062405361558, "learning_rate": 1.1159420289855074e-05, "loss": 1.1936, "step": 154 }, { "epoch": 0.07023108291798821, "grad_norm": 0.1428647402212984, "learning_rate": 1.1231884057971016e-05, "loss": 1.1776, "step": 155 }, { "epoch": 0.07068418667874943, "grad_norm": 0.1533308808183041, "learning_rate": 1.1304347826086957e-05, "loss": 1.1729, "step": 156 }, { "epoch": 0.07113729043951064, "grad_norm": 0.14670062070349094, "learning_rate": 1.1376811594202899e-05, "loss": 1.1648, "step": 157 }, { "epoch": 0.07159039420027186, "grad_norm": 0.16410231099289155, "learning_rate": 1.1449275362318842e-05, "loss": 1.1939, "step": 158 }, { "epoch": 0.07204349796103307, "grad_norm": 0.14396658873236956, "learning_rate": 1.1521739130434783e-05, "loss": 1.1758, "step": 159 }, { "epoch": 0.07249660172179428, "grad_norm": 0.171040828563285, "learning_rate": 1.1594202898550726e-05, "loss": 1.179, "step": 160 }, { "epoch": 0.0729497054825555, "grad_norm": 0.17779034248964817, "learning_rate": 1.1666666666666668e-05, "loss": 1.1835, "step": 161 }, { "epoch": 0.07340280924331671, "grad_norm": 0.20182378324784891, "learning_rate": 1.1739130434782611e-05, "loss": 1.1777, "step": 162 }, { "epoch": 0.07385591300407793, "grad_norm": 0.1847888646387923, "learning_rate": 1.181159420289855e-05, "loss": 1.1847, "step": 163 }, { "epoch": 0.07430901676483914, "grad_norm": 0.22021118923239125, "learning_rate": 1.1884057971014494e-05, "loss": 1.1688, "step": 164 }, { "epoch": 0.07476212052560036, "grad_norm": 0.18392731326891137, "learning_rate": 1.1956521739130435e-05, "loss": 1.1696, "step": 165 }, { "epoch": 0.07521522428636157, "grad_norm": 0.16715291761079498, "learning_rate": 1.2028985507246379e-05, "loss": 1.2034, "step": 166 }, { "epoch": 0.07566832804712278, "grad_norm": 0.15416757160775282, "learning_rate": 1.210144927536232e-05, "loss": 1.1802, "step": 167 }, { "epoch": 0.076121431807884, "grad_norm": 0.18254110557329295, "learning_rate": 1.2173913043478263e-05, "loss": 1.1606, "step": 168 }, { "epoch": 0.07657453556864521, "grad_norm": 0.16114009660180326, "learning_rate": 1.2246376811594205e-05, "loss": 1.1638, "step": 169 }, { "epoch": 0.07702763932940643, "grad_norm": 0.20052021177631932, "learning_rate": 1.2318840579710144e-05, "loss": 1.154, "step": 170 }, { "epoch": 0.07748074309016764, "grad_norm": 0.25964757229529967, "learning_rate": 1.2391304347826088e-05, "loss": 1.1525, "step": 171 }, { "epoch": 0.07793384685092886, "grad_norm": 0.27081952541221294, "learning_rate": 1.2463768115942029e-05, "loss": 1.1886, "step": 172 }, { "epoch": 0.07838695061169007, "grad_norm": 0.27636849045899686, "learning_rate": 1.2536231884057972e-05, "loss": 1.1861, "step": 173 }, { "epoch": 0.07884005437245128, "grad_norm": 0.20585509298814925, "learning_rate": 1.2608695652173915e-05, "loss": 1.1703, "step": 174 }, { "epoch": 0.0792931581332125, "grad_norm": 0.16705379502496914, "learning_rate": 1.2681159420289857e-05, "loss": 1.1708, "step": 175 }, { "epoch": 0.07974626189397371, "grad_norm": 0.17266273888419761, "learning_rate": 1.2753623188405797e-05, "loss": 1.1461, "step": 176 }, { "epoch": 0.08019936565473493, "grad_norm": 0.2106571714931951, "learning_rate": 1.282608695652174e-05, "loss": 1.1694, "step": 177 }, { "epoch": 0.08065246941549614, "grad_norm": 0.2730917026304194, "learning_rate": 1.2898550724637681e-05, "loss": 1.1667, "step": 178 }, { "epoch": 0.08110557317625736, "grad_norm": 0.2579174661815517, "learning_rate": 1.2971014492753624e-05, "loss": 1.1651, "step": 179 }, { "epoch": 0.08155867693701857, "grad_norm": 0.2751320594599073, "learning_rate": 1.3043478260869566e-05, "loss": 1.15, "step": 180 }, { "epoch": 0.08201178069777978, "grad_norm": 0.19955633636170586, "learning_rate": 1.3115942028985509e-05, "loss": 1.1761, "step": 181 }, { "epoch": 0.082464884458541, "grad_norm": 0.21300278274953302, "learning_rate": 1.318840579710145e-05, "loss": 1.1616, "step": 182 }, { "epoch": 0.08291798821930221, "grad_norm": 0.2847356988946747, "learning_rate": 1.3260869565217392e-05, "loss": 1.1407, "step": 183 }, { "epoch": 0.08337109198006344, "grad_norm": 0.2815201113881011, "learning_rate": 1.3333333333333333e-05, "loss": 1.1868, "step": 184 }, { "epoch": 0.08382419574082466, "grad_norm": 0.25216243350126727, "learning_rate": 1.3405797101449276e-05, "loss": 1.159, "step": 185 }, { "epoch": 0.08427729950158587, "grad_norm": 0.1680724808145423, "learning_rate": 1.3478260869565218e-05, "loss": 1.1731, "step": 186 }, { "epoch": 0.08473040326234708, "grad_norm": 0.20000955440482673, "learning_rate": 1.3550724637681161e-05, "loss": 1.1658, "step": 187 }, { "epoch": 0.0851835070231083, "grad_norm": 0.24494519216371521, "learning_rate": 1.3623188405797103e-05, "loss": 1.1777, "step": 188 }, { "epoch": 0.08563661078386951, "grad_norm": 0.27875758110067717, "learning_rate": 1.3695652173913046e-05, "loss": 1.186, "step": 189 }, { "epoch": 0.08608971454463073, "grad_norm": 0.25996633224525123, "learning_rate": 1.3768115942028985e-05, "loss": 1.1693, "step": 190 }, { "epoch": 0.08654281830539194, "grad_norm": 0.2210835087222055, "learning_rate": 1.3840579710144929e-05, "loss": 1.1731, "step": 191 }, { "epoch": 0.08699592206615316, "grad_norm": 0.22901148140350822, "learning_rate": 1.391304347826087e-05, "loss": 1.1743, "step": 192 }, { "epoch": 0.08744902582691437, "grad_norm": 0.2876824298246378, "learning_rate": 1.3985507246376813e-05, "loss": 1.1432, "step": 193 }, { "epoch": 0.08790212958767558, "grad_norm": 0.3758712186813318, "learning_rate": 1.4057971014492755e-05, "loss": 1.145, "step": 194 }, { "epoch": 0.0883552333484368, "grad_norm": 0.3323640015591156, "learning_rate": 1.4130434782608698e-05, "loss": 1.1568, "step": 195 }, { "epoch": 0.08880833710919801, "grad_norm": 0.2067787922975722, "learning_rate": 1.420289855072464e-05, "loss": 1.1438, "step": 196 }, { "epoch": 0.08926144086995923, "grad_norm": 0.2362056140950305, "learning_rate": 1.427536231884058e-05, "loss": 1.1734, "step": 197 }, { "epoch": 0.08971454463072044, "grad_norm": 0.24283726655684335, "learning_rate": 1.4347826086956522e-05, "loss": 1.1512, "step": 198 }, { "epoch": 0.09016764839148166, "grad_norm": 0.19688814877178304, "learning_rate": 1.4420289855072465e-05, "loss": 1.1527, "step": 199 }, { "epoch": 0.09062075215224287, "grad_norm": 0.19847503834468674, "learning_rate": 1.4492753623188407e-05, "loss": 1.1692, "step": 200 }, { "epoch": 0.09107385591300408, "grad_norm": 0.36759688475987085, "learning_rate": 1.456521739130435e-05, "loss": 1.1647, "step": 201 }, { "epoch": 0.0915269596737653, "grad_norm": 0.4662845518690142, "learning_rate": 1.4637681159420291e-05, "loss": 1.1521, "step": 202 }, { "epoch": 0.09198006343452651, "grad_norm": 0.39107908047851964, "learning_rate": 1.4710144927536235e-05, "loss": 1.1627, "step": 203 }, { "epoch": 0.09243316719528773, "grad_norm": 0.38390019432836325, "learning_rate": 1.4782608695652174e-05, "loss": 1.1437, "step": 204 }, { "epoch": 0.09288627095604894, "grad_norm": 0.3553420683930557, "learning_rate": 1.4855072463768116e-05, "loss": 1.131, "step": 205 }, { "epoch": 0.09333937471681016, "grad_norm": 0.2506457994360238, "learning_rate": 1.4927536231884059e-05, "loss": 1.1383, "step": 206 }, { "epoch": 0.09379247847757137, "grad_norm": 0.22845603310409918, "learning_rate": 1.5000000000000002e-05, "loss": 1.1525, "step": 207 }, { "epoch": 0.09424558223833258, "grad_norm": 0.29510790385493385, "learning_rate": 1.5072463768115944e-05, "loss": 1.1661, "step": 208 }, { "epoch": 0.0946986859990938, "grad_norm": 0.29908502831948197, "learning_rate": 1.5144927536231887e-05, "loss": 1.1512, "step": 209 }, { "epoch": 0.09515178975985501, "grad_norm": 0.18920730066617467, "learning_rate": 1.5217391304347828e-05, "loss": 1.1564, "step": 210 }, { "epoch": 0.09560489352061623, "grad_norm": 0.2026494309774227, "learning_rate": 1.5289855072463768e-05, "loss": 1.122, "step": 211 }, { "epoch": 0.09605799728137744, "grad_norm": 0.28075528177287806, "learning_rate": 1.536231884057971e-05, "loss": 1.144, "step": 212 }, { "epoch": 0.09651110104213866, "grad_norm": 0.3043481715034076, "learning_rate": 1.5434782608695654e-05, "loss": 1.1673, "step": 213 }, { "epoch": 0.09696420480289987, "grad_norm": 0.24966236259428592, "learning_rate": 1.5507246376811594e-05, "loss": 1.1451, "step": 214 }, { "epoch": 0.09741730856366108, "grad_norm": 0.258909891565715, "learning_rate": 1.5579710144927537e-05, "loss": 1.1175, "step": 215 }, { "epoch": 0.0978704123244223, "grad_norm": 0.3138270326874341, "learning_rate": 1.565217391304348e-05, "loss": 1.1544, "step": 216 }, { "epoch": 0.09832351608518351, "grad_norm": 0.28358369829195657, "learning_rate": 1.5724637681159423e-05, "loss": 1.145, "step": 217 }, { "epoch": 0.09877661984594473, "grad_norm": 0.24397672796734368, "learning_rate": 1.5797101449275363e-05, "loss": 1.1438, "step": 218 }, { "epoch": 0.09922972360670594, "grad_norm": 0.214869697964168, "learning_rate": 1.5869565217391306e-05, "loss": 1.1493, "step": 219 }, { "epoch": 0.09968282736746716, "grad_norm": 0.22443040063657702, "learning_rate": 1.5942028985507246e-05, "loss": 1.1656, "step": 220 }, { "epoch": 0.10013593112822837, "grad_norm": 0.26136127048062197, "learning_rate": 1.601449275362319e-05, "loss": 1.1515, "step": 221 }, { "epoch": 0.10058903488898958, "grad_norm": 0.2777708008850257, "learning_rate": 1.6086956521739132e-05, "loss": 1.1421, "step": 222 }, { "epoch": 0.1010421386497508, "grad_norm": 0.22570334132032477, "learning_rate": 1.6159420289855076e-05, "loss": 1.1423, "step": 223 }, { "epoch": 0.10149524241051201, "grad_norm": 0.20911759349629955, "learning_rate": 1.6231884057971015e-05, "loss": 1.1361, "step": 224 }, { "epoch": 0.10194834617127323, "grad_norm": 0.23436358920228426, "learning_rate": 1.630434782608696e-05, "loss": 1.1562, "step": 225 }, { "epoch": 0.10240144993203444, "grad_norm": 0.23455706978218666, "learning_rate": 1.6376811594202898e-05, "loss": 1.1401, "step": 226 }, { "epoch": 0.10285455369279566, "grad_norm": 0.21301388506421437, "learning_rate": 1.644927536231884e-05, "loss": 1.1308, "step": 227 }, { "epoch": 0.10330765745355687, "grad_norm": 0.2601208215659811, "learning_rate": 1.6521739130434785e-05, "loss": 1.1566, "step": 228 }, { "epoch": 0.10376076121431808, "grad_norm": 0.2742230054085892, "learning_rate": 1.6594202898550728e-05, "loss": 1.1493, "step": 229 }, { "epoch": 0.1042138649750793, "grad_norm": 0.2820438508383035, "learning_rate": 1.6666666666666667e-05, "loss": 1.1557, "step": 230 }, { "epoch": 0.10466696873584051, "grad_norm": 0.3191613154668632, "learning_rate": 1.673913043478261e-05, "loss": 1.156, "step": 231 }, { "epoch": 0.10512007249660173, "grad_norm": 0.3572254722455233, "learning_rate": 1.681159420289855e-05, "loss": 1.1538, "step": 232 }, { "epoch": 0.10557317625736294, "grad_norm": 0.3539056270052157, "learning_rate": 1.6884057971014494e-05, "loss": 1.1366, "step": 233 }, { "epoch": 0.10602628001812416, "grad_norm": 0.36590917858307137, "learning_rate": 1.6956521739130437e-05, "loss": 1.1477, "step": 234 }, { "epoch": 0.10647938377888537, "grad_norm": 0.4382147201617384, "learning_rate": 1.702898550724638e-05, "loss": 1.1526, "step": 235 }, { "epoch": 0.10693248753964658, "grad_norm": 0.46051761048974277, "learning_rate": 1.710144927536232e-05, "loss": 1.1331, "step": 236 }, { "epoch": 0.1073855913004078, "grad_norm": 0.4280122622445664, "learning_rate": 1.7173913043478263e-05, "loss": 1.1467, "step": 237 }, { "epoch": 0.10783869506116901, "grad_norm": 0.36668197190737223, "learning_rate": 1.7246376811594206e-05, "loss": 1.1664, "step": 238 }, { "epoch": 0.10829179882193023, "grad_norm": 0.33103033603721427, "learning_rate": 1.7318840579710146e-05, "loss": 1.1497, "step": 239 }, { "epoch": 0.10874490258269144, "grad_norm": 0.3156636819071045, "learning_rate": 1.739130434782609e-05, "loss": 1.1371, "step": 240 }, { "epoch": 0.10919800634345266, "grad_norm": 0.3056449799015048, "learning_rate": 1.746376811594203e-05, "loss": 1.1442, "step": 241 }, { "epoch": 0.10965111010421387, "grad_norm": 0.3003011722354453, "learning_rate": 1.7536231884057972e-05, "loss": 1.1286, "step": 242 }, { "epoch": 0.11010421386497508, "grad_norm": 0.30215534102112923, "learning_rate": 1.7608695652173915e-05, "loss": 1.1261, "step": 243 }, { "epoch": 0.1105573176257363, "grad_norm": 0.27432898298470265, "learning_rate": 1.7681159420289858e-05, "loss": 1.1314, "step": 244 }, { "epoch": 0.11101042138649751, "grad_norm": 0.28653549808927065, "learning_rate": 1.7753623188405798e-05, "loss": 1.1523, "step": 245 }, { "epoch": 0.11146352514725873, "grad_norm": 0.33840631153667167, "learning_rate": 1.782608695652174e-05, "loss": 1.1318, "step": 246 }, { "epoch": 0.11191662890801994, "grad_norm": 0.4022974004760939, "learning_rate": 1.789855072463768e-05, "loss": 1.1421, "step": 247 }, { "epoch": 0.11236973266878116, "grad_norm": 0.39707271065306265, "learning_rate": 1.7971014492753624e-05, "loss": 1.1485, "step": 248 }, { "epoch": 0.11282283642954237, "grad_norm": 0.30534167889274944, "learning_rate": 1.8043478260869567e-05, "loss": 1.1533, "step": 249 }, { "epoch": 0.11327594019030358, "grad_norm": 0.2315886361003725, "learning_rate": 1.811594202898551e-05, "loss": 1.1563, "step": 250 }, { "epoch": 0.1137290439510648, "grad_norm": 0.271006208821516, "learning_rate": 1.818840579710145e-05, "loss": 1.1097, "step": 251 }, { "epoch": 0.11418214771182601, "grad_norm": 0.3192905734997128, "learning_rate": 1.8260869565217393e-05, "loss": 1.1445, "step": 252 }, { "epoch": 0.11463525147258723, "grad_norm": 0.2988897003233889, "learning_rate": 1.8333333333333333e-05, "loss": 1.1509, "step": 253 }, { "epoch": 0.11508835523334844, "grad_norm": 0.18920261619528156, "learning_rate": 1.8405797101449276e-05, "loss": 1.153, "step": 254 }, { "epoch": 0.11554145899410966, "grad_norm": 0.21187128232655875, "learning_rate": 1.847826086956522e-05, "loss": 1.1173, "step": 255 }, { "epoch": 0.11599456275487087, "grad_norm": 0.24809439633700292, "learning_rate": 1.8550724637681162e-05, "loss": 1.1331, "step": 256 }, { "epoch": 0.11644766651563208, "grad_norm": 0.3405124657129926, "learning_rate": 1.8623188405797102e-05, "loss": 1.1553, "step": 257 }, { "epoch": 0.1169007702763933, "grad_norm": 0.4211431039726426, "learning_rate": 1.8695652173913045e-05, "loss": 1.123, "step": 258 }, { "epoch": 0.11735387403715451, "grad_norm": 0.47169324724229683, "learning_rate": 1.8768115942028985e-05, "loss": 1.1534, "step": 259 }, { "epoch": 0.11780697779791573, "grad_norm": 0.6012413102954954, "learning_rate": 1.8840579710144928e-05, "loss": 1.144, "step": 260 }, { "epoch": 0.11826008155867694, "grad_norm": 0.7269834198873751, "learning_rate": 1.891304347826087e-05, "loss": 1.1456, "step": 261 }, { "epoch": 0.11871318531943816, "grad_norm": 0.8245861495180895, "learning_rate": 1.8985507246376814e-05, "loss": 1.152, "step": 262 }, { "epoch": 0.11916628908019937, "grad_norm": 0.8064016325706005, "learning_rate": 1.9057971014492754e-05, "loss": 1.141, "step": 263 }, { "epoch": 0.11961939284096058, "grad_norm": 0.6377295638070679, "learning_rate": 1.9130434782608697e-05, "loss": 1.1575, "step": 264 }, { "epoch": 0.1200724966017218, "grad_norm": 0.5014027254716775, "learning_rate": 1.920289855072464e-05, "loss": 1.1518, "step": 265 }, { "epoch": 0.12052560036248301, "grad_norm": 0.47220397066083464, "learning_rate": 1.927536231884058e-05, "loss": 1.1387, "step": 266 }, { "epoch": 0.12097870412324423, "grad_norm": 0.4832024654476037, "learning_rate": 1.9347826086956523e-05, "loss": 1.1431, "step": 267 }, { "epoch": 0.12143180788400544, "grad_norm": 0.5125780417450365, "learning_rate": 1.9420289855072467e-05, "loss": 1.1211, "step": 268 }, { "epoch": 0.12188491164476666, "grad_norm": 0.5089292950170233, "learning_rate": 1.9492753623188406e-05, "loss": 1.1547, "step": 269 }, { "epoch": 0.12233801540552787, "grad_norm": 0.5268176004893336, "learning_rate": 1.956521739130435e-05, "loss": 1.1309, "step": 270 }, { "epoch": 0.12279111916628908, "grad_norm": 0.4520648310778846, "learning_rate": 1.9637681159420293e-05, "loss": 1.132, "step": 271 }, { "epoch": 0.1232442229270503, "grad_norm": 0.3608298224502208, "learning_rate": 1.9710144927536236e-05, "loss": 1.1473, "step": 272 }, { "epoch": 0.12369732668781151, "grad_norm": 0.40607912524572626, "learning_rate": 1.9782608695652176e-05, "loss": 1.1433, "step": 273 }, { "epoch": 0.12415043044857273, "grad_norm": 0.4252765180897805, "learning_rate": 1.9855072463768115e-05, "loss": 1.1114, "step": 274 }, { "epoch": 0.12460353420933394, "grad_norm": 0.4463380177506588, "learning_rate": 1.992753623188406e-05, "loss": 1.1168, "step": 275 }, { "epoch": 0.12505663797009514, "grad_norm": 0.370623097411747, "learning_rate": 2e-05, "loss": 1.1573, "step": 276 }, { "epoch": 0.12550974173085636, "grad_norm": 0.24811543914203152, "learning_rate": 2.0072463768115945e-05, "loss": 1.1278, "step": 277 }, { "epoch": 0.12596284549161757, "grad_norm": 0.23124511592732, "learning_rate": 2.0144927536231888e-05, "loss": 1.1384, "step": 278 }, { "epoch": 0.12641594925237878, "grad_norm": 0.404073376011677, "learning_rate": 2.0217391304347828e-05, "loss": 1.1413, "step": 279 }, { "epoch": 0.12686905301314, "grad_norm": 0.4682029047187629, "learning_rate": 2.028985507246377e-05, "loss": 1.1358, "step": 280 }, { "epoch": 0.1273221567739012, "grad_norm": 0.3836387096482354, "learning_rate": 2.0362318840579714e-05, "loss": 1.1321, "step": 281 }, { "epoch": 0.12777526053466243, "grad_norm": 0.27274353728619943, "learning_rate": 2.0434782608695657e-05, "loss": 1.1307, "step": 282 }, { "epoch": 0.12822836429542364, "grad_norm": 0.2703452787775874, "learning_rate": 2.0507246376811594e-05, "loss": 1.1458, "step": 283 }, { "epoch": 0.12868146805618486, "grad_norm": 0.32444768527350754, "learning_rate": 2.0579710144927537e-05, "loss": 1.1284, "step": 284 }, { "epoch": 0.12913457181694607, "grad_norm": 0.4059389929155608, "learning_rate": 2.065217391304348e-05, "loss": 1.1224, "step": 285 }, { "epoch": 0.12958767557770728, "grad_norm": 0.43507801010288555, "learning_rate": 2.072463768115942e-05, "loss": 1.1169, "step": 286 }, { "epoch": 0.1300407793384685, "grad_norm": 0.40309868330161214, "learning_rate": 2.0797101449275363e-05, "loss": 1.1339, "step": 287 }, { "epoch": 0.1304938830992297, "grad_norm": 0.35826616658169663, "learning_rate": 2.0869565217391306e-05, "loss": 1.1408, "step": 288 }, { "epoch": 0.13094698685999093, "grad_norm": 0.3851622133940432, "learning_rate": 2.094202898550725e-05, "loss": 1.1119, "step": 289 }, { "epoch": 0.13140009062075214, "grad_norm": 0.4410574721610812, "learning_rate": 2.101449275362319e-05, "loss": 1.107, "step": 290 }, { "epoch": 0.13185319438151336, "grad_norm": 0.46555343277223127, "learning_rate": 2.1086956521739132e-05, "loss": 1.1254, "step": 291 }, { "epoch": 0.13230629814227457, "grad_norm": 0.5625079932120484, "learning_rate": 2.1159420289855075e-05, "loss": 1.1374, "step": 292 }, { "epoch": 0.13275940190303578, "grad_norm": 0.6743000338914465, "learning_rate": 2.1231884057971018e-05, "loss": 1.149, "step": 293 }, { "epoch": 0.133212505663797, "grad_norm": 0.7250376421534288, "learning_rate": 2.1304347826086958e-05, "loss": 1.1224, "step": 294 }, { "epoch": 0.1336656094245582, "grad_norm": 0.6933176363599328, "learning_rate": 2.13768115942029e-05, "loss": 1.1284, "step": 295 }, { "epoch": 0.13411871318531943, "grad_norm": 0.43137873669114624, "learning_rate": 2.1449275362318844e-05, "loss": 1.1478, "step": 296 }, { "epoch": 0.13457181694608064, "grad_norm": 0.2858094582478082, "learning_rate": 2.152173913043478e-05, "loss": 1.1402, "step": 297 }, { "epoch": 0.13502492070684186, "grad_norm": 0.39754567805003965, "learning_rate": 2.1594202898550724e-05, "loss": 1.1157, "step": 298 }, { "epoch": 0.13547802446760307, "grad_norm": 0.48860716086344136, "learning_rate": 2.1666666666666667e-05, "loss": 1.1268, "step": 299 }, { "epoch": 0.13593112822836428, "grad_norm": 0.46380432483781153, "learning_rate": 2.173913043478261e-05, "loss": 1.1324, "step": 300 }, { "epoch": 0.1363842319891255, "grad_norm": 0.3732324977832947, "learning_rate": 2.1811594202898553e-05, "loss": 1.1312, "step": 301 }, { "epoch": 0.1368373357498867, "grad_norm": 0.35121232131400526, "learning_rate": 2.1884057971014493e-05, "loss": 1.1264, "step": 302 }, { "epoch": 0.13729043951064793, "grad_norm": 0.4293095145149619, "learning_rate": 2.1956521739130436e-05, "loss": 1.136, "step": 303 }, { "epoch": 0.13774354327140914, "grad_norm": 0.3992558788517691, "learning_rate": 2.202898550724638e-05, "loss": 1.1394, "step": 304 }, { "epoch": 0.13819664703217036, "grad_norm": 0.2832974691741076, "learning_rate": 2.2101449275362323e-05, "loss": 1.1104, "step": 305 }, { "epoch": 0.13864975079293157, "grad_norm": 0.3122541431641511, "learning_rate": 2.2173913043478262e-05, "loss": 1.1289, "step": 306 }, { "epoch": 0.13910285455369278, "grad_norm": 0.3940794203503235, "learning_rate": 2.2246376811594205e-05, "loss": 1.1071, "step": 307 }, { "epoch": 0.139555958314454, "grad_norm": 0.3889620212546529, "learning_rate": 2.231884057971015e-05, "loss": 1.1112, "step": 308 }, { "epoch": 0.1400090620752152, "grad_norm": 0.32286207658689725, "learning_rate": 2.2391304347826092e-05, "loss": 1.1246, "step": 309 }, { "epoch": 0.14046216583597643, "grad_norm": 0.35105854581343565, "learning_rate": 2.246376811594203e-05, "loss": 1.1243, "step": 310 }, { "epoch": 0.14091526959673764, "grad_norm": 0.3068386120080713, "learning_rate": 2.253623188405797e-05, "loss": 1.1249, "step": 311 }, { "epoch": 0.14136837335749886, "grad_norm": 0.3181345216097836, "learning_rate": 2.2608695652173914e-05, "loss": 1.1107, "step": 312 }, { "epoch": 0.14182147711826007, "grad_norm": 0.3688565764496039, "learning_rate": 2.2681159420289854e-05, "loss": 1.14, "step": 313 }, { "epoch": 0.14227458087902128, "grad_norm": 0.43240089287454964, "learning_rate": 2.2753623188405797e-05, "loss": 1.1317, "step": 314 }, { "epoch": 0.1427276846397825, "grad_norm": 0.43729874691473075, "learning_rate": 2.282608695652174e-05, "loss": 1.1058, "step": 315 }, { "epoch": 0.1431807884005437, "grad_norm": 0.46123462365120266, "learning_rate": 2.2898550724637684e-05, "loss": 1.141, "step": 316 }, { "epoch": 0.14363389216130493, "grad_norm": 0.5970206196529548, "learning_rate": 2.2971014492753623e-05, "loss": 1.127, "step": 317 }, { "epoch": 0.14408699592206614, "grad_norm": 0.6258668433319501, "learning_rate": 2.3043478260869567e-05, "loss": 1.1413, "step": 318 }, { "epoch": 0.14454009968282736, "grad_norm": 0.5337769250580413, "learning_rate": 2.311594202898551e-05, "loss": 1.1106, "step": 319 }, { "epoch": 0.14499320344358857, "grad_norm": 0.4710528652464373, "learning_rate": 2.3188405797101453e-05, "loss": 1.072, "step": 320 }, { "epoch": 0.14544630720434978, "grad_norm": 0.38975907651807407, "learning_rate": 2.3260869565217396e-05, "loss": 1.1298, "step": 321 }, { "epoch": 0.145899410965111, "grad_norm": 0.3203576098493583, "learning_rate": 2.3333333333333336e-05, "loss": 1.1103, "step": 322 }, { "epoch": 0.1463525147258722, "grad_norm": 0.4433738817591825, "learning_rate": 2.340579710144928e-05, "loss": 1.1455, "step": 323 }, { "epoch": 0.14680561848663343, "grad_norm": 0.6463966921943783, "learning_rate": 2.3478260869565222e-05, "loss": 1.1094, "step": 324 }, { "epoch": 0.14725872224739464, "grad_norm": 0.9207937076512789, "learning_rate": 2.355072463768116e-05, "loss": 1.1294, "step": 325 }, { "epoch": 0.14771182600815586, "grad_norm": 0.985506548709933, "learning_rate": 2.36231884057971e-05, "loss": 1.1288, "step": 326 }, { "epoch": 0.14816492976891707, "grad_norm": 0.7427914296735908, "learning_rate": 2.3695652173913045e-05, "loss": 1.1319, "step": 327 }, { "epoch": 0.14861803352967828, "grad_norm": 0.4851402470918087, "learning_rate": 2.3768115942028988e-05, "loss": 1.1292, "step": 328 }, { "epoch": 0.1490711372904395, "grad_norm": 0.4340360697464059, "learning_rate": 2.3840579710144928e-05, "loss": 1.1286, "step": 329 }, { "epoch": 0.1495242410512007, "grad_norm": 0.44665071033363524, "learning_rate": 2.391304347826087e-05, "loss": 1.1038, "step": 330 }, { "epoch": 0.14997734481196193, "grad_norm": 0.4610586728246728, "learning_rate": 2.3985507246376814e-05, "loss": 1.1194, "step": 331 }, { "epoch": 0.15043044857272314, "grad_norm": 0.4858237345622067, "learning_rate": 2.4057971014492757e-05, "loss": 1.106, "step": 332 }, { "epoch": 0.15088355233348436, "grad_norm": 0.4023776819455534, "learning_rate": 2.4130434782608697e-05, "loss": 1.1228, "step": 333 }, { "epoch": 0.15133665609424557, "grad_norm": 0.28250742492020475, "learning_rate": 2.420289855072464e-05, "loss": 1.0946, "step": 334 }, { "epoch": 0.15178975985500678, "grad_norm": 0.3285308071581642, "learning_rate": 2.4275362318840583e-05, "loss": 1.1088, "step": 335 }, { "epoch": 0.152242863615768, "grad_norm": 0.3605078235076501, "learning_rate": 2.4347826086956526e-05, "loss": 1.1244, "step": 336 }, { "epoch": 0.1526959673765292, "grad_norm": 0.42741445585427446, "learning_rate": 2.4420289855072466e-05, "loss": 1.1199, "step": 337 }, { "epoch": 0.15314907113729043, "grad_norm": 0.4938491700793613, "learning_rate": 2.449275362318841e-05, "loss": 1.1164, "step": 338 }, { "epoch": 0.15360217489805164, "grad_norm": 0.512032294048734, "learning_rate": 2.456521739130435e-05, "loss": 1.133, "step": 339 }, { "epoch": 0.15405527865881286, "grad_norm": 0.48805723300868714, "learning_rate": 2.463768115942029e-05, "loss": 1.1253, "step": 340 }, { "epoch": 0.15450838241957407, "grad_norm": 0.3993339519221479, "learning_rate": 2.4710144927536232e-05, "loss": 1.0885, "step": 341 }, { "epoch": 0.15496148618033528, "grad_norm": 0.3981693111618921, "learning_rate": 2.4782608695652175e-05, "loss": 1.1039, "step": 342 }, { "epoch": 0.1554145899410965, "grad_norm": 0.4509612301978515, "learning_rate": 2.4855072463768118e-05, "loss": 1.1193, "step": 343 }, { "epoch": 0.1558676937018577, "grad_norm": 0.5539517724072442, "learning_rate": 2.4927536231884058e-05, "loss": 1.1227, "step": 344 }, { "epoch": 0.15632079746261893, "grad_norm": 0.664006628857665, "learning_rate": 2.5e-05, "loss": 1.1187, "step": 345 }, { "epoch": 0.15677390122338014, "grad_norm": 0.6860301855510393, "learning_rate": 2.5072463768115944e-05, "loss": 1.12, "step": 346 }, { "epoch": 0.15722700498414136, "grad_norm": 0.6649179149280643, "learning_rate": 2.5144927536231888e-05, "loss": 1.1246, "step": 347 }, { "epoch": 0.15768010874490257, "grad_norm": 0.5503660645272839, "learning_rate": 2.521739130434783e-05, "loss": 1.1035, "step": 348 }, { "epoch": 0.15813321250566378, "grad_norm": 0.37899681566980026, "learning_rate": 2.528985507246377e-05, "loss": 1.1264, "step": 349 }, { "epoch": 0.158586316266425, "grad_norm": 0.2533176409497846, "learning_rate": 2.5362318840579714e-05, "loss": 1.1205, "step": 350 }, { "epoch": 0.1590394200271862, "grad_norm": 0.3127177220079576, "learning_rate": 2.5434782608695657e-05, "loss": 1.1082, "step": 351 }, { "epoch": 0.15949252378794743, "grad_norm": 0.4695675598892739, "learning_rate": 2.5507246376811593e-05, "loss": 1.0978, "step": 352 }, { "epoch": 0.15994562754870864, "grad_norm": 0.4832228482492921, "learning_rate": 2.5579710144927536e-05, "loss": 1.1251, "step": 353 }, { "epoch": 0.16039873130946986, "grad_norm": 0.4685426171260275, "learning_rate": 2.565217391304348e-05, "loss": 1.1145, "step": 354 }, { "epoch": 0.16085183507023107, "grad_norm": 0.40062355167793495, "learning_rate": 2.5724637681159423e-05, "loss": 1.1149, "step": 355 }, { "epoch": 0.16130493883099228, "grad_norm": 0.4578643838497083, "learning_rate": 2.5797101449275362e-05, "loss": 1.1134, "step": 356 }, { "epoch": 0.1617580425917535, "grad_norm": 0.615695381618595, "learning_rate": 2.5869565217391305e-05, "loss": 1.1296, "step": 357 }, { "epoch": 0.1622111463525147, "grad_norm": 0.7043072503898563, "learning_rate": 2.594202898550725e-05, "loss": 1.1109, "step": 358 }, { "epoch": 0.16266425011327593, "grad_norm": 0.7759863579549692, "learning_rate": 2.6014492753623192e-05, "loss": 1.1195, "step": 359 }, { "epoch": 0.16311735387403714, "grad_norm": 0.7641097337039678, "learning_rate": 2.608695652173913e-05, "loss": 1.1235, "step": 360 }, { "epoch": 0.16357045763479836, "grad_norm": 0.6963046776308135, "learning_rate": 2.6159420289855075e-05, "loss": 1.1301, "step": 361 }, { "epoch": 0.16402356139555957, "grad_norm": 0.5934038521367055, "learning_rate": 2.6231884057971018e-05, "loss": 1.1011, "step": 362 }, { "epoch": 0.16447666515632078, "grad_norm": 0.5526697964505245, "learning_rate": 2.630434782608696e-05, "loss": 1.0983, "step": 363 }, { "epoch": 0.164929768917082, "grad_norm": 0.49109672191230175, "learning_rate": 2.63768115942029e-05, "loss": 1.113, "step": 364 }, { "epoch": 0.1653828726778432, "grad_norm": 0.3971913795583426, "learning_rate": 2.6449275362318844e-05, "loss": 1.1049, "step": 365 }, { "epoch": 0.16583597643860443, "grad_norm": 0.5237217600527744, "learning_rate": 2.6521739130434784e-05, "loss": 1.1017, "step": 366 }, { "epoch": 0.16628908019936564, "grad_norm": 0.700028564282141, "learning_rate": 2.6594202898550723e-05, "loss": 1.118, "step": 367 }, { "epoch": 0.16674218396012688, "grad_norm": 0.7351378392007826, "learning_rate": 2.6666666666666667e-05, "loss": 1.1063, "step": 368 }, { "epoch": 0.1671952877208881, "grad_norm": 0.705317235779178, "learning_rate": 2.673913043478261e-05, "loss": 1.1171, "step": 369 }, { "epoch": 0.1676483914816493, "grad_norm": 0.5867533557722551, "learning_rate": 2.6811594202898553e-05, "loss": 1.1055, "step": 370 }, { "epoch": 0.16810149524241053, "grad_norm": 0.5115246296297361, "learning_rate": 2.6884057971014496e-05, "loss": 1.1274, "step": 371 }, { "epoch": 0.16855459900317174, "grad_norm": 0.48185872557426856, "learning_rate": 2.6956521739130436e-05, "loss": 1.1263, "step": 372 }, { "epoch": 0.16900770276393295, "grad_norm": 0.4579600744411017, "learning_rate": 2.702898550724638e-05, "loss": 1.1058, "step": 373 }, { "epoch": 0.16946080652469417, "grad_norm": 0.40439319210223557, "learning_rate": 2.7101449275362322e-05, "loss": 1.1169, "step": 374 }, { "epoch": 0.16991391028545538, "grad_norm": 0.3603368026041047, "learning_rate": 2.7173913043478265e-05, "loss": 1.1128, "step": 375 }, { "epoch": 0.1703670140462166, "grad_norm": 0.37008136559488614, "learning_rate": 2.7246376811594205e-05, "loss": 1.106, "step": 376 }, { "epoch": 0.1708201178069778, "grad_norm": 0.3682742573590215, "learning_rate": 2.7318840579710148e-05, "loss": 1.0877, "step": 377 }, { "epoch": 0.17127322156773903, "grad_norm": 0.3314733461906554, "learning_rate": 2.739130434782609e-05, "loss": 1.1055, "step": 378 }, { "epoch": 0.17172632532850024, "grad_norm": 0.31558099638654535, "learning_rate": 2.7463768115942034e-05, "loss": 1.0989, "step": 379 }, { "epoch": 0.17217942908926145, "grad_norm": 0.32871649483562043, "learning_rate": 2.753623188405797e-05, "loss": 1.1024, "step": 380 }, { "epoch": 0.17263253285002267, "grad_norm": 0.459148846179919, "learning_rate": 2.7608695652173914e-05, "loss": 1.1064, "step": 381 }, { "epoch": 0.17308563661078388, "grad_norm": 0.5649763296051244, "learning_rate": 2.7681159420289857e-05, "loss": 1.1014, "step": 382 }, { "epoch": 0.1735387403715451, "grad_norm": 0.579798703104836, "learning_rate": 2.7753623188405797e-05, "loss": 1.1115, "step": 383 }, { "epoch": 0.1739918441323063, "grad_norm": 0.587394500004828, "learning_rate": 2.782608695652174e-05, "loss": 1.0916, "step": 384 }, { "epoch": 0.17444494789306753, "grad_norm": 0.6842610113377202, "learning_rate": 2.7898550724637683e-05, "loss": 1.0833, "step": 385 }, { "epoch": 0.17489805165382874, "grad_norm": 0.8718213258268389, "learning_rate": 2.7971014492753626e-05, "loss": 1.1019, "step": 386 }, { "epoch": 0.17535115541458995, "grad_norm": 1.0106562861318078, "learning_rate": 2.8043478260869566e-05, "loss": 1.1168, "step": 387 }, { "epoch": 0.17580425917535117, "grad_norm": 0.9157708097124203, "learning_rate": 2.811594202898551e-05, "loss": 1.0914, "step": 388 }, { "epoch": 0.17625736293611238, "grad_norm": 0.6313092043944087, "learning_rate": 2.8188405797101452e-05, "loss": 1.1242, "step": 389 }, { "epoch": 0.1767104666968736, "grad_norm": 0.4333045702705286, "learning_rate": 2.8260869565217396e-05, "loss": 1.1019, "step": 390 }, { "epoch": 0.1771635704576348, "grad_norm": 0.40634950126831754, "learning_rate": 2.833333333333334e-05, "loss": 1.0953, "step": 391 }, { "epoch": 0.17761667421839603, "grad_norm": 0.501920751525023, "learning_rate": 2.840579710144928e-05, "loss": 1.1185, "step": 392 }, { "epoch": 0.17806977797915724, "grad_norm": 0.5183084148005345, "learning_rate": 2.847826086956522e-05, "loss": 1.0854, "step": 393 }, { "epoch": 0.17852288173991845, "grad_norm": 0.4266582322773056, "learning_rate": 2.855072463768116e-05, "loss": 1.1076, "step": 394 }, { "epoch": 0.17897598550067967, "grad_norm": 0.42907741339006494, "learning_rate": 2.86231884057971e-05, "loss": 1.1108, "step": 395 }, { "epoch": 0.17942908926144088, "grad_norm": 0.5811082096952249, "learning_rate": 2.8695652173913044e-05, "loss": 1.1015, "step": 396 }, { "epoch": 0.1798821930222021, "grad_norm": 0.6184904270376792, "learning_rate": 2.8768115942028988e-05, "loss": 1.1169, "step": 397 }, { "epoch": 0.1803352967829633, "grad_norm": 0.5719117901936359, "learning_rate": 2.884057971014493e-05, "loss": 1.1036, "step": 398 }, { "epoch": 0.18078840054372453, "grad_norm": 0.4843673721894502, "learning_rate": 2.891304347826087e-05, "loss": 1.1074, "step": 399 }, { "epoch": 0.18124150430448574, "grad_norm": 0.5221524381420943, "learning_rate": 2.8985507246376814e-05, "loss": 1.1156, "step": 400 }, { "epoch": 0.18169460806524695, "grad_norm": 0.5697966663842249, "learning_rate": 2.9057971014492757e-05, "loss": 1.0967, "step": 401 }, { "epoch": 0.18214771182600817, "grad_norm": 0.5814312996496311, "learning_rate": 2.91304347826087e-05, "loss": 1.1058, "step": 402 }, { "epoch": 0.18260081558676938, "grad_norm": 0.6101589193369912, "learning_rate": 2.920289855072464e-05, "loss": 1.1088, "step": 403 }, { "epoch": 0.1830539193475306, "grad_norm": 0.6656098748214163, "learning_rate": 2.9275362318840583e-05, "loss": 1.1005, "step": 404 }, { "epoch": 0.1835070231082918, "grad_norm": 0.6737849836100817, "learning_rate": 2.9347826086956526e-05, "loss": 1.0981, "step": 405 }, { "epoch": 0.18396012686905303, "grad_norm": 0.6562269067177825, "learning_rate": 2.942028985507247e-05, "loss": 1.1107, "step": 406 }, { "epoch": 0.18441323062981424, "grad_norm": 0.5708979923724022, "learning_rate": 2.949275362318841e-05, "loss": 1.1082, "step": 407 }, { "epoch": 0.18486633439057545, "grad_norm": 0.5124954152537372, "learning_rate": 2.956521739130435e-05, "loss": 1.0944, "step": 408 }, { "epoch": 0.18531943815133667, "grad_norm": 0.5042069560472442, "learning_rate": 2.9637681159420292e-05, "loss": 1.0696, "step": 409 }, { "epoch": 0.18577254191209788, "grad_norm": 0.5641130003133485, "learning_rate": 2.971014492753623e-05, "loss": 1.0922, "step": 410 }, { "epoch": 0.1862256456728591, "grad_norm": 0.668121533143444, "learning_rate": 2.9782608695652175e-05, "loss": 1.1136, "step": 411 }, { "epoch": 0.1866787494336203, "grad_norm": 0.688229558939251, "learning_rate": 2.9855072463768118e-05, "loss": 1.1013, "step": 412 }, { "epoch": 0.18713185319438153, "grad_norm": 0.7071656447143172, "learning_rate": 2.992753623188406e-05, "loss": 1.1165, "step": 413 }, { "epoch": 0.18758495695514274, "grad_norm": 0.745052727769418, "learning_rate": 3.0000000000000004e-05, "loss": 1.135, "step": 414 }, { "epoch": 0.18803806071590395, "grad_norm": 0.8220717528793962, "learning_rate": 3.0072463768115944e-05, "loss": 1.1109, "step": 415 }, { "epoch": 0.18849116447666517, "grad_norm": 0.7597777953277616, "learning_rate": 3.0144927536231887e-05, "loss": 1.1072, "step": 416 }, { "epoch": 0.18894426823742638, "grad_norm": 0.6167505058645958, "learning_rate": 3.021739130434783e-05, "loss": 1.0748, "step": 417 }, { "epoch": 0.1893973719981876, "grad_norm": 0.5071882926451573, "learning_rate": 3.0289855072463773e-05, "loss": 1.1177, "step": 418 }, { "epoch": 0.1898504757589488, "grad_norm": 0.4659063608724231, "learning_rate": 3.0362318840579713e-05, "loss": 1.1001, "step": 419 }, { "epoch": 0.19030357951971003, "grad_norm": 0.4329385133486572, "learning_rate": 3.0434782608695656e-05, "loss": 1.1159, "step": 420 }, { "epoch": 0.19075668328047124, "grad_norm": 0.47775397276257003, "learning_rate": 3.0507246376811596e-05, "loss": 1.0905, "step": 421 }, { "epoch": 0.19120978704123245, "grad_norm": 0.5819223964380112, "learning_rate": 3.0579710144927536e-05, "loss": 1.1111, "step": 422 }, { "epoch": 0.19166289080199367, "grad_norm": 0.6709461173105644, "learning_rate": 3.065217391304348e-05, "loss": 1.1228, "step": 423 }, { "epoch": 0.19211599456275488, "grad_norm": 0.5686922776503637, "learning_rate": 3.072463768115942e-05, "loss": 1.0957, "step": 424 }, { "epoch": 0.1925690983235161, "grad_norm": 0.4112652277631654, "learning_rate": 3.079710144927536e-05, "loss": 1.1052, "step": 425 }, { "epoch": 0.1930222020842773, "grad_norm": 0.4817179743705203, "learning_rate": 3.086956521739131e-05, "loss": 1.0974, "step": 426 }, { "epoch": 0.19347530584503853, "grad_norm": 0.5359970820523571, "learning_rate": 3.094202898550725e-05, "loss": 1.0956, "step": 427 }, { "epoch": 0.19392840960579974, "grad_norm": 0.5265027174446076, "learning_rate": 3.101449275362319e-05, "loss": 1.1096, "step": 428 }, { "epoch": 0.19438151336656095, "grad_norm": 0.4816125968005017, "learning_rate": 3.1086956521739134e-05, "loss": 1.0836, "step": 429 }, { "epoch": 0.19483461712732217, "grad_norm": 0.48191228166741784, "learning_rate": 3.1159420289855074e-05, "loss": 1.0745, "step": 430 }, { "epoch": 0.19528772088808338, "grad_norm": 0.5763684314289931, "learning_rate": 3.123188405797102e-05, "loss": 1.082, "step": 431 }, { "epoch": 0.1957408246488446, "grad_norm": 0.584832876047268, "learning_rate": 3.130434782608696e-05, "loss": 1.0967, "step": 432 }, { "epoch": 0.1961939284096058, "grad_norm": 0.5107134713154599, "learning_rate": 3.13768115942029e-05, "loss": 1.0925, "step": 433 }, { "epoch": 0.19664703217036703, "grad_norm": 0.484210836862659, "learning_rate": 3.144927536231885e-05, "loss": 1.0955, "step": 434 }, { "epoch": 0.19710013593112824, "grad_norm": 0.5165520665012632, "learning_rate": 3.152173913043478e-05, "loss": 1.0971, "step": 435 }, { "epoch": 0.19755323969188945, "grad_norm": 0.4963928992898546, "learning_rate": 3.1594202898550726e-05, "loss": 1.0884, "step": 436 }, { "epoch": 0.19800634345265067, "grad_norm": 0.5011065275533206, "learning_rate": 3.1666666666666666e-05, "loss": 1.0651, "step": 437 }, { "epoch": 0.19845944721341188, "grad_norm": 0.5023306093753839, "learning_rate": 3.173913043478261e-05, "loss": 1.0924, "step": 438 }, { "epoch": 0.1989125509741731, "grad_norm": 0.5601433138034749, "learning_rate": 3.181159420289855e-05, "loss": 1.0947, "step": 439 }, { "epoch": 0.1993656547349343, "grad_norm": 0.6327810321133179, "learning_rate": 3.188405797101449e-05, "loss": 1.087, "step": 440 }, { "epoch": 0.19981875849569553, "grad_norm": 0.6670073237720388, "learning_rate": 3.195652173913044e-05, "loss": 1.0918, "step": 441 }, { "epoch": 0.20027186225645674, "grad_norm": 0.5176775520601092, "learning_rate": 3.202898550724638e-05, "loss": 1.1028, "step": 442 }, { "epoch": 0.20072496601721795, "grad_norm": 0.3716186858928922, "learning_rate": 3.2101449275362325e-05, "loss": 1.1248, "step": 443 }, { "epoch": 0.20117806977797917, "grad_norm": 0.35834643082952067, "learning_rate": 3.2173913043478265e-05, "loss": 1.0967, "step": 444 }, { "epoch": 0.20163117353874038, "grad_norm": 0.47421430350293897, "learning_rate": 3.2246376811594205e-05, "loss": 1.0973, "step": 445 }, { "epoch": 0.2020842772995016, "grad_norm": 0.49588649922100925, "learning_rate": 3.231884057971015e-05, "loss": 1.0968, "step": 446 }, { "epoch": 0.2025373810602628, "grad_norm": 0.47348053719500854, "learning_rate": 3.239130434782609e-05, "loss": 1.0947, "step": 447 }, { "epoch": 0.20299048482102403, "grad_norm": 0.4059559515642778, "learning_rate": 3.246376811594203e-05, "loss": 1.0998, "step": 448 }, { "epoch": 0.20344358858178524, "grad_norm": 0.40463331028112987, "learning_rate": 3.253623188405797e-05, "loss": 1.1007, "step": 449 }, { "epoch": 0.20389669234254645, "grad_norm": 0.5099429040716631, "learning_rate": 3.260869565217392e-05, "loss": 1.0741, "step": 450 }, { "epoch": 0.20434979610330767, "grad_norm": 0.6080286311919126, "learning_rate": 3.268115942028986e-05, "loss": 1.0976, "step": 451 }, { "epoch": 0.20480289986406888, "grad_norm": 0.7398488053266316, "learning_rate": 3.2753623188405796e-05, "loss": 1.0893, "step": 452 }, { "epoch": 0.2052560036248301, "grad_norm": 0.8797715256856615, "learning_rate": 3.282608695652174e-05, "loss": 1.1197, "step": 453 }, { "epoch": 0.2057091073855913, "grad_norm": 1.002473911727178, "learning_rate": 3.289855072463768e-05, "loss": 1.1048, "step": 454 }, { "epoch": 0.20616221114635253, "grad_norm": 1.0470088439025078, "learning_rate": 3.297101449275362e-05, "loss": 1.0852, "step": 455 }, { "epoch": 0.20661531490711374, "grad_norm": 0.848320677091367, "learning_rate": 3.304347826086957e-05, "loss": 1.0775, "step": 456 }, { "epoch": 0.20706841866787495, "grad_norm": 0.5904974458080151, "learning_rate": 3.311594202898551e-05, "loss": 1.103, "step": 457 }, { "epoch": 0.20752152242863617, "grad_norm": 0.4364340924956402, "learning_rate": 3.3188405797101455e-05, "loss": 1.0944, "step": 458 }, { "epoch": 0.20797462618939738, "grad_norm": 0.53403032015375, "learning_rate": 3.3260869565217395e-05, "loss": 1.1058, "step": 459 }, { "epoch": 0.2084277299501586, "grad_norm": 0.6168024365119233, "learning_rate": 3.3333333333333335e-05, "loss": 1.0952, "step": 460 }, { "epoch": 0.2088808337109198, "grad_norm": 0.6217144526098946, "learning_rate": 3.340579710144928e-05, "loss": 1.0735, "step": 461 }, { "epoch": 0.20933393747168103, "grad_norm": 0.5740748334299386, "learning_rate": 3.347826086956522e-05, "loss": 1.0928, "step": 462 }, { "epoch": 0.20978704123244224, "grad_norm": 0.39766520106437275, "learning_rate": 3.355072463768116e-05, "loss": 1.0878, "step": 463 }, { "epoch": 0.21024014499320345, "grad_norm": 0.42421121453422583, "learning_rate": 3.36231884057971e-05, "loss": 1.0818, "step": 464 }, { "epoch": 0.21069324875396467, "grad_norm": 0.49589591518846554, "learning_rate": 3.369565217391305e-05, "loss": 1.0958, "step": 465 }, { "epoch": 0.21114635251472588, "grad_norm": 0.4170301097215545, "learning_rate": 3.376811594202899e-05, "loss": 1.1025, "step": 466 }, { "epoch": 0.2115994562754871, "grad_norm": 0.3428449515947966, "learning_rate": 3.384057971014493e-05, "loss": 1.1041, "step": 467 }, { "epoch": 0.2120525600362483, "grad_norm": 0.33890835158121735, "learning_rate": 3.391304347826087e-05, "loss": 1.1021, "step": 468 }, { "epoch": 0.21250566379700953, "grad_norm": 0.3878509534334297, "learning_rate": 3.398550724637681e-05, "loss": 1.0978, "step": 469 }, { "epoch": 0.21295876755777074, "grad_norm": 0.5541060519725876, "learning_rate": 3.405797101449276e-05, "loss": 1.0954, "step": 470 }, { "epoch": 0.21341187131853195, "grad_norm": 0.7527517406407053, "learning_rate": 3.41304347826087e-05, "loss": 1.0966, "step": 471 }, { "epoch": 0.21386497507929317, "grad_norm": 0.8661747107458745, "learning_rate": 3.420289855072464e-05, "loss": 1.0887, "step": 472 }, { "epoch": 0.21431807884005438, "grad_norm": 0.9153412177747127, "learning_rate": 3.4275362318840586e-05, "loss": 1.0886, "step": 473 }, { "epoch": 0.2147711826008156, "grad_norm": 0.9731706362809966, "learning_rate": 3.4347826086956526e-05, "loss": 1.0895, "step": 474 }, { "epoch": 0.2152242863615768, "grad_norm": 1.0461561689783863, "learning_rate": 3.4420289855072465e-05, "loss": 1.0972, "step": 475 }, { "epoch": 0.21567739012233803, "grad_norm": 0.8030176026501755, "learning_rate": 3.449275362318841e-05, "loss": 1.0853, "step": 476 }, { "epoch": 0.21613049388309924, "grad_norm": 0.7035264476542539, "learning_rate": 3.456521739130435e-05, "loss": 1.0917, "step": 477 }, { "epoch": 0.21658359764386045, "grad_norm": 0.8390663486834974, "learning_rate": 3.463768115942029e-05, "loss": 1.0732, "step": 478 }, { "epoch": 0.21703670140462167, "grad_norm": 0.9370011897270483, "learning_rate": 3.471014492753623e-05, "loss": 1.1054, "step": 479 }, { "epoch": 0.21748980516538288, "grad_norm": 0.8951443744000418, "learning_rate": 3.478260869565218e-05, "loss": 1.1006, "step": 480 }, { "epoch": 0.2179429089261441, "grad_norm": 0.6493832651622122, "learning_rate": 3.485507246376812e-05, "loss": 1.0814, "step": 481 }, { "epoch": 0.2183960126869053, "grad_norm": 0.3653463428787675, "learning_rate": 3.492753623188406e-05, "loss": 1.0924, "step": 482 }, { "epoch": 0.21884911644766653, "grad_norm": 0.6303838514521991, "learning_rate": 3.5000000000000004e-05, "loss": 1.0782, "step": 483 }, { "epoch": 0.21930222020842774, "grad_norm": 0.8153956567795502, "learning_rate": 3.5072463768115943e-05, "loss": 1.0942, "step": 484 }, { "epoch": 0.21975532396918895, "grad_norm": 0.7370351086956682, "learning_rate": 3.514492753623189e-05, "loss": 1.1014, "step": 485 }, { "epoch": 0.22020842772995017, "grad_norm": 0.5584931591184623, "learning_rate": 3.521739130434783e-05, "loss": 1.0998, "step": 486 }, { "epoch": 0.22066153149071138, "grad_norm": 0.45148985877662773, "learning_rate": 3.528985507246377e-05, "loss": 1.0731, "step": 487 }, { "epoch": 0.2211146352514726, "grad_norm": 0.3645421629313368, "learning_rate": 3.5362318840579716e-05, "loss": 1.093, "step": 488 }, { "epoch": 0.2215677390122338, "grad_norm": 0.4294012916133087, "learning_rate": 3.5434782608695656e-05, "loss": 1.0665, "step": 489 }, { "epoch": 0.22202084277299503, "grad_norm": 0.5324392517151401, "learning_rate": 3.5507246376811596e-05, "loss": 1.0852, "step": 490 }, { "epoch": 0.22247394653375624, "grad_norm": 0.579528423406382, "learning_rate": 3.5579710144927535e-05, "loss": 1.0989, "step": 491 }, { "epoch": 0.22292705029451745, "grad_norm": 0.546111707496591, "learning_rate": 3.565217391304348e-05, "loss": 1.0899, "step": 492 }, { "epoch": 0.22338015405527867, "grad_norm": 0.5335516012557613, "learning_rate": 3.572463768115942e-05, "loss": 1.0864, "step": 493 }, { "epoch": 0.22383325781603988, "grad_norm": 0.565284995769076, "learning_rate": 3.579710144927536e-05, "loss": 1.0923, "step": 494 }, { "epoch": 0.2242863615768011, "grad_norm": 0.5908051585776392, "learning_rate": 3.586956521739131e-05, "loss": 1.0961, "step": 495 }, { "epoch": 0.2247394653375623, "grad_norm": 0.5639870855092004, "learning_rate": 3.594202898550725e-05, "loss": 1.0772, "step": 496 }, { "epoch": 0.22519256909832353, "grad_norm": 0.6090935332239725, "learning_rate": 3.6014492753623194e-05, "loss": 1.1004, "step": 497 }, { "epoch": 0.22564567285908474, "grad_norm": 0.7257488531791495, "learning_rate": 3.6086956521739134e-05, "loss": 1.0936, "step": 498 }, { "epoch": 0.22609877661984595, "grad_norm": 0.7935624280192481, "learning_rate": 3.6159420289855074e-05, "loss": 1.0846, "step": 499 }, { "epoch": 0.22655188038060717, "grad_norm": 0.7739750861287609, "learning_rate": 3.623188405797102e-05, "loss": 1.0993, "step": 500 }, { "epoch": 0.22700498414136838, "grad_norm": 0.6440375210765827, "learning_rate": 3.630434782608696e-05, "loss": 1.1102, "step": 501 }, { "epoch": 0.2274580879021296, "grad_norm": 0.4824264273040651, "learning_rate": 3.63768115942029e-05, "loss": 1.0951, "step": 502 }, { "epoch": 0.2279111916628908, "grad_norm": 0.43665785943669544, "learning_rate": 3.6449275362318846e-05, "loss": 1.1022, "step": 503 }, { "epoch": 0.22836429542365203, "grad_norm": 0.601916814983532, "learning_rate": 3.6521739130434786e-05, "loss": 1.0819, "step": 504 }, { "epoch": 0.22881739918441324, "grad_norm": 0.7859744263994496, "learning_rate": 3.6594202898550726e-05, "loss": 1.0933, "step": 505 }, { "epoch": 0.22927050294517445, "grad_norm": 0.6524466304094945, "learning_rate": 3.6666666666666666e-05, "loss": 1.0942, "step": 506 }, { "epoch": 0.22972360670593567, "grad_norm": 0.50620094967601, "learning_rate": 3.673913043478261e-05, "loss": 1.0862, "step": 507 }, { "epoch": 0.23017671046669688, "grad_norm": 0.5126101904299825, "learning_rate": 3.681159420289855e-05, "loss": 1.1035, "step": 508 }, { "epoch": 0.2306298142274581, "grad_norm": 0.49677518332741155, "learning_rate": 3.68840579710145e-05, "loss": 1.0957, "step": 509 }, { "epoch": 0.2310829179882193, "grad_norm": 0.4612753440640858, "learning_rate": 3.695652173913044e-05, "loss": 1.1026, "step": 510 }, { "epoch": 0.23153602174898053, "grad_norm": 0.487096621912835, "learning_rate": 3.702898550724638e-05, "loss": 1.107, "step": 511 }, { "epoch": 0.23198912550974174, "grad_norm": 0.5510446142967612, "learning_rate": 3.7101449275362325e-05, "loss": 1.0864, "step": 512 }, { "epoch": 0.23244222927050295, "grad_norm": 0.5288742360693077, "learning_rate": 3.7173913043478264e-05, "loss": 1.0799, "step": 513 }, { "epoch": 0.23289533303126417, "grad_norm": 0.5316724090153732, "learning_rate": 3.7246376811594204e-05, "loss": 1.0898, "step": 514 }, { "epoch": 0.23334843679202538, "grad_norm": 0.5120825138939964, "learning_rate": 3.731884057971015e-05, "loss": 1.094, "step": 515 }, { "epoch": 0.2338015405527866, "grad_norm": 0.436629842714973, "learning_rate": 3.739130434782609e-05, "loss": 1.0737, "step": 516 }, { "epoch": 0.2342546443135478, "grad_norm": 0.428126122162116, "learning_rate": 3.746376811594204e-05, "loss": 1.0985, "step": 517 }, { "epoch": 0.23470774807430902, "grad_norm": 0.43317059038610833, "learning_rate": 3.753623188405797e-05, "loss": 1.0871, "step": 518 }, { "epoch": 0.23516085183507024, "grad_norm": 0.4651633251122593, "learning_rate": 3.7608695652173917e-05, "loss": 1.0745, "step": 519 }, { "epoch": 0.23561395559583145, "grad_norm": 0.43289860950924886, "learning_rate": 3.7681159420289856e-05, "loss": 1.1016, "step": 520 }, { "epoch": 0.23606705935659267, "grad_norm": 0.45032452926061944, "learning_rate": 3.7753623188405796e-05, "loss": 1.0893, "step": 521 }, { "epoch": 0.23652016311735388, "grad_norm": 0.45946983406446995, "learning_rate": 3.782608695652174e-05, "loss": 1.081, "step": 522 }, { "epoch": 0.2369732668781151, "grad_norm": 0.521873243960875, "learning_rate": 3.789855072463768e-05, "loss": 1.0977, "step": 523 }, { "epoch": 0.2374263706388763, "grad_norm": 0.5743981896118561, "learning_rate": 3.797101449275363e-05, "loss": 1.091, "step": 524 }, { "epoch": 0.23787947439963752, "grad_norm": 0.5972110590826507, "learning_rate": 3.804347826086957e-05, "loss": 1.0989, "step": 525 }, { "epoch": 0.23833257816039874, "grad_norm": 0.6364949107431003, "learning_rate": 3.811594202898551e-05, "loss": 1.0863, "step": 526 }, { "epoch": 0.23878568192115995, "grad_norm": 0.702906956021964, "learning_rate": 3.8188405797101455e-05, "loss": 1.107, "step": 527 }, { "epoch": 0.23923878568192117, "grad_norm": 0.7638191855253091, "learning_rate": 3.8260869565217395e-05, "loss": 1.0858, "step": 528 }, { "epoch": 0.23969188944268238, "grad_norm": 0.7627722828171569, "learning_rate": 3.833333333333334e-05, "loss": 1.0657, "step": 529 }, { "epoch": 0.2401449932034436, "grad_norm": 0.7779865954835913, "learning_rate": 3.840579710144928e-05, "loss": 1.0573, "step": 530 }, { "epoch": 0.2405980969642048, "grad_norm": 0.8551756279525252, "learning_rate": 3.847826086956522e-05, "loss": 1.0966, "step": 531 }, { "epoch": 0.24105120072496602, "grad_norm": 1.0146014264165442, "learning_rate": 3.855072463768116e-05, "loss": 1.0747, "step": 532 }, { "epoch": 0.24150430448572724, "grad_norm": 0.9960058548644926, "learning_rate": 3.86231884057971e-05, "loss": 1.0658, "step": 533 }, { "epoch": 0.24195740824648845, "grad_norm": 0.853369922576039, "learning_rate": 3.869565217391305e-05, "loss": 1.0754, "step": 534 }, { "epoch": 0.24241051200724967, "grad_norm": 0.7266336764170821, "learning_rate": 3.876811594202899e-05, "loss": 1.0917, "step": 535 }, { "epoch": 0.24286361576801088, "grad_norm": 0.7777172788032874, "learning_rate": 3.884057971014493e-05, "loss": 1.0911, "step": 536 }, { "epoch": 0.2433167195287721, "grad_norm": 0.9707538109455647, "learning_rate": 3.891304347826087e-05, "loss": 1.0824, "step": 537 }, { "epoch": 0.2437698232895333, "grad_norm": 0.9966505607858782, "learning_rate": 3.898550724637681e-05, "loss": 1.0703, "step": 538 }, { "epoch": 0.24422292705029452, "grad_norm": 0.9406944936945504, "learning_rate": 3.905797101449276e-05, "loss": 1.0997, "step": 539 }, { "epoch": 0.24467603081105574, "grad_norm": 0.7923601574993352, "learning_rate": 3.91304347826087e-05, "loss": 1.0922, "step": 540 }, { "epoch": 0.24512913457181695, "grad_norm": 0.5903599717715584, "learning_rate": 3.920289855072464e-05, "loss": 1.0928, "step": 541 }, { "epoch": 0.24558223833257817, "grad_norm": 0.42901471680784614, "learning_rate": 3.9275362318840585e-05, "loss": 1.0566, "step": 542 }, { "epoch": 0.24603534209333938, "grad_norm": 0.5941811997621521, "learning_rate": 3.9347826086956525e-05, "loss": 1.0826, "step": 543 }, { "epoch": 0.2464884458541006, "grad_norm": 0.5391746415732145, "learning_rate": 3.942028985507247e-05, "loss": 1.0932, "step": 544 }, { "epoch": 0.2469415496148618, "grad_norm": 0.37624278652909593, "learning_rate": 3.949275362318841e-05, "loss": 1.0761, "step": 545 }, { "epoch": 0.24739465337562302, "grad_norm": 0.45297366528819294, "learning_rate": 3.956521739130435e-05, "loss": 1.0942, "step": 546 }, { "epoch": 0.24784775713638424, "grad_norm": 0.4300991892557049, "learning_rate": 3.963768115942029e-05, "loss": 1.0711, "step": 547 }, { "epoch": 0.24830086089714545, "grad_norm": 0.3867302140377158, "learning_rate": 3.971014492753623e-05, "loss": 1.0954, "step": 548 }, { "epoch": 0.24875396465790667, "grad_norm": 0.46766967468539156, "learning_rate": 3.978260869565218e-05, "loss": 1.0665, "step": 549 }, { "epoch": 0.24920706841866788, "grad_norm": 0.4252189679423843, "learning_rate": 3.985507246376812e-05, "loss": 1.1006, "step": 550 }, { "epoch": 0.2496601721794291, "grad_norm": 0.31284703572040384, "learning_rate": 3.9927536231884064e-05, "loss": 1.0656, "step": 551 }, { "epoch": 0.2501132759401903, "grad_norm": 0.3956383571535045, "learning_rate": 4e-05, "loss": 1.0906, "step": 552 }, { "epoch": 0.2505663797009515, "grad_norm": 0.3933055193101051, "learning_rate": 4.007246376811594e-05, "loss": 1.0788, "step": 553 }, { "epoch": 0.2510194834617127, "grad_norm": 0.33629901622443015, "learning_rate": 4.014492753623189e-05, "loss": 1.0431, "step": 554 }, { "epoch": 0.25147258722247395, "grad_norm": 0.38090172060890337, "learning_rate": 4.021739130434782e-05, "loss": 1.0873, "step": 555 }, { "epoch": 0.25192569098323514, "grad_norm": 0.44029560093981, "learning_rate": 4.0289855072463776e-05, "loss": 1.075, "step": 556 }, { "epoch": 0.2523787947439964, "grad_norm": 0.4814883004244345, "learning_rate": 4.036231884057971e-05, "loss": 1.0714, "step": 557 }, { "epoch": 0.25283189850475757, "grad_norm": 0.5014544598056832, "learning_rate": 4.0434782608695655e-05, "loss": 1.0706, "step": 558 }, { "epoch": 0.2532850022655188, "grad_norm": 0.40697766361958887, "learning_rate": 4.0507246376811595e-05, "loss": 1.0745, "step": 559 }, { "epoch": 0.25373810602628, "grad_norm": 0.40095261745093214, "learning_rate": 4.057971014492754e-05, "loss": 1.0402, "step": 560 }, { "epoch": 0.25419120978704124, "grad_norm": 0.4333371262266911, "learning_rate": 4.065217391304348e-05, "loss": 1.0856, "step": 561 }, { "epoch": 0.2546443135478024, "grad_norm": 0.5125429649197069, "learning_rate": 4.072463768115943e-05, "loss": 1.077, "step": 562 }, { "epoch": 0.25509741730856367, "grad_norm": 0.5507429232819517, "learning_rate": 4.079710144927537e-05, "loss": 1.0603, "step": 563 }, { "epoch": 0.25555052106932485, "grad_norm": 0.6248644914090947, "learning_rate": 4.0869565217391314e-05, "loss": 1.0891, "step": 564 }, { "epoch": 0.2560036248300861, "grad_norm": 0.6845970752460667, "learning_rate": 4.094202898550725e-05, "loss": 1.0835, "step": 565 }, { "epoch": 0.2564567285908473, "grad_norm": 0.5342218733640833, "learning_rate": 4.101449275362319e-05, "loss": 1.1044, "step": 566 }, { "epoch": 0.2569098323516085, "grad_norm": 0.41968444096163954, "learning_rate": 4.1086956521739134e-05, "loss": 1.0756, "step": 567 }, { "epoch": 0.2573629361123697, "grad_norm": 0.5457642671728765, "learning_rate": 4.115942028985507e-05, "loss": 1.0837, "step": 568 }, { "epoch": 0.25781603987313095, "grad_norm": 0.8500279881649687, "learning_rate": 4.123188405797102e-05, "loss": 1.0762, "step": 569 }, { "epoch": 0.25826914363389214, "grad_norm": 1.2196644705920587, "learning_rate": 4.130434782608696e-05, "loss": 1.0592, "step": 570 }, { "epoch": 0.2587222473946534, "grad_norm": 0.8183370616041521, "learning_rate": 4.1376811594202906e-05, "loss": 1.0866, "step": 571 }, { "epoch": 0.25917535115541457, "grad_norm": 0.7467960549294256, "learning_rate": 4.144927536231884e-05, "loss": 1.0901, "step": 572 }, { "epoch": 0.2596284549161758, "grad_norm": 0.7684442842148952, "learning_rate": 4.1521739130434786e-05, "loss": 1.0775, "step": 573 }, { "epoch": 0.260081558676937, "grad_norm": 0.765194602197457, "learning_rate": 4.1594202898550726e-05, "loss": 1.0765, "step": 574 }, { "epoch": 0.26053466243769824, "grad_norm": 0.6852045129479882, "learning_rate": 4.166666666666667e-05, "loss": 1.098, "step": 575 }, { "epoch": 0.2609877661984594, "grad_norm": 0.5797983841364392, "learning_rate": 4.173913043478261e-05, "loss": 1.0738, "step": 576 }, { "epoch": 0.26144086995922067, "grad_norm": 0.6627042181871821, "learning_rate": 4.181159420289856e-05, "loss": 1.0798, "step": 577 }, { "epoch": 0.26189397371998185, "grad_norm": 0.7903287621581491, "learning_rate": 4.18840579710145e-05, "loss": 1.0901, "step": 578 }, { "epoch": 0.2623470774807431, "grad_norm": 0.7297627768398958, "learning_rate": 4.1956521739130445e-05, "loss": 1.0629, "step": 579 }, { "epoch": 0.2628001812415043, "grad_norm": 0.5142355886011549, "learning_rate": 4.202898550724638e-05, "loss": 1.0795, "step": 580 }, { "epoch": 0.2632532850022655, "grad_norm": 0.39982868057602344, "learning_rate": 4.210144927536232e-05, "loss": 1.053, "step": 581 }, { "epoch": 0.2637063887630267, "grad_norm": 0.45063460149475004, "learning_rate": 4.2173913043478264e-05, "loss": 1.0872, "step": 582 }, { "epoch": 0.26415949252378795, "grad_norm": 0.526146033641332, "learning_rate": 4.2246376811594204e-05, "loss": 1.0724, "step": 583 }, { "epoch": 0.26461259628454914, "grad_norm": 0.5185797773438727, "learning_rate": 4.231884057971015e-05, "loss": 1.0689, "step": 584 }, { "epoch": 0.2650657000453104, "grad_norm": 0.5373106565042397, "learning_rate": 4.239130434782609e-05, "loss": 1.0872, "step": 585 }, { "epoch": 0.26551880380607157, "grad_norm": 0.6861381083376445, "learning_rate": 4.2463768115942037e-05, "loss": 1.0732, "step": 586 }, { "epoch": 0.2659719075668328, "grad_norm": 0.8895017170903273, "learning_rate": 4.253623188405797e-05, "loss": 1.084, "step": 587 }, { "epoch": 0.266425011327594, "grad_norm": 1.0605007763168797, "learning_rate": 4.2608695652173916e-05, "loss": 1.0839, "step": 588 }, { "epoch": 0.26687811508835524, "grad_norm": 0.8146738357259502, "learning_rate": 4.2681159420289856e-05, "loss": 1.0809, "step": 589 }, { "epoch": 0.2673312188491164, "grad_norm": 0.5594191469371261, "learning_rate": 4.27536231884058e-05, "loss": 1.0773, "step": 590 }, { "epoch": 0.26778432260987767, "grad_norm": 0.7086336660911432, "learning_rate": 4.282608695652174e-05, "loss": 1.1031, "step": 591 }, { "epoch": 0.26823742637063885, "grad_norm": 0.8946620164162641, "learning_rate": 4.289855072463769e-05, "loss": 1.089, "step": 592 }, { "epoch": 0.2686905301314001, "grad_norm": 0.740257807144605, "learning_rate": 4.297101449275363e-05, "loss": 1.0222, "step": 593 }, { "epoch": 0.2691436338921613, "grad_norm": 0.6767591650340982, "learning_rate": 4.304347826086956e-05, "loss": 1.0573, "step": 594 }, { "epoch": 0.2695967376529225, "grad_norm": 0.9165299607959675, "learning_rate": 4.311594202898551e-05, "loss": 1.0864, "step": 595 }, { "epoch": 0.2700498414136837, "grad_norm": 1.1565988258420652, "learning_rate": 4.318840579710145e-05, "loss": 1.094, "step": 596 }, { "epoch": 0.27050294517444495, "grad_norm": 0.7722750452884772, "learning_rate": 4.3260869565217394e-05, "loss": 1.0663, "step": 597 }, { "epoch": 0.27095604893520614, "grad_norm": 0.7387103965710093, "learning_rate": 4.3333333333333334e-05, "loss": 1.0879, "step": 598 }, { "epoch": 0.2714091526959674, "grad_norm": 0.7141911766304255, "learning_rate": 4.340579710144928e-05, "loss": 1.0693, "step": 599 }, { "epoch": 0.27186225645672857, "grad_norm": 0.5598213198893272, "learning_rate": 4.347826086956522e-05, "loss": 1.0511, "step": 600 }, { "epoch": 0.2723153602174898, "grad_norm": 0.6376220758085563, "learning_rate": 4.355072463768117e-05, "loss": 1.0961, "step": 601 }, { "epoch": 0.272768463978251, "grad_norm": 0.6782945436640291, "learning_rate": 4.362318840579711e-05, "loss": 1.0707, "step": 602 }, { "epoch": 0.27322156773901224, "grad_norm": 0.7173700752218448, "learning_rate": 4.369565217391305e-05, "loss": 1.0732, "step": 603 }, { "epoch": 0.2736746714997734, "grad_norm": 0.7983610310363438, "learning_rate": 4.3768115942028986e-05, "loss": 1.094, "step": 604 }, { "epoch": 0.27412777526053467, "grad_norm": 0.9644161488113789, "learning_rate": 4.384057971014493e-05, "loss": 1.0829, "step": 605 }, { "epoch": 0.27458087902129585, "grad_norm": 1.147054181163614, "learning_rate": 4.391304347826087e-05, "loss": 1.0793, "step": 606 }, { "epoch": 0.2750339827820571, "grad_norm": 0.8845454433876638, "learning_rate": 4.398550724637682e-05, "loss": 1.0652, "step": 607 }, { "epoch": 0.2754870865428183, "grad_norm": 0.774010355907489, "learning_rate": 4.405797101449276e-05, "loss": 1.1043, "step": 608 }, { "epoch": 0.2759401903035795, "grad_norm": 0.5251522666585152, "learning_rate": 4.41304347826087e-05, "loss": 1.082, "step": 609 }, { "epoch": 0.2763932940643407, "grad_norm": 0.6205991150666008, "learning_rate": 4.4202898550724645e-05, "loss": 1.0885, "step": 610 }, { "epoch": 0.27684639782510195, "grad_norm": 0.7089512857823811, "learning_rate": 4.427536231884058e-05, "loss": 1.0859, "step": 611 }, { "epoch": 0.27729950158586314, "grad_norm": 0.7479919343846159, "learning_rate": 4.4347826086956525e-05, "loss": 1.065, "step": 612 }, { "epoch": 0.2777526053466244, "grad_norm": 0.7897416765456193, "learning_rate": 4.4420289855072464e-05, "loss": 1.0751, "step": 613 }, { "epoch": 0.27820570910738557, "grad_norm": 0.8600280149514313, "learning_rate": 4.449275362318841e-05, "loss": 1.0498, "step": 614 }, { "epoch": 0.2786588128681468, "grad_norm": 0.9606089758809163, "learning_rate": 4.456521739130435e-05, "loss": 1.0562, "step": 615 }, { "epoch": 0.279111916628908, "grad_norm": 0.9247306523361446, "learning_rate": 4.46376811594203e-05, "loss": 1.0662, "step": 616 }, { "epoch": 0.27956502038966924, "grad_norm": 0.8222747508301819, "learning_rate": 4.471014492753624e-05, "loss": 1.0805, "step": 617 }, { "epoch": 0.2800181241504304, "grad_norm": 0.6110087301838939, "learning_rate": 4.4782608695652184e-05, "loss": 1.0605, "step": 618 }, { "epoch": 0.28047122791119167, "grad_norm": 0.4535542261637848, "learning_rate": 4.4855072463768117e-05, "loss": 1.0764, "step": 619 }, { "epoch": 0.28092433167195285, "grad_norm": 0.49535598246158363, "learning_rate": 4.492753623188406e-05, "loss": 1.104, "step": 620 }, { "epoch": 0.2813774354327141, "grad_norm": 0.5004143495014312, "learning_rate": 4.5e-05, "loss": 1.0632, "step": 621 }, { "epoch": 0.2818305391934753, "grad_norm": 0.4807686288444202, "learning_rate": 4.507246376811594e-05, "loss": 1.0753, "step": 622 }, { "epoch": 0.2822836429542365, "grad_norm": 0.5088290472132044, "learning_rate": 4.514492753623189e-05, "loss": 1.0796, "step": 623 }, { "epoch": 0.2827367467149977, "grad_norm": 0.4510494457789166, "learning_rate": 4.521739130434783e-05, "loss": 1.067, "step": 624 }, { "epoch": 0.28318985047575895, "grad_norm": 0.4815760801221088, "learning_rate": 4.5289855072463775e-05, "loss": 1.0732, "step": 625 }, { "epoch": 0.28364295423652014, "grad_norm": 0.5821052172158913, "learning_rate": 4.536231884057971e-05, "loss": 1.087, "step": 626 }, { "epoch": 0.2840960579972814, "grad_norm": 0.5920768507893932, "learning_rate": 4.5434782608695655e-05, "loss": 1.0516, "step": 627 }, { "epoch": 0.28454916175804257, "grad_norm": 0.6247276189517333, "learning_rate": 4.5507246376811595e-05, "loss": 1.0924, "step": 628 }, { "epoch": 0.2850022655188038, "grad_norm": 0.5029817778664476, "learning_rate": 4.557971014492754e-05, "loss": 1.0731, "step": 629 }, { "epoch": 0.285455369279565, "grad_norm": 0.41784067076990283, "learning_rate": 4.565217391304348e-05, "loss": 1.0883, "step": 630 }, { "epoch": 0.28590847304032624, "grad_norm": 0.46446597349249735, "learning_rate": 4.572463768115943e-05, "loss": 1.0579, "step": 631 }, { "epoch": 0.2863615768010874, "grad_norm": 0.5044259672358967, "learning_rate": 4.579710144927537e-05, "loss": 1.0834, "step": 632 }, { "epoch": 0.28681468056184867, "grad_norm": 0.5181273792873925, "learning_rate": 4.5869565217391314e-05, "loss": 1.0694, "step": 633 }, { "epoch": 0.28726778432260985, "grad_norm": 0.5256105022669127, "learning_rate": 4.594202898550725e-05, "loss": 1.0632, "step": 634 }, { "epoch": 0.2877208880833711, "grad_norm": 0.5580763444304888, "learning_rate": 4.601449275362319e-05, "loss": 1.0475, "step": 635 }, { "epoch": 0.2881739918441323, "grad_norm": 0.5806513659655225, "learning_rate": 4.608695652173913e-05, "loss": 1.063, "step": 636 }, { "epoch": 0.2886270956048935, "grad_norm": 0.5714159113630913, "learning_rate": 4.615942028985507e-05, "loss": 1.0838, "step": 637 }, { "epoch": 0.2890801993656547, "grad_norm": 0.5243342553028678, "learning_rate": 4.623188405797102e-05, "loss": 1.091, "step": 638 }, { "epoch": 0.28953330312641595, "grad_norm": 0.6017531142299696, "learning_rate": 4.630434782608696e-05, "loss": 1.0637, "step": 639 }, { "epoch": 0.28998640688717714, "grad_norm": 0.757512836345668, "learning_rate": 4.6376811594202906e-05, "loss": 1.0709, "step": 640 }, { "epoch": 0.2904395106479384, "grad_norm": 0.9114221962671901, "learning_rate": 4.644927536231884e-05, "loss": 1.0747, "step": 641 }, { "epoch": 0.29089261440869957, "grad_norm": 1.053635089105361, "learning_rate": 4.652173913043479e-05, "loss": 1.0557, "step": 642 }, { "epoch": 0.2913457181694608, "grad_norm": 0.9000198608103896, "learning_rate": 4.6594202898550725e-05, "loss": 1.0744, "step": 643 }, { "epoch": 0.291798821930222, "grad_norm": 0.7233314551247693, "learning_rate": 4.666666666666667e-05, "loss": 1.0838, "step": 644 }, { "epoch": 0.29225192569098324, "grad_norm": 0.6347811584538552, "learning_rate": 4.673913043478261e-05, "loss": 1.0831, "step": 645 }, { "epoch": 0.2927050294517444, "grad_norm": 0.7315011412274355, "learning_rate": 4.681159420289856e-05, "loss": 1.0618, "step": 646 }, { "epoch": 0.29315813321250567, "grad_norm": 0.7876943292414821, "learning_rate": 4.68840579710145e-05, "loss": 1.058, "step": 647 }, { "epoch": 0.29361123697326685, "grad_norm": 0.6755991969088356, "learning_rate": 4.6956521739130444e-05, "loss": 1.0744, "step": 648 }, { "epoch": 0.2940643407340281, "grad_norm": 0.4340462941275464, "learning_rate": 4.7028985507246384e-05, "loss": 1.0815, "step": 649 }, { "epoch": 0.2945174444947893, "grad_norm": 0.5824339342721956, "learning_rate": 4.710144927536232e-05, "loss": 1.0631, "step": 650 }, { "epoch": 0.2949705482555505, "grad_norm": 0.7771184347568643, "learning_rate": 4.7173913043478264e-05, "loss": 1.0642, "step": 651 }, { "epoch": 0.2954236520163117, "grad_norm": 0.7391441562389865, "learning_rate": 4.72463768115942e-05, "loss": 1.055, "step": 652 }, { "epoch": 0.29587675577707295, "grad_norm": 0.6505267746775686, "learning_rate": 4.731884057971015e-05, "loss": 1.0845, "step": 653 }, { "epoch": 0.29632985953783414, "grad_norm": 0.7554049811575129, "learning_rate": 4.739130434782609e-05, "loss": 1.0899, "step": 654 }, { "epoch": 0.2967829632985954, "grad_norm": 0.8269612415017934, "learning_rate": 4.7463768115942036e-05, "loss": 1.0725, "step": 655 }, { "epoch": 0.29723606705935657, "grad_norm": 1.004340919188453, "learning_rate": 4.7536231884057976e-05, "loss": 1.0628, "step": 656 }, { "epoch": 0.2976891708201178, "grad_norm": 1.0445542848810672, "learning_rate": 4.760869565217392e-05, "loss": 1.0899, "step": 657 }, { "epoch": 0.298142274580879, "grad_norm": 0.8520549134276578, "learning_rate": 4.7681159420289855e-05, "loss": 1.0427, "step": 658 }, { "epoch": 0.29859537834164024, "grad_norm": 0.6672173137149938, "learning_rate": 4.77536231884058e-05, "loss": 1.0719, "step": 659 }, { "epoch": 0.2990484821024014, "grad_norm": 0.544522614440328, "learning_rate": 4.782608695652174e-05, "loss": 1.0755, "step": 660 }, { "epoch": 0.29950158586316267, "grad_norm": 0.5390867576362943, "learning_rate": 4.789855072463769e-05, "loss": 1.0826, "step": 661 }, { "epoch": 0.29995468962392385, "grad_norm": 0.6232662051425996, "learning_rate": 4.797101449275363e-05, "loss": 1.0621, "step": 662 }, { "epoch": 0.3004077933846851, "grad_norm": 0.6387649664701486, "learning_rate": 4.804347826086957e-05, "loss": 1.0317, "step": 663 }, { "epoch": 0.3008608971454463, "grad_norm": 0.503004936777306, "learning_rate": 4.8115942028985514e-05, "loss": 1.0671, "step": 664 }, { "epoch": 0.3013140009062075, "grad_norm": 0.3461847937754298, "learning_rate": 4.818840579710145e-05, "loss": 1.0575, "step": 665 }, { "epoch": 0.3017671046669687, "grad_norm": 0.4507772943071533, "learning_rate": 4.8260869565217394e-05, "loss": 1.0655, "step": 666 }, { "epoch": 0.30222020842772995, "grad_norm": 0.4792640045560521, "learning_rate": 4.8333333333333334e-05, "loss": 1.0726, "step": 667 }, { "epoch": 0.30267331218849114, "grad_norm": 0.4786685716989669, "learning_rate": 4.840579710144928e-05, "loss": 1.059, "step": 668 }, { "epoch": 0.3031264159492524, "grad_norm": 0.4939990038183573, "learning_rate": 4.847826086956522e-05, "loss": 1.0913, "step": 669 }, { "epoch": 0.30357951971001357, "grad_norm": 0.5466452360559007, "learning_rate": 4.8550724637681166e-05, "loss": 1.0708, "step": 670 }, { "epoch": 0.3040326234707748, "grad_norm": 0.44267573131293525, "learning_rate": 4.8623188405797106e-05, "loss": 1.0666, "step": 671 }, { "epoch": 0.304485727231536, "grad_norm": 0.37685159471780655, "learning_rate": 4.869565217391305e-05, "loss": 1.0811, "step": 672 }, { "epoch": 0.30493883099229724, "grad_norm": 0.5309639517916653, "learning_rate": 4.8768115942028986e-05, "loss": 1.0688, "step": 673 }, { "epoch": 0.3053919347530584, "grad_norm": 0.702605884560063, "learning_rate": 4.884057971014493e-05, "loss": 1.0631, "step": 674 }, { "epoch": 0.30584503851381967, "grad_norm": 0.753349563329722, "learning_rate": 4.891304347826087e-05, "loss": 1.0682, "step": 675 }, { "epoch": 0.30629814227458085, "grad_norm": 0.757801789047667, "learning_rate": 4.898550724637682e-05, "loss": 1.0793, "step": 676 }, { "epoch": 0.3067512460353421, "grad_norm": 0.7396244972938787, "learning_rate": 4.905797101449276e-05, "loss": 1.079, "step": 677 }, { "epoch": 0.3072043497961033, "grad_norm": 0.7989206404011598, "learning_rate": 4.91304347826087e-05, "loss": 1.0764, "step": 678 }, { "epoch": 0.3076574535568645, "grad_norm": 1.0270316192305808, "learning_rate": 4.9202898550724645e-05, "loss": 1.0563, "step": 679 }, { "epoch": 0.3081105573176257, "grad_norm": 1.114565676015522, "learning_rate": 4.927536231884058e-05, "loss": 1.0825, "step": 680 }, { "epoch": 0.30856366107838695, "grad_norm": 0.6930759390507214, "learning_rate": 4.9347826086956524e-05, "loss": 1.088, "step": 681 }, { "epoch": 0.30901676483914814, "grad_norm": 0.35829291816785913, "learning_rate": 4.9420289855072464e-05, "loss": 1.0646, "step": 682 }, { "epoch": 0.3094698685999094, "grad_norm": 0.47825338953067614, "learning_rate": 4.949275362318841e-05, "loss": 1.0483, "step": 683 }, { "epoch": 0.30992297236067057, "grad_norm": 0.6792747752670143, "learning_rate": 4.956521739130435e-05, "loss": 1.0908, "step": 684 }, { "epoch": 0.3103760761214318, "grad_norm": 0.8635722279753546, "learning_rate": 4.96376811594203e-05, "loss": 1.0427, "step": 685 }, { "epoch": 0.310829179882193, "grad_norm": 0.9968841192974905, "learning_rate": 4.9710144927536237e-05, "loss": 1.08, "step": 686 }, { "epoch": 0.31128228364295424, "grad_norm": 1.009579446948624, "learning_rate": 4.978260869565218e-05, "loss": 1.083, "step": 687 }, { "epoch": 0.3117353874037154, "grad_norm": 0.7832021442135422, "learning_rate": 4.9855072463768116e-05, "loss": 1.0504, "step": 688 }, { "epoch": 0.31218849116447667, "grad_norm": 0.5159211272106092, "learning_rate": 4.992753623188407e-05, "loss": 1.0755, "step": 689 }, { "epoch": 0.31264159492523785, "grad_norm": 0.7181199171702122, "learning_rate": 5e-05, "loss": 1.0701, "step": 690 }, { "epoch": 0.3130946986859991, "grad_norm": 1.0395638832891039, "learning_rate": 5.007246376811594e-05, "loss": 1.0743, "step": 691 }, { "epoch": 0.3135478024467603, "grad_norm": 1.069499743554206, "learning_rate": 5.014492753623189e-05, "loss": 1.0597, "step": 692 }, { "epoch": 0.3140009062075215, "grad_norm": 0.706608732510013, "learning_rate": 5.021739130434783e-05, "loss": 1.0654, "step": 693 }, { "epoch": 0.3144540099682827, "grad_norm": 0.44287087572492584, "learning_rate": 5.0289855072463775e-05, "loss": 1.0792, "step": 694 }, { "epoch": 0.31490711372904395, "grad_norm": 0.5474875419716723, "learning_rate": 5.0362318840579715e-05, "loss": 1.1016, "step": 695 }, { "epoch": 0.31536021748980514, "grad_norm": 0.5521320555340503, "learning_rate": 5.043478260869566e-05, "loss": 1.0608, "step": 696 }, { "epoch": 0.3158133212505664, "grad_norm": 0.5225147953484653, "learning_rate": 5.0507246376811594e-05, "loss": 1.0759, "step": 697 }, { "epoch": 0.31626642501132757, "grad_norm": 0.6348677795720319, "learning_rate": 5.057971014492754e-05, "loss": 1.0502, "step": 698 }, { "epoch": 0.3167195287720888, "grad_norm": 0.7201125639152957, "learning_rate": 5.065217391304348e-05, "loss": 1.067, "step": 699 }, { "epoch": 0.31717263253285, "grad_norm": 0.762127353798261, "learning_rate": 5.072463768115943e-05, "loss": 1.0676, "step": 700 }, { "epoch": 0.31762573629361124, "grad_norm": 0.8148107726930256, "learning_rate": 5.079710144927537e-05, "loss": 1.0453, "step": 701 }, { "epoch": 0.3180788400543724, "grad_norm": 0.8952100472200654, "learning_rate": 5.0869565217391313e-05, "loss": 1.0533, "step": 702 }, { "epoch": 0.31853194381513367, "grad_norm": 1.0601560737313216, "learning_rate": 5.094202898550725e-05, "loss": 1.079, "step": 703 }, { "epoch": 0.31898504757589485, "grad_norm": 0.9293438744665934, "learning_rate": 5.1014492753623186e-05, "loss": 1.0518, "step": 704 }, { "epoch": 0.3194381513366561, "grad_norm": 0.7285507313375482, "learning_rate": 5.108695652173913e-05, "loss": 1.0813, "step": 705 }, { "epoch": 0.3198912550974173, "grad_norm": 0.6344782963696709, "learning_rate": 5.115942028985507e-05, "loss": 1.0473, "step": 706 }, { "epoch": 0.3203443588581785, "grad_norm": 0.6508283133200116, "learning_rate": 5.123188405797102e-05, "loss": 1.0596, "step": 707 }, { "epoch": 0.3207974626189397, "grad_norm": 0.7069964966103852, "learning_rate": 5.130434782608696e-05, "loss": 1.0615, "step": 708 }, { "epoch": 0.32125056637970095, "grad_norm": 0.7468341210684187, "learning_rate": 5.1376811594202905e-05, "loss": 1.0804, "step": 709 }, { "epoch": 0.32170367014046214, "grad_norm": 0.7488615922396359, "learning_rate": 5.1449275362318845e-05, "loss": 1.0854, "step": 710 }, { "epoch": 0.3221567739012234, "grad_norm": 0.6225221841844047, "learning_rate": 5.152173913043479e-05, "loss": 1.0532, "step": 711 }, { "epoch": 0.32260987766198457, "grad_norm": 0.4757719262513073, "learning_rate": 5.1594202898550725e-05, "loss": 1.047, "step": 712 }, { "epoch": 0.3230629814227458, "grad_norm": 0.420806922221135, "learning_rate": 5.166666666666667e-05, "loss": 1.0791, "step": 713 }, { "epoch": 0.323516085183507, "grad_norm": 0.49344286145627864, "learning_rate": 5.173913043478261e-05, "loss": 1.0638, "step": 714 }, { "epoch": 0.32396918894426824, "grad_norm": 0.5729726060818425, "learning_rate": 5.181159420289856e-05, "loss": 1.0305, "step": 715 }, { "epoch": 0.3244222927050294, "grad_norm": 0.7254413390731986, "learning_rate": 5.18840579710145e-05, "loss": 1.0503, "step": 716 }, { "epoch": 0.32487539646579067, "grad_norm": 0.7632369189805258, "learning_rate": 5.1956521739130444e-05, "loss": 1.0567, "step": 717 }, { "epoch": 0.32532850022655185, "grad_norm": 0.6868128001755794, "learning_rate": 5.2028985507246384e-05, "loss": 1.0602, "step": 718 }, { "epoch": 0.3257816039873131, "grad_norm": 0.6594295203744495, "learning_rate": 5.2101449275362317e-05, "loss": 1.0543, "step": 719 }, { "epoch": 0.3262347077480743, "grad_norm": 0.690750855529275, "learning_rate": 5.217391304347826e-05, "loss": 1.0377, "step": 720 }, { "epoch": 0.3266878115088355, "grad_norm": 0.839525526523565, "learning_rate": 5.22463768115942e-05, "loss": 1.0773, "step": 721 }, { "epoch": 0.3271409152695967, "grad_norm": 1.037449089444073, "learning_rate": 5.231884057971015e-05, "loss": 1.0626, "step": 722 }, { "epoch": 0.32759401903035795, "grad_norm": 1.0190110475522725, "learning_rate": 5.239130434782609e-05, "loss": 1.0594, "step": 723 }, { "epoch": 0.32804712279111914, "grad_norm": 0.7488841168089863, "learning_rate": 5.2463768115942036e-05, "loss": 1.061, "step": 724 }, { "epoch": 0.3285002265518804, "grad_norm": 0.6543755076151874, "learning_rate": 5.2536231884057975e-05, "loss": 1.0662, "step": 725 }, { "epoch": 0.32895333031264157, "grad_norm": 0.8229912116483713, "learning_rate": 5.260869565217392e-05, "loss": 1.0404, "step": 726 }, { "epoch": 0.3294064340734028, "grad_norm": 0.969997326436284, "learning_rate": 5.2681159420289855e-05, "loss": 1.0809, "step": 727 }, { "epoch": 0.329859537834164, "grad_norm": 1.020955789468831, "learning_rate": 5.27536231884058e-05, "loss": 1.0709, "step": 728 }, { "epoch": 0.33031264159492524, "grad_norm": 0.8345414845139333, "learning_rate": 5.282608695652174e-05, "loss": 1.0466, "step": 729 }, { "epoch": 0.3307657453556864, "grad_norm": 0.6115651935472799, "learning_rate": 5.289855072463769e-05, "loss": 1.0649, "step": 730 }, { "epoch": 0.33121884911644767, "grad_norm": 0.5487684789203621, "learning_rate": 5.297101449275363e-05, "loss": 1.0611, "step": 731 }, { "epoch": 0.33167195287720885, "grad_norm": 0.5627316336288565, "learning_rate": 5.304347826086957e-05, "loss": 1.0796, "step": 732 }, { "epoch": 0.3321250566379701, "grad_norm": 0.6333688150489734, "learning_rate": 5.3115942028985514e-05, "loss": 1.0721, "step": 733 }, { "epoch": 0.3325781603987313, "grad_norm": 0.7050932521391242, "learning_rate": 5.318840579710145e-05, "loss": 1.0694, "step": 734 }, { "epoch": 0.3330312641594925, "grad_norm": 0.7294644876440662, "learning_rate": 5.32608695652174e-05, "loss": 1.0578, "step": 735 }, { "epoch": 0.33348436792025377, "grad_norm": 0.6874314482618404, "learning_rate": 5.333333333333333e-05, "loss": 1.0628, "step": 736 }, { "epoch": 0.33393747168101495, "grad_norm": 0.641519177191011, "learning_rate": 5.340579710144928e-05, "loss": 1.0636, "step": 737 }, { "epoch": 0.3343905754417762, "grad_norm": 0.6642915287741901, "learning_rate": 5.347826086956522e-05, "loss": 1.0721, "step": 738 }, { "epoch": 0.3348436792025374, "grad_norm": 0.6299728703558107, "learning_rate": 5.3550724637681166e-05, "loss": 1.0453, "step": 739 }, { "epoch": 0.3352967829632986, "grad_norm": 0.4738914493235964, "learning_rate": 5.3623188405797106e-05, "loss": 1.0421, "step": 740 }, { "epoch": 0.3357498867240598, "grad_norm": 0.48065975756941137, "learning_rate": 5.369565217391305e-05, "loss": 1.0436, "step": 741 }, { "epoch": 0.33620299048482105, "grad_norm": 0.6157300808050292, "learning_rate": 5.376811594202899e-05, "loss": 1.0763, "step": 742 }, { "epoch": 0.33665609424558224, "grad_norm": 0.5850856231209889, "learning_rate": 5.384057971014494e-05, "loss": 1.0583, "step": 743 }, { "epoch": 0.3371091980063435, "grad_norm": 0.5398577726381432, "learning_rate": 5.391304347826087e-05, "loss": 1.0793, "step": 744 }, { "epoch": 0.33756230176710467, "grad_norm": 0.5414411563677495, "learning_rate": 5.398550724637682e-05, "loss": 1.059, "step": 745 }, { "epoch": 0.3380154055278659, "grad_norm": 0.5536035039391813, "learning_rate": 5.405797101449276e-05, "loss": 1.0774, "step": 746 }, { "epoch": 0.3384685092886271, "grad_norm": 0.597618619761768, "learning_rate": 5.41304347826087e-05, "loss": 1.0815, "step": 747 }, { "epoch": 0.33892161304938834, "grad_norm": 0.6361006558775629, "learning_rate": 5.4202898550724644e-05, "loss": 1.066, "step": 748 }, { "epoch": 0.3393747168101495, "grad_norm": 0.7579744880543378, "learning_rate": 5.4275362318840584e-05, "loss": 1.0656, "step": 749 }, { "epoch": 0.33982782057091077, "grad_norm": 0.8339799707671164, "learning_rate": 5.434782608695653e-05, "loss": 1.036, "step": 750 }, { "epoch": 0.34028092433167195, "grad_norm": 0.8389029149061997, "learning_rate": 5.4420289855072464e-05, "loss": 1.0498, "step": 751 }, { "epoch": 0.3407340280924332, "grad_norm": 0.8451088221102333, "learning_rate": 5.449275362318841e-05, "loss": 1.0861, "step": 752 }, { "epoch": 0.3411871318531944, "grad_norm": 0.7503497034637959, "learning_rate": 5.456521739130435e-05, "loss": 1.0669, "step": 753 }, { "epoch": 0.3416402356139556, "grad_norm": 0.6615922667534954, "learning_rate": 5.4637681159420296e-05, "loss": 1.0791, "step": 754 }, { "epoch": 0.3420933393747168, "grad_norm": 0.6680888483767149, "learning_rate": 5.4710144927536236e-05, "loss": 1.0497, "step": 755 }, { "epoch": 0.34254644313547805, "grad_norm": 0.6609421197364006, "learning_rate": 5.478260869565218e-05, "loss": 1.0589, "step": 756 }, { "epoch": 0.34299954689623924, "grad_norm": 0.5903555356943737, "learning_rate": 5.485507246376812e-05, "loss": 1.0527, "step": 757 }, { "epoch": 0.3434526506570005, "grad_norm": 0.48863303174695094, "learning_rate": 5.492753623188407e-05, "loss": 1.0658, "step": 758 }, { "epoch": 0.34390575441776167, "grad_norm": 0.37983380539252193, "learning_rate": 5.5e-05, "loss": 1.067, "step": 759 }, { "epoch": 0.3443588581785229, "grad_norm": 0.4775635485436271, "learning_rate": 5.507246376811594e-05, "loss": 1.043, "step": 760 }, { "epoch": 0.3448119619392841, "grad_norm": 0.5122367122679161, "learning_rate": 5.514492753623189e-05, "loss": 1.0409, "step": 761 }, { "epoch": 0.34526506570004534, "grad_norm": 0.5363158180507529, "learning_rate": 5.521739130434783e-05, "loss": 1.0582, "step": 762 }, { "epoch": 0.3457181694608065, "grad_norm": 0.6902892253528404, "learning_rate": 5.5289855072463775e-05, "loss": 1.0558, "step": 763 }, { "epoch": 0.34617127322156777, "grad_norm": 0.9037305092881801, "learning_rate": 5.5362318840579714e-05, "loss": 1.0481, "step": 764 }, { "epoch": 0.34662437698232895, "grad_norm": 1.119459614858512, "learning_rate": 5.543478260869566e-05, "loss": 1.0662, "step": 765 }, { "epoch": 0.3470774807430902, "grad_norm": 0.8401856469883325, "learning_rate": 5.5507246376811594e-05, "loss": 1.0533, "step": 766 }, { "epoch": 0.3475305845038514, "grad_norm": 0.6122032148886061, "learning_rate": 5.557971014492754e-05, "loss": 1.0911, "step": 767 }, { "epoch": 0.3479836882646126, "grad_norm": 0.4630326934045826, "learning_rate": 5.565217391304348e-05, "loss": 1.036, "step": 768 }, { "epoch": 0.3484367920253738, "grad_norm": 0.442289272338855, "learning_rate": 5.572463768115943e-05, "loss": 1.0534, "step": 769 }, { "epoch": 0.34888989578613505, "grad_norm": 0.5828657505042796, "learning_rate": 5.5797101449275366e-05, "loss": 1.048, "step": 770 }, { "epoch": 0.34934299954689624, "grad_norm": 0.7873995913233393, "learning_rate": 5.586956521739131e-05, "loss": 1.0601, "step": 771 }, { "epoch": 0.3497961033076575, "grad_norm": 1.0141960466114328, "learning_rate": 5.594202898550725e-05, "loss": 1.0502, "step": 772 }, { "epoch": 0.35024920706841867, "grad_norm": 1.1130674331826642, "learning_rate": 5.6014492753623186e-05, "loss": 1.0903, "step": 773 }, { "epoch": 0.3507023108291799, "grad_norm": 0.6638106610334916, "learning_rate": 5.608695652173913e-05, "loss": 1.0523, "step": 774 }, { "epoch": 0.3511554145899411, "grad_norm": 0.40306679951507945, "learning_rate": 5.615942028985507e-05, "loss": 1.0483, "step": 775 }, { "epoch": 0.35160851835070234, "grad_norm": 0.5705228057446858, "learning_rate": 5.623188405797102e-05, "loss": 1.0527, "step": 776 }, { "epoch": 0.3520616221114635, "grad_norm": 0.721734658184513, "learning_rate": 5.630434782608696e-05, "loss": 1.0424, "step": 777 }, { "epoch": 0.35251472587222477, "grad_norm": 0.7736479299627674, "learning_rate": 5.6376811594202905e-05, "loss": 1.0458, "step": 778 }, { "epoch": 0.35296782963298595, "grad_norm": 0.728454249000286, "learning_rate": 5.6449275362318845e-05, "loss": 1.051, "step": 779 }, { "epoch": 0.3534209333937472, "grad_norm": 0.5977087427596691, "learning_rate": 5.652173913043479e-05, "loss": 1.0361, "step": 780 }, { "epoch": 0.3538740371545084, "grad_norm": 0.5429622742794915, "learning_rate": 5.6594202898550724e-05, "loss": 1.0635, "step": 781 }, { "epoch": 0.3543271409152696, "grad_norm": 0.5261710204696646, "learning_rate": 5.666666666666668e-05, "loss": 1.036, "step": 782 }, { "epoch": 0.3547802446760308, "grad_norm": 0.5556323473396069, "learning_rate": 5.673913043478261e-05, "loss": 1.0559, "step": 783 }, { "epoch": 0.35523334843679205, "grad_norm": 0.5780539758859797, "learning_rate": 5.681159420289856e-05, "loss": 1.0511, "step": 784 }, { "epoch": 0.35568645219755324, "grad_norm": 0.680135110080583, "learning_rate": 5.68840579710145e-05, "loss": 1.0663, "step": 785 }, { "epoch": 0.3561395559583145, "grad_norm": 0.7900523334494177, "learning_rate": 5.695652173913044e-05, "loss": 1.0589, "step": 786 }, { "epoch": 0.35659265971907567, "grad_norm": 0.7591250362448084, "learning_rate": 5.702898550724638e-05, "loss": 1.0464, "step": 787 }, { "epoch": 0.3570457634798369, "grad_norm": 0.5889183516134565, "learning_rate": 5.710144927536232e-05, "loss": 1.0536, "step": 788 }, { "epoch": 0.3574988672405981, "grad_norm": 0.5315198250498254, "learning_rate": 5.717391304347827e-05, "loss": 1.0676, "step": 789 }, { "epoch": 0.35795197100135934, "grad_norm": 0.5777139818672671, "learning_rate": 5.72463768115942e-05, "loss": 1.0478, "step": 790 }, { "epoch": 0.3584050747621205, "grad_norm": 0.7269923216319254, "learning_rate": 5.731884057971015e-05, "loss": 1.0452, "step": 791 }, { "epoch": 0.35885817852288177, "grad_norm": 0.9350464993178923, "learning_rate": 5.739130434782609e-05, "loss": 1.0337, "step": 792 }, { "epoch": 0.35931128228364295, "grad_norm": 1.0118952573755406, "learning_rate": 5.7463768115942035e-05, "loss": 1.0625, "step": 793 }, { "epoch": 0.3597643860444042, "grad_norm": 1.0523941603442981, "learning_rate": 5.7536231884057975e-05, "loss": 1.0615, "step": 794 }, { "epoch": 0.3602174898051654, "grad_norm": 0.9118321766075538, "learning_rate": 5.760869565217392e-05, "loss": 1.0787, "step": 795 }, { "epoch": 0.3606705935659266, "grad_norm": 0.7917240314055172, "learning_rate": 5.768115942028986e-05, "loss": 1.0916, "step": 796 }, { "epoch": 0.3611236973266878, "grad_norm": 0.6998443261659207, "learning_rate": 5.775362318840581e-05, "loss": 1.0632, "step": 797 }, { "epoch": 0.36157680108744905, "grad_norm": 0.6916813631216127, "learning_rate": 5.782608695652174e-05, "loss": 1.0463, "step": 798 }, { "epoch": 0.36202990484821024, "grad_norm": 0.564147619003302, "learning_rate": 5.789855072463769e-05, "loss": 1.0482, "step": 799 }, { "epoch": 0.3624830086089715, "grad_norm": 0.4960108216254337, "learning_rate": 5.797101449275363e-05, "loss": 1.0323, "step": 800 }, { "epoch": 0.36293611236973267, "grad_norm": 0.765451117481778, "learning_rate": 5.804347826086957e-05, "loss": 1.0497, "step": 801 }, { "epoch": 0.3633892161304939, "grad_norm": 0.9913456876221574, "learning_rate": 5.8115942028985513e-05, "loss": 1.0494, "step": 802 }, { "epoch": 0.3638423198912551, "grad_norm": 1.196734965564921, "learning_rate": 5.818840579710145e-05, "loss": 1.0502, "step": 803 }, { "epoch": 0.36429542365201634, "grad_norm": 0.6829597157289536, "learning_rate": 5.82608695652174e-05, "loss": 1.0539, "step": 804 }, { "epoch": 0.3647485274127775, "grad_norm": 0.7108514806498558, "learning_rate": 5.833333333333333e-05, "loss": 1.0586, "step": 805 }, { "epoch": 0.36520163117353877, "grad_norm": 0.9251087606491766, "learning_rate": 5.840579710144928e-05, "loss": 1.0404, "step": 806 }, { "epoch": 0.36565473493429995, "grad_norm": 1.01701559917751, "learning_rate": 5.847826086956522e-05, "loss": 1.0439, "step": 807 }, { "epoch": 0.3661078386950612, "grad_norm": 0.9332746754917329, "learning_rate": 5.8550724637681166e-05, "loss": 1.0506, "step": 808 }, { "epoch": 0.3665609424558224, "grad_norm": 0.7648363025806546, "learning_rate": 5.8623188405797105e-05, "loss": 1.0613, "step": 809 }, { "epoch": 0.3670140462165836, "grad_norm": 0.5903294123996294, "learning_rate": 5.869565217391305e-05, "loss": 1.0562, "step": 810 }, { "epoch": 0.3674671499773448, "grad_norm": 0.46046114673137184, "learning_rate": 5.876811594202899e-05, "loss": 1.0443, "step": 811 }, { "epoch": 0.36792025373810605, "grad_norm": 0.5236680991290782, "learning_rate": 5.884057971014494e-05, "loss": 1.075, "step": 812 }, { "epoch": 0.36837335749886724, "grad_norm": 0.72212445369333, "learning_rate": 5.891304347826087e-05, "loss": 1.0693, "step": 813 }, { "epoch": 0.3688264612596285, "grad_norm": 0.7705583413430515, "learning_rate": 5.898550724637682e-05, "loss": 1.0434, "step": 814 }, { "epoch": 0.36927956502038967, "grad_norm": 0.7418703891053157, "learning_rate": 5.905797101449276e-05, "loss": 1.0565, "step": 815 }, { "epoch": 0.3697326687811509, "grad_norm": 0.716180944837806, "learning_rate": 5.91304347826087e-05, "loss": 1.0446, "step": 816 }, { "epoch": 0.3701857725419121, "grad_norm": 0.7050516082461611, "learning_rate": 5.9202898550724644e-05, "loss": 1.0511, "step": 817 }, { "epoch": 0.37063887630267334, "grad_norm": 0.6895682674254866, "learning_rate": 5.9275362318840584e-05, "loss": 1.0605, "step": 818 }, { "epoch": 0.3710919800634345, "grad_norm": 0.5676316773796998, "learning_rate": 5.934782608695653e-05, "loss": 1.0558, "step": 819 }, { "epoch": 0.37154508382419577, "grad_norm": 0.42194885339492755, "learning_rate": 5.942028985507246e-05, "loss": 1.0634, "step": 820 }, { "epoch": 0.37199818758495695, "grad_norm": 0.31196909593378275, "learning_rate": 5.949275362318841e-05, "loss": 1.0438, "step": 821 }, { "epoch": 0.3724512913457182, "grad_norm": 0.44728130014391954, "learning_rate": 5.956521739130435e-05, "loss": 1.0217, "step": 822 }, { "epoch": 0.3729043951064794, "grad_norm": 0.5219963260828536, "learning_rate": 5.9637681159420296e-05, "loss": 1.0263, "step": 823 }, { "epoch": 0.3733574988672406, "grad_norm": 0.5159497792789729, "learning_rate": 5.9710144927536236e-05, "loss": 1.0478, "step": 824 }, { "epoch": 0.3738106026280018, "grad_norm": 0.5747551972382041, "learning_rate": 5.978260869565218e-05, "loss": 1.0517, "step": 825 }, { "epoch": 0.37426370638876305, "grad_norm": 0.6805353068047845, "learning_rate": 5.985507246376812e-05, "loss": 1.0416, "step": 826 }, { "epoch": 0.37471681014952424, "grad_norm": 0.8723114621842396, "learning_rate": 5.992753623188407e-05, "loss": 1.0652, "step": 827 }, { "epoch": 0.3751699139102855, "grad_norm": 1.1322453934367183, "learning_rate": 6.000000000000001e-05, "loss": 1.0507, "step": 828 }, { "epoch": 0.37562301767104667, "grad_norm": 1.002642869365044, "learning_rate": 6.007246376811594e-05, "loss": 1.0469, "step": 829 }, { "epoch": 0.3760761214318079, "grad_norm": 0.92192076638776, "learning_rate": 6.014492753623189e-05, "loss": 1.0644, "step": 830 }, { "epoch": 0.3765292251925691, "grad_norm": 0.8726355580569491, "learning_rate": 6.021739130434783e-05, "loss": 1.0429, "step": 831 }, { "epoch": 0.37698232895333034, "grad_norm": 0.699169413979365, "learning_rate": 6.0289855072463774e-05, "loss": 1.0514, "step": 832 }, { "epoch": 0.3774354327140915, "grad_norm": 0.5741993402465739, "learning_rate": 6.0362318840579714e-05, "loss": 1.043, "step": 833 }, { "epoch": 0.37788853647485277, "grad_norm": 0.6294153656500621, "learning_rate": 6.043478260869566e-05, "loss": 1.0817, "step": 834 }, { "epoch": 0.37834164023561395, "grad_norm": 0.6414956832460792, "learning_rate": 6.05072463768116e-05, "loss": 1.0543, "step": 835 }, { "epoch": 0.3787947439963752, "grad_norm": 0.5435290152590656, "learning_rate": 6.057971014492755e-05, "loss": 1.0707, "step": 836 }, { "epoch": 0.3792478477571364, "grad_norm": 0.4970218880574281, "learning_rate": 6.065217391304348e-05, "loss": 1.0492, "step": 837 }, { "epoch": 0.3797009515178976, "grad_norm": 0.5005518828210378, "learning_rate": 6.0724637681159426e-05, "loss": 1.0673, "step": 838 }, { "epoch": 0.3801540552786588, "grad_norm": 0.4265586621314692, "learning_rate": 6.0797101449275366e-05, "loss": 1.0645, "step": 839 }, { "epoch": 0.38060715903942005, "grad_norm": 0.5711602562637467, "learning_rate": 6.086956521739131e-05, "loss": 1.0602, "step": 840 }, { "epoch": 0.38106026280018124, "grad_norm": 0.5933025319390643, "learning_rate": 6.094202898550725e-05, "loss": 1.07, "step": 841 }, { "epoch": 0.3815133665609425, "grad_norm": 0.44272428059068053, "learning_rate": 6.101449275362319e-05, "loss": 1.0565, "step": 842 }, { "epoch": 0.38196647032170367, "grad_norm": 0.46865089450583675, "learning_rate": 6.108695652173914e-05, "loss": 1.0299, "step": 843 }, { "epoch": 0.3824195740824649, "grad_norm": 0.48021004430001313, "learning_rate": 6.115942028985507e-05, "loss": 1.0388, "step": 844 }, { "epoch": 0.3828726778432261, "grad_norm": 0.5323396641799173, "learning_rate": 6.123188405797102e-05, "loss": 1.0306, "step": 845 }, { "epoch": 0.38332578160398734, "grad_norm": 0.5577801182535822, "learning_rate": 6.130434782608696e-05, "loss": 1.0391, "step": 846 }, { "epoch": 0.3837788853647485, "grad_norm": 0.6595278779640518, "learning_rate": 6.137681159420291e-05, "loss": 1.0588, "step": 847 }, { "epoch": 0.38423198912550977, "grad_norm": 0.736590501633582, "learning_rate": 6.144927536231884e-05, "loss": 1.0636, "step": 848 }, { "epoch": 0.38468509288627095, "grad_norm": 0.7457930268040525, "learning_rate": 6.152173913043479e-05, "loss": 1.0465, "step": 849 }, { "epoch": 0.3851381966470322, "grad_norm": 0.8284840241690655, "learning_rate": 6.159420289855072e-05, "loss": 1.0234, "step": 850 }, { "epoch": 0.3855913004077934, "grad_norm": 0.9038500594664561, "learning_rate": 6.166666666666667e-05, "loss": 1.0562, "step": 851 }, { "epoch": 0.3860444041685546, "grad_norm": 1.0557477903328194, "learning_rate": 6.173913043478262e-05, "loss": 1.058, "step": 852 }, { "epoch": 0.3864975079293158, "grad_norm": 1.0776147419573767, "learning_rate": 6.181159420289856e-05, "loss": 1.0611, "step": 853 }, { "epoch": 0.38695061169007705, "grad_norm": 1.0162875744400872, "learning_rate": 6.18840579710145e-05, "loss": 1.0724, "step": 854 }, { "epoch": 0.38740371545083824, "grad_norm": 1.001457324300768, "learning_rate": 6.195652173913044e-05, "loss": 1.0417, "step": 855 }, { "epoch": 0.3878568192115995, "grad_norm": 0.9509578510967378, "learning_rate": 6.202898550724638e-05, "loss": 1.0635, "step": 856 }, { "epoch": 0.38830992297236067, "grad_norm": 0.7781333974670281, "learning_rate": 6.210144927536232e-05, "loss": 1.0496, "step": 857 }, { "epoch": 0.3887630267331219, "grad_norm": 0.7675940768186277, "learning_rate": 6.217391304347827e-05, "loss": 1.0518, "step": 858 }, { "epoch": 0.3892161304938831, "grad_norm": 0.8053967669820106, "learning_rate": 6.22463768115942e-05, "loss": 1.0679, "step": 859 }, { "epoch": 0.38966923425464434, "grad_norm": 0.7711198777130742, "learning_rate": 6.231884057971015e-05, "loss": 1.0581, "step": 860 }, { "epoch": 0.3901223380154055, "grad_norm": 0.8294411738821678, "learning_rate": 6.23913043478261e-05, "loss": 1.062, "step": 861 }, { "epoch": 0.39057544177616677, "grad_norm": 1.037050896885248, "learning_rate": 6.246376811594204e-05, "loss": 1.0607, "step": 862 }, { "epoch": 0.39102854553692795, "grad_norm": 1.0075428448991608, "learning_rate": 6.253623188405797e-05, "loss": 1.0797, "step": 863 }, { "epoch": 0.3914816492976892, "grad_norm": 0.7312242777158201, "learning_rate": 6.260869565217392e-05, "loss": 1.0394, "step": 864 }, { "epoch": 0.3919347530584504, "grad_norm": 0.5451568345489267, "learning_rate": 6.268115942028985e-05, "loss": 1.054, "step": 865 }, { "epoch": 0.3923878568192116, "grad_norm": 0.6410042401298008, "learning_rate": 6.27536231884058e-05, "loss": 1.0557, "step": 866 }, { "epoch": 0.3928409605799728, "grad_norm": 0.5376341206741354, "learning_rate": 6.282608695652175e-05, "loss": 1.0404, "step": 867 }, { "epoch": 0.39329406434073405, "grad_norm": 0.4760690176640781, "learning_rate": 6.28985507246377e-05, "loss": 1.0548, "step": 868 }, { "epoch": 0.39374716810149524, "grad_norm": 0.4416280450677732, "learning_rate": 6.297101449275363e-05, "loss": 1.057, "step": 869 }, { "epoch": 0.3942002718622565, "grad_norm": 0.45238530928368886, "learning_rate": 6.304347826086956e-05, "loss": 1.0571, "step": 870 }, { "epoch": 0.39465337562301767, "grad_norm": 0.4635205973533936, "learning_rate": 6.31159420289855e-05, "loss": 1.0598, "step": 871 }, { "epoch": 0.3951064793837789, "grad_norm": 0.46096881125616834, "learning_rate": 6.318840579710145e-05, "loss": 1.0522, "step": 872 }, { "epoch": 0.3955595831445401, "grad_norm": 0.4197460142409507, "learning_rate": 6.32608695652174e-05, "loss": 1.0428, "step": 873 }, { "epoch": 0.39601268690530134, "grad_norm": 0.4927959206355848, "learning_rate": 6.333333333333333e-05, "loss": 1.047, "step": 874 }, { "epoch": 0.3964657906660625, "grad_norm": 0.6448622846282847, "learning_rate": 6.340579710144928e-05, "loss": 1.0349, "step": 875 }, { "epoch": 0.39691889442682377, "grad_norm": 0.7297574664111051, "learning_rate": 6.347826086956523e-05, "loss": 1.0298, "step": 876 }, { "epoch": 0.39737199818758495, "grad_norm": 0.8983752662288029, "learning_rate": 6.355072463768117e-05, "loss": 1.0768, "step": 877 }, { "epoch": 0.3978251019483462, "grad_norm": 1.0381834482974597, "learning_rate": 6.36231884057971e-05, "loss": 1.0417, "step": 878 }, { "epoch": 0.3982782057091074, "grad_norm": 0.9917895953108649, "learning_rate": 6.369565217391305e-05, "loss": 1.0518, "step": 879 }, { "epoch": 0.3987313094698686, "grad_norm": 0.9667250542900923, "learning_rate": 6.376811594202898e-05, "loss": 1.07, "step": 880 }, { "epoch": 0.3991844132306298, "grad_norm": 0.9014465555147225, "learning_rate": 6.384057971014493e-05, "loss": 1.0236, "step": 881 }, { "epoch": 0.39963751699139105, "grad_norm": 0.7644443008031563, "learning_rate": 6.391304347826088e-05, "loss": 1.0483, "step": 882 }, { "epoch": 0.40009062075215224, "grad_norm": 0.7157358673240267, "learning_rate": 6.398550724637682e-05, "loss": 1.0332, "step": 883 }, { "epoch": 0.4005437245129135, "grad_norm": 0.6996459033818614, "learning_rate": 6.405797101449276e-05, "loss": 1.0394, "step": 884 }, { "epoch": 0.40099682827367467, "grad_norm": 0.5770420582122487, "learning_rate": 6.41304347826087e-05, "loss": 1.0574, "step": 885 }, { "epoch": 0.4014499320344359, "grad_norm": 0.5504797070068408, "learning_rate": 6.420289855072465e-05, "loss": 1.057, "step": 886 }, { "epoch": 0.4019030357951971, "grad_norm": 0.5608555243616004, "learning_rate": 6.427536231884058e-05, "loss": 1.0477, "step": 887 }, { "epoch": 0.40235613955595834, "grad_norm": 0.5057308135271467, "learning_rate": 6.434782608695653e-05, "loss": 1.0561, "step": 888 }, { "epoch": 0.4028092433167195, "grad_norm": 0.6048575547891671, "learning_rate": 6.442028985507246e-05, "loss": 1.0425, "step": 889 }, { "epoch": 0.40326234707748077, "grad_norm": 0.605612693846474, "learning_rate": 6.449275362318841e-05, "loss": 1.0321, "step": 890 }, { "epoch": 0.40371545083824195, "grad_norm": 0.5658407586628978, "learning_rate": 6.456521739130436e-05, "loss": 1.0414, "step": 891 }, { "epoch": 0.4041685545990032, "grad_norm": 0.6735563808663679, "learning_rate": 6.46376811594203e-05, "loss": 1.033, "step": 892 }, { "epoch": 0.4046216583597644, "grad_norm": 0.7384368327011808, "learning_rate": 6.471014492753624e-05, "loss": 1.0585, "step": 893 }, { "epoch": 0.4050747621205256, "grad_norm": 0.8593983504126844, "learning_rate": 6.478260869565218e-05, "loss": 1.0393, "step": 894 }, { "epoch": 0.4055278658812868, "grad_norm": 0.9194090375812336, "learning_rate": 6.485507246376811e-05, "loss": 1.0348, "step": 895 }, { "epoch": 0.40598096964204805, "grad_norm": 0.9671105496878438, "learning_rate": 6.492753623188406e-05, "loss": 1.0479, "step": 896 }, { "epoch": 0.40643407340280924, "grad_norm": 1.0880544642867835, "learning_rate": 6.500000000000001e-05, "loss": 1.0457, "step": 897 }, { "epoch": 0.4068871771635705, "grad_norm": 0.9188698607735127, "learning_rate": 6.507246376811594e-05, "loss": 1.0771, "step": 898 }, { "epoch": 0.40734028092433167, "grad_norm": 0.7897491937775787, "learning_rate": 6.514492753623189e-05, "loss": 1.0568, "step": 899 }, { "epoch": 0.4077933846850929, "grad_norm": 0.7463752639746497, "learning_rate": 6.521739130434783e-05, "loss": 1.0689, "step": 900 }, { "epoch": 0.4082464884458541, "grad_norm": 0.8495510365263993, "learning_rate": 6.528985507246378e-05, "loss": 1.0633, "step": 901 }, { "epoch": 0.40869959220661534, "grad_norm": 0.9605722100207644, "learning_rate": 6.536231884057971e-05, "loss": 1.0703, "step": 902 }, { "epoch": 0.4091526959673765, "grad_norm": 0.9135253846771476, "learning_rate": 6.543478260869566e-05, "loss": 1.0403, "step": 903 }, { "epoch": 0.40960579972813776, "grad_norm": 0.8301010150495305, "learning_rate": 6.550724637681159e-05, "loss": 1.0467, "step": 904 }, { "epoch": 0.41005890348889895, "grad_norm": 0.6952692003658697, "learning_rate": 6.557971014492754e-05, "loss": 1.0401, "step": 905 }, { "epoch": 0.4105120072496602, "grad_norm": 0.6475199707045178, "learning_rate": 6.565217391304349e-05, "loss": 1.0507, "step": 906 }, { "epoch": 0.4109651110104214, "grad_norm": 0.7261764665400339, "learning_rate": 6.572463768115943e-05, "loss": 1.0583, "step": 907 }, { "epoch": 0.4114182147711826, "grad_norm": 0.5935211100561533, "learning_rate": 6.579710144927537e-05, "loss": 1.0521, "step": 908 }, { "epoch": 0.4118713185319438, "grad_norm": 0.547922918716164, "learning_rate": 6.586956521739131e-05, "loss": 1.045, "step": 909 }, { "epoch": 0.41232442229270505, "grad_norm": 0.5963795638265788, "learning_rate": 6.594202898550725e-05, "loss": 1.0513, "step": 910 }, { "epoch": 0.41277752605346624, "grad_norm": 0.5540667117191366, "learning_rate": 6.601449275362319e-05, "loss": 1.0597, "step": 911 }, { "epoch": 0.4132306298142275, "grad_norm": 0.49532135309281483, "learning_rate": 6.608695652173914e-05, "loss": 1.0471, "step": 912 }, { "epoch": 0.41368373357498867, "grad_norm": 0.5589120722500759, "learning_rate": 6.615942028985507e-05, "loss": 1.0491, "step": 913 }, { "epoch": 0.4141368373357499, "grad_norm": 0.5302764983049771, "learning_rate": 6.623188405797102e-05, "loss": 1.0414, "step": 914 }, { "epoch": 0.4145899410965111, "grad_norm": 0.5254739372275832, "learning_rate": 6.630434782608696e-05, "loss": 1.0462, "step": 915 }, { "epoch": 0.41504304485727234, "grad_norm": 0.4312922406562473, "learning_rate": 6.637681159420291e-05, "loss": 1.0508, "step": 916 }, { "epoch": 0.4154961486180335, "grad_norm": 0.4697624136285695, "learning_rate": 6.644927536231884e-05, "loss": 1.0535, "step": 917 }, { "epoch": 0.41594925237879476, "grad_norm": 0.5790134088110307, "learning_rate": 6.652173913043479e-05, "loss": 1.0577, "step": 918 }, { "epoch": 0.41640235613955595, "grad_norm": 0.46891233827435413, "learning_rate": 6.659420289855072e-05, "loss": 1.0444, "step": 919 }, { "epoch": 0.4168554599003172, "grad_norm": 0.4036531003290714, "learning_rate": 6.666666666666667e-05, "loss": 1.0413, "step": 920 }, { "epoch": 0.4173085636610784, "grad_norm": 0.468875630957639, "learning_rate": 6.673913043478262e-05, "loss": 1.037, "step": 921 }, { "epoch": 0.4177616674218396, "grad_norm": 0.5644318281495692, "learning_rate": 6.681159420289856e-05, "loss": 1.0194, "step": 922 }, { "epoch": 0.4182147711826008, "grad_norm": 0.8425030587504226, "learning_rate": 6.68840579710145e-05, "loss": 1.0634, "step": 923 }, { "epoch": 0.41866787494336205, "grad_norm": 1.1505413237259594, "learning_rate": 6.695652173913044e-05, "loss": 1.0362, "step": 924 }, { "epoch": 0.41912097870412324, "grad_norm": 0.9362270264620902, "learning_rate": 6.702898550724639e-05, "loss": 1.0574, "step": 925 }, { "epoch": 0.4195740824648845, "grad_norm": 0.9161196090722994, "learning_rate": 6.710144927536232e-05, "loss": 1.0566, "step": 926 }, { "epoch": 0.42002718622564567, "grad_norm": 0.967112141562856, "learning_rate": 6.717391304347827e-05, "loss": 1.04, "step": 927 }, { "epoch": 0.4204802899864069, "grad_norm": 0.9348138719519306, "learning_rate": 6.72463768115942e-05, "loss": 1.0687, "step": 928 }, { "epoch": 0.4209333937471681, "grad_norm": 0.8710813418487341, "learning_rate": 6.731884057971015e-05, "loss": 1.0445, "step": 929 }, { "epoch": 0.42138649750792934, "grad_norm": 0.877457340121064, "learning_rate": 6.73913043478261e-05, "loss": 1.046, "step": 930 }, { "epoch": 0.4218396012686905, "grad_norm": 0.8809240966575627, "learning_rate": 6.746376811594204e-05, "loss": 1.0407, "step": 931 }, { "epoch": 0.42229270502945176, "grad_norm": 0.840272716202961, "learning_rate": 6.753623188405797e-05, "loss": 1.0525, "step": 932 }, { "epoch": 0.42274580879021295, "grad_norm": 0.7950608975441944, "learning_rate": 6.760869565217392e-05, "loss": 1.0466, "step": 933 }, { "epoch": 0.4231989125509742, "grad_norm": 0.7533933045733966, "learning_rate": 6.768115942028985e-05, "loss": 1.0515, "step": 934 }, { "epoch": 0.4236520163117354, "grad_norm": 0.8119271258902228, "learning_rate": 6.77536231884058e-05, "loss": 1.0523, "step": 935 }, { "epoch": 0.4241051200724966, "grad_norm": 0.8059732113436784, "learning_rate": 6.782608695652175e-05, "loss": 1.0336, "step": 936 }, { "epoch": 0.4245582238332578, "grad_norm": 0.6853579204244162, "learning_rate": 6.78985507246377e-05, "loss": 1.0476, "step": 937 }, { "epoch": 0.42501132759401905, "grad_norm": 0.5313730981833339, "learning_rate": 6.797101449275363e-05, "loss": 1.0289, "step": 938 }, { "epoch": 0.42546443135478024, "grad_norm": 0.5673829743469293, "learning_rate": 6.804347826086957e-05, "loss": 1.022, "step": 939 }, { "epoch": 0.4259175351155415, "grad_norm": 0.7233683756832778, "learning_rate": 6.811594202898552e-05, "loss": 1.02, "step": 940 }, { "epoch": 0.42637063887630267, "grad_norm": 0.8578410617748172, "learning_rate": 6.818840579710145e-05, "loss": 1.0315, "step": 941 }, { "epoch": 0.4268237426370639, "grad_norm": 0.8721926310817537, "learning_rate": 6.82608695652174e-05, "loss": 1.0372, "step": 942 }, { "epoch": 0.4272768463978251, "grad_norm": 0.7970740060701538, "learning_rate": 6.833333333333333e-05, "loss": 1.052, "step": 943 }, { "epoch": 0.42772995015858634, "grad_norm": 0.6947810122300955, "learning_rate": 6.840579710144928e-05, "loss": 1.0296, "step": 944 }, { "epoch": 0.4281830539193475, "grad_norm": 0.5875977857087723, "learning_rate": 6.847826086956522e-05, "loss": 1.0429, "step": 945 }, { "epoch": 0.42863615768010876, "grad_norm": 0.5638293724635409, "learning_rate": 6.855072463768117e-05, "loss": 1.0449, "step": 946 }, { "epoch": 0.42908926144086995, "grad_norm": 0.5667014952736253, "learning_rate": 6.86231884057971e-05, "loss": 1.04, "step": 947 }, { "epoch": 0.4295423652016312, "grad_norm": 0.6967478165543187, "learning_rate": 6.869565217391305e-05, "loss": 1.0487, "step": 948 }, { "epoch": 0.4299954689623924, "grad_norm": 0.8844596420570279, "learning_rate": 6.876811594202898e-05, "loss": 1.0541, "step": 949 }, { "epoch": 0.4304485727231536, "grad_norm": 0.9624748222050719, "learning_rate": 6.884057971014493e-05, "loss": 1.0305, "step": 950 }, { "epoch": 0.4309016764839148, "grad_norm": 0.955696960964636, "learning_rate": 6.891304347826088e-05, "loss": 1.0431, "step": 951 }, { "epoch": 0.43135478024467605, "grad_norm": 1.0798294832115816, "learning_rate": 6.898550724637682e-05, "loss": 1.0263, "step": 952 }, { "epoch": 0.43180788400543724, "grad_norm": 0.8730537184619596, "learning_rate": 6.905797101449276e-05, "loss": 1.0495, "step": 953 }, { "epoch": 0.4322609877661985, "grad_norm": 0.7671874591373968, "learning_rate": 6.91304347826087e-05, "loss": 1.0235, "step": 954 }, { "epoch": 0.43271409152695967, "grad_norm": 0.7199588405790406, "learning_rate": 6.920289855072465e-05, "loss": 1.0332, "step": 955 }, { "epoch": 0.4331671952877209, "grad_norm": 0.6333946772229584, "learning_rate": 6.927536231884058e-05, "loss": 1.0663, "step": 956 }, { "epoch": 0.4336202990484821, "grad_norm": 0.524748567136364, "learning_rate": 6.934782608695653e-05, "loss": 1.0535, "step": 957 }, { "epoch": 0.43407340280924334, "grad_norm": 0.48728536357069413, "learning_rate": 6.942028985507246e-05, "loss": 1.0714, "step": 958 }, { "epoch": 0.4345265065700045, "grad_norm": 0.5050342970767003, "learning_rate": 6.949275362318841e-05, "loss": 1.0563, "step": 959 }, { "epoch": 0.43497961033076576, "grad_norm": 0.589830318858697, "learning_rate": 6.956521739130436e-05, "loss": 1.0324, "step": 960 }, { "epoch": 0.43543271409152695, "grad_norm": 0.6278468777688327, "learning_rate": 6.96376811594203e-05, "loss": 1.0346, "step": 961 }, { "epoch": 0.4358858178522882, "grad_norm": 0.5742564796032776, "learning_rate": 6.971014492753623e-05, "loss": 1.0594, "step": 962 }, { "epoch": 0.4363389216130494, "grad_norm": 0.5433942500807342, "learning_rate": 6.978260869565218e-05, "loss": 1.0434, "step": 963 }, { "epoch": 0.4367920253738106, "grad_norm": 0.6362338726318091, "learning_rate": 6.985507246376811e-05, "loss": 1.0364, "step": 964 }, { "epoch": 0.4372451291345718, "grad_norm": 0.7572657340789991, "learning_rate": 6.992753623188407e-05, "loss": 1.0392, "step": 965 }, { "epoch": 0.43769823289533305, "grad_norm": 0.9778520384298276, "learning_rate": 7.000000000000001e-05, "loss": 1.027, "step": 966 }, { "epoch": 0.43815133665609424, "grad_norm": 1.1085746238320664, "learning_rate": 7.007246376811594e-05, "loss": 1.0431, "step": 967 }, { "epoch": 0.4386044404168555, "grad_norm": 0.7810794077524718, "learning_rate": 7.014492753623189e-05, "loss": 1.0345, "step": 968 }, { "epoch": 0.43905754417761667, "grad_norm": 0.5607812366006046, "learning_rate": 7.021739130434783e-05, "loss": 1.0326, "step": 969 }, { "epoch": 0.4395106479383779, "grad_norm": 0.4510985226572317, "learning_rate": 7.028985507246378e-05, "loss": 1.0187, "step": 970 }, { "epoch": 0.4399637516991391, "grad_norm": 0.4946753657319479, "learning_rate": 7.036231884057971e-05, "loss": 1.0229, "step": 971 }, { "epoch": 0.44041685545990034, "grad_norm": 0.6440126069811287, "learning_rate": 7.043478260869566e-05, "loss": 1.0447, "step": 972 }, { "epoch": 0.4408699592206615, "grad_norm": 0.7847868428561421, "learning_rate": 7.050724637681159e-05, "loss": 1.0509, "step": 973 }, { "epoch": 0.44132306298142276, "grad_norm": 0.8681026703993955, "learning_rate": 7.057971014492754e-05, "loss": 1.0398, "step": 974 }, { "epoch": 0.44177616674218395, "grad_norm": 0.8003194405447718, "learning_rate": 7.065217391304349e-05, "loss": 1.0538, "step": 975 }, { "epoch": 0.4422292705029452, "grad_norm": 0.6969968119349681, "learning_rate": 7.072463768115943e-05, "loss": 1.0633, "step": 976 }, { "epoch": 0.4426823742637064, "grad_norm": 0.7323162132100806, "learning_rate": 7.079710144927537e-05, "loss": 1.0611, "step": 977 }, { "epoch": 0.4431354780244676, "grad_norm": 0.7258487796385894, "learning_rate": 7.086956521739131e-05, "loss": 1.0511, "step": 978 }, { "epoch": 0.4435885817852288, "grad_norm": 0.7372804431536764, "learning_rate": 7.094202898550726e-05, "loss": 1.0384, "step": 979 }, { "epoch": 0.44404168554599005, "grad_norm": 0.8788472065941224, "learning_rate": 7.101449275362319e-05, "loss": 1.0346, "step": 980 }, { "epoch": 0.44449478930675124, "grad_norm": 1.0207965784705944, "learning_rate": 7.108695652173914e-05, "loss": 1.0273, "step": 981 }, { "epoch": 0.4449478930675125, "grad_norm": 1.0795202242181177, "learning_rate": 7.115942028985507e-05, "loss": 1.0335, "step": 982 }, { "epoch": 0.44540099682827367, "grad_norm": 0.8915280981177609, "learning_rate": 7.123188405797102e-05, "loss": 1.0472, "step": 983 }, { "epoch": 0.4458541005890349, "grad_norm": 0.7904988936130434, "learning_rate": 7.130434782608696e-05, "loss": 1.0366, "step": 984 }, { "epoch": 0.4463072043497961, "grad_norm": 0.6721841263067336, "learning_rate": 7.137681159420291e-05, "loss": 1.0481, "step": 985 }, { "epoch": 0.44676030811055734, "grad_norm": 0.5092211542152089, "learning_rate": 7.144927536231884e-05, "loss": 1.0534, "step": 986 }, { "epoch": 0.4472134118713185, "grad_norm": 0.523953682396632, "learning_rate": 7.152173913043479e-05, "loss": 1.0385, "step": 987 }, { "epoch": 0.44766651563207976, "grad_norm": 0.557593977322805, "learning_rate": 7.159420289855072e-05, "loss": 1.0399, "step": 988 }, { "epoch": 0.44811961939284095, "grad_norm": 0.44622525840607685, "learning_rate": 7.166666666666667e-05, "loss": 1.0334, "step": 989 }, { "epoch": 0.4485727231536022, "grad_norm": 0.39832973521799464, "learning_rate": 7.173913043478262e-05, "loss": 1.0322, "step": 990 }, { "epoch": 0.4490258269143634, "grad_norm": 0.5115590639521183, "learning_rate": 7.181159420289856e-05, "loss": 1.0248, "step": 991 }, { "epoch": 0.4494789306751246, "grad_norm": 0.5936249028797975, "learning_rate": 7.18840579710145e-05, "loss": 1.053, "step": 992 }, { "epoch": 0.4499320344358858, "grad_norm": 0.7505277605535544, "learning_rate": 7.195652173913044e-05, "loss": 1.0211, "step": 993 }, { "epoch": 0.45038513819664705, "grad_norm": 0.7974112656876703, "learning_rate": 7.202898550724639e-05, "loss": 1.0519, "step": 994 }, { "epoch": 0.45083824195740824, "grad_norm": 0.7473742836547216, "learning_rate": 7.210144927536232e-05, "loss": 1.038, "step": 995 }, { "epoch": 0.4512913457181695, "grad_norm": 0.7263574357184208, "learning_rate": 7.217391304347827e-05, "loss": 1.0335, "step": 996 }, { "epoch": 0.45174444947893067, "grad_norm": 0.7034316422945831, "learning_rate": 7.22463768115942e-05, "loss": 1.0473, "step": 997 }, { "epoch": 0.4521975532396919, "grad_norm": 0.7371716243611999, "learning_rate": 7.231884057971015e-05, "loss": 1.0669, "step": 998 }, { "epoch": 0.4526506570004531, "grad_norm": 0.8758988449804557, "learning_rate": 7.23913043478261e-05, "loss": 1.0661, "step": 999 }, { "epoch": 0.45310376076121434, "grad_norm": 0.9961833894519668, "learning_rate": 7.246376811594204e-05, "loss": 1.019, "step": 1000 }, { "epoch": 0.4535568645219755, "grad_norm": 1.146991013313742, "learning_rate": 7.253623188405797e-05, "loss": 1.0348, "step": 1001 }, { "epoch": 0.45400996828273676, "grad_norm": 0.7454737383199079, "learning_rate": 7.260869565217392e-05, "loss": 1.0262, "step": 1002 }, { "epoch": 0.45446307204349795, "grad_norm": 0.3826111993460407, "learning_rate": 7.268115942028985e-05, "loss": 1.0499, "step": 1003 }, { "epoch": 0.4549161758042592, "grad_norm": 0.4842033762195332, "learning_rate": 7.27536231884058e-05, "loss": 1.0317, "step": 1004 }, { "epoch": 0.4553692795650204, "grad_norm": 0.6958897973104038, "learning_rate": 7.282608695652175e-05, "loss": 1.0244, "step": 1005 }, { "epoch": 0.4558223833257816, "grad_norm": 0.8117338411963315, "learning_rate": 7.289855072463769e-05, "loss": 1.0115, "step": 1006 }, { "epoch": 0.4562754870865428, "grad_norm": 0.9268583358925833, "learning_rate": 7.297101449275363e-05, "loss": 1.032, "step": 1007 }, { "epoch": 0.45672859084730405, "grad_norm": 1.020838915321635, "learning_rate": 7.304347826086957e-05, "loss": 1.0469, "step": 1008 }, { "epoch": 0.45718169460806524, "grad_norm": 1.0000477627131994, "learning_rate": 7.311594202898552e-05, "loss": 1.0422, "step": 1009 }, { "epoch": 0.4576347983688265, "grad_norm": 1.0828925459289305, "learning_rate": 7.318840579710145e-05, "loss": 1.057, "step": 1010 }, { "epoch": 0.45808790212958767, "grad_norm": 0.7734084484600081, "learning_rate": 7.32608695652174e-05, "loss": 1.0321, "step": 1011 }, { "epoch": 0.4585410058903489, "grad_norm": 0.5911697030285394, "learning_rate": 7.333333333333333e-05, "loss": 1.047, "step": 1012 }, { "epoch": 0.4589941096511101, "grad_norm": 0.669970834638163, "learning_rate": 7.340579710144928e-05, "loss": 1.0303, "step": 1013 }, { "epoch": 0.45944721341187134, "grad_norm": 0.7846841409147159, "learning_rate": 7.347826086956522e-05, "loss": 1.051, "step": 1014 }, { "epoch": 0.4599003171726325, "grad_norm": 0.8001013593278896, "learning_rate": 7.355072463768117e-05, "loss": 1.0487, "step": 1015 }, { "epoch": 0.46035342093339376, "grad_norm": 0.7285973365301782, "learning_rate": 7.36231884057971e-05, "loss": 1.037, "step": 1016 }, { "epoch": 0.46080652469415495, "grad_norm": 0.6851626196402256, "learning_rate": 7.369565217391305e-05, "loss": 1.0469, "step": 1017 }, { "epoch": 0.4612596284549162, "grad_norm": 0.6024493893131486, "learning_rate": 7.3768115942029e-05, "loss": 1.0467, "step": 1018 }, { "epoch": 0.4617127322156774, "grad_norm": 0.5802846454871345, "learning_rate": 7.384057971014494e-05, "loss": 1.0499, "step": 1019 }, { "epoch": 0.4621658359764386, "grad_norm": 0.5435141247028327, "learning_rate": 7.391304347826088e-05, "loss": 1.0103, "step": 1020 }, { "epoch": 0.4626189397371998, "grad_norm": 0.5502572893234073, "learning_rate": 7.398550724637682e-05, "loss": 1.0376, "step": 1021 }, { "epoch": 0.46307204349796105, "grad_norm": 0.5714862062539426, "learning_rate": 7.405797101449276e-05, "loss": 1.0488, "step": 1022 }, { "epoch": 0.46352514725872224, "grad_norm": 0.5147192217757677, "learning_rate": 7.41304347826087e-05, "loss": 1.0394, "step": 1023 }, { "epoch": 0.4639782510194835, "grad_norm": 0.5513486591877739, "learning_rate": 7.420289855072465e-05, "loss": 1.0314, "step": 1024 }, { "epoch": 0.46443135478024467, "grad_norm": 0.6091310436433124, "learning_rate": 7.427536231884058e-05, "loss": 1.035, "step": 1025 }, { "epoch": 0.4648844585410059, "grad_norm": 0.6211135810846541, "learning_rate": 7.434782608695653e-05, "loss": 1.0543, "step": 1026 }, { "epoch": 0.4653375623017671, "grad_norm": 0.6168726844597251, "learning_rate": 7.442028985507246e-05, "loss": 1.0341, "step": 1027 }, { "epoch": 0.46579066606252834, "grad_norm": 0.7324034807922628, "learning_rate": 7.449275362318841e-05, "loss": 1.0577, "step": 1028 }, { "epoch": 0.4662437698232895, "grad_norm": 0.9668430733449924, "learning_rate": 7.456521739130435e-05, "loss": 1.0713, "step": 1029 }, { "epoch": 0.46669687358405076, "grad_norm": 1.1392300823321582, "learning_rate": 7.46376811594203e-05, "loss": 1.0436, "step": 1030 }, { "epoch": 0.46714997734481195, "grad_norm": 0.8673337345583584, "learning_rate": 7.471014492753623e-05, "loss": 1.0258, "step": 1031 }, { "epoch": 0.4676030811055732, "grad_norm": 0.771726607887614, "learning_rate": 7.478260869565218e-05, "loss": 1.0459, "step": 1032 }, { "epoch": 0.4680561848663344, "grad_norm": 0.8025453986515559, "learning_rate": 7.485507246376813e-05, "loss": 1.021, "step": 1033 }, { "epoch": 0.4685092886270956, "grad_norm": 0.8867561495849974, "learning_rate": 7.492753623188407e-05, "loss": 1.0121, "step": 1034 }, { "epoch": 0.4689623923878568, "grad_norm": 0.8410073741060616, "learning_rate": 7.500000000000001e-05, "loss": 1.0322, "step": 1035 }, { "epoch": 0.46941549614861805, "grad_norm": 0.7513237572938335, "learning_rate": 7.507246376811594e-05, "loss": 1.0213, "step": 1036 }, { "epoch": 0.46986859990937924, "grad_norm": 0.7044293628046868, "learning_rate": 7.514492753623189e-05, "loss": 1.0402, "step": 1037 }, { "epoch": 0.4703217036701405, "grad_norm": 0.7475266030744413, "learning_rate": 7.521739130434783e-05, "loss": 1.073, "step": 1038 }, { "epoch": 0.47077480743090167, "grad_norm": 0.8316739421226873, "learning_rate": 7.528985507246378e-05, "loss": 1.037, "step": 1039 }, { "epoch": 0.4712279111916629, "grad_norm": 0.8632006093441521, "learning_rate": 7.536231884057971e-05, "loss": 1.0516, "step": 1040 }, { "epoch": 0.4716810149524241, "grad_norm": 0.9921012231019771, "learning_rate": 7.543478260869566e-05, "loss": 1.0587, "step": 1041 }, { "epoch": 0.47213411871318534, "grad_norm": 1.1552740780018422, "learning_rate": 7.550724637681159e-05, "loss": 1.0452, "step": 1042 }, { "epoch": 0.4725872224739465, "grad_norm": 0.7705106274711769, "learning_rate": 7.557971014492754e-05, "loss": 1.0385, "step": 1043 }, { "epoch": 0.47304032623470776, "grad_norm": 0.6629629928411389, "learning_rate": 7.565217391304349e-05, "loss": 1.0248, "step": 1044 }, { "epoch": 0.47349342999546895, "grad_norm": 0.6226975532943384, "learning_rate": 7.572463768115943e-05, "loss": 1.0501, "step": 1045 }, { "epoch": 0.4739465337562302, "grad_norm": 0.5206648100949298, "learning_rate": 7.579710144927536e-05, "loss": 1.0261, "step": 1046 }, { "epoch": 0.4743996375169914, "grad_norm": 0.5229955587443459, "learning_rate": 7.586956521739131e-05, "loss": 1.0148, "step": 1047 }, { "epoch": 0.4748527412777526, "grad_norm": 0.653507774145783, "learning_rate": 7.594202898550726e-05, "loss": 1.039, "step": 1048 }, { "epoch": 0.4753058450385138, "grad_norm": 0.7334844379196753, "learning_rate": 7.601449275362319e-05, "loss": 1.018, "step": 1049 }, { "epoch": 0.47575894879927505, "grad_norm": 0.7264509127668619, "learning_rate": 7.608695652173914e-05, "loss": 1.0175, "step": 1050 }, { "epoch": 0.47621205256003624, "grad_norm": 0.759393732377215, "learning_rate": 7.615942028985507e-05, "loss": 1.0352, "step": 1051 }, { "epoch": 0.4766651563207975, "grad_norm": 0.7087771788624839, "learning_rate": 7.623188405797102e-05, "loss": 1.043, "step": 1052 }, { "epoch": 0.47711826008155866, "grad_norm": 0.5948132939256898, "learning_rate": 7.630434782608696e-05, "loss": 1.0238, "step": 1053 }, { "epoch": 0.4775713638423199, "grad_norm": 0.6068368730597038, "learning_rate": 7.637681159420291e-05, "loss": 1.0502, "step": 1054 }, { "epoch": 0.4780244676030811, "grad_norm": 0.5826640466791965, "learning_rate": 7.644927536231884e-05, "loss": 1.0243, "step": 1055 }, { "epoch": 0.47847757136384234, "grad_norm": 0.4863689170734788, "learning_rate": 7.652173913043479e-05, "loss": 1.0323, "step": 1056 }, { "epoch": 0.4789306751246035, "grad_norm": 0.44894288944177585, "learning_rate": 7.659420289855072e-05, "loss": 1.0288, "step": 1057 }, { "epoch": 0.47938377888536476, "grad_norm": 0.483132621254299, "learning_rate": 7.666666666666668e-05, "loss": 1.0346, "step": 1058 }, { "epoch": 0.47983688264612595, "grad_norm": 0.5491444722536396, "learning_rate": 7.673913043478262e-05, "loss": 1.0541, "step": 1059 }, { "epoch": 0.4802899864068872, "grad_norm": 0.5985860129723741, "learning_rate": 7.681159420289856e-05, "loss": 1.0292, "step": 1060 }, { "epoch": 0.4807430901676484, "grad_norm": 0.6579202715470125, "learning_rate": 7.68840579710145e-05, "loss": 1.0269, "step": 1061 }, { "epoch": 0.4811961939284096, "grad_norm": 0.6354894826950389, "learning_rate": 7.695652173913044e-05, "loss": 1.0377, "step": 1062 }, { "epoch": 0.4816492976891708, "grad_norm": 0.7346089330096801, "learning_rate": 7.702898550724639e-05, "loss": 1.0171, "step": 1063 }, { "epoch": 0.48210240144993205, "grad_norm": 0.8732165746373935, "learning_rate": 7.710144927536232e-05, "loss": 1.0103, "step": 1064 }, { "epoch": 0.48255550521069324, "grad_norm": 0.8749138250149588, "learning_rate": 7.717391304347827e-05, "loss": 1.0149, "step": 1065 }, { "epoch": 0.4830086089714545, "grad_norm": 0.8096448634958864, "learning_rate": 7.72463768115942e-05, "loss": 1.0191, "step": 1066 }, { "epoch": 0.48346171273221566, "grad_norm": 0.7842592595590672, "learning_rate": 7.731884057971015e-05, "loss": 1.0351, "step": 1067 }, { "epoch": 0.4839148164929769, "grad_norm": 0.7293361190790575, "learning_rate": 7.73913043478261e-05, "loss": 1.0493, "step": 1068 }, { "epoch": 0.4843679202537381, "grad_norm": 0.7292219026241858, "learning_rate": 7.746376811594204e-05, "loss": 1.0298, "step": 1069 }, { "epoch": 0.48482102401449934, "grad_norm": 0.7791258114553521, "learning_rate": 7.753623188405797e-05, "loss": 1.0299, "step": 1070 }, { "epoch": 0.4852741277752605, "grad_norm": 0.68363165817219, "learning_rate": 7.760869565217392e-05, "loss": 1.041, "step": 1071 }, { "epoch": 0.48572723153602176, "grad_norm": 0.6167092108351728, "learning_rate": 7.768115942028987e-05, "loss": 1.014, "step": 1072 }, { "epoch": 0.48618033529678295, "grad_norm": 0.7591966777073307, "learning_rate": 7.775362318840581e-05, "loss": 1.064, "step": 1073 }, { "epoch": 0.4866334390575442, "grad_norm": 0.8158736889287136, "learning_rate": 7.782608695652175e-05, "loss": 1.041, "step": 1074 }, { "epoch": 0.4870865428183054, "grad_norm": 0.6321927588077079, "learning_rate": 7.789855072463769e-05, "loss": 1.065, "step": 1075 }, { "epoch": 0.4875396465790666, "grad_norm": 0.6128875494025281, "learning_rate": 7.797101449275363e-05, "loss": 1.0403, "step": 1076 }, { "epoch": 0.4879927503398278, "grad_norm": 0.8018497114949825, "learning_rate": 7.804347826086957e-05, "loss": 1.0426, "step": 1077 }, { "epoch": 0.48844585410058905, "grad_norm": 0.933889271430776, "learning_rate": 7.811594202898552e-05, "loss": 1.0519, "step": 1078 }, { "epoch": 0.48889895786135024, "grad_norm": 1.0943492368302825, "learning_rate": 7.818840579710145e-05, "loss": 1.0421, "step": 1079 }, { "epoch": 0.4893520616221115, "grad_norm": 1.000332989793235, "learning_rate": 7.82608695652174e-05, "loss": 1.0212, "step": 1080 }, { "epoch": 0.48980516538287266, "grad_norm": 1.05286780284306, "learning_rate": 7.833333333333333e-05, "loss": 1.0478, "step": 1081 }, { "epoch": 0.4902582691436339, "grad_norm": 0.9343281567547387, "learning_rate": 7.840579710144928e-05, "loss": 1.0475, "step": 1082 }, { "epoch": 0.4907113729043951, "grad_norm": 0.8347296073430865, "learning_rate": 7.847826086956522e-05, "loss": 1.0309, "step": 1083 }, { "epoch": 0.49116447666515634, "grad_norm": 0.8115535892371672, "learning_rate": 7.855072463768117e-05, "loss": 1.049, "step": 1084 }, { "epoch": 0.4916175804259175, "grad_norm": 0.7933065030479503, "learning_rate": 7.86231884057971e-05, "loss": 1.038, "step": 1085 }, { "epoch": 0.49207068418667876, "grad_norm": 0.8698447238392416, "learning_rate": 7.869565217391305e-05, "loss": 1.0356, "step": 1086 }, { "epoch": 0.49252378794743995, "grad_norm": 0.9885266876025843, "learning_rate": 7.8768115942029e-05, "loss": 1.0242, "step": 1087 }, { "epoch": 0.4929768917082012, "grad_norm": 0.9339351072120061, "learning_rate": 7.884057971014494e-05, "loss": 1.0492, "step": 1088 }, { "epoch": 0.4934299954689624, "grad_norm": 0.7685596178072077, "learning_rate": 7.891304347826088e-05, "loss": 1.0614, "step": 1089 }, { "epoch": 0.4938830992297236, "grad_norm": 0.6027631597446408, "learning_rate": 7.898550724637682e-05, "loss": 1.0103, "step": 1090 }, { "epoch": 0.4943362029904848, "grad_norm": 0.5801226344317478, "learning_rate": 7.905797101449276e-05, "loss": 1.0416, "step": 1091 }, { "epoch": 0.49478930675124605, "grad_norm": 0.513913691859463, "learning_rate": 7.91304347826087e-05, "loss": 1.0335, "step": 1092 }, { "epoch": 0.49524241051200724, "grad_norm": 0.4288469061148328, "learning_rate": 7.920289855072465e-05, "loss": 1.0529, "step": 1093 }, { "epoch": 0.4956955142727685, "grad_norm": 0.4760091089955674, "learning_rate": 7.927536231884058e-05, "loss": 1.0415, "step": 1094 }, { "epoch": 0.49614861803352966, "grad_norm": 0.45804473185226674, "learning_rate": 7.934782608695653e-05, "loss": 1.0188, "step": 1095 }, { "epoch": 0.4966017217942909, "grad_norm": 0.4367719824522219, "learning_rate": 7.942028985507246e-05, "loss": 1.0357, "step": 1096 }, { "epoch": 0.4970548255550521, "grad_norm": 0.5875023765010227, "learning_rate": 7.949275362318841e-05, "loss": 1.0242, "step": 1097 }, { "epoch": 0.49750792931581334, "grad_norm": 0.7703473089661725, "learning_rate": 7.956521739130435e-05, "loss": 1.0355, "step": 1098 }, { "epoch": 0.4979610330765745, "grad_norm": 1.0395558607515227, "learning_rate": 7.96376811594203e-05, "loss": 1.0387, "step": 1099 }, { "epoch": 0.49841413683733576, "grad_norm": 1.2230081189454132, "learning_rate": 7.971014492753623e-05, "loss": 1.0165, "step": 1100 }, { "epoch": 0.49886724059809695, "grad_norm": 0.6614189710717384, "learning_rate": 7.978260869565218e-05, "loss": 1.0337, "step": 1101 }, { "epoch": 0.4993203443588582, "grad_norm": 0.5092214276358354, "learning_rate": 7.985507246376813e-05, "loss": 1.0351, "step": 1102 }, { "epoch": 0.4997734481196194, "grad_norm": 0.7460532952678652, "learning_rate": 7.992753623188407e-05, "loss": 1.0606, "step": 1103 }, { "epoch": 0.5002265518803806, "grad_norm": 1.0346051577411288, "learning_rate": 8e-05, "loss": 1.0403, "step": 1104 }, { "epoch": 0.5006796556411418, "grad_norm": 1.1271299234567023, "learning_rate": 7.999999799855449e-05, "loss": 1.0243, "step": 1105 }, { "epoch": 0.501132759401903, "grad_norm": 0.7331734828663232, "learning_rate": 7.999999199421812e-05, "loss": 1.0168, "step": 1106 }, { "epoch": 0.5015858631626643, "grad_norm": 0.5394774516234627, "learning_rate": 7.99999819869915e-05, "loss": 1.0229, "step": 1107 }, { "epoch": 0.5020389669234254, "grad_norm": 0.5713945333229701, "learning_rate": 7.999996797687565e-05, "loss": 1.0236, "step": 1108 }, { "epoch": 0.5024920706841867, "grad_norm": 0.599382135764971, "learning_rate": 7.999994996387195e-05, "loss": 1.0111, "step": 1109 }, { "epoch": 0.5029451744449479, "grad_norm": 0.643174535928619, "learning_rate": 7.999992794798221e-05, "loss": 1.0342, "step": 1110 }, { "epoch": 0.5033982782057091, "grad_norm": 0.7728594691961052, "learning_rate": 7.999990192920862e-05, "loss": 1.0573, "step": 1111 }, { "epoch": 0.5038513819664703, "grad_norm": 0.8429443426039722, "learning_rate": 7.999987190755382e-05, "loss": 1.0134, "step": 1112 }, { "epoch": 0.5043044857272315, "grad_norm": 0.9188035126333576, "learning_rate": 7.999983788302078e-05, "loss": 1.036, "step": 1113 }, { "epoch": 0.5047575894879928, "grad_norm": 0.9697803737384354, "learning_rate": 7.999979985561294e-05, "loss": 1.0281, "step": 1114 }, { "epoch": 0.505210693248754, "grad_norm": 0.9954136595936456, "learning_rate": 7.999975782533405e-05, "loss": 1.0411, "step": 1115 }, { "epoch": 0.5056637970095151, "grad_norm": 0.9921024543561481, "learning_rate": 7.999971179218836e-05, "loss": 1.0284, "step": 1116 }, { "epoch": 0.5061169007702764, "grad_norm": 1.0614303880648628, "learning_rate": 7.999966175618047e-05, "loss": 1.0439, "step": 1117 }, { "epoch": 0.5065700045310376, "grad_norm": 0.8963714131969625, "learning_rate": 7.99996077173154e-05, "loss": 1.032, "step": 1118 }, { "epoch": 0.5070231082917989, "grad_norm": 0.7466338619203818, "learning_rate": 7.99995496755985e-05, "loss": 1.0252, "step": 1119 }, { "epoch": 0.50747621205256, "grad_norm": 0.6994194085564243, "learning_rate": 7.999948763103565e-05, "loss": 1.0359, "step": 1120 }, { "epoch": 0.5079293158133212, "grad_norm": 0.6365721407473129, "learning_rate": 7.999942158363303e-05, "loss": 1.0226, "step": 1121 }, { "epoch": 0.5083824195740825, "grad_norm": 0.6201291475471515, "learning_rate": 7.999935153339723e-05, "loss": 1.0363, "step": 1122 }, { "epoch": 0.5088355233348437, "grad_norm": 0.6850395547171272, "learning_rate": 7.999927748033528e-05, "loss": 1.0256, "step": 1123 }, { "epoch": 0.5092886270956049, "grad_norm": 0.7928548089882755, "learning_rate": 7.999919942445459e-05, "loss": 1.0314, "step": 1124 }, { "epoch": 0.5097417308563661, "grad_norm": 0.9331021118468223, "learning_rate": 7.999911736576296e-05, "loss": 1.0483, "step": 1125 }, { "epoch": 0.5101948346171273, "grad_norm": 0.9763084900599104, "learning_rate": 7.999903130426863e-05, "loss": 1.0406, "step": 1126 }, { "epoch": 0.5106479383778886, "grad_norm": 0.9499011555740294, "learning_rate": 7.999894123998018e-05, "loss": 1.0385, "step": 1127 }, { "epoch": 0.5111010421386497, "grad_norm": 0.88527504741301, "learning_rate": 7.999884717290664e-05, "loss": 1.0583, "step": 1128 }, { "epoch": 0.511554145899411, "grad_norm": 0.7578859906818105, "learning_rate": 7.999874910305743e-05, "loss": 1.0375, "step": 1129 }, { "epoch": 0.5120072496601722, "grad_norm": 0.5703387720139756, "learning_rate": 7.999864703044234e-05, "loss": 1.0243, "step": 1130 }, { "epoch": 0.5124603534209334, "grad_norm": 0.4498483750323068, "learning_rate": 7.999854095507159e-05, "loss": 1.0504, "step": 1131 }, { "epoch": 0.5129134571816946, "grad_norm": 0.40924862905323933, "learning_rate": 7.999843087695582e-05, "loss": 1.0415, "step": 1132 }, { "epoch": 0.5133665609424558, "grad_norm": 0.5161058987857556, "learning_rate": 7.999831679610602e-05, "loss": 1.0161, "step": 1133 }, { "epoch": 0.513819664703217, "grad_norm": 0.6041051046642323, "learning_rate": 7.999819871253363e-05, "loss": 1.0077, "step": 1134 }, { "epoch": 0.5142727684639783, "grad_norm": 0.4703399324201906, "learning_rate": 7.999807662625045e-05, "loss": 1.0242, "step": 1135 }, { "epoch": 0.5147258722247394, "grad_norm": 0.43859817653187794, "learning_rate": 7.999795053726867e-05, "loss": 1.0375, "step": 1136 }, { "epoch": 0.5151789759855007, "grad_norm": 0.5143597895129192, "learning_rate": 7.999782044560097e-05, "loss": 1.007, "step": 1137 }, { "epoch": 0.5156320797462619, "grad_norm": 0.5345171789314354, "learning_rate": 7.999768635126033e-05, "loss": 1.032, "step": 1138 }, { "epoch": 0.5160851835070231, "grad_norm": 0.5820233304144337, "learning_rate": 7.999754825426017e-05, "loss": 1.0419, "step": 1139 }, { "epoch": 0.5165382872677843, "grad_norm": 0.557048880357325, "learning_rate": 7.999740615461432e-05, "loss": 1.0346, "step": 1140 }, { "epoch": 0.5169913910285455, "grad_norm": 0.5011765413393698, "learning_rate": 7.9997260052337e-05, "loss": 1.0374, "step": 1141 }, { "epoch": 0.5174444947893068, "grad_norm": 0.473962198796823, "learning_rate": 7.999710994744282e-05, "loss": 1.0337, "step": 1142 }, { "epoch": 0.517897598550068, "grad_norm": 0.425701393805719, "learning_rate": 7.999695583994682e-05, "loss": 1.0201, "step": 1143 }, { "epoch": 0.5183507023108291, "grad_norm": 0.40908390052979815, "learning_rate": 7.999679772986438e-05, "loss": 1.0204, "step": 1144 }, { "epoch": 0.5188038060715904, "grad_norm": 0.45557885728743924, "learning_rate": 7.999663561721138e-05, "loss": 1.0265, "step": 1145 }, { "epoch": 0.5192569098323516, "grad_norm": 0.5811167016413247, "learning_rate": 7.9996469502004e-05, "loss": 1.0189, "step": 1146 }, { "epoch": 0.5197100135931129, "grad_norm": 0.6812286760973083, "learning_rate": 7.999629938425889e-05, "loss": 1.0382, "step": 1147 }, { "epoch": 0.520163117353874, "grad_norm": 0.5949132106255988, "learning_rate": 7.999612526399306e-05, "loss": 1.0244, "step": 1148 }, { "epoch": 0.5206162211146352, "grad_norm": 0.409770822039131, "learning_rate": 7.999594714122394e-05, "loss": 1.0452, "step": 1149 }, { "epoch": 0.5210693248753965, "grad_norm": 0.5183945891910413, "learning_rate": 7.999576501596935e-05, "loss": 1.0417, "step": 1150 }, { "epoch": 0.5215224286361577, "grad_norm": 0.5706159201098238, "learning_rate": 7.99955788882475e-05, "loss": 1.0286, "step": 1151 }, { "epoch": 0.5219755323969189, "grad_norm": 0.512486959009842, "learning_rate": 7.999538875807705e-05, "loss": 1.0259, "step": 1152 }, { "epoch": 0.5224286361576801, "grad_norm": 0.40023721602471335, "learning_rate": 7.999519462547701e-05, "loss": 1.0362, "step": 1153 }, { "epoch": 0.5228817399184413, "grad_norm": 0.44619253909311396, "learning_rate": 7.999499649046682e-05, "loss": 1.0263, "step": 1154 }, { "epoch": 0.5233348436792026, "grad_norm": 0.6107577053411937, "learning_rate": 7.999479435306628e-05, "loss": 1.0174, "step": 1155 }, { "epoch": 0.5237879474399637, "grad_norm": 0.7036878425411313, "learning_rate": 7.999458821329565e-05, "loss": 1.0233, "step": 1156 }, { "epoch": 0.524241051200725, "grad_norm": 0.7573483111542322, "learning_rate": 7.999437807117553e-05, "loss": 1.0253, "step": 1157 }, { "epoch": 0.5246941549614862, "grad_norm": 0.9334261094598572, "learning_rate": 7.999416392672696e-05, "loss": 1.0235, "step": 1158 }, { "epoch": 0.5251472587222474, "grad_norm": 1.1407361868657449, "learning_rate": 7.99939457799714e-05, "loss": 1.0185, "step": 1159 }, { "epoch": 0.5256003624830086, "grad_norm": 0.8337982272986525, "learning_rate": 7.999372363093063e-05, "loss": 1.031, "step": 1160 }, { "epoch": 0.5260534662437698, "grad_norm": 0.6419787764983016, "learning_rate": 7.99934974796269e-05, "loss": 1.017, "step": 1161 }, { "epoch": 0.526506570004531, "grad_norm": 0.546080042071012, "learning_rate": 7.999326732608285e-05, "loss": 1.024, "step": 1162 }, { "epoch": 0.5269596737652923, "grad_norm": 0.4734937496453547, "learning_rate": 7.999303317032152e-05, "loss": 1.0386, "step": 1163 }, { "epoch": 0.5274127775260534, "grad_norm": 0.4250620500636937, "learning_rate": 7.999279501236631e-05, "loss": 1.0265, "step": 1164 }, { "epoch": 0.5278658812868147, "grad_norm": 0.5004320178233093, "learning_rate": 7.999255285224108e-05, "loss": 1.033, "step": 1165 }, { "epoch": 0.5283189850475759, "grad_norm": 0.5602386874642422, "learning_rate": 7.999230668997007e-05, "loss": 1.0177, "step": 1166 }, { "epoch": 0.5287720888083371, "grad_norm": 0.6014375715088095, "learning_rate": 7.99920565255779e-05, "loss": 1.0409, "step": 1167 }, { "epoch": 0.5292251925690983, "grad_norm": 0.6095291652555817, "learning_rate": 7.999180235908958e-05, "loss": 1.0372, "step": 1168 }, { "epoch": 0.5296782963298595, "grad_norm": 0.5500631293376748, "learning_rate": 7.99915441905306e-05, "loss": 1.0207, "step": 1169 }, { "epoch": 0.5301314000906208, "grad_norm": 0.580136877674314, "learning_rate": 7.999128201992674e-05, "loss": 1.0324, "step": 1170 }, { "epoch": 0.530584503851382, "grad_norm": 0.5825411429491368, "learning_rate": 7.999101584730428e-05, "loss": 0.9971, "step": 1171 }, { "epoch": 0.5310376076121431, "grad_norm": 0.6461377022380741, "learning_rate": 7.999074567268984e-05, "loss": 1.0303, "step": 1172 }, { "epoch": 0.5314907113729044, "grad_norm": 0.7457727899688372, "learning_rate": 7.999047149611044e-05, "loss": 1.0552, "step": 1173 }, { "epoch": 0.5319438151336656, "grad_norm": 0.8180058949933293, "learning_rate": 7.999019331759354e-05, "loss": 1.0179, "step": 1174 }, { "epoch": 0.5323969188944269, "grad_norm": 0.8284040394328297, "learning_rate": 7.998991113716697e-05, "loss": 1.0123, "step": 1175 }, { "epoch": 0.532850022655188, "grad_norm": 0.7923554749431125, "learning_rate": 7.998962495485898e-05, "loss": 1.0212, "step": 1176 }, { "epoch": 0.5333031264159492, "grad_norm": 0.7710331061010077, "learning_rate": 7.998933477069818e-05, "loss": 1.0386, "step": 1177 }, { "epoch": 0.5337562301767105, "grad_norm": 0.8521742759690789, "learning_rate": 7.998904058471365e-05, "loss": 1.0372, "step": 1178 }, { "epoch": 0.5342093339374717, "grad_norm": 0.9197953844681924, "learning_rate": 7.998874239693479e-05, "loss": 1.0109, "step": 1179 }, { "epoch": 0.5346624376982329, "grad_norm": 0.9357708334475308, "learning_rate": 7.998844020739146e-05, "loss": 1.0508, "step": 1180 }, { "epoch": 0.5351155414589941, "grad_norm": 0.9254661342059657, "learning_rate": 7.99881340161139e-05, "loss": 1.0121, "step": 1181 }, { "epoch": 0.5355686452197553, "grad_norm": 0.8195862624160726, "learning_rate": 7.998782382313275e-05, "loss": 1.0441, "step": 1182 }, { "epoch": 0.5360217489805166, "grad_norm": 0.6106258320722989, "learning_rate": 7.998750962847904e-05, "loss": 1.0306, "step": 1183 }, { "epoch": 0.5364748527412777, "grad_norm": 0.5195142176472523, "learning_rate": 7.998719143218424e-05, "loss": 1.0104, "step": 1184 }, { "epoch": 0.536927956502039, "grad_norm": 0.5536536973417522, "learning_rate": 7.998686923428018e-05, "loss": 1.0395, "step": 1185 }, { "epoch": 0.5373810602628002, "grad_norm": 0.6537831499757324, "learning_rate": 7.998654303479908e-05, "loss": 1.039, "step": 1186 }, { "epoch": 0.5378341640235614, "grad_norm": 0.7069536262842647, "learning_rate": 7.998621283377361e-05, "loss": 1.0225, "step": 1187 }, { "epoch": 0.5382872677843226, "grad_norm": 0.6909714921932185, "learning_rate": 7.998587863123682e-05, "loss": 1.0051, "step": 1188 }, { "epoch": 0.5387403715450838, "grad_norm": 0.6635845935642078, "learning_rate": 7.998554042722213e-05, "loss": 1.0376, "step": 1189 }, { "epoch": 0.539193475305845, "grad_norm": 0.7152550333238266, "learning_rate": 7.99851982217634e-05, "loss": 1.0034, "step": 1190 }, { "epoch": 0.5396465790666063, "grad_norm": 0.7899875659710719, "learning_rate": 7.998485201489487e-05, "loss": 1.0418, "step": 1191 }, { "epoch": 0.5400996828273674, "grad_norm": 0.6709646000487871, "learning_rate": 7.998450180665118e-05, "loss": 1.0219, "step": 1192 }, { "epoch": 0.5405527865881287, "grad_norm": 0.46282934900689526, "learning_rate": 7.998414759706739e-05, "loss": 1.0048, "step": 1193 }, { "epoch": 0.5410058903488899, "grad_norm": 0.4473566604279943, "learning_rate": 7.998378938617894e-05, "loss": 1.0225, "step": 1194 }, { "epoch": 0.5414589941096511, "grad_norm": 0.5212484622079214, "learning_rate": 7.998342717402168e-05, "loss": 1.03, "step": 1195 }, { "epoch": 0.5419120978704123, "grad_norm": 0.5318025072319275, "learning_rate": 7.998306096063184e-05, "loss": 1.0178, "step": 1196 }, { "epoch": 0.5423652016311735, "grad_norm": 0.4873535216773402, "learning_rate": 7.99826907460461e-05, "loss": 1.0276, "step": 1197 }, { "epoch": 0.5428183053919348, "grad_norm": 0.413248353384605, "learning_rate": 7.998231653030148e-05, "loss": 1.0345, "step": 1198 }, { "epoch": 0.543271409152696, "grad_norm": 0.4097158866560782, "learning_rate": 7.998193831343543e-05, "loss": 1.0201, "step": 1199 }, { "epoch": 0.5437245129134571, "grad_norm": 0.427425153055543, "learning_rate": 7.998155609548581e-05, "loss": 1.025, "step": 1200 }, { "epoch": 0.5441776166742184, "grad_norm": 0.3427221042936924, "learning_rate": 7.998116987649089e-05, "loss": 1.0304, "step": 1201 }, { "epoch": 0.5446307204349796, "grad_norm": 0.4089941315363874, "learning_rate": 7.998077965648927e-05, "loss": 1.0268, "step": 1202 }, { "epoch": 0.5450838241957409, "grad_norm": 0.579876602293775, "learning_rate": 7.998038543552005e-05, "loss": 1.0321, "step": 1203 }, { "epoch": 0.545536927956502, "grad_norm": 0.5830065995168041, "learning_rate": 7.997998721362265e-05, "loss": 1.0348, "step": 1204 }, { "epoch": 0.5459900317172632, "grad_norm": 0.4833527871245349, "learning_rate": 7.997958499083691e-05, "loss": 1.043, "step": 1205 }, { "epoch": 0.5464431354780245, "grad_norm": 0.5815806378388118, "learning_rate": 7.997917876720312e-05, "loss": 1.0337, "step": 1206 }, { "epoch": 0.5468962392387857, "grad_norm": 0.6055297128056428, "learning_rate": 7.997876854276191e-05, "loss": 1.0104, "step": 1207 }, { "epoch": 0.5473493429995469, "grad_norm": 0.46809586300501566, "learning_rate": 7.997835431755434e-05, "loss": 1.0177, "step": 1208 }, { "epoch": 0.5478024467603081, "grad_norm": 0.5152588372863512, "learning_rate": 7.997793609162184e-05, "loss": 1.0304, "step": 1209 }, { "epoch": 0.5482555505210693, "grad_norm": 0.49666787586611516, "learning_rate": 7.997751386500629e-05, "loss": 1.0077, "step": 1210 }, { "epoch": 0.5487086542818306, "grad_norm": 0.5505448450068152, "learning_rate": 7.997708763774993e-05, "loss": 1.0308, "step": 1211 }, { "epoch": 0.5491617580425917, "grad_norm": 0.615473800778557, "learning_rate": 7.997665740989542e-05, "loss": 1.0097, "step": 1212 }, { "epoch": 0.549614861803353, "grad_norm": 0.6137749860185625, "learning_rate": 7.99762231814858e-05, "loss": 1.0268, "step": 1213 }, { "epoch": 0.5500679655641142, "grad_norm": 0.5662494996695859, "learning_rate": 7.997578495256454e-05, "loss": 1.0451, "step": 1214 }, { "epoch": 0.5505210693248754, "grad_norm": 0.5492412815062004, "learning_rate": 7.997534272317549e-05, "loss": 1.0499, "step": 1215 }, { "epoch": 0.5509741730856366, "grad_norm": 0.5922299199209929, "learning_rate": 7.99748964933629e-05, "loss": 1.0498, "step": 1216 }, { "epoch": 0.5514272768463978, "grad_norm": 0.5356408710357502, "learning_rate": 7.997444626317144e-05, "loss": 1.013, "step": 1217 }, { "epoch": 0.551880380607159, "grad_norm": 0.57341782856597, "learning_rate": 7.997399203264615e-05, "loss": 1.0252, "step": 1218 }, { "epoch": 0.5523334843679203, "grad_norm": 0.8464183922154396, "learning_rate": 7.997353380183248e-05, "loss": 1.0106, "step": 1219 }, { "epoch": 0.5527865881286814, "grad_norm": 1.1596220469596072, "learning_rate": 7.997307157077628e-05, "loss": 1.0467, "step": 1220 }, { "epoch": 0.5532396918894427, "grad_norm": 0.8573063683837182, "learning_rate": 7.997260533952385e-05, "loss": 1.008, "step": 1221 }, { "epoch": 0.5536927956502039, "grad_norm": 0.7219472060286635, "learning_rate": 7.997213510812182e-05, "loss": 1.0168, "step": 1222 }, { "epoch": 0.5541458994109651, "grad_norm": 0.6208241109522304, "learning_rate": 7.997166087661724e-05, "loss": 1.0226, "step": 1223 }, { "epoch": 0.5545990031717263, "grad_norm": 0.5378763367037768, "learning_rate": 7.997118264505758e-05, "loss": 1.0322, "step": 1224 }, { "epoch": 0.5550521069324875, "grad_norm": 0.47249866135523294, "learning_rate": 7.997070041349068e-05, "loss": 1.0483, "step": 1225 }, { "epoch": 0.5555052106932488, "grad_norm": 0.3892861486177576, "learning_rate": 7.997021418196483e-05, "loss": 1.0071, "step": 1226 }, { "epoch": 0.55595831445401, "grad_norm": 0.4062535196097057, "learning_rate": 7.996972395052865e-05, "loss": 1.0283, "step": 1227 }, { "epoch": 0.5564114182147711, "grad_norm": 0.4880528033542733, "learning_rate": 7.996922971923122e-05, "loss": 1.0312, "step": 1228 }, { "epoch": 0.5568645219755324, "grad_norm": 0.4748119628816332, "learning_rate": 7.996873148812199e-05, "loss": 1.0226, "step": 1229 }, { "epoch": 0.5573176257362936, "grad_norm": 0.5469481155800627, "learning_rate": 7.996822925725084e-05, "loss": 1.0339, "step": 1230 }, { "epoch": 0.5577707294970549, "grad_norm": 0.6632808041709943, "learning_rate": 7.996772302666801e-05, "loss": 1.0087, "step": 1231 }, { "epoch": 0.558223833257816, "grad_norm": 0.8120664816479919, "learning_rate": 7.996721279642416e-05, "loss": 1.0364, "step": 1232 }, { "epoch": 0.5586769370185772, "grad_norm": 0.9414688804116748, "learning_rate": 7.996669856657035e-05, "loss": 1.0345, "step": 1233 }, { "epoch": 0.5591300407793385, "grad_norm": 1.0923501162747409, "learning_rate": 7.996618033715805e-05, "loss": 0.9894, "step": 1234 }, { "epoch": 0.5595831445400997, "grad_norm": 0.8690879582186088, "learning_rate": 7.996565810823911e-05, "loss": 1.0388, "step": 1235 }, { "epoch": 0.5600362483008609, "grad_norm": 0.709951212952395, "learning_rate": 7.99651318798658e-05, "loss": 1.0398, "step": 1236 }, { "epoch": 0.5604893520616221, "grad_norm": 0.5256449579338882, "learning_rate": 7.996460165209079e-05, "loss": 1.023, "step": 1237 }, { "epoch": 0.5609424558223833, "grad_norm": 0.3474301003864474, "learning_rate": 7.99640674249671e-05, "loss": 1.0253, "step": 1238 }, { "epoch": 0.5613955595831446, "grad_norm": 0.39684222277923, "learning_rate": 7.996352919854822e-05, "loss": 1.0321, "step": 1239 }, { "epoch": 0.5618486633439057, "grad_norm": 0.4252769565948308, "learning_rate": 7.996298697288802e-05, "loss": 1.0359, "step": 1240 }, { "epoch": 0.562301767104667, "grad_norm": 0.48391131115190605, "learning_rate": 7.996244074804076e-05, "loss": 1.0173, "step": 1241 }, { "epoch": 0.5627548708654282, "grad_norm": 0.5722857244731819, "learning_rate": 7.996189052406108e-05, "loss": 1.0343, "step": 1242 }, { "epoch": 0.5632079746261894, "grad_norm": 0.6430744883371875, "learning_rate": 7.996133630100406e-05, "loss": 1.0096, "step": 1243 }, { "epoch": 0.5636610783869506, "grad_norm": 0.7268020858152191, "learning_rate": 7.996077807892516e-05, "loss": 1.0046, "step": 1244 }, { "epoch": 0.5641141821477118, "grad_norm": 0.7999531667533266, "learning_rate": 7.996021585788024e-05, "loss": 1.0233, "step": 1245 }, { "epoch": 0.564567285908473, "grad_norm": 0.9252344888333061, "learning_rate": 7.995964963792555e-05, "loss": 1.0245, "step": 1246 }, { "epoch": 0.5650203896692343, "grad_norm": 1.0101292476664485, "learning_rate": 7.995907941911778e-05, "loss": 1.0228, "step": 1247 }, { "epoch": 0.5654734934299954, "grad_norm": 0.8449018496651457, "learning_rate": 7.995850520151397e-05, "loss": 1.0278, "step": 1248 }, { "epoch": 0.5659265971907567, "grad_norm": 0.7647360883297393, "learning_rate": 7.995792698517161e-05, "loss": 1.0195, "step": 1249 }, { "epoch": 0.5663797009515179, "grad_norm": 0.8727322467544308, "learning_rate": 7.995734477014854e-05, "loss": 1.0206, "step": 1250 }, { "epoch": 0.5668328047122791, "grad_norm": 1.055655635240549, "learning_rate": 7.995675855650302e-05, "loss": 1.0497, "step": 1251 }, { "epoch": 0.5672859084730403, "grad_norm": 1.0064160887372249, "learning_rate": 7.995616834429372e-05, "loss": 1.0259, "step": 1252 }, { "epoch": 0.5677390122338015, "grad_norm": 0.8970853362783442, "learning_rate": 7.995557413357972e-05, "loss": 1.0205, "step": 1253 }, { "epoch": 0.5681921159945628, "grad_norm": 0.8909980098152634, "learning_rate": 7.995497592442046e-05, "loss": 1.036, "step": 1254 }, { "epoch": 0.568645219755324, "grad_norm": 0.6237259501329525, "learning_rate": 7.995437371687583e-05, "loss": 1.0256, "step": 1255 }, { "epoch": 0.5690983235160851, "grad_norm": 0.47254518663375533, "learning_rate": 7.995376751100607e-05, "loss": 1.0122, "step": 1256 }, { "epoch": 0.5695514272768464, "grad_norm": 0.4455985889000617, "learning_rate": 7.995315730687186e-05, "loss": 1.0192, "step": 1257 }, { "epoch": 0.5700045310376076, "grad_norm": 0.504104480517931, "learning_rate": 7.995254310453427e-05, "loss": 1.0409, "step": 1258 }, { "epoch": 0.5704576347983689, "grad_norm": 0.484839709850666, "learning_rate": 7.995192490405473e-05, "loss": 1.0309, "step": 1259 }, { "epoch": 0.57091073855913, "grad_norm": 0.5054474697670646, "learning_rate": 7.995130270549513e-05, "loss": 1.0272, "step": 1260 }, { "epoch": 0.5713638423198912, "grad_norm": 0.5446173635422935, "learning_rate": 7.995067650891774e-05, "loss": 1.0419, "step": 1261 }, { "epoch": 0.5718169460806525, "grad_norm": 0.4268727039339721, "learning_rate": 7.995004631438522e-05, "loss": 1.0305, "step": 1262 }, { "epoch": 0.5722700498414137, "grad_norm": 0.46874919016621774, "learning_rate": 7.994941212196063e-05, "loss": 1.0328, "step": 1263 }, { "epoch": 0.5727231536021749, "grad_norm": 0.6055163713454689, "learning_rate": 7.994877393170743e-05, "loss": 1.0111, "step": 1264 }, { "epoch": 0.5731762573629361, "grad_norm": 0.6471076040532604, "learning_rate": 7.994813174368951e-05, "loss": 1.0207, "step": 1265 }, { "epoch": 0.5736293611236973, "grad_norm": 0.650097629833748, "learning_rate": 7.99474855579711e-05, "loss": 1.0444, "step": 1266 }, { "epoch": 0.5740824648844586, "grad_norm": 0.6065401437504047, "learning_rate": 7.994683537461689e-05, "loss": 1.0315, "step": 1267 }, { "epoch": 0.5745355686452197, "grad_norm": 0.6082613335524404, "learning_rate": 7.994618119369195e-05, "loss": 1.0161, "step": 1268 }, { "epoch": 0.574988672405981, "grad_norm": 0.5555329116015045, "learning_rate": 7.994552301526172e-05, "loss": 1.0152, "step": 1269 }, { "epoch": 0.5754417761667422, "grad_norm": 0.5782430931908541, "learning_rate": 7.994486083939208e-05, "loss": 1.0539, "step": 1270 }, { "epoch": 0.5758948799275034, "grad_norm": 0.6137424824624652, "learning_rate": 7.994419466614931e-05, "loss": 1.011, "step": 1271 }, { "epoch": 0.5763479836882646, "grad_norm": 0.5918828598642002, "learning_rate": 7.994352449560004e-05, "loss": 1.0369, "step": 1272 }, { "epoch": 0.5768010874490258, "grad_norm": 0.5741869306524292, "learning_rate": 7.994285032781137e-05, "loss": 1.019, "step": 1273 }, { "epoch": 0.577254191209787, "grad_norm": 0.6248222458162763, "learning_rate": 7.994217216285075e-05, "loss": 1.0398, "step": 1274 }, { "epoch": 0.5777072949705483, "grad_norm": 0.6840747426002871, "learning_rate": 7.994149000078605e-05, "loss": 1.0057, "step": 1275 }, { "epoch": 0.5781603987313094, "grad_norm": 0.676752091272638, "learning_rate": 7.994080384168553e-05, "loss": 1.0224, "step": 1276 }, { "epoch": 0.5786135024920707, "grad_norm": 0.655379331070566, "learning_rate": 7.994011368561784e-05, "loss": 0.986, "step": 1277 }, { "epoch": 0.5790666062528319, "grad_norm": 0.7538077178261188, "learning_rate": 7.993941953265208e-05, "loss": 1.0211, "step": 1278 }, { "epoch": 0.5795197100135931, "grad_norm": 0.8496758664637832, "learning_rate": 7.99387213828577e-05, "loss": 1.0052, "step": 1279 }, { "epoch": 0.5799728137743543, "grad_norm": 0.8164468592195955, "learning_rate": 7.993801923630456e-05, "loss": 1.0202, "step": 1280 }, { "epoch": 0.5804259175351155, "grad_norm": 0.716740369284908, "learning_rate": 7.993731309306294e-05, "loss": 1.0192, "step": 1281 }, { "epoch": 0.5808790212958768, "grad_norm": 0.6283456375521487, "learning_rate": 7.993660295320348e-05, "loss": 1.0239, "step": 1282 }, { "epoch": 0.581332125056638, "grad_norm": 0.48270698292799064, "learning_rate": 7.993588881679726e-05, "loss": 1.0236, "step": 1283 }, { "epoch": 0.5817852288173991, "grad_norm": 0.3799000728284925, "learning_rate": 7.993517068391576e-05, "loss": 1.0248, "step": 1284 }, { "epoch": 0.5822383325781604, "grad_norm": 0.43175488553917113, "learning_rate": 7.993444855463083e-05, "loss": 1.0262, "step": 1285 }, { "epoch": 0.5826914363389216, "grad_norm": 0.49778135208143626, "learning_rate": 7.993372242901472e-05, "loss": 1.0123, "step": 1286 }, { "epoch": 0.5831445400996829, "grad_norm": 0.5129515579345615, "learning_rate": 7.993299230714012e-05, "loss": 1.0072, "step": 1287 }, { "epoch": 0.583597643860444, "grad_norm": 0.5929687352454699, "learning_rate": 7.993225818908008e-05, "loss": 1.0294, "step": 1288 }, { "epoch": 0.5840507476212052, "grad_norm": 0.7413585657157983, "learning_rate": 7.993152007490808e-05, "loss": 1.0092, "step": 1289 }, { "epoch": 0.5845038513819665, "grad_norm": 0.8709692203852452, "learning_rate": 7.993077796469798e-05, "loss": 1.0358, "step": 1290 }, { "epoch": 0.5849569551427277, "grad_norm": 0.9687089059151951, "learning_rate": 7.993003185852403e-05, "loss": 1.0021, "step": 1291 }, { "epoch": 0.5854100589034889, "grad_norm": 0.9127188750128729, "learning_rate": 7.992928175646091e-05, "loss": 1.0438, "step": 1292 }, { "epoch": 0.5858631626642501, "grad_norm": 0.7973383358781634, "learning_rate": 7.992852765858368e-05, "loss": 1.0085, "step": 1293 }, { "epoch": 0.5863162664250113, "grad_norm": 0.6723313781507059, "learning_rate": 7.99277695649678e-05, "loss": 1.0293, "step": 1294 }, { "epoch": 0.5867693701857726, "grad_norm": 0.46896674861795995, "learning_rate": 7.992700747568913e-05, "loss": 1.0202, "step": 1295 }, { "epoch": 0.5872224739465337, "grad_norm": 0.3467992106455086, "learning_rate": 7.992624139082395e-05, "loss": 1.027, "step": 1296 }, { "epoch": 0.587675577707295, "grad_norm": 0.43622755569301624, "learning_rate": 7.992547131044893e-05, "loss": 1.0169, "step": 1297 }, { "epoch": 0.5881286814680562, "grad_norm": 0.4899051618099983, "learning_rate": 7.99246972346411e-05, "loss": 1.02, "step": 1298 }, { "epoch": 0.5885817852288174, "grad_norm": 0.43648430382572845, "learning_rate": 7.992391916347794e-05, "loss": 1.0151, "step": 1299 }, { "epoch": 0.5890348889895786, "grad_norm": 0.4281097371194378, "learning_rate": 7.992313709703734e-05, "loss": 1.0136, "step": 1300 }, { "epoch": 0.5894879927503398, "grad_norm": 0.44075070691105056, "learning_rate": 7.992235103539751e-05, "loss": 1.0167, "step": 1301 }, { "epoch": 0.589941096511101, "grad_norm": 0.4593112605719641, "learning_rate": 7.992156097863717e-05, "loss": 1.0313, "step": 1302 }, { "epoch": 0.5903942002718623, "grad_norm": 0.499086721213773, "learning_rate": 7.992076692683533e-05, "loss": 1.0164, "step": 1303 }, { "epoch": 0.5908473040326234, "grad_norm": 0.4901908087818816, "learning_rate": 7.99199688800715e-05, "loss": 1.0222, "step": 1304 }, { "epoch": 0.5913004077933847, "grad_norm": 0.49174259927844927, "learning_rate": 7.99191668384255e-05, "loss": 1.011, "step": 1305 }, { "epoch": 0.5917535115541459, "grad_norm": 0.5460476803086022, "learning_rate": 7.991836080197763e-05, "loss": 1.0126, "step": 1306 }, { "epoch": 0.5922066153149071, "grad_norm": 0.5979466509868409, "learning_rate": 7.991755077080852e-05, "loss": 1.0087, "step": 1307 }, { "epoch": 0.5926597190756683, "grad_norm": 0.6761311010393153, "learning_rate": 7.991673674499925e-05, "loss": 1.0245, "step": 1308 }, { "epoch": 0.5931128228364295, "grad_norm": 0.7036487372567146, "learning_rate": 7.991591872463128e-05, "loss": 1.0292, "step": 1309 }, { "epoch": 0.5935659265971908, "grad_norm": 0.7911082361851407, "learning_rate": 7.991509670978647e-05, "loss": 0.9986, "step": 1310 }, { "epoch": 0.594019030357952, "grad_norm": 0.8851819825654075, "learning_rate": 7.991427070054707e-05, "loss": 1.0374, "step": 1311 }, { "epoch": 0.5944721341187131, "grad_norm": 0.93283973351068, "learning_rate": 7.991344069699576e-05, "loss": 1.0575, "step": 1312 }, { "epoch": 0.5949252378794744, "grad_norm": 0.8229185560256355, "learning_rate": 7.991260669921558e-05, "loss": 1.0089, "step": 1313 }, { "epoch": 0.5953783416402356, "grad_norm": 0.5896782475295145, "learning_rate": 7.991176870729002e-05, "loss": 1.0307, "step": 1314 }, { "epoch": 0.5958314454009969, "grad_norm": 0.5110879640518794, "learning_rate": 7.99109267213029e-05, "loss": 1.0134, "step": 1315 }, { "epoch": 0.596284549161758, "grad_norm": 0.585352474377031, "learning_rate": 7.99100807413385e-05, "loss": 1.0343, "step": 1316 }, { "epoch": 0.5967376529225192, "grad_norm": 0.5016266440067714, "learning_rate": 7.990923076748148e-05, "loss": 1.0371, "step": 1317 }, { "epoch": 0.5971907566832805, "grad_norm": 0.41592145123123814, "learning_rate": 7.99083767998169e-05, "loss": 1.0223, "step": 1318 }, { "epoch": 0.5976438604440417, "grad_norm": 0.48117331003584146, "learning_rate": 7.990751883843022e-05, "loss": 1.0275, "step": 1319 }, { "epoch": 0.5980969642048029, "grad_norm": 0.5389727249043007, "learning_rate": 7.990665688340729e-05, "loss": 1.0189, "step": 1320 }, { "epoch": 0.5985500679655641, "grad_norm": 0.4873807646674866, "learning_rate": 7.990579093483437e-05, "loss": 1.0358, "step": 1321 }, { "epoch": 0.5990031717263253, "grad_norm": 0.4428473382088798, "learning_rate": 7.990492099279814e-05, "loss": 1.0141, "step": 1322 }, { "epoch": 0.5994562754870866, "grad_norm": 0.4932162584315563, "learning_rate": 7.99040470573856e-05, "loss": 1.0286, "step": 1323 }, { "epoch": 0.5999093792478477, "grad_norm": 0.5129294300592919, "learning_rate": 7.990316912868427e-05, "loss": 1.0562, "step": 1324 }, { "epoch": 0.600362483008609, "grad_norm": 0.429642177923022, "learning_rate": 7.990228720678197e-05, "loss": 1.0228, "step": 1325 }, { "epoch": 0.6008155867693702, "grad_norm": 0.3892090441595682, "learning_rate": 7.990140129176697e-05, "loss": 0.9967, "step": 1326 }, { "epoch": 0.6012686905301314, "grad_norm": 0.44013250356914635, "learning_rate": 7.99005113837279e-05, "loss": 1.014, "step": 1327 }, { "epoch": 0.6017217942908926, "grad_norm": 0.5678644769103107, "learning_rate": 7.989961748275387e-05, "loss": 1.0079, "step": 1328 }, { "epoch": 0.6021748980516538, "grad_norm": 0.5677223932996805, "learning_rate": 7.989871958893428e-05, "loss": 1.0334, "step": 1329 }, { "epoch": 0.602628001812415, "grad_norm": 0.4978967981168205, "learning_rate": 7.989781770235901e-05, "loss": 1.0179, "step": 1330 }, { "epoch": 0.6030811055731763, "grad_norm": 0.3748828765776205, "learning_rate": 7.989691182311831e-05, "loss": 1.0137, "step": 1331 }, { "epoch": 0.6035342093339374, "grad_norm": 0.34808938166685394, "learning_rate": 7.989600195130283e-05, "loss": 1.0041, "step": 1332 }, { "epoch": 0.6039873130946987, "grad_norm": 0.47858919928772975, "learning_rate": 7.989508808700361e-05, "loss": 0.994, "step": 1333 }, { "epoch": 0.6044404168554599, "grad_norm": 0.5058599517480309, "learning_rate": 7.989417023031215e-05, "loss": 1.0103, "step": 1334 }, { "epoch": 0.6048935206162211, "grad_norm": 0.47071921751345075, "learning_rate": 7.989324838132025e-05, "loss": 1.0247, "step": 1335 }, { "epoch": 0.6053466243769823, "grad_norm": 0.48898087688606806, "learning_rate": 7.989232254012018e-05, "loss": 1.0312, "step": 1336 }, { "epoch": 0.6057997281377435, "grad_norm": 0.6531039978670422, "learning_rate": 7.98913927068046e-05, "loss": 1.0062, "step": 1337 }, { "epoch": 0.6062528318985048, "grad_norm": 0.9059461428424687, "learning_rate": 7.989045888146654e-05, "loss": 1.018, "step": 1338 }, { "epoch": 0.606705935659266, "grad_norm": 1.1088555870667667, "learning_rate": 7.988952106419947e-05, "loss": 1.0114, "step": 1339 }, { "epoch": 0.6071590394200271, "grad_norm": 0.8619997071817314, "learning_rate": 7.988857925509724e-05, "loss": 1.0078, "step": 1340 }, { "epoch": 0.6076121431807884, "grad_norm": 0.7234462856250524, "learning_rate": 7.988763345425409e-05, "loss": 1.0207, "step": 1341 }, { "epoch": 0.6080652469415496, "grad_norm": 0.6162467171545714, "learning_rate": 7.988668366176465e-05, "loss": 1.0151, "step": 1342 }, { "epoch": 0.6085183507023109, "grad_norm": 0.5444858546470046, "learning_rate": 7.988572987772401e-05, "loss": 1.0137, "step": 1343 }, { "epoch": 0.608971454463072, "grad_norm": 0.48361940757438027, "learning_rate": 7.988477210222758e-05, "loss": 1.0057, "step": 1344 }, { "epoch": 0.6094245582238332, "grad_norm": 0.4300538111286079, "learning_rate": 7.988381033537124e-05, "loss": 1.0175, "step": 1345 }, { "epoch": 0.6098776619845945, "grad_norm": 0.3998522888025513, "learning_rate": 7.98828445772512e-05, "loss": 1.0064, "step": 1346 }, { "epoch": 0.6103307657453557, "grad_norm": 0.400219899796057, "learning_rate": 7.988187482796413e-05, "loss": 1.0009, "step": 1347 }, { "epoch": 0.6107838695061168, "grad_norm": 0.4832184130660549, "learning_rate": 7.988090108760706e-05, "loss": 1.0177, "step": 1348 }, { "epoch": 0.6112369732668781, "grad_norm": 0.5612012271900654, "learning_rate": 7.987992335627746e-05, "loss": 1.0011, "step": 1349 }, { "epoch": 0.6116900770276393, "grad_norm": 0.5965644099632761, "learning_rate": 7.987894163407315e-05, "loss": 1.0385, "step": 1350 }, { "epoch": 0.6121431807884006, "grad_norm": 0.6069489076567568, "learning_rate": 7.987795592109237e-05, "loss": 1.0077, "step": 1351 }, { "epoch": 0.6125962845491617, "grad_norm": 0.5757679844453336, "learning_rate": 7.987696621743379e-05, "loss": 1.0043, "step": 1352 }, { "epoch": 0.613049388309923, "grad_norm": 0.6222927340123223, "learning_rate": 7.987597252319642e-05, "loss": 0.9974, "step": 1353 }, { "epoch": 0.6135024920706842, "grad_norm": 0.6881882876303184, "learning_rate": 7.987497483847972e-05, "loss": 1.0275, "step": 1354 }, { "epoch": 0.6139555958314454, "grad_norm": 0.7338144937651409, "learning_rate": 7.987397316338354e-05, "loss": 1.0068, "step": 1355 }, { "epoch": 0.6144086995922066, "grad_norm": 0.6908693019570364, "learning_rate": 7.987296749800807e-05, "loss": 1.0088, "step": 1356 }, { "epoch": 0.6148618033529678, "grad_norm": 0.623204866078906, "learning_rate": 7.987195784245402e-05, "loss": 1.0162, "step": 1357 }, { "epoch": 0.615314907113729, "grad_norm": 0.5691714850171953, "learning_rate": 7.987094419682238e-05, "loss": 1.0187, "step": 1358 }, { "epoch": 0.6157680108744903, "grad_norm": 0.5298111369093415, "learning_rate": 7.986992656121461e-05, "loss": 1.0178, "step": 1359 }, { "epoch": 0.6162211146352514, "grad_norm": 0.4586928747994298, "learning_rate": 7.986890493573252e-05, "loss": 1.033, "step": 1360 }, { "epoch": 0.6166742183960127, "grad_norm": 0.44834114921050844, "learning_rate": 7.986787932047838e-05, "loss": 1.0072, "step": 1361 }, { "epoch": 0.6171273221567739, "grad_norm": 0.5368293919216797, "learning_rate": 7.98668497155548e-05, "loss": 1.0174, "step": 1362 }, { "epoch": 0.6175804259175351, "grad_norm": 0.5911885180548977, "learning_rate": 7.986581612106485e-05, "loss": 1.0257, "step": 1363 }, { "epoch": 0.6180335296782963, "grad_norm": 0.5422017864925156, "learning_rate": 7.986477853711191e-05, "loss": 1.006, "step": 1364 }, { "epoch": 0.6184866334390575, "grad_norm": 0.4708699594940483, "learning_rate": 7.986373696379987e-05, "loss": 1.0033, "step": 1365 }, { "epoch": 0.6189397371998188, "grad_norm": 0.4822948503199424, "learning_rate": 7.986269140123291e-05, "loss": 1.0159, "step": 1366 }, { "epoch": 0.61939284096058, "grad_norm": 0.5102938142812308, "learning_rate": 7.98616418495157e-05, "loss": 1.0312, "step": 1367 }, { "epoch": 0.6198459447213411, "grad_norm": 0.5403025116762755, "learning_rate": 7.986058830875326e-05, "loss": 1.0047, "step": 1368 }, { "epoch": 0.6202990484821024, "grad_norm": 0.5987299485013236, "learning_rate": 7.985953077905102e-05, "loss": 1.0172, "step": 1369 }, { "epoch": 0.6207521522428636, "grad_norm": 0.7005403546242173, "learning_rate": 7.985846926051483e-05, "loss": 1.025, "step": 1370 }, { "epoch": 0.6212052560036249, "grad_norm": 0.7590792483030188, "learning_rate": 7.985740375325085e-05, "loss": 1.0319, "step": 1371 }, { "epoch": 0.621658359764386, "grad_norm": 0.7662655260489889, "learning_rate": 7.98563342573658e-05, "loss": 1.0343, "step": 1372 }, { "epoch": 0.6221114635251472, "grad_norm": 0.716861153445679, "learning_rate": 7.985526077296663e-05, "loss": 1.0114, "step": 1373 }, { "epoch": 0.6225645672859085, "grad_norm": 0.5853911529891288, "learning_rate": 7.98541833001608e-05, "loss": 1.0145, "step": 1374 }, { "epoch": 0.6230176710466697, "grad_norm": 0.4785948521637449, "learning_rate": 7.985310183905615e-05, "loss": 0.9955, "step": 1375 }, { "epoch": 0.6234707748074308, "grad_norm": 0.5422265743313313, "learning_rate": 7.985201638976087e-05, "loss": 1.0135, "step": 1376 }, { "epoch": 0.6239238785681921, "grad_norm": 0.6856075960216567, "learning_rate": 7.98509269523836e-05, "loss": 1.0203, "step": 1377 }, { "epoch": 0.6243769823289533, "grad_norm": 0.6416195737734478, "learning_rate": 7.984983352703338e-05, "loss": 1.0111, "step": 1378 }, { "epoch": 0.6248300860897146, "grad_norm": 0.4768221361443679, "learning_rate": 7.98487361138196e-05, "loss": 1.0165, "step": 1379 }, { "epoch": 0.6252831898504757, "grad_norm": 0.5167669705029995, "learning_rate": 7.98476347128521e-05, "loss": 1.0258, "step": 1380 }, { "epoch": 0.625736293611237, "grad_norm": 0.5659122207073963, "learning_rate": 7.98465293242411e-05, "loss": 1.0135, "step": 1381 }, { "epoch": 0.6261893973719982, "grad_norm": 0.5481313024025625, "learning_rate": 7.98454199480972e-05, "loss": 1.0266, "step": 1382 }, { "epoch": 0.6266425011327594, "grad_norm": 0.5794169977971602, "learning_rate": 7.984430658453145e-05, "loss": 1.016, "step": 1383 }, { "epoch": 0.6270956048935206, "grad_norm": 0.7378291952583701, "learning_rate": 7.984318923365524e-05, "loss": 1.0093, "step": 1384 }, { "epoch": 0.6275487086542818, "grad_norm": 0.8302625056707188, "learning_rate": 7.984206789558039e-05, "loss": 1.0378, "step": 1385 }, { "epoch": 0.628001812415043, "grad_norm": 0.8517045338643516, "learning_rate": 7.984094257041912e-05, "loss": 1.0282, "step": 1386 }, { "epoch": 0.6284549161758043, "grad_norm": 0.785645308726276, "learning_rate": 7.983981325828405e-05, "loss": 1.002, "step": 1387 }, { "epoch": 0.6289080199365654, "grad_norm": 0.8341408101244278, "learning_rate": 7.983867995928818e-05, "loss": 1.0013, "step": 1388 }, { "epoch": 0.6293611236973267, "grad_norm": 0.9893138172968772, "learning_rate": 7.983754267354491e-05, "loss": 1.0489, "step": 1389 }, { "epoch": 0.6298142274580879, "grad_norm": 1.0257878950300188, "learning_rate": 7.98364014011681e-05, "loss": 1.0152, "step": 1390 }, { "epoch": 0.6302673312188491, "grad_norm": 0.8716210422657986, "learning_rate": 7.983525614227191e-05, "loss": 1.0114, "step": 1391 }, { "epoch": 0.6307204349796103, "grad_norm": 0.7529685849048984, "learning_rate": 7.983410689697095e-05, "loss": 1.0146, "step": 1392 }, { "epoch": 0.6311735387403715, "grad_norm": 0.6486227516250451, "learning_rate": 7.983295366538027e-05, "loss": 1.0248, "step": 1393 }, { "epoch": 0.6316266425011328, "grad_norm": 0.538444617487835, "learning_rate": 7.983179644761524e-05, "loss": 1.0103, "step": 1394 }, { "epoch": 0.632079746261894, "grad_norm": 0.41957085049011034, "learning_rate": 7.983063524379169e-05, "loss": 1.0456, "step": 1395 }, { "epoch": 0.6325328500226551, "grad_norm": 0.4213208873438223, "learning_rate": 7.98294700540258e-05, "loss": 1.0138, "step": 1396 }, { "epoch": 0.6329859537834164, "grad_norm": 0.5433448569887083, "learning_rate": 7.982830087843418e-05, "loss": 1.0103, "step": 1397 }, { "epoch": 0.6334390575441776, "grad_norm": 0.6091832660374707, "learning_rate": 7.982712771713385e-05, "loss": 0.9894, "step": 1398 }, { "epoch": 0.6338921613049389, "grad_norm": 0.5948748749286283, "learning_rate": 7.982595057024216e-05, "loss": 1.0232, "step": 1399 }, { "epoch": 0.6343452650657, "grad_norm": 0.5858002005562055, "learning_rate": 7.982476943787699e-05, "loss": 1.0242, "step": 1400 }, { "epoch": 0.6347983688264612, "grad_norm": 0.5429060625026636, "learning_rate": 7.982358432015648e-05, "loss": 1.0034, "step": 1401 }, { "epoch": 0.6352514725872225, "grad_norm": 0.3764163114638043, "learning_rate": 7.982239521719923e-05, "loss": 1.0275, "step": 1402 }, { "epoch": 0.6357045763479837, "grad_norm": 0.3948730192486164, "learning_rate": 7.982120212912426e-05, "loss": 1.0503, "step": 1403 }, { "epoch": 0.6361576801087448, "grad_norm": 0.44933348074965873, "learning_rate": 7.982000505605096e-05, "loss": 0.9894, "step": 1404 }, { "epoch": 0.6366107838695061, "grad_norm": 0.4904710480870958, "learning_rate": 7.981880399809911e-05, "loss": 0.9958, "step": 1405 }, { "epoch": 0.6370638876302673, "grad_norm": 0.486840133464061, "learning_rate": 7.98175989553889e-05, "loss": 1.0242, "step": 1406 }, { "epoch": 0.6375169913910286, "grad_norm": 0.49997499732120987, "learning_rate": 7.981638992804094e-05, "loss": 1.0312, "step": 1407 }, { "epoch": 0.6379700951517897, "grad_norm": 0.4654750659629424, "learning_rate": 7.981517691617622e-05, "loss": 1.0214, "step": 1408 }, { "epoch": 0.638423198912551, "grad_norm": 0.4575455009493585, "learning_rate": 7.98139599199161e-05, "loss": 1.048, "step": 1409 }, { "epoch": 0.6388763026733122, "grad_norm": 0.4880839352787064, "learning_rate": 7.981273893938241e-05, "loss": 1.0229, "step": 1410 }, { "epoch": 0.6393294064340734, "grad_norm": 0.5285315038053539, "learning_rate": 7.98115139746973e-05, "loss": 1.0084, "step": 1411 }, { "epoch": 0.6397825101948346, "grad_norm": 0.579835830509641, "learning_rate": 7.981028502598338e-05, "loss": 0.9914, "step": 1412 }, { "epoch": 0.6402356139555958, "grad_norm": 0.6176435240058339, "learning_rate": 7.980905209336361e-05, "loss": 0.9926, "step": 1413 }, { "epoch": 0.640688717716357, "grad_norm": 0.6884080540775496, "learning_rate": 7.980781517696139e-05, "loss": 1.0289, "step": 1414 }, { "epoch": 0.6411418214771183, "grad_norm": 0.7106171782808312, "learning_rate": 7.98065742769005e-05, "loss": 0.9943, "step": 1415 }, { "epoch": 0.6415949252378794, "grad_norm": 0.7500094289917739, "learning_rate": 7.980532939330511e-05, "loss": 1.0057, "step": 1416 }, { "epoch": 0.6420480289986407, "grad_norm": 0.7925763824753216, "learning_rate": 7.980408052629982e-05, "loss": 1.0255, "step": 1417 }, { "epoch": 0.6425011327594019, "grad_norm": 0.7964215210627832, "learning_rate": 7.980282767600956e-05, "loss": 1.0092, "step": 1418 }, { "epoch": 0.6429542365201631, "grad_norm": 0.7236551001456291, "learning_rate": 7.980157084255978e-05, "loss": 1.0354, "step": 1419 }, { "epoch": 0.6434073402809243, "grad_norm": 0.5157072278793445, "learning_rate": 7.980031002607618e-05, "loss": 1.0193, "step": 1420 }, { "epoch": 0.6438604440416855, "grad_norm": 0.4284077007962374, "learning_rate": 7.979904522668497e-05, "loss": 1.0198, "step": 1421 }, { "epoch": 0.6443135478024468, "grad_norm": 0.46252288769261857, "learning_rate": 7.979777644451272e-05, "loss": 1.0016, "step": 1422 }, { "epoch": 0.644766651563208, "grad_norm": 0.45510068598042314, "learning_rate": 7.979650367968638e-05, "loss": 0.9973, "step": 1423 }, { "epoch": 0.6452197553239691, "grad_norm": 0.4945544608617784, "learning_rate": 7.979522693233336e-05, "loss": 1.001, "step": 1424 }, { "epoch": 0.6456728590847304, "grad_norm": 0.5613131500744093, "learning_rate": 7.979394620258137e-05, "loss": 1.0301, "step": 1425 }, { "epoch": 0.6461259628454916, "grad_norm": 0.6253926631855871, "learning_rate": 7.979266149055863e-05, "loss": 0.9836, "step": 1426 }, { "epoch": 0.6465790666062529, "grad_norm": 0.7038358082175951, "learning_rate": 7.979137279639369e-05, "loss": 1.0067, "step": 1427 }, { "epoch": 0.647032170367014, "grad_norm": 0.8050780347652264, "learning_rate": 7.979008012021547e-05, "loss": 1.0256, "step": 1428 }, { "epoch": 0.6474852741277752, "grad_norm": 0.8819924085605778, "learning_rate": 7.978878346215338e-05, "loss": 0.9881, "step": 1429 }, { "epoch": 0.6479383778885365, "grad_norm": 0.7612973427634823, "learning_rate": 7.978748282233717e-05, "loss": 1.0352, "step": 1430 }, { "epoch": 0.6483914816492977, "grad_norm": 0.688304885128866, "learning_rate": 7.978617820089698e-05, "loss": 1.0353, "step": 1431 }, { "epoch": 0.6488445854100588, "grad_norm": 0.7667150265832894, "learning_rate": 7.978486959796339e-05, "loss": 1.0111, "step": 1432 }, { "epoch": 0.6492976891708201, "grad_norm": 0.6100057360379778, "learning_rate": 7.978355701366733e-05, "loss": 1.0108, "step": 1433 }, { "epoch": 0.6497507929315813, "grad_norm": 0.676258899909195, "learning_rate": 7.978224044814017e-05, "loss": 1.0314, "step": 1434 }, { "epoch": 0.6502038966923426, "grad_norm": 0.8080622763214208, "learning_rate": 7.978091990151366e-05, "loss": 1.0113, "step": 1435 }, { "epoch": 0.6506570004531037, "grad_norm": 0.7717412776418393, "learning_rate": 7.977959537391995e-05, "loss": 1.0332, "step": 1436 }, { "epoch": 0.651110104213865, "grad_norm": 0.7980299663934964, "learning_rate": 7.977826686549159e-05, "loss": 1.0234, "step": 1437 }, { "epoch": 0.6515632079746262, "grad_norm": 0.6902859915339009, "learning_rate": 7.977693437636151e-05, "loss": 0.9984, "step": 1438 }, { "epoch": 0.6520163117353874, "grad_norm": 0.4530997431954452, "learning_rate": 7.977559790666308e-05, "loss": 1.0075, "step": 1439 }, { "epoch": 0.6524694154961486, "grad_norm": 0.5030047080835661, "learning_rate": 7.977425745653001e-05, "loss": 0.9935, "step": 1440 }, { "epoch": 0.6529225192569098, "grad_norm": 0.558406523220189, "learning_rate": 7.97729130260965e-05, "loss": 1.0277, "step": 1441 }, { "epoch": 0.653375623017671, "grad_norm": 0.5508795102004916, "learning_rate": 7.977156461549702e-05, "loss": 1.0293, "step": 1442 }, { "epoch": 0.6538287267784323, "grad_norm": 0.5684815438744267, "learning_rate": 7.977021222486656e-05, "loss": 1.0236, "step": 1443 }, { "epoch": 0.6542818305391934, "grad_norm": 0.5930160557213442, "learning_rate": 7.976885585434044e-05, "loss": 1.0156, "step": 1444 }, { "epoch": 0.6547349342999547, "grad_norm": 0.6344511153619422, "learning_rate": 7.97674955040544e-05, "loss": 1.0369, "step": 1445 }, { "epoch": 0.6551880380607159, "grad_norm": 0.7290720088172509, "learning_rate": 7.976613117414454e-05, "loss": 1.0344, "step": 1446 }, { "epoch": 0.6556411418214771, "grad_norm": 0.6768855675780302, "learning_rate": 7.976476286474745e-05, "loss": 1.016, "step": 1447 }, { "epoch": 0.6560942455822383, "grad_norm": 0.5882984099992403, "learning_rate": 7.976339057600001e-05, "loss": 1.0324, "step": 1448 }, { "epoch": 0.6565473493429995, "grad_norm": 0.47846208034996407, "learning_rate": 7.976201430803958e-05, "loss": 1.0354, "step": 1449 }, { "epoch": 0.6570004531037608, "grad_norm": 0.34808906211246454, "learning_rate": 7.976063406100386e-05, "loss": 1.0258, "step": 1450 }, { "epoch": 0.657453556864522, "grad_norm": 0.31026923056061106, "learning_rate": 7.9759249835031e-05, "loss": 1.0021, "step": 1451 }, { "epoch": 0.6579066606252831, "grad_norm": 0.32725334214645163, "learning_rate": 7.975786163025951e-05, "loss": 1.0015, "step": 1452 }, { "epoch": 0.6583597643860444, "grad_norm": 0.3861794901511407, "learning_rate": 7.975646944682831e-05, "loss": 0.9893, "step": 1453 }, { "epoch": 0.6588128681468056, "grad_norm": 0.42195524936126183, "learning_rate": 7.975507328487671e-05, "loss": 1.0261, "step": 1454 }, { "epoch": 0.6592659719075669, "grad_norm": 0.4841677738909684, "learning_rate": 7.975367314454445e-05, "loss": 0.999, "step": 1455 }, { "epoch": 0.659719075668328, "grad_norm": 0.5541820120797204, "learning_rate": 7.975226902597163e-05, "loss": 0.9984, "step": 1456 }, { "epoch": 0.6601721794290892, "grad_norm": 0.5903146826143885, "learning_rate": 7.975086092929877e-05, "loss": 0.9907, "step": 1457 }, { "epoch": 0.6606252831898505, "grad_norm": 0.6176963964114568, "learning_rate": 7.974944885466677e-05, "loss": 1.0059, "step": 1458 }, { "epoch": 0.6610783869506117, "grad_norm": 0.6035256849782081, "learning_rate": 7.974803280221695e-05, "loss": 0.9967, "step": 1459 }, { "epoch": 0.6615314907113728, "grad_norm": 0.5501080530987686, "learning_rate": 7.974661277209102e-05, "loss": 1.0259, "step": 1460 }, { "epoch": 0.6619845944721341, "grad_norm": 0.5378034474677492, "learning_rate": 7.974518876443106e-05, "loss": 0.9795, "step": 1461 }, { "epoch": 0.6624376982328953, "grad_norm": 0.4908680046801233, "learning_rate": 7.974376077937962e-05, "loss": 1.0013, "step": 1462 }, { "epoch": 0.6628908019936566, "grad_norm": 0.4588868033927309, "learning_rate": 7.974232881707956e-05, "loss": 1.0148, "step": 1463 }, { "epoch": 0.6633439057544177, "grad_norm": 0.39519947441106223, "learning_rate": 7.97408928776742e-05, "loss": 1.0245, "step": 1464 }, { "epoch": 0.663797009515179, "grad_norm": 0.3442208011404144, "learning_rate": 7.973945296130723e-05, "loss": 1.0263, "step": 1465 }, { "epoch": 0.6642501132759402, "grad_norm": 0.3430407504385823, "learning_rate": 7.973800906812276e-05, "loss": 1.0328, "step": 1466 }, { "epoch": 0.6647032170367014, "grad_norm": 0.3352696664447572, "learning_rate": 7.973656119826526e-05, "loss": 1.0228, "step": 1467 }, { "epoch": 0.6651563207974626, "grad_norm": 0.29174993136196353, "learning_rate": 7.973510935187964e-05, "loss": 1.0108, "step": 1468 }, { "epoch": 0.6656094245582238, "grad_norm": 0.3926895119600163, "learning_rate": 7.973365352911119e-05, "loss": 1.0024, "step": 1469 }, { "epoch": 0.666062528318985, "grad_norm": 0.5555446527056933, "learning_rate": 7.973219373010558e-05, "loss": 1.039, "step": 1470 }, { "epoch": 0.6665156320797463, "grad_norm": 0.9966734010588625, "learning_rate": 7.973072995500891e-05, "loss": 1.046, "step": 1471 }, { "epoch": 0.6669687358405075, "grad_norm": 0.9062030900329278, "learning_rate": 7.972926220396765e-05, "loss": 1.0247, "step": 1472 }, { "epoch": 0.6674218396012687, "grad_norm": 1.0199335951946666, "learning_rate": 7.972779047712872e-05, "loss": 1.0205, "step": 1473 }, { "epoch": 0.6678749433620299, "grad_norm": 0.9644120906218749, "learning_rate": 7.972631477463935e-05, "loss": 1.0251, "step": 1474 }, { "epoch": 0.6683280471227911, "grad_norm": 0.751110082749679, "learning_rate": 7.972483509664724e-05, "loss": 1.0219, "step": 1475 }, { "epoch": 0.6687811508835524, "grad_norm": 0.5745862381596225, "learning_rate": 7.972335144330046e-05, "loss": 1.029, "step": 1476 }, { "epoch": 0.6692342546443135, "grad_norm": 0.4490352755433728, "learning_rate": 7.972186381474748e-05, "loss": 1.0242, "step": 1477 }, { "epoch": 0.6696873584050748, "grad_norm": 0.3376601280723612, "learning_rate": 7.972037221113719e-05, "loss": 1.0168, "step": 1478 }, { "epoch": 0.670140462165836, "grad_norm": 0.3867229009139686, "learning_rate": 7.971887663261882e-05, "loss": 1.0183, "step": 1479 }, { "epoch": 0.6705935659265972, "grad_norm": 0.5408313872825935, "learning_rate": 7.971737707934208e-05, "loss": 0.9988, "step": 1480 }, { "epoch": 0.6710466696873584, "grad_norm": 0.5848075736100454, "learning_rate": 7.9715873551457e-05, "loss": 1.0128, "step": 1481 }, { "epoch": 0.6714997734481196, "grad_norm": 0.6050124050505322, "learning_rate": 7.971436604911406e-05, "loss": 1.033, "step": 1482 }, { "epoch": 0.6719528772088809, "grad_norm": 0.6789941735661497, "learning_rate": 7.971285457246412e-05, "loss": 1.016, "step": 1483 }, { "epoch": 0.6724059809696421, "grad_norm": 0.7430580548901236, "learning_rate": 7.971133912165842e-05, "loss": 1.0072, "step": 1484 }, { "epoch": 0.6728590847304032, "grad_norm": 0.8614883209050451, "learning_rate": 7.970981969684862e-05, "loss": 1.0083, "step": 1485 }, { "epoch": 0.6733121884911645, "grad_norm": 0.8593312786041576, "learning_rate": 7.97082962981868e-05, "loss": 1.0159, "step": 1486 }, { "epoch": 0.6737652922519257, "grad_norm": 0.7776893694304029, "learning_rate": 7.970676892582535e-05, "loss": 1.0196, "step": 1487 }, { "epoch": 0.674218396012687, "grad_norm": 0.676694910644676, "learning_rate": 7.970523757991719e-05, "loss": 1.0244, "step": 1488 }, { "epoch": 0.6746714997734481, "grad_norm": 0.6430344277781613, "learning_rate": 7.970370226061551e-05, "loss": 1.0109, "step": 1489 }, { "epoch": 0.6751246035342093, "grad_norm": 0.5281702422660048, "learning_rate": 7.970216296807399e-05, "loss": 1.0085, "step": 1490 }, { "epoch": 0.6755777072949706, "grad_norm": 0.40504746238916184, "learning_rate": 7.970061970244664e-05, "loss": 1.0133, "step": 1491 }, { "epoch": 0.6760308110557318, "grad_norm": 0.4158042180394529, "learning_rate": 7.969907246388791e-05, "loss": 1.03, "step": 1492 }, { "epoch": 0.676483914816493, "grad_norm": 0.5262653539859083, "learning_rate": 7.969752125255264e-05, "loss": 1.0257, "step": 1493 }, { "epoch": 0.6769370185772542, "grad_norm": 0.7300888073488593, "learning_rate": 7.969596606859607e-05, "loss": 1.0046, "step": 1494 }, { "epoch": 0.6773901223380154, "grad_norm": 0.8686189162828565, "learning_rate": 7.969440691217381e-05, "loss": 1.0242, "step": 1495 }, { "epoch": 0.6778432260987767, "grad_norm": 0.9013532787646564, "learning_rate": 7.96928437834419e-05, "loss": 1.0204, "step": 1496 }, { "epoch": 0.6782963298595378, "grad_norm": 0.8481508360161993, "learning_rate": 7.969127668255677e-05, "loss": 1.0338, "step": 1497 }, { "epoch": 0.678749433620299, "grad_norm": 0.687733945519034, "learning_rate": 7.968970560967524e-05, "loss": 1.0052, "step": 1498 }, { "epoch": 0.6792025373810603, "grad_norm": 0.5109736669454543, "learning_rate": 7.968813056495452e-05, "loss": 1.031, "step": 1499 }, { "epoch": 0.6796556411418215, "grad_norm": 0.5062810655744734, "learning_rate": 7.968655154855224e-05, "loss": 1.0114, "step": 1500 }, { "epoch": 0.6801087449025827, "grad_norm": 0.5033597209277646, "learning_rate": 7.968496856062643e-05, "loss": 1.0194, "step": 1501 }, { "epoch": 0.6805618486633439, "grad_norm": 0.5620045640542983, "learning_rate": 7.968338160133547e-05, "loss": 0.989, "step": 1502 }, { "epoch": 0.6810149524241051, "grad_norm": 0.567104167744797, "learning_rate": 7.96817906708382e-05, "loss": 1.0205, "step": 1503 }, { "epoch": 0.6814680561848664, "grad_norm": 0.5133389303814233, "learning_rate": 7.96801957692938e-05, "loss": 1.0108, "step": 1504 }, { "epoch": 0.6819211599456275, "grad_norm": 0.45181786121808776, "learning_rate": 7.96785968968619e-05, "loss": 1.0292, "step": 1505 }, { "epoch": 0.6823742637063888, "grad_norm": 0.36636974260208216, "learning_rate": 7.96769940537025e-05, "loss": 1.0013, "step": 1506 }, { "epoch": 0.68282736746715, "grad_norm": 0.3134994185659703, "learning_rate": 7.967538723997598e-05, "loss": 1.0125, "step": 1507 }, { "epoch": 0.6832804712279112, "grad_norm": 0.3022020102884248, "learning_rate": 7.967377645584316e-05, "loss": 0.9733, "step": 1508 }, { "epoch": 0.6837335749886724, "grad_norm": 0.35911322749855856, "learning_rate": 7.967216170146523e-05, "loss": 1.006, "step": 1509 }, { "epoch": 0.6841866787494336, "grad_norm": 0.4139522894751992, "learning_rate": 7.967054297700377e-05, "loss": 1.0165, "step": 1510 }, { "epoch": 0.6846397825101949, "grad_norm": 0.43139380608039424, "learning_rate": 7.966892028262079e-05, "loss": 1.0429, "step": 1511 }, { "epoch": 0.6850928862709561, "grad_norm": 0.48160364539741546, "learning_rate": 7.966729361847865e-05, "loss": 0.9861, "step": 1512 }, { "epoch": 0.6855459900317172, "grad_norm": 0.5351580811964385, "learning_rate": 7.966566298474015e-05, "loss": 1.0077, "step": 1513 }, { "epoch": 0.6859990937924785, "grad_norm": 0.6166802343406751, "learning_rate": 7.966402838156849e-05, "loss": 1.0039, "step": 1514 }, { "epoch": 0.6864521975532397, "grad_norm": 0.6754396019150058, "learning_rate": 7.96623898091272e-05, "loss": 1.0247, "step": 1515 }, { "epoch": 0.686905301314001, "grad_norm": 0.6719682359049793, "learning_rate": 7.96607472675803e-05, "loss": 0.9929, "step": 1516 }, { "epoch": 0.6873584050747621, "grad_norm": 0.6671275038069883, "learning_rate": 7.965910075709214e-05, "loss": 0.988, "step": 1517 }, { "epoch": 0.6878115088355233, "grad_norm": 0.6404921109169001, "learning_rate": 7.965745027782749e-05, "loss": 1.0207, "step": 1518 }, { "epoch": 0.6882646125962846, "grad_norm": 0.5419683640805226, "learning_rate": 7.965579582995154e-05, "loss": 0.9946, "step": 1519 }, { "epoch": 0.6887177163570458, "grad_norm": 0.5088826960483392, "learning_rate": 7.965413741362984e-05, "loss": 1.0072, "step": 1520 }, { "epoch": 0.689170820117807, "grad_norm": 0.43088868022965177, "learning_rate": 7.965247502902834e-05, "loss": 0.9935, "step": 1521 }, { "epoch": 0.6896239238785682, "grad_norm": 0.4210416177436171, "learning_rate": 7.96508086763134e-05, "loss": 1.024, "step": 1522 }, { "epoch": 0.6900770276393294, "grad_norm": 0.5011229017825451, "learning_rate": 7.96491383556518e-05, "loss": 1.0288, "step": 1523 }, { "epoch": 0.6905301314000907, "grad_norm": 0.5057001158885828, "learning_rate": 7.964746406721068e-05, "loss": 1.0188, "step": 1524 }, { "epoch": 0.6909832351608518, "grad_norm": 0.5379334449043298, "learning_rate": 7.964578581115757e-05, "loss": 1.0216, "step": 1525 }, { "epoch": 0.691436338921613, "grad_norm": 0.542425782611315, "learning_rate": 7.964410358766045e-05, "loss": 1.0273, "step": 1526 }, { "epoch": 0.6918894426823743, "grad_norm": 0.4657057342167314, "learning_rate": 7.964241739688764e-05, "loss": 0.9705, "step": 1527 }, { "epoch": 0.6923425464431355, "grad_norm": 0.4369883807074941, "learning_rate": 7.964072723900788e-05, "loss": 1.0241, "step": 1528 }, { "epoch": 0.6927956502038967, "grad_norm": 0.44154601682350353, "learning_rate": 7.963903311419033e-05, "loss": 1.0139, "step": 1529 }, { "epoch": 0.6932487539646579, "grad_norm": 0.41763801418709073, "learning_rate": 7.963733502260452e-05, "loss": 0.9988, "step": 1530 }, { "epoch": 0.6937018577254191, "grad_norm": 0.5074514812923638, "learning_rate": 7.963563296442034e-05, "loss": 1.0177, "step": 1531 }, { "epoch": 0.6941549614861804, "grad_norm": 0.6281339911222484, "learning_rate": 7.96339269398082e-05, "loss": 1.0137, "step": 1532 }, { "epoch": 0.6946080652469415, "grad_norm": 0.7097083178529444, "learning_rate": 7.963221694893874e-05, "loss": 1.0211, "step": 1533 }, { "epoch": 0.6950611690077028, "grad_norm": 0.6892548871377133, "learning_rate": 7.963050299198313e-05, "loss": 0.9941, "step": 1534 }, { "epoch": 0.695514272768464, "grad_norm": 0.701067910178001, "learning_rate": 7.962878506911289e-05, "loss": 1.0273, "step": 1535 }, { "epoch": 0.6959673765292252, "grad_norm": 0.7588886896693601, "learning_rate": 7.962706318049991e-05, "loss": 1.0123, "step": 1536 }, { "epoch": 0.6964204802899864, "grad_norm": 0.7818004794863563, "learning_rate": 7.962533732631652e-05, "loss": 1.0053, "step": 1537 }, { "epoch": 0.6968735840507476, "grad_norm": 0.8043568632432971, "learning_rate": 7.962360750673545e-05, "loss": 1.0252, "step": 1538 }, { "epoch": 0.6973266878115089, "grad_norm": 0.7762580159610217, "learning_rate": 7.962187372192978e-05, "loss": 1.0143, "step": 1539 }, { "epoch": 0.6977797915722701, "grad_norm": 0.7275693558309244, "learning_rate": 7.962013597207302e-05, "loss": 1.023, "step": 1540 }, { "epoch": 0.6982328953330312, "grad_norm": 0.6602031234752906, "learning_rate": 7.961839425733907e-05, "loss": 1.0219, "step": 1541 }, { "epoch": 0.6986859990937925, "grad_norm": 0.521059811213846, "learning_rate": 7.961664857790222e-05, "loss": 1.0164, "step": 1542 }, { "epoch": 0.6991391028545537, "grad_norm": 0.4123806625670966, "learning_rate": 7.961489893393718e-05, "loss": 1.0214, "step": 1543 }, { "epoch": 0.699592206615315, "grad_norm": 0.3606772226593471, "learning_rate": 7.961314532561905e-05, "loss": 1.0135, "step": 1544 }, { "epoch": 0.7000453103760761, "grad_norm": 0.32851746243102536, "learning_rate": 7.961138775312326e-05, "loss": 1.0079, "step": 1545 }, { "epoch": 0.7004984141368373, "grad_norm": 0.35675188316390893, "learning_rate": 7.960962621662578e-05, "loss": 0.9885, "step": 1546 }, { "epoch": 0.7009515178975986, "grad_norm": 0.37259061467292554, "learning_rate": 7.960786071630282e-05, "loss": 1.0268, "step": 1547 }, { "epoch": 0.7014046216583598, "grad_norm": 0.42921845382985846, "learning_rate": 7.96060912523311e-05, "loss": 1.0139, "step": 1548 }, { "epoch": 0.701857725419121, "grad_norm": 0.47494075290477933, "learning_rate": 7.960431782488767e-05, "loss": 1.0108, "step": 1549 }, { "epoch": 0.7023108291798822, "grad_norm": 0.4400543184707477, "learning_rate": 7.960254043415002e-05, "loss": 0.9986, "step": 1550 }, { "epoch": 0.7027639329406434, "grad_norm": 0.3608930026046516, "learning_rate": 7.9600759080296e-05, "loss": 1.0305, "step": 1551 }, { "epoch": 0.7032170367014047, "grad_norm": 0.29263241912513344, "learning_rate": 7.959897376350388e-05, "loss": 1.0104, "step": 1552 }, { "epoch": 0.7036701404621658, "grad_norm": 0.35286532971831575, "learning_rate": 7.959718448395232e-05, "loss": 1.0337, "step": 1553 }, { "epoch": 0.704123244222927, "grad_norm": 0.41349079432868535, "learning_rate": 7.959539124182038e-05, "loss": 1.0106, "step": 1554 }, { "epoch": 0.7045763479836883, "grad_norm": 0.40389480395603344, "learning_rate": 7.959359403728752e-05, "loss": 1.0002, "step": 1555 }, { "epoch": 0.7050294517444495, "grad_norm": 0.3754726333646763, "learning_rate": 7.959179287053357e-05, "loss": 1.0102, "step": 1556 }, { "epoch": 0.7054825555052107, "grad_norm": 0.4524930341796494, "learning_rate": 7.958998774173881e-05, "loss": 1.034, "step": 1557 }, { "epoch": 0.7059356592659719, "grad_norm": 0.5051807583291633, "learning_rate": 7.958817865108383e-05, "loss": 1.0056, "step": 1558 }, { "epoch": 0.7063887630267331, "grad_norm": 0.515989067731342, "learning_rate": 7.958636559874974e-05, "loss": 1.023, "step": 1559 }, { "epoch": 0.7068418667874944, "grad_norm": 0.5516877790878058, "learning_rate": 7.958454858491792e-05, "loss": 1.0066, "step": 1560 }, { "epoch": 0.7072949705482555, "grad_norm": 0.6066823338684836, "learning_rate": 7.958272760977023e-05, "loss": 0.9996, "step": 1561 }, { "epoch": 0.7077480743090168, "grad_norm": 0.6468167620629018, "learning_rate": 7.958090267348889e-05, "loss": 1.0169, "step": 1562 }, { "epoch": 0.708201178069778, "grad_norm": 0.7238130863291455, "learning_rate": 7.957907377625653e-05, "loss": 1.012, "step": 1563 }, { "epoch": 0.7086542818305392, "grad_norm": 0.7880921242733888, "learning_rate": 7.957724091825616e-05, "loss": 0.9962, "step": 1564 }, { "epoch": 0.7091073855913004, "grad_norm": 0.8771921558002355, "learning_rate": 7.957540409967122e-05, "loss": 0.993, "step": 1565 }, { "epoch": 0.7095604893520616, "grad_norm": 0.8784818173110891, "learning_rate": 7.95735633206855e-05, "loss": 0.9981, "step": 1566 }, { "epoch": 0.7100135931128229, "grad_norm": 0.8251155035513219, "learning_rate": 7.957171858148323e-05, "loss": 0.9926, "step": 1567 }, { "epoch": 0.7104666968735841, "grad_norm": 0.716981219706255, "learning_rate": 7.9569869882249e-05, "loss": 0.994, "step": 1568 }, { "epoch": 0.7109198006343452, "grad_norm": 0.5916347155339621, "learning_rate": 7.956801722316783e-05, "loss": 1.0236, "step": 1569 }, { "epoch": 0.7113729043951065, "grad_norm": 0.5947854862160126, "learning_rate": 7.956616060442512e-05, "loss": 1.0112, "step": 1570 }, { "epoch": 0.7118260081558677, "grad_norm": 0.6124566238646623, "learning_rate": 7.956430002620665e-05, "loss": 0.983, "step": 1571 }, { "epoch": 0.712279111916629, "grad_norm": 0.6146922218907788, "learning_rate": 7.956243548869863e-05, "loss": 1.0122, "step": 1572 }, { "epoch": 0.7127322156773901, "grad_norm": 0.5621448585699824, "learning_rate": 7.956056699208763e-05, "loss": 0.9981, "step": 1573 }, { "epoch": 0.7131853194381513, "grad_norm": 0.43542270766852975, "learning_rate": 7.955869453656066e-05, "loss": 1.0234, "step": 1574 }, { "epoch": 0.7136384231989126, "grad_norm": 0.3338372748709061, "learning_rate": 7.955681812230507e-05, "loss": 0.9882, "step": 1575 }, { "epoch": 0.7140915269596738, "grad_norm": 0.32745996994220444, "learning_rate": 7.955493774950865e-05, "loss": 1.0029, "step": 1576 }, { "epoch": 0.714544630720435, "grad_norm": 0.33352842696574325, "learning_rate": 7.955305341835961e-05, "loss": 1.026, "step": 1577 }, { "epoch": 0.7149977344811962, "grad_norm": 0.3958782275040566, "learning_rate": 7.955116512904645e-05, "loss": 1.0245, "step": 1578 }, { "epoch": 0.7154508382419574, "grad_norm": 0.4377245556886684, "learning_rate": 7.954927288175817e-05, "loss": 1.029, "step": 1579 }, { "epoch": 0.7159039420027187, "grad_norm": 0.46646017185266625, "learning_rate": 7.954737667668415e-05, "loss": 0.9895, "step": 1580 }, { "epoch": 0.7163570457634798, "grad_norm": 0.5163044789083415, "learning_rate": 7.954547651401413e-05, "loss": 1.0072, "step": 1581 }, { "epoch": 0.716810149524241, "grad_norm": 0.6023295203255714, "learning_rate": 7.954357239393826e-05, "loss": 1.0069, "step": 1582 }, { "epoch": 0.7172632532850023, "grad_norm": 0.5963351373151251, "learning_rate": 7.954166431664709e-05, "loss": 0.9957, "step": 1583 }, { "epoch": 0.7177163570457635, "grad_norm": 0.5336138977112196, "learning_rate": 7.953975228233157e-05, "loss": 0.999, "step": 1584 }, { "epoch": 0.7181694608065247, "grad_norm": 0.47217710916324324, "learning_rate": 7.953783629118304e-05, "loss": 0.9837, "step": 1585 }, { "epoch": 0.7186225645672859, "grad_norm": 0.3622263309521871, "learning_rate": 7.953591634339324e-05, "loss": 1.0142, "step": 1586 }, { "epoch": 0.7190756683280471, "grad_norm": 0.31989757643978195, "learning_rate": 7.95339924391543e-05, "loss": 1.0204, "step": 1587 }, { "epoch": 0.7195287720888084, "grad_norm": 0.36203206707883484, "learning_rate": 7.953206457865876e-05, "loss": 0.98, "step": 1588 }, { "epoch": 0.7199818758495695, "grad_norm": 0.4061456438355696, "learning_rate": 7.953013276209952e-05, "loss": 1.0081, "step": 1589 }, { "epoch": 0.7204349796103308, "grad_norm": 0.4180686643324238, "learning_rate": 7.952819698966991e-05, "loss": 1.0073, "step": 1590 }, { "epoch": 0.720888083371092, "grad_norm": 0.3731159747227544, "learning_rate": 7.952625726156368e-05, "loss": 0.9942, "step": 1591 }, { "epoch": 0.7213411871318532, "grad_norm": 0.4040229383569668, "learning_rate": 7.952431357797491e-05, "loss": 1.0122, "step": 1592 }, { "epoch": 0.7217942908926144, "grad_norm": 0.4521783382855033, "learning_rate": 7.952236593909812e-05, "loss": 1.0098, "step": 1593 }, { "epoch": 0.7222473946533756, "grad_norm": 0.5453185936179477, "learning_rate": 7.952041434512819e-05, "loss": 1.0001, "step": 1594 }, { "epoch": 0.7227004984141369, "grad_norm": 0.6809164988822938, "learning_rate": 7.951845879626046e-05, "loss": 1.0191, "step": 1595 }, { "epoch": 0.7231536021748981, "grad_norm": 0.8243861589966297, "learning_rate": 7.951649929269061e-05, "loss": 1.0129, "step": 1596 }, { "epoch": 0.7236067059356592, "grad_norm": 0.9105047638494694, "learning_rate": 7.951453583461473e-05, "loss": 1.028, "step": 1597 }, { "epoch": 0.7240598096964205, "grad_norm": 0.8495244854176752, "learning_rate": 7.95125684222293e-05, "loss": 1.0249, "step": 1598 }, { "epoch": 0.7245129134571817, "grad_norm": 0.662049537250751, "learning_rate": 7.951059705573122e-05, "loss": 0.9923, "step": 1599 }, { "epoch": 0.724966017217943, "grad_norm": 0.5299971409330597, "learning_rate": 7.950862173531776e-05, "loss": 1.016, "step": 1600 }, { "epoch": 0.7254191209787041, "grad_norm": 0.5231015314015639, "learning_rate": 7.950664246118658e-05, "loss": 1.0167, "step": 1601 }, { "epoch": 0.7258722247394653, "grad_norm": 0.5532449720503836, "learning_rate": 7.950465923353576e-05, "loss": 1.0074, "step": 1602 }, { "epoch": 0.7263253285002266, "grad_norm": 0.5085413196294509, "learning_rate": 7.95026720525638e-05, "loss": 1.0128, "step": 1603 }, { "epoch": 0.7267784322609878, "grad_norm": 0.43238829849945826, "learning_rate": 7.950068091846953e-05, "loss": 1.0022, "step": 1604 }, { "epoch": 0.727231536021749, "grad_norm": 0.45117305539097835, "learning_rate": 7.949868583145218e-05, "loss": 1.0227, "step": 1605 }, { "epoch": 0.7276846397825102, "grad_norm": 0.48545717529877674, "learning_rate": 7.949668679171146e-05, "loss": 0.9857, "step": 1606 }, { "epoch": 0.7281377435432714, "grad_norm": 0.4038879375634923, "learning_rate": 7.949468379944738e-05, "loss": 1.0025, "step": 1607 }, { "epoch": 0.7285908473040327, "grad_norm": 0.34668600995878457, "learning_rate": 7.949267685486041e-05, "loss": 1.0154, "step": 1608 }, { "epoch": 0.7290439510647938, "grad_norm": 0.37008849164333835, "learning_rate": 7.949066595815136e-05, "loss": 0.9894, "step": 1609 }, { "epoch": 0.729497054825555, "grad_norm": 0.37078634078569883, "learning_rate": 7.94886511095215e-05, "loss": 1.0145, "step": 1610 }, { "epoch": 0.7299501585863163, "grad_norm": 0.3184963789043455, "learning_rate": 7.948663230917243e-05, "loss": 1.0221, "step": 1611 }, { "epoch": 0.7304032623470775, "grad_norm": 0.3195299129713063, "learning_rate": 7.948460955730619e-05, "loss": 0.9909, "step": 1612 }, { "epoch": 0.7308563661078387, "grad_norm": 0.3731718969250711, "learning_rate": 7.948258285412519e-05, "loss": 1.0071, "step": 1613 }, { "epoch": 0.7313094698685999, "grad_norm": 0.3852357645154974, "learning_rate": 7.948055219983226e-05, "loss": 0.9975, "step": 1614 }, { "epoch": 0.7317625736293611, "grad_norm": 0.4217827941488213, "learning_rate": 7.947851759463061e-05, "loss": 0.9967, "step": 1615 }, { "epoch": 0.7322156773901224, "grad_norm": 0.5180223667307684, "learning_rate": 7.947647903872385e-05, "loss": 1.0186, "step": 1616 }, { "epoch": 0.7326687811508835, "grad_norm": 0.6224300500826725, "learning_rate": 7.947443653231597e-05, "loss": 1.0206, "step": 1617 }, { "epoch": 0.7331218849116448, "grad_norm": 0.6858490436395657, "learning_rate": 7.947239007561138e-05, "loss": 1.0087, "step": 1618 }, { "epoch": 0.733574988672406, "grad_norm": 0.7668482042581413, "learning_rate": 7.947033966881485e-05, "loss": 1.016, "step": 1619 }, { "epoch": 0.7340280924331672, "grad_norm": 0.8181620155594636, "learning_rate": 7.94682853121316e-05, "loss": 0.9924, "step": 1620 }, { "epoch": 0.7344811961939284, "grad_norm": 0.8035014464623123, "learning_rate": 7.946622700576723e-05, "loss": 0.9884, "step": 1621 }, { "epoch": 0.7349342999546896, "grad_norm": 0.6835200238930198, "learning_rate": 7.946416474992767e-05, "loss": 1.0117, "step": 1622 }, { "epoch": 0.7353874037154509, "grad_norm": 0.5769116999151161, "learning_rate": 7.946209854481931e-05, "loss": 1.0163, "step": 1623 }, { "epoch": 0.7358405074762121, "grad_norm": 0.6218922881745422, "learning_rate": 7.946002839064892e-05, "loss": 1.017, "step": 1624 }, { "epoch": 0.7362936112369732, "grad_norm": 0.5683586611325129, "learning_rate": 7.94579542876237e-05, "loss": 0.9758, "step": 1625 }, { "epoch": 0.7367467149977345, "grad_norm": 0.4794631262082956, "learning_rate": 7.945587623595118e-05, "loss": 0.9821, "step": 1626 }, { "epoch": 0.7371998187584957, "grad_norm": 0.4137033713767733, "learning_rate": 7.945379423583929e-05, "loss": 1.007, "step": 1627 }, { "epoch": 0.737652922519257, "grad_norm": 0.3643094602271056, "learning_rate": 7.945170828749643e-05, "loss": 1.0181, "step": 1628 }, { "epoch": 0.7381060262800181, "grad_norm": 0.4081266760845532, "learning_rate": 7.944961839113132e-05, "loss": 1.0225, "step": 1629 }, { "epoch": 0.7385591300407793, "grad_norm": 0.41624449755302784, "learning_rate": 7.944752454695309e-05, "loss": 1.0, "step": 1630 }, { "epoch": 0.7390122338015406, "grad_norm": 0.39916789171583894, "learning_rate": 7.944542675517129e-05, "loss": 1.0163, "step": 1631 }, { "epoch": 0.7394653375623018, "grad_norm": 0.47462394450858686, "learning_rate": 7.944332501599587e-05, "loss": 1.0179, "step": 1632 }, { "epoch": 0.739918441323063, "grad_norm": 0.5352625449480276, "learning_rate": 7.944121932963714e-05, "loss": 0.9966, "step": 1633 }, { "epoch": 0.7403715450838242, "grad_norm": 0.5863197783894135, "learning_rate": 7.943910969630579e-05, "loss": 1.0136, "step": 1634 }, { "epoch": 0.7408246488445854, "grad_norm": 0.7010118017311967, "learning_rate": 7.943699611621297e-05, "loss": 1.0133, "step": 1635 }, { "epoch": 0.7412777526053467, "grad_norm": 0.8991661596429811, "learning_rate": 7.94348785895702e-05, "loss": 0.9943, "step": 1636 }, { "epoch": 0.7417308563661078, "grad_norm": 1.1023737608516182, "learning_rate": 7.943275711658936e-05, "loss": 1.0273, "step": 1637 }, { "epoch": 0.742183960126869, "grad_norm": 0.8474879808785131, "learning_rate": 7.943063169748275e-05, "loss": 0.9919, "step": 1638 }, { "epoch": 0.7426370638876303, "grad_norm": 0.5551403234316276, "learning_rate": 7.94285023324631e-05, "loss": 1.0236, "step": 1639 }, { "epoch": 0.7430901676483915, "grad_norm": 0.3906005696854374, "learning_rate": 7.942636902174346e-05, "loss": 1.0329, "step": 1640 }, { "epoch": 0.7435432714091527, "grad_norm": 0.47285114575359405, "learning_rate": 7.942423176553733e-05, "loss": 1.0125, "step": 1641 }, { "epoch": 0.7439963751699139, "grad_norm": 0.6647369097098993, "learning_rate": 7.94220905640586e-05, "loss": 0.9983, "step": 1642 }, { "epoch": 0.7444494789306751, "grad_norm": 0.7729741179176136, "learning_rate": 7.941994541752152e-05, "loss": 0.9985, "step": 1643 }, { "epoch": 0.7449025826914364, "grad_norm": 0.7812480803585076, "learning_rate": 7.94177963261408e-05, "loss": 1.0037, "step": 1644 }, { "epoch": 0.7453556864521975, "grad_norm": 0.7053576546967346, "learning_rate": 7.941564329013146e-05, "loss": 0.9935, "step": 1645 }, { "epoch": 0.7458087902129588, "grad_norm": 0.5964423692422425, "learning_rate": 7.9413486309709e-05, "loss": 0.9986, "step": 1646 }, { "epoch": 0.74626189397372, "grad_norm": 0.47584370376517915, "learning_rate": 7.941132538508924e-05, "loss": 1.0066, "step": 1647 }, { "epoch": 0.7467149977344812, "grad_norm": 0.3605203890301983, "learning_rate": 7.940916051648843e-05, "loss": 1.0019, "step": 1648 }, { "epoch": 0.7471681014952424, "grad_norm": 0.5284585558437289, "learning_rate": 7.940699170412325e-05, "loss": 0.991, "step": 1649 }, { "epoch": 0.7476212052560036, "grad_norm": 0.6828239096040992, "learning_rate": 7.940481894821069e-05, "loss": 1.0327, "step": 1650 }, { "epoch": 0.7480743090167649, "grad_norm": 0.6957461424452519, "learning_rate": 7.940264224896822e-05, "loss": 0.995, "step": 1651 }, { "epoch": 0.7485274127775261, "grad_norm": 0.5733268133507373, "learning_rate": 7.940046160661367e-05, "loss": 1.0153, "step": 1652 }, { "epoch": 0.7489805165382872, "grad_norm": 0.4190226224617187, "learning_rate": 7.939827702136521e-05, "loss": 1.0135, "step": 1653 }, { "epoch": 0.7494336202990485, "grad_norm": 0.39154063173411524, "learning_rate": 7.939608849344151e-05, "loss": 1.0, "step": 1654 }, { "epoch": 0.7498867240598097, "grad_norm": 0.4232942933720083, "learning_rate": 7.939389602306156e-05, "loss": 0.9922, "step": 1655 }, { "epoch": 0.750339827820571, "grad_norm": 0.3853842869027918, "learning_rate": 7.939169961044478e-05, "loss": 1.0056, "step": 1656 }, { "epoch": 0.7507929315813321, "grad_norm": 0.34020250708601546, "learning_rate": 7.938949925581095e-05, "loss": 1.0173, "step": 1657 }, { "epoch": 0.7512460353420933, "grad_norm": 0.3584745609909878, "learning_rate": 7.938729495938028e-05, "loss": 1.0279, "step": 1658 }, { "epoch": 0.7516991391028546, "grad_norm": 0.37908909031999344, "learning_rate": 7.938508672137334e-05, "loss": 0.9892, "step": 1659 }, { "epoch": 0.7521522428636158, "grad_norm": 0.31027248563240706, "learning_rate": 7.938287454201112e-05, "loss": 1.0135, "step": 1660 }, { "epoch": 0.752605346624377, "grad_norm": 0.2771381662451795, "learning_rate": 7.938065842151501e-05, "loss": 1.0111, "step": 1661 }, { "epoch": 0.7530584503851382, "grad_norm": 0.3011509684204425, "learning_rate": 7.937843836010678e-05, "loss": 0.9928, "step": 1662 }, { "epoch": 0.7535115541458994, "grad_norm": 0.3119470116572116, "learning_rate": 7.937621435800859e-05, "loss": 0.9935, "step": 1663 }, { "epoch": 0.7539646579066607, "grad_norm": 0.3471780401831567, "learning_rate": 7.937398641544301e-05, "loss": 0.9956, "step": 1664 }, { "epoch": 0.7544177616674218, "grad_norm": 0.3814470298079084, "learning_rate": 7.937175453263297e-05, "loss": 1.0097, "step": 1665 }, { "epoch": 0.754870865428183, "grad_norm": 0.351877048562439, "learning_rate": 7.936951870980186e-05, "loss": 0.9928, "step": 1666 }, { "epoch": 0.7553239691889443, "grad_norm": 0.37984735378096407, "learning_rate": 7.936727894717337e-05, "loss": 1.007, "step": 1667 }, { "epoch": 0.7557770729497055, "grad_norm": 0.42020507108204785, "learning_rate": 7.93650352449717e-05, "loss": 1.0098, "step": 1668 }, { "epoch": 0.7562301767104667, "grad_norm": 0.49512093084156217, "learning_rate": 7.936278760342134e-05, "loss": 1.0141, "step": 1669 }, { "epoch": 0.7566832804712279, "grad_norm": 0.7075754301413149, "learning_rate": 7.936053602274722e-05, "loss": 1.0284, "step": 1670 }, { "epoch": 0.7571363842319891, "grad_norm": 0.7563261228417391, "learning_rate": 7.935828050317468e-05, "loss": 1.0255, "step": 1671 }, { "epoch": 0.7575894879927504, "grad_norm": 0.8985410597604552, "learning_rate": 7.935602104492942e-05, "loss": 1.0299, "step": 1672 }, { "epoch": 0.7580425917535115, "grad_norm": 1.0406718380325013, "learning_rate": 7.935375764823756e-05, "loss": 1.0079, "step": 1673 }, { "epoch": 0.7584956955142728, "grad_norm": 0.9877485955057882, "learning_rate": 7.935149031332559e-05, "loss": 1.004, "step": 1674 }, { "epoch": 0.758948799275034, "grad_norm": 0.9528538114612987, "learning_rate": 7.934921904042042e-05, "loss": 1.0062, "step": 1675 }, { "epoch": 0.7594019030357952, "grad_norm": 0.8565317894396598, "learning_rate": 7.934694382974932e-05, "loss": 1.0119, "step": 1676 }, { "epoch": 0.7598550067965564, "grad_norm": 0.6741811067071641, "learning_rate": 7.934466468154002e-05, "loss": 1.0231, "step": 1677 }, { "epoch": 0.7603081105573176, "grad_norm": 0.45100095989740235, "learning_rate": 7.934238159602056e-05, "loss": 0.999, "step": 1678 }, { "epoch": 0.7607612143180789, "grad_norm": 0.4227246289115686, "learning_rate": 7.93400945734194e-05, "loss": 1.0151, "step": 1679 }, { "epoch": 0.7612143180788401, "grad_norm": 0.5703107548042767, "learning_rate": 7.933780361396545e-05, "loss": 1.0088, "step": 1680 }, { "epoch": 0.7616674218396012, "grad_norm": 0.6378023236102821, "learning_rate": 7.933550871788796e-05, "loss": 1.0217, "step": 1681 }, { "epoch": 0.7621205256003625, "grad_norm": 0.6353649123905634, "learning_rate": 7.933320988541658e-05, "loss": 0.9967, "step": 1682 }, { "epoch": 0.7625736293611237, "grad_norm": 0.6440240958846031, "learning_rate": 7.933090711678134e-05, "loss": 1.0007, "step": 1683 }, { "epoch": 0.763026733121885, "grad_norm": 0.6559303908369104, "learning_rate": 7.932860041221272e-05, "loss": 1.0156, "step": 1684 }, { "epoch": 0.7634798368826461, "grad_norm": 0.5790447263957664, "learning_rate": 7.932628977194153e-05, "loss": 1.0143, "step": 1685 }, { "epoch": 0.7639329406434073, "grad_norm": 0.3786316956205154, "learning_rate": 7.932397519619902e-05, "loss": 1.0205, "step": 1686 }, { "epoch": 0.7643860444041686, "grad_norm": 0.3896127134842547, "learning_rate": 7.932165668521679e-05, "loss": 1.0002, "step": 1687 }, { "epoch": 0.7648391481649298, "grad_norm": 0.5358144654371881, "learning_rate": 7.931933423922686e-05, "loss": 1.0257, "step": 1688 }, { "epoch": 0.765292251925691, "grad_norm": 0.5186935928666013, "learning_rate": 7.931700785846168e-05, "loss": 1.0109, "step": 1689 }, { "epoch": 0.7657453556864522, "grad_norm": 0.4893581546687109, "learning_rate": 7.931467754315403e-05, "loss": 0.9954, "step": 1690 }, { "epoch": 0.7661984594472134, "grad_norm": 0.5479367685712077, "learning_rate": 7.931234329353711e-05, "loss": 1.0131, "step": 1691 }, { "epoch": 0.7666515632079747, "grad_norm": 0.5208895808642376, "learning_rate": 7.931000510984451e-05, "loss": 1.029, "step": 1692 }, { "epoch": 0.7671046669687358, "grad_norm": 0.4186016676486666, "learning_rate": 7.930766299231023e-05, "loss": 1.0165, "step": 1693 }, { "epoch": 0.767557770729497, "grad_norm": 0.35464820829206023, "learning_rate": 7.930531694116864e-05, "loss": 0.9833, "step": 1694 }, { "epoch": 0.7680108744902583, "grad_norm": 0.36014186109030066, "learning_rate": 7.930296695665451e-05, "loss": 0.9828, "step": 1695 }, { "epoch": 0.7684639782510195, "grad_norm": 0.460482606238101, "learning_rate": 7.930061303900303e-05, "loss": 0.9911, "step": 1696 }, { "epoch": 0.7689170820117807, "grad_norm": 0.5615832940303308, "learning_rate": 7.929825518844974e-05, "loss": 0.9858, "step": 1697 }, { "epoch": 0.7693701857725419, "grad_norm": 0.5912377053355609, "learning_rate": 7.92958934052306e-05, "loss": 0.9989, "step": 1698 }, { "epoch": 0.7698232895333031, "grad_norm": 0.5806259067723458, "learning_rate": 7.929352768958199e-05, "loss": 1.0007, "step": 1699 }, { "epoch": 0.7702763932940644, "grad_norm": 0.5752337483363245, "learning_rate": 7.929115804174061e-05, "loss": 1.0273, "step": 1700 }, { "epoch": 0.7707294970548255, "grad_norm": 0.7005596732727789, "learning_rate": 7.928878446194361e-05, "loss": 1.0251, "step": 1701 }, { "epoch": 0.7711826008155868, "grad_norm": 0.7964393911977502, "learning_rate": 7.928640695042851e-05, "loss": 1.0235, "step": 1702 }, { "epoch": 0.771635704576348, "grad_norm": 0.7668375380670055, "learning_rate": 7.928402550743325e-05, "loss": 1.0028, "step": 1703 }, { "epoch": 0.7720888083371092, "grad_norm": 0.6914133898121193, "learning_rate": 7.928164013319615e-05, "loss": 1.0208, "step": 1704 }, { "epoch": 0.7725419120978704, "grad_norm": 0.5656961279583747, "learning_rate": 7.927925082795591e-05, "loss": 1.0064, "step": 1705 }, { "epoch": 0.7729950158586316, "grad_norm": 0.426067368477409, "learning_rate": 7.927685759195163e-05, "loss": 1.0161, "step": 1706 }, { "epoch": 0.7734481196193929, "grad_norm": 0.4371127051021416, "learning_rate": 7.927446042542282e-05, "loss": 1.0047, "step": 1707 }, { "epoch": 0.7739012233801541, "grad_norm": 0.4786832861299591, "learning_rate": 7.927205932860935e-05, "loss": 1.0006, "step": 1708 }, { "epoch": 0.7743543271409152, "grad_norm": 0.45751172043113464, "learning_rate": 7.926965430175151e-05, "loss": 0.9915, "step": 1709 }, { "epoch": 0.7748074309016765, "grad_norm": 0.3887446163849064, "learning_rate": 7.926724534508997e-05, "loss": 0.9866, "step": 1710 }, { "epoch": 0.7752605346624377, "grad_norm": 0.40734868682953446, "learning_rate": 7.926483245886584e-05, "loss": 0.9982, "step": 1711 }, { "epoch": 0.775713638423199, "grad_norm": 0.40750578368018625, "learning_rate": 7.926241564332054e-05, "loss": 0.9863, "step": 1712 }, { "epoch": 0.7761667421839601, "grad_norm": 0.3578690474446747, "learning_rate": 7.925999489869593e-05, "loss": 1.0003, "step": 1713 }, { "epoch": 0.7766198459447213, "grad_norm": 0.3618136045792735, "learning_rate": 7.925757022523428e-05, "loss": 0.9949, "step": 1714 }, { "epoch": 0.7770729497054826, "grad_norm": 0.3876030847072061, "learning_rate": 7.925514162317822e-05, "loss": 0.9953, "step": 1715 }, { "epoch": 0.7775260534662438, "grad_norm": 0.446280369400854, "learning_rate": 7.925270909277078e-05, "loss": 0.9962, "step": 1716 }, { "epoch": 0.777979157227005, "grad_norm": 0.5458024834106662, "learning_rate": 7.92502726342554e-05, "loss": 0.996, "step": 1717 }, { "epoch": 0.7784322609877662, "grad_norm": 0.5784754460479163, "learning_rate": 7.92478322478759e-05, "loss": 1.0276, "step": 1718 }, { "epoch": 0.7788853647485274, "grad_norm": 0.6465590196119311, "learning_rate": 7.924538793387649e-05, "loss": 1.0001, "step": 1719 }, { "epoch": 0.7793384685092887, "grad_norm": 0.7155578598966205, "learning_rate": 7.924293969250179e-05, "loss": 0.9847, "step": 1720 }, { "epoch": 0.7797915722700498, "grad_norm": 0.6992954467576794, "learning_rate": 7.924048752399677e-05, "loss": 1.0074, "step": 1721 }, { "epoch": 0.780244676030811, "grad_norm": 0.5614438871944243, "learning_rate": 7.923803142860686e-05, "loss": 0.9981, "step": 1722 }, { "epoch": 0.7806977797915723, "grad_norm": 0.47000054524059875, "learning_rate": 7.923557140657783e-05, "loss": 1.0046, "step": 1723 }, { "epoch": 0.7811508835523335, "grad_norm": 0.4648954135407203, "learning_rate": 7.923310745815586e-05, "loss": 1.0049, "step": 1724 }, { "epoch": 0.7816039873130947, "grad_norm": 0.42893116891789485, "learning_rate": 7.923063958358753e-05, "loss": 1.023, "step": 1725 }, { "epoch": 0.7820570910738559, "grad_norm": 0.43320509085905456, "learning_rate": 7.92281677831198e-05, "loss": 0.9996, "step": 1726 }, { "epoch": 0.7825101948346171, "grad_norm": 0.35106887496805816, "learning_rate": 7.922569205700003e-05, "loss": 1.002, "step": 1727 }, { "epoch": 0.7829632985953784, "grad_norm": 0.3640231197046711, "learning_rate": 7.922321240547599e-05, "loss": 1.0123, "step": 1728 }, { "epoch": 0.7834164023561395, "grad_norm": 0.4142714603986028, "learning_rate": 7.922072882879578e-05, "loss": 1.0224, "step": 1729 }, { "epoch": 0.7838695061169008, "grad_norm": 0.41870775375759983, "learning_rate": 7.921824132720797e-05, "loss": 0.9922, "step": 1730 }, { "epoch": 0.784322609877662, "grad_norm": 0.42731481270642707, "learning_rate": 7.92157499009615e-05, "loss": 0.9739, "step": 1731 }, { "epoch": 0.7847757136384232, "grad_norm": 0.491297782554837, "learning_rate": 7.921325455030567e-05, "loss": 1.019, "step": 1732 }, { "epoch": 0.7852288173991844, "grad_norm": 0.582516077004279, "learning_rate": 7.921075527549019e-05, "loss": 0.9965, "step": 1733 }, { "epoch": 0.7856819211599456, "grad_norm": 0.7064779955219811, "learning_rate": 7.920825207676517e-05, "loss": 1.0005, "step": 1734 }, { "epoch": 0.7861350249207069, "grad_norm": 0.8407785015631494, "learning_rate": 7.920574495438114e-05, "loss": 1.0185, "step": 1735 }, { "epoch": 0.7865881286814681, "grad_norm": 0.944956751530791, "learning_rate": 7.920323390858897e-05, "loss": 1.0036, "step": 1736 }, { "epoch": 0.7870412324422292, "grad_norm": 0.9761581979035553, "learning_rate": 7.920071893963995e-05, "loss": 0.9813, "step": 1737 }, { "epoch": 0.7874943362029905, "grad_norm": 0.9388008921688037, "learning_rate": 7.919820004778576e-05, "loss": 0.9847, "step": 1738 }, { "epoch": 0.7879474399637517, "grad_norm": 0.7183119856011982, "learning_rate": 7.919567723327847e-05, "loss": 1.0092, "step": 1739 }, { "epoch": 0.788400543724513, "grad_norm": 0.48062871847479677, "learning_rate": 7.919315049637055e-05, "loss": 1.0185, "step": 1740 }, { "epoch": 0.7888536474852741, "grad_norm": 0.4353240493968448, "learning_rate": 7.919061983731484e-05, "loss": 1.0049, "step": 1741 }, { "epoch": 0.7893067512460353, "grad_norm": 0.44624179648010465, "learning_rate": 7.918808525636459e-05, "loss": 1.0192, "step": 1742 }, { "epoch": 0.7897598550067966, "grad_norm": 0.49617270548398656, "learning_rate": 7.918554675377346e-05, "loss": 1.011, "step": 1743 }, { "epoch": 0.7902129587675578, "grad_norm": 0.5622422419815405, "learning_rate": 7.918300432979548e-05, "loss": 1.0024, "step": 1744 }, { "epoch": 0.790666062528319, "grad_norm": 0.6686846797685868, "learning_rate": 7.918045798468506e-05, "loss": 1.0081, "step": 1745 }, { "epoch": 0.7911191662890802, "grad_norm": 0.6993204848242809, "learning_rate": 7.917790771869705e-05, "loss": 1.0043, "step": 1746 }, { "epoch": 0.7915722700498414, "grad_norm": 0.6741614213226226, "learning_rate": 7.917535353208661e-05, "loss": 0.9769, "step": 1747 }, { "epoch": 0.7920253738106027, "grad_norm": 0.6816942480119196, "learning_rate": 7.917279542510939e-05, "loss": 1.0852, "step": 1748 }, { "epoch": 0.7924784775713638, "grad_norm": 0.4538196132781188, "learning_rate": 7.917023339802136e-05, "loss": 1.0147, "step": 1749 }, { "epoch": 0.792931581332125, "grad_norm": 0.4828902176884733, "learning_rate": 7.916766745107891e-05, "loss": 0.9765, "step": 1750 }, { "epoch": 0.7933846850928863, "grad_norm": 0.5600110803353898, "learning_rate": 7.916509758453884e-05, "loss": 1.0093, "step": 1751 }, { "epoch": 0.7938377888536475, "grad_norm": 0.4692574411135353, "learning_rate": 7.916252379865832e-05, "loss": 1.0223, "step": 1752 }, { "epoch": 0.7942908926144087, "grad_norm": 0.4332950805587387, "learning_rate": 7.915994609369489e-05, "loss": 1.0098, "step": 1753 }, { "epoch": 0.7947439963751699, "grad_norm": 0.4630825972245691, "learning_rate": 7.915736446990651e-05, "loss": 0.9893, "step": 1754 }, { "epoch": 0.7951971001359311, "grad_norm": 0.457090636430734, "learning_rate": 7.915477892755155e-05, "loss": 1.0233, "step": 1755 }, { "epoch": 0.7956502038966924, "grad_norm": 0.4371060429624419, "learning_rate": 7.915218946688874e-05, "loss": 0.9993, "step": 1756 }, { "epoch": 0.7961033076574535, "grad_norm": 0.3787090268547609, "learning_rate": 7.914959608817722e-05, "loss": 1.0046, "step": 1757 }, { "epoch": 0.7965564114182148, "grad_norm": 0.308517371622529, "learning_rate": 7.91469987916765e-05, "loss": 1.0231, "step": 1758 }, { "epoch": 0.797009515178976, "grad_norm": 0.34505234775846944, "learning_rate": 7.914439757764649e-05, "loss": 1.0292, "step": 1759 }, { "epoch": 0.7974626189397372, "grad_norm": 0.368712677780817, "learning_rate": 7.914179244634755e-05, "loss": 0.9943, "step": 1760 }, { "epoch": 0.7979157227004984, "grad_norm": 0.42968436978019486, "learning_rate": 7.913918339804031e-05, "loss": 1.0015, "step": 1761 }, { "epoch": 0.7983688264612596, "grad_norm": 0.47941532347431415, "learning_rate": 7.913657043298593e-05, "loss": 1.0152, "step": 1762 }, { "epoch": 0.7988219302220209, "grad_norm": 0.41935974810767634, "learning_rate": 7.913395355144585e-05, "loss": 0.9951, "step": 1763 }, { "epoch": 0.7992750339827821, "grad_norm": 0.4306821454611544, "learning_rate": 7.913133275368197e-05, "loss": 0.997, "step": 1764 }, { "epoch": 0.7997281377435432, "grad_norm": 0.4745832001968623, "learning_rate": 7.912870803995656e-05, "loss": 1.0152, "step": 1765 }, { "epoch": 0.8001812415043045, "grad_norm": 0.5706517677536564, "learning_rate": 7.912607941053227e-05, "loss": 1.0104, "step": 1766 }, { "epoch": 0.8006343452650657, "grad_norm": 0.698760877606769, "learning_rate": 7.912344686567213e-05, "loss": 0.9947, "step": 1767 }, { "epoch": 0.801087449025827, "grad_norm": 0.7664047256174077, "learning_rate": 7.912081040563964e-05, "loss": 1.0025, "step": 1768 }, { "epoch": 0.8015405527865881, "grad_norm": 0.7768980826281738, "learning_rate": 7.911817003069861e-05, "loss": 1.011, "step": 1769 }, { "epoch": 0.8019936565473493, "grad_norm": 0.845716064293413, "learning_rate": 7.911552574111325e-05, "loss": 0.9845, "step": 1770 }, { "epoch": 0.8024467603081106, "grad_norm": 0.8690539931970847, "learning_rate": 7.91128775371482e-05, "loss": 1.0059, "step": 1771 }, { "epoch": 0.8028998640688718, "grad_norm": 0.846728643074039, "learning_rate": 7.911022541906847e-05, "loss": 1.0033, "step": 1772 }, { "epoch": 0.803352967829633, "grad_norm": 0.7383913333783318, "learning_rate": 7.910756938713946e-05, "loss": 1.0163, "step": 1773 }, { "epoch": 0.8038060715903942, "grad_norm": 0.5706071751723576, "learning_rate": 7.910490944162697e-05, "loss": 0.997, "step": 1774 }, { "epoch": 0.8042591753511554, "grad_norm": 0.49513584650436626, "learning_rate": 7.910224558279718e-05, "loss": 0.9836, "step": 1775 }, { "epoch": 0.8047122791119167, "grad_norm": 0.4919071969444239, "learning_rate": 7.909957781091668e-05, "loss": 1.0115, "step": 1776 }, { "epoch": 0.8051653828726778, "grad_norm": 0.49514317750350473, "learning_rate": 7.909690612625242e-05, "loss": 0.9864, "step": 1777 }, { "epoch": 0.805618486633439, "grad_norm": 0.5429322421816768, "learning_rate": 7.909423052907177e-05, "loss": 1.0161, "step": 1778 }, { "epoch": 0.8060715903942003, "grad_norm": 0.5581671191016393, "learning_rate": 7.909155101964249e-05, "loss": 1.0064, "step": 1779 }, { "epoch": 0.8065246941549615, "grad_norm": 0.5377349793389113, "learning_rate": 7.908886759823272e-05, "loss": 1.0007, "step": 1780 }, { "epoch": 0.8069777979157227, "grad_norm": 0.558773606301179, "learning_rate": 7.9086180265111e-05, "loss": 0.9894, "step": 1781 }, { "epoch": 0.8074309016764839, "grad_norm": 0.5314829065429638, "learning_rate": 7.908348902054625e-05, "loss": 0.9847, "step": 1782 }, { "epoch": 0.8078840054372451, "grad_norm": 0.4493365538269386, "learning_rate": 7.90807938648078e-05, "loss": 0.9889, "step": 1783 }, { "epoch": 0.8083371091980064, "grad_norm": 0.41843370006374675, "learning_rate": 7.907809479816536e-05, "loss": 0.9957, "step": 1784 }, { "epoch": 0.8087902129587675, "grad_norm": 0.5699852564319579, "learning_rate": 7.9075391820889e-05, "loss": 1.0191, "step": 1785 }, { "epoch": 0.8092433167195288, "grad_norm": 0.6436748295883203, "learning_rate": 7.907268493324927e-05, "loss": 1.015, "step": 1786 }, { "epoch": 0.80969642048029, "grad_norm": 0.5640774683019906, "learning_rate": 7.9069974135517e-05, "loss": 1.0134, "step": 1787 }, { "epoch": 0.8101495242410512, "grad_norm": 0.5107424562785331, "learning_rate": 7.906725942796352e-05, "loss": 0.9843, "step": 1788 }, { "epoch": 0.8106026280018124, "grad_norm": 0.5075727888046286, "learning_rate": 7.906454081086044e-05, "loss": 1.0143, "step": 1789 }, { "epoch": 0.8110557317625736, "grad_norm": 0.5393413610389007, "learning_rate": 7.906181828447984e-05, "loss": 0.9819, "step": 1790 }, { "epoch": 0.8115088355233349, "grad_norm": 0.49267080485197595, "learning_rate": 7.905909184909418e-05, "loss": 0.9748, "step": 1791 }, { "epoch": 0.8119619392840961, "grad_norm": 0.4587915283517099, "learning_rate": 7.90563615049763e-05, "loss": 0.999, "step": 1792 }, { "epoch": 0.8124150430448572, "grad_norm": 0.5031334504474669, "learning_rate": 7.905362725239943e-05, "loss": 0.982, "step": 1793 }, { "epoch": 0.8128681468056185, "grad_norm": 0.5399758024970294, "learning_rate": 7.905088909163717e-05, "loss": 1.0053, "step": 1794 }, { "epoch": 0.8133212505663797, "grad_norm": 0.503605679353082, "learning_rate": 7.904814702296357e-05, "loss": 1.0085, "step": 1795 }, { "epoch": 0.813774354327141, "grad_norm": 0.46925666949539013, "learning_rate": 7.904540104665301e-05, "loss": 1.0032, "step": 1796 }, { "epoch": 0.8142274580879021, "grad_norm": 0.4688656675728691, "learning_rate": 7.904265116298029e-05, "loss": 0.9717, "step": 1797 }, { "epoch": 0.8146805618486633, "grad_norm": 0.4885557022140133, "learning_rate": 7.903989737222061e-05, "loss": 0.9979, "step": 1798 }, { "epoch": 0.8151336656094246, "grad_norm": 0.39616521298073915, "learning_rate": 7.903713967464953e-05, "loss": 0.9862, "step": 1799 }, { "epoch": 0.8155867693701858, "grad_norm": 0.3762262429082591, "learning_rate": 7.903437807054303e-05, "loss": 1.0156, "step": 1800 }, { "epoch": 0.816039873130947, "grad_norm": 0.4554658535110703, "learning_rate": 7.903161256017746e-05, "loss": 0.9869, "step": 1801 }, { "epoch": 0.8164929768917082, "grad_norm": 0.5267949788072804, "learning_rate": 7.90288431438296e-05, "loss": 0.9958, "step": 1802 }, { "epoch": 0.8169460806524694, "grad_norm": 0.5549384098053076, "learning_rate": 7.902606982177656e-05, "loss": 1.0178, "step": 1803 }, { "epoch": 0.8173991844132307, "grad_norm": 0.6137419435288811, "learning_rate": 7.90232925942959e-05, "loss": 1.0031, "step": 1804 }, { "epoch": 0.8178522881739918, "grad_norm": 0.6415422080368998, "learning_rate": 7.90205114616655e-05, "loss": 0.9949, "step": 1805 }, { "epoch": 0.818305391934753, "grad_norm": 0.6627521623183902, "learning_rate": 7.901772642416371e-05, "loss": 0.9948, "step": 1806 }, { "epoch": 0.8187584956955143, "grad_norm": 0.6563183136585526, "learning_rate": 7.901493748206923e-05, "loss": 1.0033, "step": 1807 }, { "epoch": 0.8192115994562755, "grad_norm": 0.6901602806768219, "learning_rate": 7.901214463566115e-05, "loss": 1.0092, "step": 1808 }, { "epoch": 0.8196647032170367, "grad_norm": 0.7337935905701471, "learning_rate": 7.900934788521895e-05, "loss": 0.9976, "step": 1809 }, { "epoch": 0.8201178069777979, "grad_norm": 0.8086062509738247, "learning_rate": 7.900654723102254e-05, "loss": 1.0071, "step": 1810 }, { "epoch": 0.8205709107385591, "grad_norm": 0.8221491064183633, "learning_rate": 7.900374267335214e-05, "loss": 1.0066, "step": 1811 }, { "epoch": 0.8210240144993204, "grad_norm": 0.7811585973991464, "learning_rate": 7.900093421248845e-05, "loss": 0.9881, "step": 1812 }, { "epoch": 0.8214771182600815, "grad_norm": 0.7033417203585226, "learning_rate": 7.899812184871248e-05, "loss": 1.0126, "step": 1813 }, { "epoch": 0.8219302220208428, "grad_norm": 0.6152652828760458, "learning_rate": 7.89953055823057e-05, "loss": 1.0264, "step": 1814 }, { "epoch": 0.822383325781604, "grad_norm": 0.5614738922230755, "learning_rate": 7.899248541354994e-05, "loss": 1.0109, "step": 1815 }, { "epoch": 0.8228364295423652, "grad_norm": 0.5229225030743664, "learning_rate": 7.898966134272739e-05, "loss": 1.0135, "step": 1816 }, { "epoch": 0.8232895333031264, "grad_norm": 0.43002717309565414, "learning_rate": 7.898683337012071e-05, "loss": 0.9987, "step": 1817 }, { "epoch": 0.8237426370638876, "grad_norm": 0.34129142286432507, "learning_rate": 7.898400149601286e-05, "loss": 0.9944, "step": 1818 }, { "epoch": 0.8241957408246489, "grad_norm": 0.3446409407657762, "learning_rate": 7.898116572068725e-05, "loss": 0.9865, "step": 1819 }, { "epoch": 0.8246488445854101, "grad_norm": 0.38602309970636617, "learning_rate": 7.897832604442766e-05, "loss": 1.0161, "step": 1820 }, { "epoch": 0.8251019483461712, "grad_norm": 0.33626791508457804, "learning_rate": 7.897548246751826e-05, "loss": 1.0031, "step": 1821 }, { "epoch": 0.8255550521069325, "grad_norm": 0.33312719948616387, "learning_rate": 7.89726349902436e-05, "loss": 1.0117, "step": 1822 }, { "epoch": 0.8260081558676937, "grad_norm": 0.3302647602920906, "learning_rate": 7.896978361288867e-05, "loss": 1.0174, "step": 1823 }, { "epoch": 0.826461259628455, "grad_norm": 0.3378311999025536, "learning_rate": 7.896692833573878e-05, "loss": 1.0011, "step": 1824 }, { "epoch": 0.8269143633892161, "grad_norm": 0.377168016163107, "learning_rate": 7.896406915907968e-05, "loss": 0.988, "step": 1825 }, { "epoch": 0.8273674671499773, "grad_norm": 0.446544325388909, "learning_rate": 7.896120608319751e-05, "loss": 1.0051, "step": 1826 }, { "epoch": 0.8278205709107386, "grad_norm": 0.44856077854180465, "learning_rate": 7.895833910837873e-05, "loss": 1.0009, "step": 1827 }, { "epoch": 0.8282736746714998, "grad_norm": 0.4775678976953656, "learning_rate": 7.89554682349103e-05, "loss": 0.9917, "step": 1828 }, { "epoch": 0.828726778432261, "grad_norm": 0.4893089940086447, "learning_rate": 7.89525934630795e-05, "loss": 0.9932, "step": 1829 }, { "epoch": 0.8291798821930222, "grad_norm": 0.4970484399240108, "learning_rate": 7.894971479317401e-05, "loss": 1.0149, "step": 1830 }, { "epoch": 0.8296329859537834, "grad_norm": 0.4957128288110956, "learning_rate": 7.894683222548189e-05, "loss": 1.011, "step": 1831 }, { "epoch": 0.8300860897145447, "grad_norm": 0.48105052093831796, "learning_rate": 7.894394576029163e-05, "loss": 0.992, "step": 1832 }, { "epoch": 0.8305391934753058, "grad_norm": 0.45429369281584897, "learning_rate": 7.894105539789207e-05, "loss": 1.0084, "step": 1833 }, { "epoch": 0.830992297236067, "grad_norm": 0.4250169518440908, "learning_rate": 7.893816113857246e-05, "loss": 1.007, "step": 1834 }, { "epoch": 0.8314454009968283, "grad_norm": 0.3987223489754311, "learning_rate": 7.893526298262244e-05, "loss": 1.0186, "step": 1835 }, { "epoch": 0.8318985047575895, "grad_norm": 0.37735484104406164, "learning_rate": 7.893236093033203e-05, "loss": 0.993, "step": 1836 }, { "epoch": 0.8323516085183507, "grad_norm": 0.3527359842113097, "learning_rate": 7.892945498199166e-05, "loss": 0.9749, "step": 1837 }, { "epoch": 0.8328047122791119, "grad_norm": 0.3503336052185402, "learning_rate": 7.89265451378921e-05, "loss": 0.9853, "step": 1838 }, { "epoch": 0.8332578160398731, "grad_norm": 0.343984421406376, "learning_rate": 7.892363139832456e-05, "loss": 1.0055, "step": 1839 }, { "epoch": 0.8337109198006344, "grad_norm": 0.3586518274665539, "learning_rate": 7.892071376358065e-05, "loss": 0.9908, "step": 1840 }, { "epoch": 0.8341640235613955, "grad_norm": 0.3764062091758345, "learning_rate": 7.891779223395231e-05, "loss": 1.019, "step": 1841 }, { "epoch": 0.8346171273221568, "grad_norm": 0.3379532929196838, "learning_rate": 7.891486680973191e-05, "loss": 1.0207, "step": 1842 }, { "epoch": 0.835070231082918, "grad_norm": 0.38728446007297423, "learning_rate": 7.891193749121224e-05, "loss": 0.9943, "step": 1843 }, { "epoch": 0.8355233348436792, "grad_norm": 0.46519838324847823, "learning_rate": 7.890900427868639e-05, "loss": 1.0031, "step": 1844 }, { "epoch": 0.8359764386044404, "grad_norm": 0.5300077294644543, "learning_rate": 7.890606717244794e-05, "loss": 1.008, "step": 1845 }, { "epoch": 0.8364295423652016, "grad_norm": 0.5794041650836942, "learning_rate": 7.890312617279078e-05, "loss": 1.0069, "step": 1846 }, { "epoch": 0.8368826461259629, "grad_norm": 0.6403996113673124, "learning_rate": 7.890018128000924e-05, "loss": 0.9857, "step": 1847 }, { "epoch": 0.8373357498867241, "grad_norm": 0.6852135136708155, "learning_rate": 7.8897232494398e-05, "loss": 0.9897, "step": 1848 }, { "epoch": 0.8377888536474852, "grad_norm": 0.6740580985045481, "learning_rate": 7.889427981625218e-05, "loss": 1.0029, "step": 1849 }, { "epoch": 0.8382419574082465, "grad_norm": 0.6879820311639377, "learning_rate": 7.889132324586725e-05, "loss": 1.0005, "step": 1850 }, { "epoch": 0.8386950611690077, "grad_norm": 0.7619720765368011, "learning_rate": 7.888836278353908e-05, "loss": 1.0093, "step": 1851 }, { "epoch": 0.839148164929769, "grad_norm": 0.7621526962426177, "learning_rate": 7.888539842956393e-05, "loss": 1.0025, "step": 1852 }, { "epoch": 0.8396012686905301, "grad_norm": 0.6813186187241086, "learning_rate": 7.888243018423845e-05, "loss": 0.9997, "step": 1853 }, { "epoch": 0.8400543724512913, "grad_norm": 0.5885961887874978, "learning_rate": 7.887945804785967e-05, "loss": 1.0344, "step": 1854 }, { "epoch": 0.8405074762120526, "grad_norm": 0.528169232716149, "learning_rate": 7.887648202072503e-05, "loss": 1.0089, "step": 1855 }, { "epoch": 0.8409605799728138, "grad_norm": 0.4990009012206636, "learning_rate": 7.887350210313236e-05, "loss": 1.0118, "step": 1856 }, { "epoch": 0.841413683733575, "grad_norm": 0.5179005915133846, "learning_rate": 7.887051829537984e-05, "loss": 1.0093, "step": 1857 }, { "epoch": 0.8418667874943362, "grad_norm": 0.525951129656698, "learning_rate": 7.886753059776608e-05, "loss": 0.9761, "step": 1858 }, { "epoch": 0.8423198912550974, "grad_norm": 0.5614166950514986, "learning_rate": 7.886453901059008e-05, "loss": 0.9787, "step": 1859 }, { "epoch": 0.8427729950158587, "grad_norm": 0.6006701146822967, "learning_rate": 7.886154353415118e-05, "loss": 1.0098, "step": 1860 }, { "epoch": 0.8432260987766198, "grad_norm": 0.5774729909693486, "learning_rate": 7.885854416874916e-05, "loss": 0.9849, "step": 1861 }, { "epoch": 0.843679202537381, "grad_norm": 0.6262857399400916, "learning_rate": 7.885554091468418e-05, "loss": 0.992, "step": 1862 }, { "epoch": 0.8441323062981423, "grad_norm": 0.7282629382210395, "learning_rate": 7.885253377225679e-05, "loss": 0.9791, "step": 1863 }, { "epoch": 0.8445854100589035, "grad_norm": 0.7690208858694954, "learning_rate": 7.884952274176791e-05, "loss": 1.0306, "step": 1864 }, { "epoch": 0.8450385138196647, "grad_norm": 0.6816015156923737, "learning_rate": 7.884650782351886e-05, "loss": 0.9995, "step": 1865 }, { "epoch": 0.8454916175804259, "grad_norm": 0.5296382085491577, "learning_rate": 7.884348901781134e-05, "loss": 0.998, "step": 1866 }, { "epoch": 0.8459447213411871, "grad_norm": 0.475146529418782, "learning_rate": 7.884046632494747e-05, "loss": 0.9774, "step": 1867 }, { "epoch": 0.8463978251019484, "grad_norm": 0.4899880591505279, "learning_rate": 7.883743974522972e-05, "loss": 0.9956, "step": 1868 }, { "epoch": 0.8468509288627095, "grad_norm": 0.522780708755454, "learning_rate": 7.883440927896098e-05, "loss": 0.9893, "step": 1869 }, { "epoch": 0.8473040326234708, "grad_norm": 0.5155549394232641, "learning_rate": 7.883137492644452e-05, "loss": 1.0054, "step": 1870 }, { "epoch": 0.847757136384232, "grad_norm": 0.4622275618812002, "learning_rate": 7.882833668798397e-05, "loss": 1.0115, "step": 1871 }, { "epoch": 0.8482102401449932, "grad_norm": 0.48337712888865764, "learning_rate": 7.88252945638834e-05, "loss": 1.0172, "step": 1872 }, { "epoch": 0.8486633439057544, "grad_norm": 0.4476415366529871, "learning_rate": 7.882224855444723e-05, "loss": 0.9799, "step": 1873 }, { "epoch": 0.8491164476665156, "grad_norm": 1.6998771090498546, "learning_rate": 7.881919865998026e-05, "loss": 1.0438, "step": 1874 }, { "epoch": 0.8495695514272769, "grad_norm": 0.42601389072540025, "learning_rate": 7.881614488078775e-05, "loss": 1.0235, "step": 1875 }, { "epoch": 0.8500226551880381, "grad_norm": 0.6951386116013771, "learning_rate": 7.881308721717525e-05, "loss": 1.0196, "step": 1876 }, { "epoch": 0.8504757589487992, "grad_norm": 0.5925834987604119, "learning_rate": 7.881002566944878e-05, "loss": 0.9918, "step": 1877 }, { "epoch": 0.8509288627095605, "grad_norm": 0.5309925812222615, "learning_rate": 7.88069602379147e-05, "loss": 1.0, "step": 1878 }, { "epoch": 0.8513819664703217, "grad_norm": 0.5379729772217, "learning_rate": 7.880389092287977e-05, "loss": 0.9885, "step": 1879 }, { "epoch": 0.851835070231083, "grad_norm": 0.44157522692622925, "learning_rate": 7.880081772465117e-05, "loss": 0.9872, "step": 1880 }, { "epoch": 0.8522881739918441, "grad_norm": 0.5275620301551198, "learning_rate": 7.879774064353641e-05, "loss": 1.001, "step": 1881 }, { "epoch": 0.8527412777526053, "grad_norm": 0.6246611955805937, "learning_rate": 7.879465967984343e-05, "loss": 1.0006, "step": 1882 }, { "epoch": 0.8531943815133666, "grad_norm": 0.5576197021239777, "learning_rate": 7.879157483388055e-05, "loss": 0.9912, "step": 1883 }, { "epoch": 0.8536474852741278, "grad_norm": 0.4572849449994135, "learning_rate": 7.87884861059565e-05, "loss": 0.994, "step": 1884 }, { "epoch": 0.854100589034889, "grad_norm": 0.4241884921416474, "learning_rate": 7.878539349638035e-05, "loss": 0.995, "step": 1885 }, { "epoch": 0.8545536927956502, "grad_norm": 0.44845347019955467, "learning_rate": 7.878229700546159e-05, "loss": 1.0055, "step": 1886 }, { "epoch": 0.8550067965564114, "grad_norm": 0.4660937900684248, "learning_rate": 7.877919663351009e-05, "loss": 1.0121, "step": 1887 }, { "epoch": 0.8554599003171727, "grad_norm": 0.5249749537215752, "learning_rate": 7.877609238083611e-05, "loss": 1.0099, "step": 1888 }, { "epoch": 0.8559130040779338, "grad_norm": 0.5181899724956033, "learning_rate": 7.877298424775032e-05, "loss": 1.0172, "step": 1889 }, { "epoch": 0.856366107838695, "grad_norm": 0.4438507775230076, "learning_rate": 7.876987223456371e-05, "loss": 1.0157, "step": 1890 }, { "epoch": 0.8568192115994563, "grad_norm": 0.5208323359078278, "learning_rate": 7.876675634158779e-05, "loss": 1.0171, "step": 1891 }, { "epoch": 0.8572723153602175, "grad_norm": 0.6174994756315758, "learning_rate": 7.876363656913428e-05, "loss": 0.9941, "step": 1892 }, { "epoch": 0.8577254191209787, "grad_norm": 0.6873270403633773, "learning_rate": 7.876051291751545e-05, "loss": 1.0193, "step": 1893 }, { "epoch": 0.8581785228817399, "grad_norm": 0.682704138710227, "learning_rate": 7.875738538704387e-05, "loss": 0.9872, "step": 1894 }, { "epoch": 0.8586316266425011, "grad_norm": 0.6881588784297674, "learning_rate": 7.87542539780325e-05, "loss": 1.018, "step": 1895 }, { "epoch": 0.8590847304032624, "grad_norm": 0.7180466043341026, "learning_rate": 7.875111869079472e-05, "loss": 1.0092, "step": 1896 }, { "epoch": 0.8595378341640235, "grad_norm": 0.6447304695972348, "learning_rate": 7.874797952564431e-05, "loss": 0.9938, "step": 1897 }, { "epoch": 0.8599909379247848, "grad_norm": 0.5520551905870082, "learning_rate": 7.874483648289538e-05, "loss": 0.9843, "step": 1898 }, { "epoch": 0.860444041685546, "grad_norm": 0.39277017260239455, "learning_rate": 7.874168956286248e-05, "loss": 0.9989, "step": 1899 }, { "epoch": 0.8608971454463072, "grad_norm": 0.2827975420084032, "learning_rate": 7.873853876586052e-05, "loss": 1.0155, "step": 1900 }, { "epoch": 0.8613502492070684, "grad_norm": 0.3510786743290551, "learning_rate": 7.873538409220482e-05, "loss": 0.9956, "step": 1901 }, { "epoch": 0.8618033529678296, "grad_norm": 0.5250044613128058, "learning_rate": 7.873222554221105e-05, "loss": 0.987, "step": 1902 }, { "epoch": 0.8622564567285909, "grad_norm": 0.6697692341061444, "learning_rate": 7.87290631161953e-05, "loss": 0.9819, "step": 1903 }, { "epoch": 0.8627095604893521, "grad_norm": 0.8375931462932719, "learning_rate": 7.872589681447407e-05, "loss": 1.0193, "step": 1904 }, { "epoch": 0.8631626642501132, "grad_norm": 0.9086326491292269, "learning_rate": 7.87227266373642e-05, "loss": 0.9901, "step": 1905 }, { "epoch": 0.8636157680108745, "grad_norm": 0.8947834475838599, "learning_rate": 7.871955258518294e-05, "loss": 1.002, "step": 1906 }, { "epoch": 0.8640688717716357, "grad_norm": 0.8096962914147138, "learning_rate": 7.87163746582479e-05, "loss": 0.9991, "step": 1907 }, { "epoch": 0.864521975532397, "grad_norm": 0.6643973289845967, "learning_rate": 7.871319285687715e-05, "loss": 0.9957, "step": 1908 }, { "epoch": 0.8649750792931581, "grad_norm": 0.5228492896982254, "learning_rate": 7.871000718138905e-05, "loss": 1.0094, "step": 1909 }, { "epoch": 0.8654281830539193, "grad_norm": 0.3670920447901226, "learning_rate": 7.870681763210243e-05, "loss": 1.0133, "step": 1910 }, { "epoch": 0.8658812868146806, "grad_norm": 0.27463672800027933, "learning_rate": 7.870362420933648e-05, "loss": 1.0232, "step": 1911 }, { "epoch": 0.8663343905754418, "grad_norm": 0.3296668804292145, "learning_rate": 7.870042691341074e-05, "loss": 0.9967, "step": 1912 }, { "epoch": 0.866787494336203, "grad_norm": 0.3942364727847778, "learning_rate": 7.869722574464521e-05, "loss": 0.9913, "step": 1913 }, { "epoch": 0.8672405980969642, "grad_norm": 0.3985898855305173, "learning_rate": 7.86940207033602e-05, "loss": 0.9904, "step": 1914 }, { "epoch": 0.8676937018577254, "grad_norm": 0.39013527615088806, "learning_rate": 7.869081178987647e-05, "loss": 0.9829, "step": 1915 }, { "epoch": 0.8681468056184867, "grad_norm": 0.4080385746794164, "learning_rate": 7.868759900451514e-05, "loss": 1.0145, "step": 1916 }, { "epoch": 0.8685999093792478, "grad_norm": 0.39018918290649174, "learning_rate": 7.868438234759772e-05, "loss": 0.9926, "step": 1917 }, { "epoch": 0.869053013140009, "grad_norm": 0.36304547934425085, "learning_rate": 7.86811618194461e-05, "loss": 1.005, "step": 1918 }, { "epoch": 0.8695061169007703, "grad_norm": 0.35360879880270507, "learning_rate": 7.867793742038258e-05, "loss": 0.999, "step": 1919 }, { "epoch": 0.8699592206615315, "grad_norm": 0.33300428083486466, "learning_rate": 7.867470915072982e-05, "loss": 0.9942, "step": 1920 }, { "epoch": 0.8704123244222927, "grad_norm": 0.3224925457224455, "learning_rate": 7.867147701081088e-05, "loss": 0.9957, "step": 1921 }, { "epoch": 0.8708654281830539, "grad_norm": 0.3129564597452825, "learning_rate": 7.866824100094923e-05, "loss": 1.0028, "step": 1922 }, { "epoch": 0.8713185319438151, "grad_norm": 0.3104609279782206, "learning_rate": 7.866500112146867e-05, "loss": 1.0095, "step": 1923 }, { "epoch": 0.8717716357045764, "grad_norm": 0.33855440359288047, "learning_rate": 7.866175737269345e-05, "loss": 1.0225, "step": 1924 }, { "epoch": 0.8722247394653375, "grad_norm": 0.35979077664446263, "learning_rate": 7.865850975494818e-05, "loss": 0.9911, "step": 1925 }, { "epoch": 0.8726778432260988, "grad_norm": 0.40369598419147606, "learning_rate": 7.865525826855782e-05, "loss": 0.9999, "step": 1926 }, { "epoch": 0.87313094698686, "grad_norm": 0.4584402621258206, "learning_rate": 7.86520029138478e-05, "loss": 0.9957, "step": 1927 }, { "epoch": 0.8735840507476212, "grad_norm": 0.9625828601593942, "learning_rate": 7.864874369114386e-05, "loss": 1.0281, "step": 1928 }, { "epoch": 0.8740371545083824, "grad_norm": 0.6464473530297169, "learning_rate": 7.864548060077218e-05, "loss": 0.9892, "step": 1929 }, { "epoch": 0.8744902582691436, "grad_norm": 0.7597117377269464, "learning_rate": 7.86422136430593e-05, "loss": 1.0052, "step": 1930 }, { "epoch": 0.8749433620299049, "grad_norm": 0.8892977326897675, "learning_rate": 7.863894281833212e-05, "loss": 0.9929, "step": 1931 }, { "epoch": 0.8753964657906661, "grad_norm": 0.9029557725353057, "learning_rate": 7.863566812691799e-05, "loss": 1.0053, "step": 1932 }, { "epoch": 0.8758495695514272, "grad_norm": 0.7467490406636332, "learning_rate": 7.863238956914462e-05, "loss": 1.0007, "step": 1933 }, { "epoch": 0.8763026733121885, "grad_norm": 0.5474482966161008, "learning_rate": 7.862910714534008e-05, "loss": 1.0016, "step": 1934 }, { "epoch": 0.8767557770729497, "grad_norm": 0.5102883689840326, "learning_rate": 7.862582085583286e-05, "loss": 1.0099, "step": 1935 }, { "epoch": 0.877208880833711, "grad_norm": 0.6093885353144222, "learning_rate": 7.862253070095185e-05, "loss": 1.0156, "step": 1936 }, { "epoch": 0.8776619845944721, "grad_norm": 1.1216571637116286, "learning_rate": 7.861923668102626e-05, "loss": 0.9954, "step": 1937 }, { "epoch": 0.8781150883552333, "grad_norm": 0.33760421917878375, "learning_rate": 7.861593879638577e-05, "loss": 1.01, "step": 1938 }, { "epoch": 0.8785681921159946, "grad_norm": 0.3787378450804937, "learning_rate": 7.861263704736038e-05, "loss": 1.0015, "step": 1939 }, { "epoch": 0.8790212958767558, "grad_norm": 0.47716610127745057, "learning_rate": 7.860933143428051e-05, "loss": 1.0066, "step": 1940 }, { "epoch": 0.879474399637517, "grad_norm": 0.5012408093777181, "learning_rate": 7.860602195747696e-05, "loss": 0.9871, "step": 1941 }, { "epoch": 0.8799275033982782, "grad_norm": 0.4342757498139315, "learning_rate": 7.860270861728092e-05, "loss": 0.99, "step": 1942 }, { "epoch": 0.8803806071590394, "grad_norm": 0.5341186577727081, "learning_rate": 7.859939141402397e-05, "loss": 1.0116, "step": 1943 }, { "epoch": 0.8808337109198007, "grad_norm": 0.6977314317097666, "learning_rate": 7.859607034803805e-05, "loss": 1.0028, "step": 1944 }, { "epoch": 0.8812868146805618, "grad_norm": 0.7872879661412624, "learning_rate": 7.859274541965554e-05, "loss": 0.9926, "step": 1945 }, { "epoch": 0.881739918441323, "grad_norm": 0.7976109331257298, "learning_rate": 7.858941662920914e-05, "loss": 0.9928, "step": 1946 }, { "epoch": 0.8821930222020843, "grad_norm": 0.7310686913106118, "learning_rate": 7.858608397703198e-05, "loss": 1.0068, "step": 1947 }, { "epoch": 0.8826461259628455, "grad_norm": 0.6231944846062322, "learning_rate": 7.858274746345757e-05, "loss": 0.9951, "step": 1948 }, { "epoch": 0.8830992297236067, "grad_norm": 0.5336565134135821, "learning_rate": 7.85794070888198e-05, "loss": 1.0068, "step": 1949 }, { "epoch": 0.8835523334843679, "grad_norm": 0.4249967706854565, "learning_rate": 7.857606285345295e-05, "loss": 0.9869, "step": 1950 }, { "epoch": 0.8840054372451291, "grad_norm": 0.3736033343864738, "learning_rate": 7.857271475769169e-05, "loss": 0.9976, "step": 1951 }, { "epoch": 0.8844585410058904, "grad_norm": 0.4776658270657853, "learning_rate": 7.856936280187108e-05, "loss": 1.0071, "step": 1952 }, { "epoch": 0.8849116447666515, "grad_norm": 0.5086522225284822, "learning_rate": 7.856600698632651e-05, "loss": 1.0098, "step": 1953 }, { "epoch": 0.8853647485274128, "grad_norm": 0.4758581330434052, "learning_rate": 7.856264731139387e-05, "loss": 0.9801, "step": 1954 }, { "epoch": 0.885817852288174, "grad_norm": 0.4796922216226913, "learning_rate": 7.855928377740932e-05, "loss": 1.0045, "step": 1955 }, { "epoch": 0.8862709560489352, "grad_norm": 0.5160547945643172, "learning_rate": 7.855591638470948e-05, "loss": 1.008, "step": 1956 }, { "epoch": 0.8867240598096964, "grad_norm": 0.4316922013335362, "learning_rate": 7.855254513363134e-05, "loss": 0.9927, "step": 1957 }, { "epoch": 0.8871771635704576, "grad_norm": 0.3083503962305714, "learning_rate": 7.854917002451225e-05, "loss": 1.0125, "step": 1958 }, { "epoch": 0.8876302673312189, "grad_norm": 0.3337257681335021, "learning_rate": 7.854579105768995e-05, "loss": 0.9813, "step": 1959 }, { "epoch": 0.8880833710919801, "grad_norm": 0.4257864159106086, "learning_rate": 7.854240823350262e-05, "loss": 1.0031, "step": 1960 }, { "epoch": 0.8885364748527412, "grad_norm": 0.47493197687602834, "learning_rate": 7.853902155228878e-05, "loss": 0.9698, "step": 1961 }, { "epoch": 0.8889895786135025, "grad_norm": 0.5070107409414635, "learning_rate": 7.853563101438733e-05, "loss": 1.0009, "step": 1962 }, { "epoch": 0.8894426823742637, "grad_norm": 0.5034404563248486, "learning_rate": 7.853223662013756e-05, "loss": 1.0203, "step": 1963 }, { "epoch": 0.889895786135025, "grad_norm": 0.5302093406980669, "learning_rate": 7.852883836987917e-05, "loss": 0.9885, "step": 1964 }, { "epoch": 0.8903488898957861, "grad_norm": 0.5021806633729046, "learning_rate": 7.852543626395222e-05, "loss": 1.032, "step": 1965 }, { "epoch": 0.8908019936565473, "grad_norm": 0.4240519125365953, "learning_rate": 7.852203030269718e-05, "loss": 0.9935, "step": 1966 }, { "epoch": 0.8912550974173086, "grad_norm": 0.3779073611651617, "learning_rate": 7.851862048645488e-05, "loss": 0.9904, "step": 1967 }, { "epoch": 0.8917082011780698, "grad_norm": 0.5167112224203235, "learning_rate": 7.851520681556656e-05, "loss": 1.0086, "step": 1968 }, { "epoch": 0.892161304938831, "grad_norm": 0.6187612379497455, "learning_rate": 7.851178929037383e-05, "loss": 1.0207, "step": 1969 }, { "epoch": 0.8926144086995922, "grad_norm": 0.5871527458850951, "learning_rate": 7.850836791121869e-05, "loss": 0.9939, "step": 1970 }, { "epoch": 0.8930675124603534, "grad_norm": 0.5617467406554079, "learning_rate": 7.85049426784435e-05, "loss": 0.9942, "step": 1971 }, { "epoch": 0.8935206162211147, "grad_norm": 0.4758218659844952, "learning_rate": 7.850151359239107e-05, "loss": 1.008, "step": 1972 }, { "epoch": 0.8939737199818758, "grad_norm": 0.44938859362087796, "learning_rate": 7.849808065340453e-05, "loss": 0.9803, "step": 1973 }, { "epoch": 0.894426823742637, "grad_norm": 0.45564163691419096, "learning_rate": 7.849464386182744e-05, "loss": 1.0047, "step": 1974 }, { "epoch": 0.8948799275033983, "grad_norm": 0.521847377910822, "learning_rate": 7.849120321800372e-05, "loss": 0.9864, "step": 1975 }, { "epoch": 0.8953330312641595, "grad_norm": 0.5697407043340399, "learning_rate": 7.848775872227768e-05, "loss": 1.0032, "step": 1976 }, { "epoch": 0.8957861350249207, "grad_norm": 0.553080925014953, "learning_rate": 7.848431037499402e-05, "loss": 0.9902, "step": 1977 }, { "epoch": 0.8962392387856819, "grad_norm": 0.5478861370546282, "learning_rate": 7.848085817649782e-05, "loss": 1.0351, "step": 1978 }, { "epoch": 0.8966923425464431, "grad_norm": 0.5227046266152102, "learning_rate": 7.847740212713456e-05, "loss": 1.0088, "step": 1979 }, { "epoch": 0.8971454463072044, "grad_norm": 0.49963690784380965, "learning_rate": 7.847394222725009e-05, "loss": 0.9959, "step": 1980 }, { "epoch": 0.8975985500679655, "grad_norm": 0.44391879244418975, "learning_rate": 7.847047847719063e-05, "loss": 0.9857, "step": 1981 }, { "epoch": 0.8980516538287268, "grad_norm": 0.34831822115497924, "learning_rate": 7.846701087730285e-05, "loss": 1.0075, "step": 1982 }, { "epoch": 0.898504757589488, "grad_norm": 0.438365852831001, "learning_rate": 7.846353942793372e-05, "loss": 1.0114, "step": 1983 }, { "epoch": 0.8989578613502492, "grad_norm": 0.5785317495345688, "learning_rate": 7.846006412943065e-05, "loss": 1.0234, "step": 1984 }, { "epoch": 0.8994109651110104, "grad_norm": 0.7037360190248301, "learning_rate": 7.845658498214143e-05, "loss": 0.9856, "step": 1985 }, { "epoch": 0.8998640688717716, "grad_norm": 0.8572360417353982, "learning_rate": 7.84531019864142e-05, "loss": 1.0135, "step": 1986 }, { "epoch": 0.9003171726325329, "grad_norm": 0.881839561402847, "learning_rate": 7.844961514259755e-05, "loss": 0.9906, "step": 1987 }, { "epoch": 0.9007702763932941, "grad_norm": 0.8497708258886764, "learning_rate": 7.844612445104039e-05, "loss": 0.9617, "step": 1988 }, { "epoch": 0.9012233801540552, "grad_norm": 0.8278029906178168, "learning_rate": 7.844262991209204e-05, "loss": 0.9764, "step": 1989 }, { "epoch": 0.9016764839148165, "grad_norm": 0.749071517145521, "learning_rate": 7.843913152610222e-05, "loss": 1.0001, "step": 1990 }, { "epoch": 0.9021295876755777, "grad_norm": 0.617594606973515, "learning_rate": 7.8435629293421e-05, "loss": 0.9805, "step": 1991 }, { "epoch": 0.902582691436339, "grad_norm": 0.46510046010633405, "learning_rate": 7.843212321439889e-05, "loss": 0.9824, "step": 1992 }, { "epoch": 0.9030357951971001, "grad_norm": 0.3847908548841719, "learning_rate": 7.842861328938671e-05, "loss": 1.0012, "step": 1993 }, { "epoch": 0.9034888989578613, "grad_norm": 0.45555509833612634, "learning_rate": 7.842509951873576e-05, "loss": 1.0013, "step": 1994 }, { "epoch": 0.9039420027186226, "grad_norm": 0.579152029070868, "learning_rate": 7.842158190279762e-05, "loss": 1.0187, "step": 1995 }, { "epoch": 0.9043951064793838, "grad_norm": 0.6604977316250528, "learning_rate": 7.841806044192433e-05, "loss": 1.0104, "step": 1996 }, { "epoch": 0.904848210240145, "grad_norm": 0.6717895133695969, "learning_rate": 7.841453513646829e-05, "loss": 0.9856, "step": 1997 }, { "epoch": 0.9053013140009062, "grad_norm": 0.6390298935952569, "learning_rate": 7.841100598678227e-05, "loss": 0.9864, "step": 1998 }, { "epoch": 0.9057544177616674, "grad_norm": 0.5879158866193978, "learning_rate": 7.840747299321946e-05, "loss": 0.9896, "step": 1999 }, { "epoch": 0.9062075215224287, "grad_norm": 0.5060514907768945, "learning_rate": 7.84039361561334e-05, "loss": 1.0006, "step": 2000 }, { "epoch": 0.9066606252831898, "grad_norm": 0.45771478037576113, "learning_rate": 7.840039547587803e-05, "loss": 0.9843, "step": 2001 }, { "epoch": 0.907113729043951, "grad_norm": 0.4102666152880805, "learning_rate": 7.839685095280769e-05, "loss": 1.0031, "step": 2002 }, { "epoch": 0.9075668328047123, "grad_norm": 0.37444644133711846, "learning_rate": 7.839330258727707e-05, "loss": 0.9993, "step": 2003 }, { "epoch": 0.9080199365654735, "grad_norm": 0.3574929331145775, "learning_rate": 7.838975037964127e-05, "loss": 0.9892, "step": 2004 }, { "epoch": 0.9084730403262347, "grad_norm": 0.3544195075149952, "learning_rate": 7.838619433025578e-05, "loss": 0.9764, "step": 2005 }, { "epoch": 0.9089261440869959, "grad_norm": 0.3895834115068427, "learning_rate": 7.838263443947646e-05, "loss": 1.0079, "step": 2006 }, { "epoch": 0.9093792478477571, "grad_norm": 0.4141839609374459, "learning_rate": 7.837907070765951e-05, "loss": 1.019, "step": 2007 }, { "epoch": 0.9098323516085184, "grad_norm": 0.3711755560686412, "learning_rate": 7.837550313516163e-05, "loss": 1.0151, "step": 2008 }, { "epoch": 0.9102854553692795, "grad_norm": 0.3351984948943602, "learning_rate": 7.83719317223398e-05, "loss": 0.9963, "step": 2009 }, { "epoch": 0.9107385591300408, "grad_norm": 0.42377379460925324, "learning_rate": 7.83683564695514e-05, "loss": 1.0046, "step": 2010 }, { "epoch": 0.911191662890802, "grad_norm": 0.3148260830008473, "learning_rate": 7.836477737715426e-05, "loss": 0.9837, "step": 2011 }, { "epoch": 0.9116447666515632, "grad_norm": 0.3011266185565036, "learning_rate": 7.836119444550651e-05, "loss": 0.994, "step": 2012 }, { "epoch": 0.9120978704123244, "grad_norm": 0.3102893983640567, "learning_rate": 7.835760767496672e-05, "loss": 0.9912, "step": 2013 }, { "epoch": 0.9125509741730856, "grad_norm": 0.31516994425292594, "learning_rate": 7.835401706589381e-05, "loss": 0.9757, "step": 2014 }, { "epoch": 0.9130040779338469, "grad_norm": 0.3321506027102069, "learning_rate": 7.835042261864713e-05, "loss": 1.0054, "step": 2015 }, { "epoch": 0.9134571816946081, "grad_norm": 0.35386601868951545, "learning_rate": 7.834682433358636e-05, "loss": 0.9891, "step": 2016 }, { "epoch": 0.9139102854553692, "grad_norm": 0.37968490155522017, "learning_rate": 7.83432222110716e-05, "loss": 0.9912, "step": 2017 }, { "epoch": 0.9143633892161305, "grad_norm": 0.33345273442695483, "learning_rate": 7.833961625146331e-05, "loss": 0.9987, "step": 2018 }, { "epoch": 0.9148164929768917, "grad_norm": 0.2845394971669567, "learning_rate": 7.833600645512236e-05, "loss": 1.0082, "step": 2019 }, { "epoch": 0.915269596737653, "grad_norm": 0.46610072709256267, "learning_rate": 7.833239282240998e-05, "loss": 0.9759, "step": 2020 }, { "epoch": 0.9157227004984141, "grad_norm": 0.6640961636484622, "learning_rate": 7.832877535368782e-05, "loss": 1.0098, "step": 2021 }, { "epoch": 0.9161758042591753, "grad_norm": 0.7587460516165241, "learning_rate": 7.832515404931784e-05, "loss": 0.988, "step": 2022 }, { "epoch": 0.9166289080199366, "grad_norm": 0.7028442106607444, "learning_rate": 7.832152890966246e-05, "loss": 0.9736, "step": 2023 }, { "epoch": 0.9170820117806978, "grad_norm": 0.6489474589654937, "learning_rate": 7.831789993508448e-05, "loss": 1.0158, "step": 2024 }, { "epoch": 0.917535115541459, "grad_norm": 0.5890829027655057, "learning_rate": 7.831426712594702e-05, "loss": 0.999, "step": 2025 }, { "epoch": 0.9179882193022202, "grad_norm": 0.48183497116956625, "learning_rate": 7.831063048261364e-05, "loss": 0.9802, "step": 2026 }, { "epoch": 0.9184413230629814, "grad_norm": 0.3655968396322037, "learning_rate": 7.830699000544826e-05, "loss": 1.0104, "step": 2027 }, { "epoch": 0.9188944268237427, "grad_norm": 0.29198312137595495, "learning_rate": 7.830334569481521e-05, "loss": 0.9664, "step": 2028 }, { "epoch": 0.9193475305845038, "grad_norm": 0.3141085815753943, "learning_rate": 7.829969755107917e-05, "loss": 1.001, "step": 2029 }, { "epoch": 0.919800634345265, "grad_norm": 0.37084135443990174, "learning_rate": 7.829604557460523e-05, "loss": 0.9843, "step": 2030 }, { "epoch": 0.9202537381060263, "grad_norm": 0.49817263382139615, "learning_rate": 7.829238976575883e-05, "loss": 1.0081, "step": 2031 }, { "epoch": 0.9207068418667875, "grad_norm": 0.6129588936488894, "learning_rate": 7.828873012490581e-05, "loss": 1.0019, "step": 2032 }, { "epoch": 0.9211599456275487, "grad_norm": 0.6851031451716078, "learning_rate": 7.828506665241244e-05, "loss": 1.0065, "step": 2033 }, { "epoch": 0.9216130493883099, "grad_norm": 0.711280305619024, "learning_rate": 7.82813993486453e-05, "loss": 1.0315, "step": 2034 }, { "epoch": 0.9220661531490711, "grad_norm": 0.7170363142225513, "learning_rate": 7.827772821397139e-05, "loss": 1.016, "step": 2035 }, { "epoch": 0.9225192569098324, "grad_norm": 0.6368101863216217, "learning_rate": 7.82740532487581e-05, "loss": 1.0105, "step": 2036 }, { "epoch": 0.9229723606705935, "grad_norm": 0.5374633716615991, "learning_rate": 7.827037445337317e-05, "loss": 1.0068, "step": 2037 }, { "epoch": 0.9234254644313548, "grad_norm": 0.45090767136941956, "learning_rate": 7.826669182818476e-05, "loss": 0.9972, "step": 2038 }, { "epoch": 0.923878568192116, "grad_norm": 0.37135053959910536, "learning_rate": 7.826300537356141e-05, "loss": 0.9887, "step": 2039 }, { "epoch": 0.9243316719528772, "grad_norm": 0.37841922480272205, "learning_rate": 7.825931508987202e-05, "loss": 0.9979, "step": 2040 }, { "epoch": 0.9247847757136384, "grad_norm": 0.3755961952833791, "learning_rate": 7.825562097748588e-05, "loss": 0.9865, "step": 2041 }, { "epoch": 0.9252378794743996, "grad_norm": 0.35384915208488543, "learning_rate": 7.825192303677266e-05, "loss": 0.9845, "step": 2042 }, { "epoch": 0.9256909832351609, "grad_norm": 0.3622754850648786, "learning_rate": 7.824822126810245e-05, "loss": 0.9761, "step": 2043 }, { "epoch": 0.9261440869959221, "grad_norm": 0.3765108049262093, "learning_rate": 7.824451567184567e-05, "loss": 0.9914, "step": 2044 }, { "epoch": 0.9265971907566832, "grad_norm": 0.3923515451486628, "learning_rate": 7.824080624837317e-05, "loss": 0.9953, "step": 2045 }, { "epoch": 0.9270502945174445, "grad_norm": 0.422955704662956, "learning_rate": 7.823709299805613e-05, "loss": 0.9726, "step": 2046 }, { "epoch": 0.9275033982782057, "grad_norm": 0.5644346142307348, "learning_rate": 7.823337592126617e-05, "loss": 0.9985, "step": 2047 }, { "epoch": 0.927956502038967, "grad_norm": 0.7545002253020594, "learning_rate": 7.822965501837525e-05, "loss": 0.9717, "step": 2048 }, { "epoch": 0.9284096057997281, "grad_norm": 0.913103532150095, "learning_rate": 7.822593028975574e-05, "loss": 0.9832, "step": 2049 }, { "epoch": 0.9288627095604893, "grad_norm": 0.9679708701127155, "learning_rate": 7.822220173578038e-05, "loss": 0.9863, "step": 2050 }, { "epoch": 0.9293158133212506, "grad_norm": 0.8497764014316461, "learning_rate": 7.821846935682228e-05, "loss": 1.0057, "step": 2051 }, { "epoch": 0.9297689170820118, "grad_norm": 0.6855155645305441, "learning_rate": 7.821473315325498e-05, "loss": 0.9999, "step": 2052 }, { "epoch": 0.930222020842773, "grad_norm": 0.6365933339637798, "learning_rate": 7.821099312545233e-05, "loss": 0.9951, "step": 2053 }, { "epoch": 0.9306751246035342, "grad_norm": 0.6571024079360772, "learning_rate": 7.820724927378864e-05, "loss": 1.0081, "step": 2054 }, { "epoch": 0.9311282283642954, "grad_norm": 0.5384668596834687, "learning_rate": 7.820350159863854e-05, "loss": 0.9961, "step": 2055 }, { "epoch": 0.9315813321250567, "grad_norm": 0.4250329223293304, "learning_rate": 7.819975010037708e-05, "loss": 0.9975, "step": 2056 }, { "epoch": 0.9320344358858178, "grad_norm": 0.4391801256062363, "learning_rate": 7.819599477937967e-05, "loss": 1.0156, "step": 2057 }, { "epoch": 0.932487539646579, "grad_norm": 0.41252229589877143, "learning_rate": 7.819223563602214e-05, "loss": 0.992, "step": 2058 }, { "epoch": 0.9329406434073403, "grad_norm": 0.3925819723289104, "learning_rate": 7.818847267068065e-05, "loss": 1.0011, "step": 2059 }, { "epoch": 0.9333937471681015, "grad_norm": 0.4103954436407144, "learning_rate": 7.818470588373178e-05, "loss": 0.9594, "step": 2060 }, { "epoch": 0.9338468509288627, "grad_norm": 0.3916539436287158, "learning_rate": 7.818093527555248e-05, "loss": 1.0238, "step": 2061 }, { "epoch": 0.9342999546896239, "grad_norm": 0.3701422015267218, "learning_rate": 7.817716084652007e-05, "loss": 0.9912, "step": 2062 }, { "epoch": 0.9347530584503851, "grad_norm": 0.43232231607187654, "learning_rate": 7.817338259701227e-05, "loss": 0.9976, "step": 2063 }, { "epoch": 0.9352061622111464, "grad_norm": 0.40829557558126833, "learning_rate": 7.816960052740721e-05, "loss": 1.016, "step": 2064 }, { "epoch": 0.9356592659719075, "grad_norm": 0.298843233895758, "learning_rate": 7.816581463808334e-05, "loss": 1.0109, "step": 2065 }, { "epoch": 0.9361123697326688, "grad_norm": 0.30342268781011744, "learning_rate": 7.816202492941952e-05, "loss": 0.9825, "step": 2066 }, { "epoch": 0.93656547349343, "grad_norm": 0.31937538919844555, "learning_rate": 7.8158231401795e-05, "loss": 0.9688, "step": 2067 }, { "epoch": 0.9370185772541912, "grad_norm": 0.3008958685433031, "learning_rate": 7.815443405558942e-05, "loss": 0.9959, "step": 2068 }, { "epoch": 0.9374716810149524, "grad_norm": 0.302961712064914, "learning_rate": 7.815063289118276e-05, "loss": 1.0037, "step": 2069 }, { "epoch": 0.9379247847757136, "grad_norm": 0.3396327406994233, "learning_rate": 7.814682790895544e-05, "loss": 1.0032, "step": 2070 }, { "epoch": 0.9383778885364749, "grad_norm": 0.3036089168082906, "learning_rate": 7.814301910928824e-05, "loss": 1.0175, "step": 2071 }, { "epoch": 0.9388309922972361, "grad_norm": 0.3204587358714023, "learning_rate": 7.813920649256227e-05, "loss": 0.9914, "step": 2072 }, { "epoch": 0.9392840960579972, "grad_norm": 0.44021018604866297, "learning_rate": 7.813539005915911e-05, "loss": 0.9954, "step": 2073 }, { "epoch": 0.9397371998187585, "grad_norm": 0.44361136301317217, "learning_rate": 7.813156980946066e-05, "loss": 0.9833, "step": 2074 }, { "epoch": 0.9401903035795197, "grad_norm": 0.4756889486079688, "learning_rate": 7.812774574384922e-05, "loss": 0.9984, "step": 2075 }, { "epoch": 0.940643407340281, "grad_norm": 0.5521161787991091, "learning_rate": 7.812391786270748e-05, "loss": 0.9688, "step": 2076 }, { "epoch": 0.9410965111010421, "grad_norm": 0.5755923837444415, "learning_rate": 7.812008616641851e-05, "loss": 0.9795, "step": 2077 }, { "epoch": 0.9415496148618033, "grad_norm": 0.5350527493331929, "learning_rate": 7.811625065536576e-05, "loss": 0.9828, "step": 2078 }, { "epoch": 0.9420027186225646, "grad_norm": 0.5201166767579753, "learning_rate": 7.811241132993305e-05, "loss": 0.983, "step": 2079 }, { "epoch": 0.9424558223833258, "grad_norm": 0.5216295095177013, "learning_rate": 7.810856819050457e-05, "loss": 0.9812, "step": 2080 }, { "epoch": 0.942908926144087, "grad_norm": 0.5439823296263803, "learning_rate": 7.810472123746493e-05, "loss": 0.9896, "step": 2081 }, { "epoch": 0.9433620299048482, "grad_norm": 0.5297298385377766, "learning_rate": 7.810087047119911e-05, "loss": 1.0103, "step": 2082 }, { "epoch": 0.9438151336656094, "grad_norm": 0.48977321431781573, "learning_rate": 7.809701589209246e-05, "loss": 1.0014, "step": 2083 }, { "epoch": 0.9442682374263707, "grad_norm": 0.5085775801175768, "learning_rate": 7.809315750053071e-05, "loss": 0.9989, "step": 2084 }, { "epoch": 0.9447213411871318, "grad_norm": 0.5837197054248623, "learning_rate": 7.808929529689999e-05, "loss": 0.9655, "step": 2085 }, { "epoch": 0.945174444947893, "grad_norm": 0.6635867069740374, "learning_rate": 7.808542928158678e-05, "loss": 0.9935, "step": 2086 }, { "epoch": 0.9456275487086543, "grad_norm": 0.7977268451244774, "learning_rate": 7.808155945497799e-05, "loss": 0.964, "step": 2087 }, { "epoch": 0.9460806524694155, "grad_norm": 0.8610750282581193, "learning_rate": 7.807768581746086e-05, "loss": 0.9853, "step": 2088 }, { "epoch": 0.9465337562301767, "grad_norm": 0.8621285948175907, "learning_rate": 7.807380836942303e-05, "loss": 0.99, "step": 2089 }, { "epoch": 0.9469868599909379, "grad_norm": 0.7365563489301574, "learning_rate": 7.806992711125254e-05, "loss": 0.9864, "step": 2090 }, { "epoch": 0.9474399637516991, "grad_norm": 0.5365597822647672, "learning_rate": 7.80660420433378e-05, "loss": 0.9941, "step": 2091 }, { "epoch": 0.9478930675124604, "grad_norm": 0.4083698031415168, "learning_rate": 7.806215316606758e-05, "loss": 0.9885, "step": 2092 }, { "epoch": 0.9483461712732215, "grad_norm": 0.380355751129671, "learning_rate": 7.805826047983105e-05, "loss": 0.9838, "step": 2093 }, { "epoch": 0.9487992750339828, "grad_norm": 0.5110345550848984, "learning_rate": 7.805436398501778e-05, "loss": 1.0033, "step": 2094 }, { "epoch": 0.949252378794744, "grad_norm": 0.6306621147908984, "learning_rate": 7.80504636820177e-05, "loss": 0.9995, "step": 2095 }, { "epoch": 0.9497054825555052, "grad_norm": 0.6442146867403175, "learning_rate": 7.804655957122108e-05, "loss": 1.0067, "step": 2096 }, { "epoch": 0.9501585863162664, "grad_norm": 0.5226923342567326, "learning_rate": 7.804265165301866e-05, "loss": 0.9858, "step": 2097 }, { "epoch": 0.9506116900770276, "grad_norm": 0.3687482190741135, "learning_rate": 7.80387399278015e-05, "loss": 0.9882, "step": 2098 }, { "epoch": 0.9510647938377889, "grad_norm": 0.4242447616689821, "learning_rate": 7.803482439596105e-05, "loss": 0.9878, "step": 2099 }, { "epoch": 0.9515178975985501, "grad_norm": 0.4157911046097607, "learning_rate": 7.803090505788914e-05, "loss": 1.0016, "step": 2100 }, { "epoch": 0.9519710013593112, "grad_norm": 0.3581375669878615, "learning_rate": 7.802698191397801e-05, "loss": 0.9981, "step": 2101 }, { "epoch": 0.9524241051200725, "grad_norm": 0.44192574554839625, "learning_rate": 7.802305496462024e-05, "loss": 0.9883, "step": 2102 }, { "epoch": 0.9528772088808337, "grad_norm": 0.5540542289520467, "learning_rate": 7.801912421020881e-05, "loss": 0.9982, "step": 2103 }, { "epoch": 0.953330312641595, "grad_norm": 0.6155027059016724, "learning_rate": 7.801518965113708e-05, "loss": 0.9854, "step": 2104 }, { "epoch": 0.9537834164023561, "grad_norm": 0.5798910998779804, "learning_rate": 7.80112512877988e-05, "loss": 0.9745, "step": 2105 }, { "epoch": 0.9542365201631173, "grad_norm": 0.5219405667234036, "learning_rate": 7.800730912058808e-05, "loss": 1.0, "step": 2106 }, { "epoch": 0.9546896239238786, "grad_norm": 0.4956557919556331, "learning_rate": 7.800336314989944e-05, "loss": 0.9887, "step": 2107 }, { "epoch": 0.9551427276846398, "grad_norm": 0.4375075961544017, "learning_rate": 7.799941337612773e-05, "loss": 0.9734, "step": 2108 }, { "epoch": 0.955595831445401, "grad_norm": 0.4048801602478982, "learning_rate": 7.799545979966823e-05, "loss": 0.9705, "step": 2109 }, { "epoch": 0.9560489352061622, "grad_norm": 0.38334510663971855, "learning_rate": 7.799150242091659e-05, "loss": 0.9823, "step": 2110 }, { "epoch": 0.9565020389669234, "grad_norm": 0.40210355220153243, "learning_rate": 7.798754124026883e-05, "loss": 0.9903, "step": 2111 }, { "epoch": 0.9569551427276847, "grad_norm": 0.40070907723663945, "learning_rate": 7.798357625812136e-05, "loss": 1.0006, "step": 2112 }, { "epoch": 0.9574082464884458, "grad_norm": 0.42795031086258617, "learning_rate": 7.797960747487095e-05, "loss": 0.989, "step": 2113 }, { "epoch": 0.957861350249207, "grad_norm": 0.4669370867666925, "learning_rate": 7.797563489091477e-05, "loss": 1.0079, "step": 2114 }, { "epoch": 0.9583144540099683, "grad_norm": 0.470333675265646, "learning_rate": 7.797165850665036e-05, "loss": 0.981, "step": 2115 }, { "epoch": 0.9587675577707295, "grad_norm": 0.43288387979303256, "learning_rate": 7.796767832247567e-05, "loss": 0.9938, "step": 2116 }, { "epoch": 0.9592206615314907, "grad_norm": 0.4017591922243532, "learning_rate": 7.796369433878898e-05, "loss": 1.0084, "step": 2117 }, { "epoch": 0.9596737652922519, "grad_norm": 0.3360632671509012, "learning_rate": 7.795970655598899e-05, "loss": 0.9648, "step": 2118 }, { "epoch": 0.9601268690530131, "grad_norm": 0.3072133758686045, "learning_rate": 7.795571497447476e-05, "loss": 0.9811, "step": 2119 }, { "epoch": 0.9605799728137744, "grad_norm": 0.32802197786303117, "learning_rate": 7.795171959464574e-05, "loss": 0.9911, "step": 2120 }, { "epoch": 0.9610330765745355, "grad_norm": 0.3575543379570858, "learning_rate": 7.794772041690176e-05, "loss": 0.9882, "step": 2121 }, { "epoch": 0.9614861803352968, "grad_norm": 0.37553480725698923, "learning_rate": 7.794371744164302e-05, "loss": 1.0013, "step": 2122 }, { "epoch": 0.961939284096058, "grad_norm": 0.44614877473160675, "learning_rate": 7.793971066927011e-05, "loss": 0.9941, "step": 2123 }, { "epoch": 0.9623923878568192, "grad_norm": 0.5205182870964682, "learning_rate": 7.7935700100184e-05, "loss": 0.9921, "step": 2124 }, { "epoch": 0.9628454916175804, "grad_norm": 0.5592260710904154, "learning_rate": 7.793168573478602e-05, "loss": 0.9928, "step": 2125 }, { "epoch": 0.9632985953783416, "grad_norm": 0.5741950423661091, "learning_rate": 7.792766757347793e-05, "loss": 0.9813, "step": 2126 }, { "epoch": 0.9637516991391029, "grad_norm": 0.6253050726844158, "learning_rate": 7.79236456166618e-05, "loss": 0.9948, "step": 2127 }, { "epoch": 0.9642048028998641, "grad_norm": 0.6418333756672492, "learning_rate": 7.791961986474014e-05, "loss": 1.0214, "step": 2128 }, { "epoch": 0.9646579066606252, "grad_norm": 0.6696006201973774, "learning_rate": 7.791559031811582e-05, "loss": 0.9839, "step": 2129 }, { "epoch": 0.9651110104213865, "grad_norm": 0.7232101798479087, "learning_rate": 7.791155697719205e-05, "loss": 0.9893, "step": 2130 }, { "epoch": 0.9655641141821477, "grad_norm": 0.6594774412488007, "learning_rate": 7.790751984237249e-05, "loss": 0.9656, "step": 2131 }, { "epoch": 0.966017217942909, "grad_norm": 0.49851776908417367, "learning_rate": 7.790347891406115e-05, "loss": 0.9982, "step": 2132 }, { "epoch": 0.9664703217036701, "grad_norm": 0.419392304519247, "learning_rate": 7.789943419266238e-05, "loss": 0.9746, "step": 2133 }, { "epoch": 0.9669234254644313, "grad_norm": 0.43497253525155083, "learning_rate": 7.789538567858098e-05, "loss": 0.9766, "step": 2134 }, { "epoch": 0.9673765292251926, "grad_norm": 0.5117151509542426, "learning_rate": 7.789133337222208e-05, "loss": 0.9794, "step": 2135 }, { "epoch": 0.9678296329859538, "grad_norm": 0.5963983125719138, "learning_rate": 7.78872772739912e-05, "loss": 0.9968, "step": 2136 }, { "epoch": 0.968282736746715, "grad_norm": 0.6366404624099207, "learning_rate": 7.788321738429423e-05, "loss": 0.9766, "step": 2137 }, { "epoch": 0.9687358405074762, "grad_norm": 0.6150712923965966, "learning_rate": 7.787915370353748e-05, "loss": 0.9698, "step": 2138 }, { "epoch": 0.9691889442682374, "grad_norm": 0.5620605873604473, "learning_rate": 7.787508623212761e-05, "loss": 1.0113, "step": 2139 }, { "epoch": 0.9696420480289987, "grad_norm": 0.48246864978786924, "learning_rate": 7.787101497047164e-05, "loss": 0.9974, "step": 2140 }, { "epoch": 0.9700951517897598, "grad_norm": 0.42851991225856234, "learning_rate": 7.7866939918977e-05, "loss": 1.0002, "step": 2141 }, { "epoch": 0.970548255550521, "grad_norm": 0.41198295713209065, "learning_rate": 7.78628610780515e-05, "loss": 0.9932, "step": 2142 }, { "epoch": 0.9710013593112823, "grad_norm": 0.32302229600631865, "learning_rate": 7.78587784481033e-05, "loss": 0.9878, "step": 2143 }, { "epoch": 0.9714544630720435, "grad_norm": 0.29030071762804044, "learning_rate": 7.785469202954098e-05, "loss": 1.0091, "step": 2144 }, { "epoch": 0.9719075668328047, "grad_norm": 0.3420647262335002, "learning_rate": 7.785060182277346e-05, "loss": 1.0011, "step": 2145 }, { "epoch": 0.9723606705935659, "grad_norm": 0.34533468199268247, "learning_rate": 7.784650782821007e-05, "loss": 1.0022, "step": 2146 }, { "epoch": 0.9728137743543271, "grad_norm": 0.32885051539104754, "learning_rate": 7.784241004626047e-05, "loss": 0.9925, "step": 2147 }, { "epoch": 0.9732668781150884, "grad_norm": 0.3763547945367274, "learning_rate": 7.783830847733479e-05, "loss": 0.9613, "step": 2148 }, { "epoch": 0.9737199818758495, "grad_norm": 0.4184017458778367, "learning_rate": 7.783420312184343e-05, "loss": 1.0201, "step": 2149 }, { "epoch": 0.9741730856366108, "grad_norm": 0.4603283385500579, "learning_rate": 7.783009398019726e-05, "loss": 1.0002, "step": 2150 }, { "epoch": 0.974626189397372, "grad_norm": 0.48466393885374326, "learning_rate": 7.782598105280748e-05, "loss": 0.9814, "step": 2151 }, { "epoch": 0.9750792931581332, "grad_norm": 0.565665446593682, "learning_rate": 7.782186434008568e-05, "loss": 0.9563, "step": 2152 }, { "epoch": 0.9755323969188944, "grad_norm": 0.6486815444891957, "learning_rate": 7.781774384244382e-05, "loss": 0.9882, "step": 2153 }, { "epoch": 0.9759855006796556, "grad_norm": 0.6639937831547987, "learning_rate": 7.781361956029426e-05, "loss": 0.9953, "step": 2154 }, { "epoch": 0.9764386044404169, "grad_norm": 0.6234715507644025, "learning_rate": 7.780949149404972e-05, "loss": 1.0098, "step": 2155 }, { "epoch": 0.9768917082011781, "grad_norm": 0.5790328071689506, "learning_rate": 7.780535964412331e-05, "loss": 0.9855, "step": 2156 }, { "epoch": 0.9773448119619392, "grad_norm": 0.495286526827741, "learning_rate": 7.780122401092851e-05, "loss": 0.9983, "step": 2157 }, { "epoch": 0.9777979157227005, "grad_norm": 0.4205650580843999, "learning_rate": 7.779708459487916e-05, "loss": 0.9939, "step": 2158 }, { "epoch": 0.9782510194834617, "grad_norm": 0.411595369618661, "learning_rate": 7.779294139638954e-05, "loss": 0.9859, "step": 2159 }, { "epoch": 0.978704123244223, "grad_norm": 0.45200648265705107, "learning_rate": 7.778879441587426e-05, "loss": 1.0046, "step": 2160 }, { "epoch": 0.9791572270049841, "grad_norm": 0.460568896204747, "learning_rate": 7.778464365374829e-05, "loss": 0.993, "step": 2161 }, { "epoch": 0.9796103307657453, "grad_norm": 0.36931240320234293, "learning_rate": 7.778048911042704e-05, "loss": 0.9851, "step": 2162 }, { "epoch": 0.9800634345265066, "grad_norm": 0.3459295139086134, "learning_rate": 7.777633078632625e-05, "loss": 1.0105, "step": 2163 }, { "epoch": 0.9805165382872678, "grad_norm": 0.451273527345173, "learning_rate": 7.777216868186204e-05, "loss": 0.9967, "step": 2164 }, { "epoch": 0.980969642048029, "grad_norm": 0.491683876183606, "learning_rate": 7.776800279745096e-05, "loss": 0.9862, "step": 2165 }, { "epoch": 0.9814227458087902, "grad_norm": 0.464624225082488, "learning_rate": 7.776383313350985e-05, "loss": 0.9864, "step": 2166 }, { "epoch": 0.9818758495695514, "grad_norm": 0.42659827701809383, "learning_rate": 7.775965969045601e-05, "loss": 0.9715, "step": 2167 }, { "epoch": 0.9823289533303127, "grad_norm": 0.38047002915177586, "learning_rate": 7.775548246870708e-05, "loss": 0.9724, "step": 2168 }, { "epoch": 0.9827820570910738, "grad_norm": 0.3122482835908092, "learning_rate": 7.775130146868108e-05, "loss": 1.0012, "step": 2169 }, { "epoch": 0.983235160851835, "grad_norm": 0.2840013759574264, "learning_rate": 7.774711669079642e-05, "loss": 0.9914, "step": 2170 }, { "epoch": 0.9836882646125963, "grad_norm": 0.30588141851934963, "learning_rate": 7.774292813547185e-05, "loss": 0.9959, "step": 2171 }, { "epoch": 0.9841413683733575, "grad_norm": 0.32648403356098665, "learning_rate": 7.773873580312658e-05, "loss": 0.9869, "step": 2172 }, { "epoch": 0.9845944721341187, "grad_norm": 0.32648614182137103, "learning_rate": 7.773453969418012e-05, "loss": 1.003, "step": 2173 }, { "epoch": 0.9850475758948799, "grad_norm": 0.3274014678423587, "learning_rate": 7.773033980905237e-05, "loss": 0.9795, "step": 2174 }, { "epoch": 0.9855006796556411, "grad_norm": 0.4146302132729202, "learning_rate": 7.772613614816363e-05, "loss": 0.9768, "step": 2175 }, { "epoch": 0.9859537834164024, "grad_norm": 0.5149947462598622, "learning_rate": 7.772192871193458e-05, "loss": 0.9979, "step": 2176 }, { "epoch": 0.9864068871771635, "grad_norm": 0.552898282225141, "learning_rate": 7.771771750078627e-05, "loss": 0.989, "step": 2177 }, { "epoch": 0.9868599909379248, "grad_norm": 0.6146263538343587, "learning_rate": 7.771350251514011e-05, "loss": 1.0006, "step": 2178 }, { "epoch": 0.987313094698686, "grad_norm": 0.7799688737367118, "learning_rate": 7.770928375541792e-05, "loss": 0.9742, "step": 2179 }, { "epoch": 0.9877661984594472, "grad_norm": 0.8510586398898786, "learning_rate": 7.770506122204186e-05, "loss": 0.9912, "step": 2180 }, { "epoch": 0.9882193022202084, "grad_norm": 0.9285863749689631, "learning_rate": 7.770083491543451e-05, "loss": 0.9914, "step": 2181 }, { "epoch": 0.9886724059809696, "grad_norm": 0.9771362145421928, "learning_rate": 7.76966048360188e-05, "loss": 0.9763, "step": 2182 }, { "epoch": 0.9891255097417309, "grad_norm": 0.9525473915044447, "learning_rate": 7.769237098421804e-05, "loss": 0.9755, "step": 2183 }, { "epoch": 0.9895786135024921, "grad_norm": 0.7223123127560324, "learning_rate": 7.768813336045592e-05, "loss": 0.9986, "step": 2184 }, { "epoch": 0.9900317172632532, "grad_norm": 0.5062798965803557, "learning_rate": 7.76838919651565e-05, "loss": 0.9979, "step": 2185 }, { "epoch": 0.9904848210240145, "grad_norm": 0.40995514215835865, "learning_rate": 7.767964679874426e-05, "loss": 0.9991, "step": 2186 }, { "epoch": 0.9909379247847757, "grad_norm": 0.3940827291707012, "learning_rate": 7.767539786164397e-05, "loss": 0.9779, "step": 2187 }, { "epoch": 0.991391028545537, "grad_norm": 0.45931130913900076, "learning_rate": 7.767114515428088e-05, "loss": 0.9856, "step": 2188 }, { "epoch": 0.9918441323062981, "grad_norm": 0.5299982456285304, "learning_rate": 7.766688867708054e-05, "loss": 1.0006, "step": 2189 }, { "epoch": 0.9922972360670593, "grad_norm": 0.4956953455948334, "learning_rate": 7.766262843046892e-05, "loss": 1.0046, "step": 2190 }, { "epoch": 0.9927503398278206, "grad_norm": 0.4498482074027127, "learning_rate": 7.765836441487235e-05, "loss": 1.0036, "step": 2191 }, { "epoch": 0.9932034435885818, "grad_norm": 0.39345639705093377, "learning_rate": 7.765409663071753e-05, "loss": 0.9873, "step": 2192 }, { "epoch": 0.993656547349343, "grad_norm": 0.42239199857475396, "learning_rate": 7.764982507843155e-05, "loss": 1.0157, "step": 2193 }, { "epoch": 0.9941096511101042, "grad_norm": 0.42960941244512546, "learning_rate": 7.764554975844189e-05, "loss": 0.9835, "step": 2194 }, { "epoch": 0.9945627548708654, "grad_norm": 0.41732270795223697, "learning_rate": 7.764127067117637e-05, "loss": 1.0121, "step": 2195 }, { "epoch": 0.9950158586316267, "grad_norm": 0.4136414923215798, "learning_rate": 7.763698781706321e-05, "loss": 0.9945, "step": 2196 }, { "epoch": 0.9954689623923878, "grad_norm": 0.4070368241679993, "learning_rate": 7.763270119653102e-05, "loss": 0.9767, "step": 2197 }, { "epoch": 0.995922066153149, "grad_norm": 0.4222928692637354, "learning_rate": 7.762841081000878e-05, "loss": 0.9825, "step": 2198 }, { "epoch": 0.9963751699139103, "grad_norm": 0.432135672770369, "learning_rate": 7.762411665792579e-05, "loss": 0.9923, "step": 2199 }, { "epoch": 0.9968282736746715, "grad_norm": 0.4861915504479678, "learning_rate": 7.761981874071183e-05, "loss": 0.998, "step": 2200 }, { "epoch": 0.9972813774354327, "grad_norm": 0.4272798498372898, "learning_rate": 7.761551705879698e-05, "loss": 1.0032, "step": 2201 }, { "epoch": 0.9977344811961939, "grad_norm": 0.36490646092558954, "learning_rate": 7.76112116126117e-05, "loss": 1.0097, "step": 2202 }, { "epoch": 0.9981875849569551, "grad_norm": 0.30452239573027595, "learning_rate": 7.760690240258689e-05, "loss": 1.0017, "step": 2203 }, { "epoch": 0.9986406887177164, "grad_norm": 0.36359824011491354, "learning_rate": 7.760258942915374e-05, "loss": 0.9897, "step": 2204 }, { "epoch": 0.9990937924784775, "grad_norm": 0.36449451525039905, "learning_rate": 7.759827269274389e-05, "loss": 0.9968, "step": 2205 }, { "epoch": 0.9995468962392388, "grad_norm": 0.3137590848421144, "learning_rate": 7.759395219378928e-05, "loss": 1.0083, "step": 2206 }, { "epoch": 1.0, "grad_norm": 0.33854613468427275, "learning_rate": 7.758962793272233e-05, "loss": 0.9947, "step": 2207 }, { "epoch": 1.0004531037607611, "grad_norm": 0.34868047939797986, "learning_rate": 7.758529990997576e-05, "loss": 0.9954, "step": 2208 }, { "epoch": 1.0009062075215225, "grad_norm": 0.33706033197586055, "learning_rate": 7.758096812598266e-05, "loss": 0.9807, "step": 2209 }, { "epoch": 1.0013593112822836, "grad_norm": 0.3242363588635295, "learning_rate": 7.757663258117655e-05, "loss": 0.9782, "step": 2210 }, { "epoch": 1.001812415043045, "grad_norm": 0.9145773551272771, "learning_rate": 7.757229327599128e-05, "loss": 0.9915, "step": 2211 }, { "epoch": 1.002265518803806, "grad_norm": 0.4135734638509482, "learning_rate": 7.756795021086111e-05, "loss": 0.9569, "step": 2212 }, { "epoch": 1.0027186225645672, "grad_norm": 0.4429488315985266, "learning_rate": 7.756360338622064e-05, "loss": 0.9619, "step": 2213 }, { "epoch": 1.0031717263253286, "grad_norm": 0.4472325581604767, "learning_rate": 7.75592528025049e-05, "loss": 0.9727, "step": 2214 }, { "epoch": 1.0036248300860897, "grad_norm": 0.5170511046808863, "learning_rate": 7.755489846014921e-05, "loss": 0.9714, "step": 2215 }, { "epoch": 1.0040779338468508, "grad_norm": 0.6136046203547717, "learning_rate": 7.755054035958938e-05, "loss": 0.9966, "step": 2216 }, { "epoch": 1.0045310376076122, "grad_norm": 0.7046880849882647, "learning_rate": 7.75461785012615e-05, "loss": 0.9778, "step": 2217 }, { "epoch": 1.0049841413683733, "grad_norm": 0.7097113924675079, "learning_rate": 7.754181288560206e-05, "loss": 0.9765, "step": 2218 }, { "epoch": 1.0054372451291347, "grad_norm": 0.6703507072783582, "learning_rate": 7.753744351304796e-05, "loss": 0.9854, "step": 2219 }, { "epoch": 1.0058903488898958, "grad_norm": 0.6272544075531048, "learning_rate": 7.753307038403645e-05, "loss": 0.9825, "step": 2220 }, { "epoch": 1.006343452650657, "grad_norm": 0.5497130799421454, "learning_rate": 7.752869349900515e-05, "loss": 1.0042, "step": 2221 }, { "epoch": 1.0067965564114183, "grad_norm": 0.5099457264673898, "learning_rate": 7.752431285839208e-05, "loss": 0.9816, "step": 2222 }, { "epoch": 1.0072496601721794, "grad_norm": 0.4578168673672096, "learning_rate": 7.75199284626356e-05, "loss": 0.9821, "step": 2223 }, { "epoch": 1.0077027639329406, "grad_norm": 0.3954766322026191, "learning_rate": 7.751554031217448e-05, "loss": 0.978, "step": 2224 }, { "epoch": 1.008155867693702, "grad_norm": 0.4227671002150135, "learning_rate": 7.751114840744785e-05, "loss": 0.973, "step": 2225 }, { "epoch": 1.008608971454463, "grad_norm": 0.4591903245018078, "learning_rate": 7.750675274889523e-05, "loss": 0.9732, "step": 2226 }, { "epoch": 1.0090620752152244, "grad_norm": 0.43934960967547326, "learning_rate": 7.750235333695647e-05, "loss": 0.9824, "step": 2227 }, { "epoch": 1.0095151789759855, "grad_norm": 0.40440339871913566, "learning_rate": 7.749795017207188e-05, "loss": 0.9918, "step": 2228 }, { "epoch": 1.0099682827367467, "grad_norm": 0.4367711430588934, "learning_rate": 7.749354325468204e-05, "loss": 0.9808, "step": 2229 }, { "epoch": 1.010421386497508, "grad_norm": 0.5006388495476066, "learning_rate": 7.748913258522798e-05, "loss": 0.9739, "step": 2230 }, { "epoch": 1.0108744902582691, "grad_norm": 0.4743007393283478, "learning_rate": 7.74847181641511e-05, "loss": 1.0012, "step": 2231 }, { "epoch": 1.0113275940190303, "grad_norm": 0.4206582107866583, "learning_rate": 7.748029999189317e-05, "loss": 0.987, "step": 2232 }, { "epoch": 1.0117806977797916, "grad_norm": 0.38518435584565347, "learning_rate": 7.747587806889629e-05, "loss": 0.9773, "step": 2233 }, { "epoch": 1.0122338015405528, "grad_norm": 0.3424445244324865, "learning_rate": 7.747145239560298e-05, "loss": 0.9866, "step": 2234 }, { "epoch": 1.012686905301314, "grad_norm": 0.35210254008615366, "learning_rate": 7.746702297245617e-05, "loss": 1.0075, "step": 2235 }, { "epoch": 1.0131400090620752, "grad_norm": 0.353473580280082, "learning_rate": 7.746258979989908e-05, "loss": 1.002, "step": 2236 }, { "epoch": 1.0135931128228364, "grad_norm": 0.368810827961444, "learning_rate": 7.745815287837534e-05, "loss": 0.9785, "step": 2237 }, { "epoch": 1.0140462165835977, "grad_norm": 0.3904413244278006, "learning_rate": 7.745371220832899e-05, "loss": 0.9963, "step": 2238 }, { "epoch": 1.0144993203443589, "grad_norm": 0.40426702306697593, "learning_rate": 7.744926779020441e-05, "loss": 0.984, "step": 2239 }, { "epoch": 1.01495242410512, "grad_norm": 0.4063753651792182, "learning_rate": 7.744481962444637e-05, "loss": 0.9906, "step": 2240 }, { "epoch": 1.0154055278658813, "grad_norm": 0.4296703244051078, "learning_rate": 7.744036771149999e-05, "loss": 0.9893, "step": 2241 }, { "epoch": 1.0158586316266425, "grad_norm": 0.4178912746065843, "learning_rate": 7.743591205181079e-05, "loss": 0.9795, "step": 2242 }, { "epoch": 1.0163117353874038, "grad_norm": 0.39843986201199405, "learning_rate": 7.743145264582467e-05, "loss": 0.9895, "step": 2243 }, { "epoch": 1.016764839148165, "grad_norm": 0.3844070757498156, "learning_rate": 7.742698949398789e-05, "loss": 0.9714, "step": 2244 }, { "epoch": 1.017217942908926, "grad_norm": 0.4618941043920855, "learning_rate": 7.742252259674706e-05, "loss": 0.9923, "step": 2245 }, { "epoch": 1.0176710466696874, "grad_norm": 0.5684396861256167, "learning_rate": 7.741805195454922e-05, "loss": 0.9948, "step": 2246 }, { "epoch": 1.0181241504304486, "grad_norm": 0.6762289061239887, "learning_rate": 7.741357756784176e-05, "loss": 0.9888, "step": 2247 }, { "epoch": 1.0185772541912097, "grad_norm": 0.760738289953106, "learning_rate": 7.740909943707243e-05, "loss": 0.9799, "step": 2248 }, { "epoch": 1.019030357951971, "grad_norm": 0.7754103165743162, "learning_rate": 7.740461756268937e-05, "loss": 0.9793, "step": 2249 }, { "epoch": 1.0194834617127322, "grad_norm": 0.764887623490782, "learning_rate": 7.74001319451411e-05, "loss": 0.9918, "step": 2250 }, { "epoch": 1.0199365654734933, "grad_norm": 0.7042553358807158, "learning_rate": 7.739564258487649e-05, "loss": 0.9806, "step": 2251 }, { "epoch": 1.0203896692342547, "grad_norm": 0.6078249514333471, "learning_rate": 7.73911494823448e-05, "loss": 0.9788, "step": 2252 }, { "epoch": 1.0208427729950158, "grad_norm": 0.5484356224797128, "learning_rate": 7.738665263799568e-05, "loss": 0.9851, "step": 2253 }, { "epoch": 1.0212958767557772, "grad_norm": 0.5030879241494938, "learning_rate": 7.738215205227913e-05, "loss": 0.9881, "step": 2254 }, { "epoch": 1.0217489805165383, "grad_norm": 0.5205086023636416, "learning_rate": 7.737764772564555e-05, "loss": 0.9541, "step": 2255 }, { "epoch": 1.0222020842772994, "grad_norm": 0.512573059148289, "learning_rate": 7.737313965854568e-05, "loss": 0.9934, "step": 2256 }, { "epoch": 1.0226551880380608, "grad_norm": 0.3831300551329167, "learning_rate": 7.736862785143064e-05, "loss": 0.9767, "step": 2257 }, { "epoch": 1.023108291798822, "grad_norm": 0.35360858980673054, "learning_rate": 7.736411230475198e-05, "loss": 0.9908, "step": 2258 }, { "epoch": 1.0235613955595833, "grad_norm": 0.38342915781922726, "learning_rate": 7.735959301896155e-05, "loss": 0.9886, "step": 2259 }, { "epoch": 1.0240144993203444, "grad_norm": 0.3139747729542949, "learning_rate": 7.735506999451159e-05, "loss": 1.007, "step": 2260 }, { "epoch": 1.0244676030811055, "grad_norm": 0.2855494979875749, "learning_rate": 7.735054323185476e-05, "loss": 0.982, "step": 2261 }, { "epoch": 1.0249207068418669, "grad_norm": 0.33358053626981976, "learning_rate": 7.734601273144405e-05, "loss": 0.9795, "step": 2262 }, { "epoch": 1.025373810602628, "grad_norm": 0.3613950334906593, "learning_rate": 7.734147849373285e-05, "loss": 0.9827, "step": 2263 }, { "epoch": 1.0258269143633891, "grad_norm": 0.39368315621367156, "learning_rate": 7.733694051917489e-05, "loss": 0.9645, "step": 2264 }, { "epoch": 1.0262800181241505, "grad_norm": 0.40061333806930344, "learning_rate": 7.733239880822432e-05, "loss": 0.9658, "step": 2265 }, { "epoch": 1.0267331218849116, "grad_norm": 0.3892878661617279, "learning_rate": 7.732785336133563e-05, "loss": 1.005, "step": 2266 }, { "epoch": 1.027186225645673, "grad_norm": 0.3710084149249515, "learning_rate": 7.732330417896368e-05, "loss": 0.9971, "step": 2267 }, { "epoch": 1.027639329406434, "grad_norm": 0.3642425738152861, "learning_rate": 7.731875126156372e-05, "loss": 0.9759, "step": 2268 }, { "epoch": 1.0280924331671952, "grad_norm": 0.39207829884474615, "learning_rate": 7.73141946095914e-05, "loss": 0.9752, "step": 2269 }, { "epoch": 1.0285455369279566, "grad_norm": 0.44370199024272017, "learning_rate": 7.730963422350267e-05, "loss": 0.9701, "step": 2270 }, { "epoch": 1.0289986406887177, "grad_norm": 0.47327232988742746, "learning_rate": 7.730507010375393e-05, "loss": 0.9964, "step": 2271 }, { "epoch": 1.0294517444494788, "grad_norm": 0.48832067496135173, "learning_rate": 7.730050225080191e-05, "loss": 0.991, "step": 2272 }, { "epoch": 1.0299048482102402, "grad_norm": 0.509939202567841, "learning_rate": 7.729593066510372e-05, "loss": 0.9764, "step": 2273 }, { "epoch": 1.0303579519710013, "grad_norm": 0.527898769904739, "learning_rate": 7.729135534711689e-05, "loss": 0.9916, "step": 2274 }, { "epoch": 1.0308110557317627, "grad_norm": 0.4999062501815995, "learning_rate": 7.728677629729922e-05, "loss": 0.9712, "step": 2275 }, { "epoch": 1.0312641594925238, "grad_norm": 0.4692412813896213, "learning_rate": 7.728219351610898e-05, "loss": 0.9742, "step": 2276 }, { "epoch": 1.031717263253285, "grad_norm": 0.40577418195856557, "learning_rate": 7.727760700400477e-05, "loss": 0.9957, "step": 2277 }, { "epoch": 1.0321703670140463, "grad_norm": 0.37710485061921933, "learning_rate": 7.727301676144559e-05, "loss": 0.9698, "step": 2278 }, { "epoch": 1.0326234707748074, "grad_norm": 0.3498189798242626, "learning_rate": 7.726842278889078e-05, "loss": 1.017, "step": 2279 }, { "epoch": 1.0330765745355686, "grad_norm": 0.35374270326124346, "learning_rate": 7.726382508680007e-05, "loss": 0.9982, "step": 2280 }, { "epoch": 1.03352967829633, "grad_norm": 0.3361228453782643, "learning_rate": 7.725922365563355e-05, "loss": 0.9644, "step": 2281 }, { "epoch": 1.033982782057091, "grad_norm": 0.3442111354514974, "learning_rate": 7.725461849585173e-05, "loss": 0.9963, "step": 2282 }, { "epoch": 1.0344358858178522, "grad_norm": 0.3298023587774501, "learning_rate": 7.725000960791545e-05, "loss": 0.9905, "step": 2283 }, { "epoch": 1.0348889895786135, "grad_norm": 0.3077483546766694, "learning_rate": 7.72453969922859e-05, "loss": 0.9883, "step": 2284 }, { "epoch": 1.0353420933393747, "grad_norm": 0.3381666786869418, "learning_rate": 7.72407806494247e-05, "loss": 0.9724, "step": 2285 }, { "epoch": 1.035795197100136, "grad_norm": 0.3617619669777764, "learning_rate": 7.723616057979382e-05, "loss": 0.9885, "step": 2286 }, { "epoch": 1.0362483008608971, "grad_norm": 0.3972486808964251, "learning_rate": 7.72315367838556e-05, "loss": 0.9964, "step": 2287 }, { "epoch": 1.0367014046216583, "grad_norm": 0.6888313860021061, "learning_rate": 7.722690926207273e-05, "loss": 1.0021, "step": 2288 }, { "epoch": 1.0371545083824196, "grad_norm": 0.4019672505278212, "learning_rate": 7.722227801490834e-05, "loss": 1.0122, "step": 2289 }, { "epoch": 1.0376076121431808, "grad_norm": 0.4258090329116165, "learning_rate": 7.721764304282585e-05, "loss": 0.9807, "step": 2290 }, { "epoch": 1.038060715903942, "grad_norm": 0.4415439996244747, "learning_rate": 7.721300434628911e-05, "loss": 0.9838, "step": 2291 }, { "epoch": 1.0385138196647032, "grad_norm": 0.47497004834351503, "learning_rate": 7.720836192576232e-05, "loss": 0.9924, "step": 2292 }, { "epoch": 1.0389669234254644, "grad_norm": 0.49807427603839605, "learning_rate": 7.720371578171007e-05, "loss": 0.9745, "step": 2293 }, { "epoch": 1.0394200271862257, "grad_norm": 0.4683041713334128, "learning_rate": 7.71990659145973e-05, "loss": 0.9961, "step": 2294 }, { "epoch": 1.0398731309469869, "grad_norm": 0.4624773313227435, "learning_rate": 7.719441232488931e-05, "loss": 0.9862, "step": 2295 }, { "epoch": 1.040326234707748, "grad_norm": 0.5674276866285614, "learning_rate": 7.718975501305184e-05, "loss": 1.0019, "step": 2296 }, { "epoch": 1.0407793384685093, "grad_norm": 0.6633107676157595, "learning_rate": 7.718509397955093e-05, "loss": 0.9665, "step": 2297 }, { "epoch": 1.0412324422292705, "grad_norm": 0.6863394303878921, "learning_rate": 7.718042922485304e-05, "loss": 0.9883, "step": 2298 }, { "epoch": 1.0416855459900318, "grad_norm": 0.698457469364977, "learning_rate": 7.717576074942495e-05, "loss": 0.9886, "step": 2299 }, { "epoch": 1.042138649750793, "grad_norm": 0.7215575761437343, "learning_rate": 7.717108855373387e-05, "loss": 0.9927, "step": 2300 }, { "epoch": 1.042591753511554, "grad_norm": 0.7910191738748871, "learning_rate": 7.716641263824736e-05, "loss": 0.9759, "step": 2301 }, { "epoch": 1.0430448572723154, "grad_norm": 0.77114729204893, "learning_rate": 7.716173300343333e-05, "loss": 0.9865, "step": 2302 }, { "epoch": 1.0434979610330766, "grad_norm": 0.7313715061695734, "learning_rate": 7.71570496497601e-05, "loss": 0.9917, "step": 2303 }, { "epoch": 1.0439510647938377, "grad_norm": 0.9002022652031891, "learning_rate": 7.715236257769632e-05, "loss": 1.0143, "step": 2304 }, { "epoch": 1.044404168554599, "grad_norm": 1.1820127931564444, "learning_rate": 7.714767178771106e-05, "loss": 0.9873, "step": 2305 }, { "epoch": 1.0448572723153602, "grad_norm": 0.28927422473207504, "learning_rate": 7.714297728027374e-05, "loss": 0.9922, "step": 2306 }, { "epoch": 1.0453103760761215, "grad_norm": 0.38651042866696567, "learning_rate": 7.713827905585414e-05, "loss": 0.9941, "step": 2307 }, { "epoch": 1.0457634798368827, "grad_norm": 0.7218867439496548, "learning_rate": 7.71335771149224e-05, "loss": 1.0034, "step": 2308 }, { "epoch": 1.0462165835976438, "grad_norm": 0.9480053673544542, "learning_rate": 7.71288714579491e-05, "loss": 0.9852, "step": 2309 }, { "epoch": 1.0466696873584052, "grad_norm": 0.9156272267439997, "learning_rate": 7.71241620854051e-05, "loss": 0.9845, "step": 2310 }, { "epoch": 1.0471227911191663, "grad_norm": 0.7187758057872362, "learning_rate": 7.711944899776173e-05, "loss": 0.987, "step": 2311 }, { "epoch": 1.0475758948799274, "grad_norm": 0.6435575085927, "learning_rate": 7.711473219549059e-05, "loss": 0.9901, "step": 2312 }, { "epoch": 1.0480289986406888, "grad_norm": 0.7449166952623635, "learning_rate": 7.711001167906372e-05, "loss": 0.9684, "step": 2313 }, { "epoch": 1.04848210240145, "grad_norm": 0.6995219770116153, "learning_rate": 7.710528744895352e-05, "loss": 0.9854, "step": 2314 }, { "epoch": 1.0489352061622113, "grad_norm": 0.5424044185982163, "learning_rate": 7.710055950563275e-05, "loss": 0.9833, "step": 2315 }, { "epoch": 1.0493883099229724, "grad_norm": 0.5799184462051742, "learning_rate": 7.709582784957454e-05, "loss": 1.0063, "step": 2316 }, { "epoch": 1.0498414136837335, "grad_norm": 0.6065776334914154, "learning_rate": 7.70910924812524e-05, "loss": 0.993, "step": 2317 }, { "epoch": 1.0502945174444949, "grad_norm": 0.4197223984932703, "learning_rate": 7.708635340114021e-05, "loss": 1.0178, "step": 2318 }, { "epoch": 1.050747621205256, "grad_norm": 0.47037722034242063, "learning_rate": 7.708161060971221e-05, "loss": 0.9698, "step": 2319 }, { "epoch": 1.0512007249660171, "grad_norm": 0.5379393459505475, "learning_rate": 7.707686410744303e-05, "loss": 0.9959, "step": 2320 }, { "epoch": 1.0516538287267785, "grad_norm": 0.4817107285363248, "learning_rate": 7.70721138948077e-05, "loss": 0.9587, "step": 2321 }, { "epoch": 1.0521069324875396, "grad_norm": 0.42132735768228036, "learning_rate": 7.706735997228151e-05, "loss": 0.9856, "step": 2322 }, { "epoch": 1.052560036248301, "grad_norm": 0.3894595037844726, "learning_rate": 7.706260234034027e-05, "loss": 0.9983, "step": 2323 }, { "epoch": 1.053013140009062, "grad_norm": 0.34119453425625385, "learning_rate": 7.705784099946003e-05, "loss": 0.9907, "step": 2324 }, { "epoch": 1.0534662437698232, "grad_norm": 0.38807651890598693, "learning_rate": 7.705307595011731e-05, "loss": 0.994, "step": 2325 }, { "epoch": 1.0539193475305846, "grad_norm": 0.4059563515867244, "learning_rate": 7.704830719278892e-05, "loss": 0.9875, "step": 2326 }, { "epoch": 1.0543724512913457, "grad_norm": 0.3730007617291975, "learning_rate": 7.704353472795212e-05, "loss": 1.0017, "step": 2327 }, { "epoch": 1.0548255550521068, "grad_norm": 0.35489979973985375, "learning_rate": 7.703875855608447e-05, "loss": 1.0063, "step": 2328 }, { "epoch": 1.0552786588128682, "grad_norm": 0.3329908509833306, "learning_rate": 7.703397867766395e-05, "loss": 0.9875, "step": 2329 }, { "epoch": 1.0557317625736293, "grad_norm": 0.3413101149336497, "learning_rate": 7.70291950931689e-05, "loss": 0.9719, "step": 2330 }, { "epoch": 1.0561848663343907, "grad_norm": 0.38795069233886137, "learning_rate": 7.7024407803078e-05, "loss": 0.9732, "step": 2331 }, { "epoch": 1.0566379700951518, "grad_norm": 0.4117669178165828, "learning_rate": 7.701961680787034e-05, "loss": 0.9833, "step": 2332 }, { "epoch": 1.057091073855913, "grad_norm": 0.37011862950732993, "learning_rate": 7.701482210802535e-05, "loss": 0.9705, "step": 2333 }, { "epoch": 1.0575441776166743, "grad_norm": 0.36390662981399946, "learning_rate": 7.701002370402289e-05, "loss": 0.9815, "step": 2334 }, { "epoch": 1.0579972813774354, "grad_norm": 0.32244715299858734, "learning_rate": 7.70052215963431e-05, "loss": 0.9786, "step": 2335 }, { "epoch": 1.0584503851381966, "grad_norm": 0.3087304917213242, "learning_rate": 7.700041578546655e-05, "loss": 0.9773, "step": 2336 }, { "epoch": 1.058903488898958, "grad_norm": 0.35424720257223596, "learning_rate": 7.699560627187418e-05, "loss": 0.996, "step": 2337 }, { "epoch": 1.059356592659719, "grad_norm": 0.348820694802652, "learning_rate": 7.699079305604728e-05, "loss": 0.9768, "step": 2338 }, { "epoch": 1.0598096964204804, "grad_norm": 0.3731937160424326, "learning_rate": 7.698597613846754e-05, "loss": 0.996, "step": 2339 }, { "epoch": 1.0602628001812415, "grad_norm": 0.5012875601140542, "learning_rate": 7.698115551961695e-05, "loss": 0.9858, "step": 2340 }, { "epoch": 1.0607159039420027, "grad_norm": 0.5464942552256514, "learning_rate": 7.697633119997797e-05, "loss": 1.0042, "step": 2341 }, { "epoch": 1.061169007702764, "grad_norm": 0.4951464545051961, "learning_rate": 7.697150318003337e-05, "loss": 0.9909, "step": 2342 }, { "epoch": 1.0616221114635251, "grad_norm": 0.45566859514288005, "learning_rate": 7.696667146026628e-05, "loss": 0.9714, "step": 2343 }, { "epoch": 1.0620752152242863, "grad_norm": 0.476844363523204, "learning_rate": 7.696183604116024e-05, "loss": 0.9952, "step": 2344 }, { "epoch": 1.0625283189850476, "grad_norm": 0.4896829062335543, "learning_rate": 7.695699692319913e-05, "loss": 0.9712, "step": 2345 }, { "epoch": 1.0629814227458088, "grad_norm": 0.461027154075929, "learning_rate": 7.695215410686723e-05, "loss": 1.0002, "step": 2346 }, { "epoch": 1.0634345265065699, "grad_norm": 0.3482480163085047, "learning_rate": 7.694730759264915e-05, "loss": 0.9784, "step": 2347 }, { "epoch": 1.0638876302673312, "grad_norm": 0.2488895282494781, "learning_rate": 7.694245738102991e-05, "loss": 1.0003, "step": 2348 }, { "epoch": 1.0643407340280924, "grad_norm": 0.2713726150501195, "learning_rate": 7.693760347249486e-05, "loss": 0.9838, "step": 2349 }, { "epoch": 1.0647938377888537, "grad_norm": 0.41832804913479277, "learning_rate": 7.693274586752977e-05, "loss": 0.9606, "step": 2350 }, { "epoch": 1.0652469415496149, "grad_norm": 0.4978222185792617, "learning_rate": 7.692788456662072e-05, "loss": 0.9765, "step": 2351 }, { "epoch": 1.065700045310376, "grad_norm": 0.4898004772575809, "learning_rate": 7.692301957025423e-05, "loss": 0.9953, "step": 2352 }, { "epoch": 1.0661531490711373, "grad_norm": 0.48249400687973276, "learning_rate": 7.69181508789171e-05, "loss": 0.9836, "step": 2353 }, { "epoch": 1.0666062528318985, "grad_norm": 0.5596052240612132, "learning_rate": 7.691327849309661e-05, "loss": 0.9899, "step": 2354 }, { "epoch": 1.0670593565926598, "grad_norm": 0.41315898802215945, "learning_rate": 7.69084024132803e-05, "loss": 0.9956, "step": 2355 }, { "epoch": 1.067512460353421, "grad_norm": 0.3764383753905227, "learning_rate": 7.690352263995617e-05, "loss": 0.9824, "step": 2356 }, { "epoch": 1.067965564114182, "grad_norm": 0.35417240552693185, "learning_rate": 7.689863917361251e-05, "loss": 0.9864, "step": 2357 }, { "epoch": 1.0684186678749434, "grad_norm": 0.3082258859741951, "learning_rate": 7.689375201473805e-05, "loss": 0.968, "step": 2358 }, { "epoch": 1.0688717716357046, "grad_norm": 0.29571171540269814, "learning_rate": 7.688886116382185e-05, "loss": 0.9648, "step": 2359 }, { "epoch": 1.0693248753964657, "grad_norm": 0.3163394979380561, "learning_rate": 7.688396662135334e-05, "loss": 0.9662, "step": 2360 }, { "epoch": 1.069777979157227, "grad_norm": 0.3907054121503886, "learning_rate": 7.687906838782235e-05, "loss": 1.0076, "step": 2361 }, { "epoch": 1.0702310829179882, "grad_norm": 0.42705967850433213, "learning_rate": 7.687416646371904e-05, "loss": 0.9859, "step": 2362 }, { "epoch": 1.0706841866787493, "grad_norm": 0.4920307562275912, "learning_rate": 7.686926084953396e-05, "loss": 0.9857, "step": 2363 }, { "epoch": 1.0711372904395107, "grad_norm": 0.6228829094140255, "learning_rate": 7.686435154575803e-05, "loss": 0.9707, "step": 2364 }, { "epoch": 1.0715903942002718, "grad_norm": 0.7395352131284834, "learning_rate": 7.685943855288253e-05, "loss": 0.9963, "step": 2365 }, { "epoch": 1.0720434979610332, "grad_norm": 0.8006764222885261, "learning_rate": 7.685452187139912e-05, "loss": 0.9546, "step": 2366 }, { "epoch": 1.0724966017217943, "grad_norm": 0.8013523759438416, "learning_rate": 7.68496015017998e-05, "loss": 0.9741, "step": 2367 }, { "epoch": 1.0729497054825554, "grad_norm": 0.7527387826644998, "learning_rate": 7.6844677444577e-05, "loss": 0.9794, "step": 2368 }, { "epoch": 1.0734028092433168, "grad_norm": 0.6707593382062679, "learning_rate": 7.683974970022345e-05, "loss": 0.9832, "step": 2369 }, { "epoch": 1.073855913004078, "grad_norm": 0.5146939663728316, "learning_rate": 7.683481826923231e-05, "loss": 0.9858, "step": 2370 }, { "epoch": 1.0743090167648393, "grad_norm": 0.4262606672725893, "learning_rate": 7.682988315209706e-05, "loss": 0.9826, "step": 2371 }, { "epoch": 1.0747621205256004, "grad_norm": 0.40682015855525744, "learning_rate": 7.682494434931157e-05, "loss": 0.9663, "step": 2372 }, { "epoch": 1.0752152242863615, "grad_norm": 0.3963143692312314, "learning_rate": 7.682000186137007e-05, "loss": 0.967, "step": 2373 }, { "epoch": 1.0756683280471229, "grad_norm": 0.3335557935809195, "learning_rate": 7.681505568876719e-05, "loss": 0.9809, "step": 2374 }, { "epoch": 1.076121431807884, "grad_norm": 0.3326533539411316, "learning_rate": 7.681010583199789e-05, "loss": 0.9756, "step": 2375 }, { "epoch": 1.0765745355686451, "grad_norm": 0.36254237774014275, "learning_rate": 7.68051522915575e-05, "loss": 0.9784, "step": 2376 }, { "epoch": 1.0770276393294065, "grad_norm": 0.3659782186699314, "learning_rate": 7.680019506794176e-05, "loss": 0.9596, "step": 2377 }, { "epoch": 1.0774807430901676, "grad_norm": 0.3977663546810698, "learning_rate": 7.679523416164672e-05, "loss": 0.9688, "step": 2378 }, { "epoch": 1.0779338468509287, "grad_norm": 0.39359910099806084, "learning_rate": 7.679026957316886e-05, "loss": 0.9866, "step": 2379 }, { "epoch": 1.07838695061169, "grad_norm": 0.37839872564458904, "learning_rate": 7.678530130300498e-05, "loss": 0.9794, "step": 2380 }, { "epoch": 1.0788400543724512, "grad_norm": 0.4592867013059008, "learning_rate": 7.678032935165228e-05, "loss": 0.9585, "step": 2381 }, { "epoch": 1.0792931581332126, "grad_norm": 0.49710413916411667, "learning_rate": 7.67753537196083e-05, "loss": 0.9725, "step": 2382 }, { "epoch": 1.0797462618939737, "grad_norm": 0.4932480323541638, "learning_rate": 7.677037440737096e-05, "loss": 0.9795, "step": 2383 }, { "epoch": 1.0801993656547348, "grad_norm": 0.4951347909675037, "learning_rate": 7.676539141543854e-05, "loss": 0.9805, "step": 2384 }, { "epoch": 1.0806524694154962, "grad_norm": 0.49415072377329355, "learning_rate": 7.676040474430974e-05, "loss": 0.9619, "step": 2385 }, { "epoch": 1.0811055731762573, "grad_norm": 0.5064712860140123, "learning_rate": 7.675541439448356e-05, "loss": 0.9982, "step": 2386 }, { "epoch": 1.0815586769370187, "grad_norm": 0.5369104799507026, "learning_rate": 7.675042036645941e-05, "loss": 0.9683, "step": 2387 }, { "epoch": 1.0820117806977798, "grad_norm": 0.5389496381428438, "learning_rate": 7.674542266073703e-05, "loss": 0.9782, "step": 2388 }, { "epoch": 1.082464884458541, "grad_norm": 0.5508926663635413, "learning_rate": 7.674042127781657e-05, "loss": 0.9926, "step": 2389 }, { "epoch": 1.0829179882193023, "grad_norm": 0.5556962832573733, "learning_rate": 7.673541621819853e-05, "loss": 0.9902, "step": 2390 }, { "epoch": 1.0833710919800634, "grad_norm": 0.5232655862497213, "learning_rate": 7.673040748238376e-05, "loss": 0.9744, "step": 2391 }, { "epoch": 1.0838241957408246, "grad_norm": 0.44509274934909066, "learning_rate": 7.672539507087353e-05, "loss": 0.9617, "step": 2392 }, { "epoch": 1.084277299501586, "grad_norm": 0.365415340230986, "learning_rate": 7.672037898416942e-05, "loss": 0.9851, "step": 2393 }, { "epoch": 1.084730403262347, "grad_norm": 0.29512361275654, "learning_rate": 7.671535922277338e-05, "loss": 0.978, "step": 2394 }, { "epoch": 1.0851835070231084, "grad_norm": 0.34745620609803785, "learning_rate": 7.67103357871878e-05, "loss": 0.9914, "step": 2395 }, { "epoch": 1.0856366107838695, "grad_norm": 0.3955223655577131, "learning_rate": 7.670530867791535e-05, "loss": 0.9673, "step": 2396 }, { "epoch": 1.0860897145446307, "grad_norm": 0.3446320227571404, "learning_rate": 7.670027789545911e-05, "loss": 1.01, "step": 2397 }, { "epoch": 1.086542818305392, "grad_norm": 0.32282634038310676, "learning_rate": 7.669524344032252e-05, "loss": 0.9896, "step": 2398 }, { "epoch": 1.0869959220661531, "grad_norm": 0.35093226797976595, "learning_rate": 7.669020531300939e-05, "loss": 0.9946, "step": 2399 }, { "epoch": 1.0874490258269143, "grad_norm": 0.33203442788380766, "learning_rate": 7.668516351402392e-05, "loss": 0.9786, "step": 2400 }, { "epoch": 1.0879021295876756, "grad_norm": 0.29813116103200454, "learning_rate": 7.668011804387064e-05, "loss": 1.0006, "step": 2401 }, { "epoch": 1.0883552333484368, "grad_norm": 0.3045767176760953, "learning_rate": 7.667506890305444e-05, "loss": 0.979, "step": 2402 }, { "epoch": 1.088808337109198, "grad_norm": 0.36288242730751047, "learning_rate": 7.667001609208061e-05, "loss": 0.9818, "step": 2403 }, { "epoch": 1.0892614408699592, "grad_norm": 0.415163933605958, "learning_rate": 7.666495961145483e-05, "loss": 0.9734, "step": 2404 }, { "epoch": 1.0897145446307204, "grad_norm": 0.43450767338473295, "learning_rate": 7.665989946168307e-05, "loss": 0.9706, "step": 2405 }, { "epoch": 1.0901676483914817, "grad_norm": 0.39216796187442204, "learning_rate": 7.665483564327172e-05, "loss": 0.9772, "step": 2406 }, { "epoch": 1.0906207521522429, "grad_norm": 0.35084464884570754, "learning_rate": 7.664976815672755e-05, "loss": 0.9706, "step": 2407 }, { "epoch": 1.091073855913004, "grad_norm": 0.337015548022755, "learning_rate": 7.664469700255766e-05, "loss": 0.982, "step": 2408 }, { "epoch": 1.0915269596737653, "grad_norm": 0.3642325925870671, "learning_rate": 7.663962218126951e-05, "loss": 0.9715, "step": 2409 }, { "epoch": 1.0919800634345265, "grad_norm": 0.4318955662541273, "learning_rate": 7.6634543693371e-05, "loss": 0.9971, "step": 2410 }, { "epoch": 1.0924331671952878, "grad_norm": 0.5505323787193458, "learning_rate": 7.662946153937031e-05, "loss": 0.9971, "step": 2411 }, { "epoch": 1.092886270956049, "grad_norm": 0.639848633625471, "learning_rate": 7.662437571977601e-05, "loss": 0.9715, "step": 2412 }, { "epoch": 1.09333937471681, "grad_norm": 0.6695777245300125, "learning_rate": 7.661928623509709e-05, "loss": 0.9875, "step": 2413 }, { "epoch": 1.0937924784775714, "grad_norm": 0.7349743976371135, "learning_rate": 7.661419308584283e-05, "loss": 0.9824, "step": 2414 }, { "epoch": 1.0942455822383326, "grad_norm": 0.788906815897849, "learning_rate": 7.660909627252294e-05, "loss": 0.9647, "step": 2415 }, { "epoch": 1.0946986859990937, "grad_norm": 0.7614404495965472, "learning_rate": 7.660399579564745e-05, "loss": 0.9953, "step": 2416 }, { "epoch": 1.095151789759855, "grad_norm": 0.6676132719911634, "learning_rate": 7.659889165572678e-05, "loss": 0.9795, "step": 2417 }, { "epoch": 1.0956048935206162, "grad_norm": 0.4877123538221517, "learning_rate": 7.659378385327174e-05, "loss": 0.9969, "step": 2418 }, { "epoch": 1.0960579972813775, "grad_norm": 0.3946957593202744, "learning_rate": 7.658867238879344e-05, "loss": 0.9728, "step": 2419 }, { "epoch": 1.0965111010421387, "grad_norm": 0.3325167185887406, "learning_rate": 7.65835572628034e-05, "loss": 0.9966, "step": 2420 }, { "epoch": 1.0969642048028998, "grad_norm": 0.3210365952526539, "learning_rate": 7.657843847581353e-05, "loss": 0.9629, "step": 2421 }, { "epoch": 1.0974173085636612, "grad_norm": 0.3353444982675422, "learning_rate": 7.657331602833607e-05, "loss": 0.9797, "step": 2422 }, { "epoch": 1.0978704123244223, "grad_norm": 0.37158811164984856, "learning_rate": 7.656818992088362e-05, "loss": 0.9909, "step": 2423 }, { "epoch": 1.0983235160851834, "grad_norm": 0.505056228769612, "learning_rate": 7.656306015396917e-05, "loss": 0.9812, "step": 2424 }, { "epoch": 1.0987766198459448, "grad_norm": 0.5954444448990883, "learning_rate": 7.655792672810606e-05, "loss": 0.9785, "step": 2425 }, { "epoch": 1.099229723606706, "grad_norm": 0.6294243341527463, "learning_rate": 7.655278964380803e-05, "loss": 0.9913, "step": 2426 }, { "epoch": 1.0996828273674673, "grad_norm": 0.6045968464746331, "learning_rate": 7.654764890158912e-05, "loss": 1.0022, "step": 2427 }, { "epoch": 1.1001359311282284, "grad_norm": 0.566630566220259, "learning_rate": 7.654250450196379e-05, "loss": 0.9733, "step": 2428 }, { "epoch": 1.1005890348889895, "grad_norm": 0.5359213099460115, "learning_rate": 7.653735644544689e-05, "loss": 0.999, "step": 2429 }, { "epoch": 1.1010421386497509, "grad_norm": 0.5007174433350086, "learning_rate": 7.653220473255355e-05, "loss": 0.9896, "step": 2430 }, { "epoch": 1.101495242410512, "grad_norm": 0.4630250076814972, "learning_rate": 7.652704936379932e-05, "loss": 0.9673, "step": 2431 }, { "epoch": 1.1019483461712731, "grad_norm": 0.41971177892447453, "learning_rate": 7.652189033970013e-05, "loss": 0.9687, "step": 2432 }, { "epoch": 1.1024014499320345, "grad_norm": 0.4152841097929698, "learning_rate": 7.651672766077224e-05, "loss": 0.9722, "step": 2433 }, { "epoch": 1.1028545536927956, "grad_norm": 0.4565342248652505, "learning_rate": 7.65115613275323e-05, "loss": 0.957, "step": 2434 }, { "epoch": 1.103307657453557, "grad_norm": 0.4569440569397193, "learning_rate": 7.65063913404973e-05, "loss": 0.9882, "step": 2435 }, { "epoch": 1.103760761214318, "grad_norm": 0.4337589036699544, "learning_rate": 7.650121770018463e-05, "loss": 0.9884, "step": 2436 }, { "epoch": 1.1042138649750792, "grad_norm": 0.3831120886951879, "learning_rate": 7.649604040711202e-05, "loss": 0.985, "step": 2437 }, { "epoch": 1.1046669687358406, "grad_norm": 0.3664671293797935, "learning_rate": 7.649085946179758e-05, "loss": 0.997, "step": 2438 }, { "epoch": 1.1051200724966017, "grad_norm": 0.3594297175623974, "learning_rate": 7.648567486475979e-05, "loss": 0.9707, "step": 2439 }, { "epoch": 1.1055731762573628, "grad_norm": 0.2848363712291444, "learning_rate": 7.648048661651744e-05, "loss": 0.9702, "step": 2440 }, { "epoch": 1.1060262800181242, "grad_norm": 0.30012705059216827, "learning_rate": 7.647529471758977e-05, "loss": 0.9931, "step": 2441 }, { "epoch": 1.1064793837788853, "grad_norm": 0.346813731854995, "learning_rate": 7.647009916849634e-05, "loss": 0.9773, "step": 2442 }, { "epoch": 1.1069324875396467, "grad_norm": 0.396228935527594, "learning_rate": 7.646489996975708e-05, "loss": 0.9859, "step": 2443 }, { "epoch": 1.1073855913004078, "grad_norm": 0.4241780696737915, "learning_rate": 7.645969712189227e-05, "loss": 0.9801, "step": 2444 }, { "epoch": 1.107838695061169, "grad_norm": 0.4176298338959358, "learning_rate": 7.645449062542257e-05, "loss": 0.9882, "step": 2445 }, { "epoch": 1.1082917988219303, "grad_norm": 0.4108499205842145, "learning_rate": 7.644928048086903e-05, "loss": 0.9782, "step": 2446 }, { "epoch": 1.1087449025826914, "grad_norm": 0.4356653305632001, "learning_rate": 7.644406668875302e-05, "loss": 0.9676, "step": 2447 }, { "epoch": 1.1091980063434526, "grad_norm": 0.45085434757878473, "learning_rate": 7.643884924959631e-05, "loss": 1.001, "step": 2448 }, { "epoch": 1.109651110104214, "grad_norm": 0.5110254913057515, "learning_rate": 7.643362816392103e-05, "loss": 0.9959, "step": 2449 }, { "epoch": 1.110104213864975, "grad_norm": 0.5677112343193285, "learning_rate": 7.642840343224964e-05, "loss": 0.9724, "step": 2450 }, { "epoch": 1.1105573176257364, "grad_norm": 0.6173994639211757, "learning_rate": 7.642317505510498e-05, "loss": 0.9809, "step": 2451 }, { "epoch": 1.1110104213864975, "grad_norm": 0.6786913189482117, "learning_rate": 7.641794303301032e-05, "loss": 0.9593, "step": 2452 }, { "epoch": 1.1114635251472587, "grad_norm": 0.6648114867801375, "learning_rate": 7.64127073664892e-05, "loss": 0.9841, "step": 2453 }, { "epoch": 1.11191662890802, "grad_norm": 0.5692991673771419, "learning_rate": 7.640746805606556e-05, "loss": 0.9886, "step": 2454 }, { "epoch": 1.1123697326687811, "grad_norm": 0.5050659024206452, "learning_rate": 7.640222510226372e-05, "loss": 0.9715, "step": 2455 }, { "epoch": 1.1128228364295423, "grad_norm": 0.47834784273531755, "learning_rate": 7.639697850560836e-05, "loss": 0.9938, "step": 2456 }, { "epoch": 1.1132759401903036, "grad_norm": 0.4557731214811346, "learning_rate": 7.639172826662453e-05, "loss": 0.9825, "step": 2457 }, { "epoch": 1.1137290439510648, "grad_norm": 0.3921962945406824, "learning_rate": 7.63864743858376e-05, "loss": 0.9941, "step": 2458 }, { "epoch": 1.114182147711826, "grad_norm": 0.3147965678782001, "learning_rate": 7.638121686377337e-05, "loss": 0.983, "step": 2459 }, { "epoch": 1.1146352514725872, "grad_norm": 0.28984547095857904, "learning_rate": 7.637595570095796e-05, "loss": 0.9898, "step": 2460 }, { "epoch": 1.1150883552333484, "grad_norm": 0.27766322533301396, "learning_rate": 7.637069089791786e-05, "loss": 0.9726, "step": 2461 }, { "epoch": 1.1155414589941097, "grad_norm": 0.3231144313260778, "learning_rate": 7.636542245517995e-05, "loss": 0.9808, "step": 2462 }, { "epoch": 1.1159945627548709, "grad_norm": 0.35383098022179454, "learning_rate": 7.636015037327142e-05, "loss": 0.9871, "step": 2463 }, { "epoch": 1.116447666515632, "grad_norm": 0.3607640019325352, "learning_rate": 7.63548746527199e-05, "loss": 0.9751, "step": 2464 }, { "epoch": 1.1169007702763933, "grad_norm": 0.3842554590183816, "learning_rate": 7.63495952940533e-05, "loss": 0.987, "step": 2465 }, { "epoch": 1.1173538740371545, "grad_norm": 0.48133877412999415, "learning_rate": 7.634431229779998e-05, "loss": 0.9841, "step": 2466 }, { "epoch": 1.1178069777979158, "grad_norm": 0.529264282186466, "learning_rate": 7.63390256644886e-05, "loss": 0.9948, "step": 2467 }, { "epoch": 1.118260081558677, "grad_norm": 0.9687274467499214, "learning_rate": 7.63337353946482e-05, "loss": 1.0231, "step": 2468 }, { "epoch": 1.118713185319438, "grad_norm": 0.5181899279743798, "learning_rate": 7.632844148880822e-05, "loss": 0.9882, "step": 2469 }, { "epoch": 1.1191662890801994, "grad_norm": 0.44337873362191516, "learning_rate": 7.632314394749839e-05, "loss": 0.9799, "step": 2470 }, { "epoch": 1.1196193928409606, "grad_norm": 0.3798867660296729, "learning_rate": 7.631784277124887e-05, "loss": 0.9806, "step": 2471 }, { "epoch": 1.1200724966017217, "grad_norm": 0.34901450542573836, "learning_rate": 7.631253796059016e-05, "loss": 0.9932, "step": 2472 }, { "epoch": 1.120525600362483, "grad_norm": 0.322967870217302, "learning_rate": 7.630722951605314e-05, "loss": 0.9987, "step": 2473 }, { "epoch": 1.1209787041232442, "grad_norm": 0.3763401552088881, "learning_rate": 7.6301917438169e-05, "loss": 0.9797, "step": 2474 }, { "epoch": 1.1214318078840055, "grad_norm": 0.44247306001625036, "learning_rate": 7.629660172746937e-05, "loss": 0.9733, "step": 2475 }, { "epoch": 1.1218849116447667, "grad_norm": 0.5109078133817909, "learning_rate": 7.629128238448617e-05, "loss": 0.9789, "step": 2476 }, { "epoch": 1.1223380154055278, "grad_norm": 0.4246372262318805, "learning_rate": 7.628595940975175e-05, "loss": 0.9948, "step": 2477 }, { "epoch": 1.1227911191662892, "grad_norm": 0.36406497624105827, "learning_rate": 7.628063280379878e-05, "loss": 0.9855, "step": 2478 }, { "epoch": 1.1232442229270503, "grad_norm": 0.3418545003710715, "learning_rate": 7.627530256716029e-05, "loss": 1.0058, "step": 2479 }, { "epoch": 1.1236973266878114, "grad_norm": 0.35158446111714464, "learning_rate": 7.626996870036971e-05, "loss": 0.9882, "step": 2480 }, { "epoch": 1.1241504304485728, "grad_norm": 0.3227320589475805, "learning_rate": 7.626463120396082e-05, "loss": 0.9789, "step": 2481 }, { "epoch": 1.124603534209334, "grad_norm": 0.3794734305066578, "learning_rate": 7.625929007846773e-05, "loss": 0.9861, "step": 2482 }, { "epoch": 1.1250566379700953, "grad_norm": 0.3885691357200379, "learning_rate": 7.625394532442494e-05, "loss": 0.9782, "step": 2483 }, { "epoch": 1.1255097417308564, "grad_norm": 0.4750651509145684, "learning_rate": 7.624859694236734e-05, "loss": 1.0086, "step": 2484 }, { "epoch": 1.1259628454916175, "grad_norm": 0.3733711262492507, "learning_rate": 7.624324493283013e-05, "loss": 0.9715, "step": 2485 }, { "epoch": 1.1264159492523789, "grad_norm": 0.3589282216714168, "learning_rate": 7.62378892963489e-05, "loss": 0.995, "step": 2486 }, { "epoch": 1.12686905301314, "grad_norm": 0.365070739398397, "learning_rate": 7.623253003345961e-05, "loss": 0.9824, "step": 2487 }, { "epoch": 1.1273221567739011, "grad_norm": 0.36209371195396983, "learning_rate": 7.622716714469857e-05, "loss": 1.0037, "step": 2488 }, { "epoch": 1.1277752605346625, "grad_norm": 0.4195067323806626, "learning_rate": 7.622180063060245e-05, "loss": 0.9756, "step": 2489 }, { "epoch": 1.1282283642954236, "grad_norm": 0.5103907171265322, "learning_rate": 7.621643049170831e-05, "loss": 1.0034, "step": 2490 }, { "epoch": 1.1286814680561847, "grad_norm": 0.5586325756115907, "learning_rate": 7.62110567285535e-05, "loss": 0.972, "step": 2491 }, { "epoch": 1.129134571816946, "grad_norm": 0.5299107786492889, "learning_rate": 7.620567934167586e-05, "loss": 0.9962, "step": 2492 }, { "epoch": 1.1295876755777072, "grad_norm": 0.4513121675309715, "learning_rate": 7.620029833161345e-05, "loss": 0.965, "step": 2493 }, { "epoch": 1.1300407793384686, "grad_norm": 0.39914828717096823, "learning_rate": 7.619491369890482e-05, "loss": 0.9865, "step": 2494 }, { "epoch": 1.1304938830992297, "grad_norm": 0.42707003011909794, "learning_rate": 7.618952544408876e-05, "loss": 0.9716, "step": 2495 }, { "epoch": 1.1309469868599908, "grad_norm": 0.4298735051059118, "learning_rate": 7.618413356770452e-05, "loss": 0.9707, "step": 2496 }, { "epoch": 1.1314000906207522, "grad_norm": 0.40765757402964853, "learning_rate": 7.61787380702917e-05, "loss": 0.9811, "step": 2497 }, { "epoch": 1.1318531943815133, "grad_norm": 0.42825950856543327, "learning_rate": 7.617333895239018e-05, "loss": 0.9848, "step": 2498 }, { "epoch": 1.1323062981422747, "grad_norm": 0.44189724580638196, "learning_rate": 7.616793621454032e-05, "loss": 0.9799, "step": 2499 }, { "epoch": 1.1327594019030358, "grad_norm": 0.4814696322251123, "learning_rate": 7.616252985728274e-05, "loss": 0.9816, "step": 2500 }, { "epoch": 1.133212505663797, "grad_norm": 0.5165037726726442, "learning_rate": 7.615711988115849e-05, "loss": 0.9808, "step": 2501 }, { "epoch": 1.1336656094245583, "grad_norm": 0.5441997454405592, "learning_rate": 7.615170628670896e-05, "loss": 0.9784, "step": 2502 }, { "epoch": 1.1341187131853194, "grad_norm": 0.5590839571536307, "learning_rate": 7.614628907447589e-05, "loss": 1.0004, "step": 2503 }, { "epoch": 1.1345718169460806, "grad_norm": 0.6065912562750794, "learning_rate": 7.614086824500141e-05, "loss": 0.9866, "step": 2504 }, { "epoch": 1.135024920706842, "grad_norm": 0.6554539973424408, "learning_rate": 7.613544379882796e-05, "loss": 0.9834, "step": 2505 }, { "epoch": 1.135478024467603, "grad_norm": 0.7140256047054072, "learning_rate": 7.613001573649841e-05, "loss": 0.9953, "step": 2506 }, { "epoch": 1.1359311282283642, "grad_norm": 0.7297118158485031, "learning_rate": 7.612458405855595e-05, "loss": 0.9867, "step": 2507 }, { "epoch": 1.1363842319891255, "grad_norm": 0.6650584497386307, "learning_rate": 7.611914876554415e-05, "loss": 0.9971, "step": 2508 }, { "epoch": 1.1368373357498867, "grad_norm": 0.5343310126890661, "learning_rate": 7.611370985800688e-05, "loss": 0.9901, "step": 2509 }, { "epoch": 1.137290439510648, "grad_norm": 0.4513299596380962, "learning_rate": 7.61082673364885e-05, "loss": 0.9772, "step": 2510 }, { "epoch": 1.1377435432714091, "grad_norm": 0.3688843140988566, "learning_rate": 7.610282120153361e-05, "loss": 0.976, "step": 2511 }, { "epoch": 1.1381966470321703, "grad_norm": 0.31056450293745724, "learning_rate": 7.609737145368723e-05, "loss": 1.0039, "step": 2512 }, { "epoch": 1.1386497507929316, "grad_norm": 0.25720957850646686, "learning_rate": 7.609191809349473e-05, "loss": 0.9859, "step": 2513 }, { "epoch": 1.1391028545536928, "grad_norm": 0.32638093372424215, "learning_rate": 7.608646112150183e-05, "loss": 0.9714, "step": 2514 }, { "epoch": 1.139555958314454, "grad_norm": 0.47084963216081266, "learning_rate": 7.608100053825463e-05, "loss": 0.9649, "step": 2515 }, { "epoch": 1.1400090620752152, "grad_norm": 0.6018568908213243, "learning_rate": 7.607553634429956e-05, "loss": 0.9868, "step": 2516 }, { "epoch": 1.1404621658359764, "grad_norm": 0.625356562062222, "learning_rate": 7.607006854018347e-05, "loss": 0.9867, "step": 2517 }, { "epoch": 1.1409152695967377, "grad_norm": 0.5578735652022416, "learning_rate": 7.606459712645352e-05, "loss": 0.9859, "step": 2518 }, { "epoch": 1.1413683733574989, "grad_norm": 0.5143544542976242, "learning_rate": 7.605912210365725e-05, "loss": 0.9828, "step": 2519 }, { "epoch": 1.14182147711826, "grad_norm": 0.443490406702864, "learning_rate": 7.605364347234255e-05, "loss": 0.9778, "step": 2520 }, { "epoch": 1.1422745808790213, "grad_norm": 0.35971853524995134, "learning_rate": 7.604816123305768e-05, "loss": 0.9788, "step": 2521 }, { "epoch": 1.1427276846397825, "grad_norm": 0.29670534847934027, "learning_rate": 7.604267538635127e-05, "loss": 0.9984, "step": 2522 }, { "epoch": 1.1431807884005436, "grad_norm": 0.2773825744297022, "learning_rate": 7.603718593277229e-05, "loss": 0.9711, "step": 2523 }, { "epoch": 1.143633892161305, "grad_norm": 0.33540964138885176, "learning_rate": 7.60316928728701e-05, "loss": 0.9816, "step": 2524 }, { "epoch": 1.144086995922066, "grad_norm": 0.31827577295211557, "learning_rate": 7.602619620719438e-05, "loss": 0.9765, "step": 2525 }, { "epoch": 1.1445400996828274, "grad_norm": 0.28319722277693754, "learning_rate": 7.602069593629521e-05, "loss": 0.9845, "step": 2526 }, { "epoch": 1.1449932034435886, "grad_norm": 0.32160784953712646, "learning_rate": 7.601519206072301e-05, "loss": 0.9849, "step": 2527 }, { "epoch": 1.1454463072043497, "grad_norm": 0.3493008609345628, "learning_rate": 7.600968458102855e-05, "loss": 1.004, "step": 2528 }, { "epoch": 1.145899410965111, "grad_norm": 0.42052172412792727, "learning_rate": 7.6004173497763e-05, "loss": 0.976, "step": 2529 }, { "epoch": 1.1463525147258722, "grad_norm": 0.40010436950392947, "learning_rate": 7.599865881147786e-05, "loss": 0.966, "step": 2530 }, { "epoch": 1.1468056184866335, "grad_norm": 0.3798056837395429, "learning_rate": 7.5993140522725e-05, "loss": 1.0003, "step": 2531 }, { "epoch": 1.1472587222473947, "grad_norm": 0.48324446415745803, "learning_rate": 7.598761863205662e-05, "loss": 0.9708, "step": 2532 }, { "epoch": 1.1477118260081558, "grad_norm": 0.5827331710615489, "learning_rate": 7.598209314002535e-05, "loss": 0.9687, "step": 2533 }, { "epoch": 1.1481649297689172, "grad_norm": 0.68552286276728, "learning_rate": 7.597656404718411e-05, "loss": 0.9921, "step": 2534 }, { "epoch": 1.1486180335296783, "grad_norm": 0.8181569205455452, "learning_rate": 7.597103135408622e-05, "loss": 0.9812, "step": 2535 }, { "epoch": 1.1490711372904394, "grad_norm": 0.9427206970180544, "learning_rate": 7.596549506128534e-05, "loss": 0.9881, "step": 2536 }, { "epoch": 1.1495242410512008, "grad_norm": 0.9867394355430641, "learning_rate": 7.595995516933551e-05, "loss": 0.9846, "step": 2537 }, { "epoch": 1.149977344811962, "grad_norm": 0.9151397119882004, "learning_rate": 7.595441167879111e-05, "loss": 0.9764, "step": 2538 }, { "epoch": 1.150430448572723, "grad_norm": 0.687969500765408, "learning_rate": 7.594886459020691e-05, "loss": 0.9732, "step": 2539 }, { "epoch": 1.1508835523334844, "grad_norm": 0.40357266449116663, "learning_rate": 7.5943313904138e-05, "loss": 0.968, "step": 2540 }, { "epoch": 1.1513366560942455, "grad_norm": 0.46194361465943445, "learning_rate": 7.593775962113985e-05, "loss": 0.9684, "step": 2541 }, { "epoch": 1.1517897598550069, "grad_norm": 0.6982789973645712, "learning_rate": 7.59322017417683e-05, "loss": 0.9947, "step": 2542 }, { "epoch": 1.152242863615768, "grad_norm": 0.7878279032336817, "learning_rate": 7.592664026657953e-05, "loss": 0.9868, "step": 2543 }, { "epoch": 1.1526959673765291, "grad_norm": 0.7403330401793615, "learning_rate": 7.592107519613009e-05, "loss": 0.9661, "step": 2544 }, { "epoch": 1.1531490711372905, "grad_norm": 0.629346152492884, "learning_rate": 7.59155065309769e-05, "loss": 0.9814, "step": 2545 }, { "epoch": 1.1536021748980516, "grad_norm": 0.5002816300350051, "learning_rate": 7.590993427167723e-05, "loss": 0.9868, "step": 2546 }, { "epoch": 1.154055278658813, "grad_norm": 0.42935147923442196, "learning_rate": 7.590435841878869e-05, "loss": 0.9728, "step": 2547 }, { "epoch": 1.154508382419574, "grad_norm": 0.3772588836932535, "learning_rate": 7.589877897286928e-05, "loss": 0.9663, "step": 2548 }, { "epoch": 1.1549614861803352, "grad_norm": 0.2988499516084099, "learning_rate": 7.589319593447734e-05, "loss": 1.0134, "step": 2549 }, { "epoch": 1.1554145899410966, "grad_norm": 0.31621104729393334, "learning_rate": 7.58876093041716e-05, "loss": 0.9824, "step": 2550 }, { "epoch": 1.1558676937018577, "grad_norm": 0.48053727596197354, "learning_rate": 7.58820190825111e-05, "loss": 0.9951, "step": 2551 }, { "epoch": 1.1563207974626188, "grad_norm": 0.4990357868964069, "learning_rate": 7.587642527005527e-05, "loss": 1.0043, "step": 2552 }, { "epoch": 1.1567739012233802, "grad_norm": 0.4665755858938584, "learning_rate": 7.587082786736391e-05, "loss": 0.9902, "step": 2553 }, { "epoch": 1.1572270049841413, "grad_norm": 0.5120935474708824, "learning_rate": 7.586522687499718e-05, "loss": 1.0014, "step": 2554 }, { "epoch": 1.1576801087449025, "grad_norm": 0.5494860281225715, "learning_rate": 7.585962229351554e-05, "loss": 0.9753, "step": 2555 }, { "epoch": 1.1581332125056638, "grad_norm": 0.5235926392078417, "learning_rate": 7.585401412347986e-05, "loss": 1.0039, "step": 2556 }, { "epoch": 1.158586316266425, "grad_norm": 0.44384911911524855, "learning_rate": 7.58484023654514e-05, "loss": 0.9773, "step": 2557 }, { "epoch": 1.1590394200271863, "grad_norm": 0.36832679658549006, "learning_rate": 7.584278701999172e-05, "loss": 0.9663, "step": 2558 }, { "epoch": 1.1594925237879474, "grad_norm": 0.3402358884343581, "learning_rate": 7.583716808766275e-05, "loss": 0.9666, "step": 2559 }, { "epoch": 1.1599456275487086, "grad_norm": 0.43265722581748595, "learning_rate": 7.583154556902679e-05, "loss": 0.9489, "step": 2560 }, { "epoch": 1.16039873130947, "grad_norm": 0.559269770988186, "learning_rate": 7.582591946464652e-05, "loss": 0.9732, "step": 2561 }, { "epoch": 1.160851835070231, "grad_norm": 0.6148746422305191, "learning_rate": 7.582028977508495e-05, "loss": 0.9671, "step": 2562 }, { "epoch": 1.1613049388309924, "grad_norm": 0.5991511917764619, "learning_rate": 7.581465650090544e-05, "loss": 0.9699, "step": 2563 }, { "epoch": 1.1617580425917535, "grad_norm": 0.5357849403048932, "learning_rate": 7.580901964267173e-05, "loss": 0.9741, "step": 2564 }, { "epoch": 1.1622111463525147, "grad_norm": 0.4477246328271389, "learning_rate": 7.580337920094792e-05, "loss": 0.9879, "step": 2565 }, { "epoch": 1.162664250113276, "grad_norm": 0.4114544506655787, "learning_rate": 7.579773517629847e-05, "loss": 0.9857, "step": 2566 }, { "epoch": 1.1631173538740371, "grad_norm": 0.36797174495983204, "learning_rate": 7.579208756928816e-05, "loss": 1.0, "step": 2567 }, { "epoch": 1.1635704576347983, "grad_norm": 0.3395612413432217, "learning_rate": 7.578643638048219e-05, "loss": 0.9779, "step": 2568 }, { "epoch": 1.1640235613955596, "grad_norm": 0.4054541745222927, "learning_rate": 7.578078161044608e-05, "loss": 0.9904, "step": 2569 }, { "epoch": 1.1644766651563208, "grad_norm": 0.3613628110470456, "learning_rate": 7.577512325974571e-05, "loss": 0.9768, "step": 2570 }, { "epoch": 1.1649297689170819, "grad_norm": 0.37846705869860575, "learning_rate": 7.576946132894732e-05, "loss": 0.9629, "step": 2571 }, { "epoch": 1.1653828726778432, "grad_norm": 0.3991956071057166, "learning_rate": 7.576379581861752e-05, "loss": 0.98, "step": 2572 }, { "epoch": 1.1658359764386044, "grad_norm": 0.4083277631251796, "learning_rate": 7.575812672932326e-05, "loss": 0.9688, "step": 2573 }, { "epoch": 1.1662890801993657, "grad_norm": 0.3557608897999762, "learning_rate": 7.575245406163188e-05, "loss": 0.9839, "step": 2574 }, { "epoch": 1.1667421839601269, "grad_norm": 0.3691489707997441, "learning_rate": 7.574677781611104e-05, "loss": 0.9703, "step": 2575 }, { "epoch": 1.167195287720888, "grad_norm": 0.4821624233926667, "learning_rate": 7.574109799332878e-05, "loss": 0.9759, "step": 2576 }, { "epoch": 1.1676483914816493, "grad_norm": 0.5632372597254504, "learning_rate": 7.573541459385349e-05, "loss": 0.9877, "step": 2577 }, { "epoch": 1.1681014952424105, "grad_norm": 0.6080311868198225, "learning_rate": 7.572972761825393e-05, "loss": 0.9672, "step": 2578 }, { "epoch": 1.1685545990031718, "grad_norm": 0.6113269043371918, "learning_rate": 7.572403706709918e-05, "loss": 0.9846, "step": 2579 }, { "epoch": 1.169007702763933, "grad_norm": 0.5510450123592654, "learning_rate": 7.571834294095875e-05, "loss": 0.9832, "step": 2580 }, { "epoch": 1.169460806524694, "grad_norm": 0.4487437162188544, "learning_rate": 7.571264524040242e-05, "loss": 0.9572, "step": 2581 }, { "epoch": 1.1699139102854554, "grad_norm": 0.37170727081967353, "learning_rate": 7.570694396600041e-05, "loss": 1.0183, "step": 2582 }, { "epoch": 1.1703670140462166, "grad_norm": 0.36483943240779093, "learning_rate": 7.570123911832323e-05, "loss": 0.9764, "step": 2583 }, { "epoch": 1.170820117806978, "grad_norm": 0.3770114130572402, "learning_rate": 7.569553069794181e-05, "loss": 0.9826, "step": 2584 }, { "epoch": 1.171273221567739, "grad_norm": 0.4021299985024194, "learning_rate": 7.568981870542736e-05, "loss": 0.9786, "step": 2585 }, { "epoch": 1.1717263253285002, "grad_norm": 0.43833637979777856, "learning_rate": 7.568410314135153e-05, "loss": 0.9841, "step": 2586 }, { "epoch": 1.1721794290892615, "grad_norm": 0.44224834103580285, "learning_rate": 7.567838400628627e-05, "loss": 0.9862, "step": 2587 }, { "epoch": 1.1726325328500227, "grad_norm": 0.37352959611001674, "learning_rate": 7.567266130080392e-05, "loss": 0.9732, "step": 2588 }, { "epoch": 1.1730856366107838, "grad_norm": 0.3102618010905045, "learning_rate": 7.566693502547715e-05, "loss": 0.9901, "step": 2589 }, { "epoch": 1.1735387403715452, "grad_norm": 0.26621149720327364, "learning_rate": 7.566120518087902e-05, "loss": 0.9787, "step": 2590 }, { "epoch": 1.1739918441323063, "grad_norm": 0.21398626941949558, "learning_rate": 7.56554717675829e-05, "loss": 0.9773, "step": 2591 }, { "epoch": 1.1744449478930674, "grad_norm": 0.22748993960426175, "learning_rate": 7.564973478616257e-05, "loss": 0.9701, "step": 2592 }, { "epoch": 1.1748980516538288, "grad_norm": 0.2605348337265936, "learning_rate": 7.564399423719214e-05, "loss": 0.9702, "step": 2593 }, { "epoch": 1.17535115541459, "grad_norm": 0.2747341455708, "learning_rate": 7.563825012124607e-05, "loss": 0.9778, "step": 2594 }, { "epoch": 1.1758042591753513, "grad_norm": 0.3098935121104697, "learning_rate": 7.56325024388992e-05, "loss": 0.9795, "step": 2595 }, { "epoch": 1.1762573629361124, "grad_norm": 0.36048297473703966, "learning_rate": 7.56267511907267e-05, "loss": 0.9834, "step": 2596 }, { "epoch": 1.1767104666968735, "grad_norm": 0.3850699220829136, "learning_rate": 7.56209963773041e-05, "loss": 0.9718, "step": 2597 }, { "epoch": 1.1771635704576349, "grad_norm": 0.40262571507016026, "learning_rate": 7.561523799920734e-05, "loss": 0.9776, "step": 2598 }, { "epoch": 1.177616674218396, "grad_norm": 0.4005136480571103, "learning_rate": 7.560947605701265e-05, "loss": 1.0021, "step": 2599 }, { "epoch": 1.1780697779791574, "grad_norm": 0.5004965680340077, "learning_rate": 7.560371055129661e-05, "loss": 0.9666, "step": 2600 }, { "epoch": 1.1785228817399185, "grad_norm": 0.5702276705160897, "learning_rate": 7.559794148263624e-05, "loss": 0.9757, "step": 2601 }, { "epoch": 1.1789759855006796, "grad_norm": 0.649186386838457, "learning_rate": 7.559216885160883e-05, "loss": 0.978, "step": 2602 }, { "epoch": 1.179429089261441, "grad_norm": 0.7098238572729949, "learning_rate": 7.558639265879207e-05, "loss": 0.9824, "step": 2603 }, { "epoch": 1.179882193022202, "grad_norm": 0.7214986746371697, "learning_rate": 7.5580612904764e-05, "loss": 0.9953, "step": 2604 }, { "epoch": 1.1803352967829632, "grad_norm": 0.6363889827784596, "learning_rate": 7.5574829590103e-05, "loss": 0.9797, "step": 2605 }, { "epoch": 1.1807884005437246, "grad_norm": 0.5452837417509165, "learning_rate": 7.556904271538783e-05, "loss": 0.9715, "step": 2606 }, { "epoch": 1.1812415043044857, "grad_norm": 0.46834317271674564, "learning_rate": 7.556325228119761e-05, "loss": 0.9602, "step": 2607 }, { "epoch": 1.1816946080652468, "grad_norm": 0.4007924223657978, "learning_rate": 7.555745828811177e-05, "loss": 0.9883, "step": 2608 }, { "epoch": 1.1821477118260082, "grad_norm": 0.31901376400246245, "learning_rate": 7.555166073671015e-05, "loss": 0.9755, "step": 2609 }, { "epoch": 1.1826008155867693, "grad_norm": 0.3183490861080296, "learning_rate": 7.554585962757292e-05, "loss": 0.9693, "step": 2610 }, { "epoch": 1.1830539193475307, "grad_norm": 0.3886718464968299, "learning_rate": 7.554005496128062e-05, "loss": 0.9807, "step": 2611 }, { "epoch": 1.1835070231082918, "grad_norm": 0.4306634017659485, "learning_rate": 7.553424673841411e-05, "loss": 0.9793, "step": 2612 }, { "epoch": 1.183960126869053, "grad_norm": 0.4576988378946155, "learning_rate": 7.552843495955465e-05, "loss": 0.9902, "step": 2613 }, { "epoch": 1.1844132306298143, "grad_norm": 0.4719928412100822, "learning_rate": 7.552261962528383e-05, "loss": 0.9527, "step": 2614 }, { "epoch": 1.1848663343905754, "grad_norm": 0.4963430183739181, "learning_rate": 7.551680073618362e-05, "loss": 0.9894, "step": 2615 }, { "epoch": 1.1853194381513368, "grad_norm": 0.5648732969371314, "learning_rate": 7.551097829283631e-05, "loss": 0.9959, "step": 2616 }, { "epoch": 1.185772541912098, "grad_norm": 0.6171098008689008, "learning_rate": 7.550515229582458e-05, "loss": 1.0055, "step": 2617 }, { "epoch": 1.186225645672859, "grad_norm": 0.5947632621028823, "learning_rate": 7.549932274573144e-05, "loss": 0.9882, "step": 2618 }, { "epoch": 1.1866787494336204, "grad_norm": 0.5480402999775308, "learning_rate": 7.549348964314027e-05, "loss": 0.9943, "step": 2619 }, { "epoch": 1.1871318531943815, "grad_norm": 0.5105029402949444, "learning_rate": 7.548765298863482e-05, "loss": 0.9661, "step": 2620 }, { "epoch": 1.1875849569551427, "grad_norm": 0.4032684120121399, "learning_rate": 7.548181278279914e-05, "loss": 0.9965, "step": 2621 }, { "epoch": 1.188038060715904, "grad_norm": 0.3147238800726987, "learning_rate": 7.54759690262177e-05, "loss": 0.966, "step": 2622 }, { "epoch": 1.1884911644766651, "grad_norm": 0.37874526832035027, "learning_rate": 7.54701217194753e-05, "loss": 0.9909, "step": 2623 }, { "epoch": 1.1889442682374263, "grad_norm": 0.39769938405136923, "learning_rate": 7.546427086315707e-05, "loss": 0.9683, "step": 2624 }, { "epoch": 1.1893973719981876, "grad_norm": 0.3177883140642906, "learning_rate": 7.545841645784853e-05, "loss": 0.9756, "step": 2625 }, { "epoch": 1.1898504757589488, "grad_norm": 0.3269685102233105, "learning_rate": 7.545255850413554e-05, "loss": 0.9933, "step": 2626 }, { "epoch": 1.19030357951971, "grad_norm": 0.39973793487569287, "learning_rate": 7.544669700260434e-05, "loss": 0.997, "step": 2627 }, { "epoch": 1.1907566832804712, "grad_norm": 0.4102187615230679, "learning_rate": 7.54408319538415e-05, "loss": 0.9919, "step": 2628 }, { "epoch": 1.1912097870412324, "grad_norm": 0.3315260418608983, "learning_rate": 7.543496335843392e-05, "loss": 0.9859, "step": 2629 }, { "epoch": 1.1916628908019937, "grad_norm": 0.31143517359513795, "learning_rate": 7.54290912169689e-05, "loss": 0.9772, "step": 2630 }, { "epoch": 1.1921159945627549, "grad_norm": 0.36143225950608077, "learning_rate": 7.542321553003409e-05, "loss": 0.9796, "step": 2631 }, { "epoch": 1.1925690983235162, "grad_norm": 0.3387024833975662, "learning_rate": 7.541733629821745e-05, "loss": 1.0177, "step": 2632 }, { "epoch": 1.1930222020842773, "grad_norm": 0.29628384336754227, "learning_rate": 7.541145352210737e-05, "loss": 0.9586, "step": 2633 }, { "epoch": 1.1934753058450385, "grad_norm": 0.30983565491179427, "learning_rate": 7.540556720229253e-05, "loss": 0.9773, "step": 2634 }, { "epoch": 1.1939284096057998, "grad_norm": 0.34027391792048955, "learning_rate": 7.539967733936201e-05, "loss": 0.9744, "step": 2635 }, { "epoch": 1.194381513366561, "grad_norm": 0.35824264571261893, "learning_rate": 7.539378393390518e-05, "loss": 0.984, "step": 2636 }, { "epoch": 1.194834617127322, "grad_norm": 0.35619788458933704, "learning_rate": 7.538788698651183e-05, "loss": 1.0014, "step": 2637 }, { "epoch": 1.1952877208880834, "grad_norm": 0.4028243909445975, "learning_rate": 7.538198649777208e-05, "loss": 0.985, "step": 2638 }, { "epoch": 1.1957408246488446, "grad_norm": 0.46241404099752087, "learning_rate": 7.537608246827642e-05, "loss": 0.9821, "step": 2639 }, { "epoch": 1.1961939284096057, "grad_norm": 0.49717466134099736, "learning_rate": 7.537017489861566e-05, "loss": 0.9731, "step": 2640 }, { "epoch": 1.196647032170367, "grad_norm": 0.4834939696710326, "learning_rate": 7.536426378938098e-05, "loss": 0.9809, "step": 2641 }, { "epoch": 1.1971001359311282, "grad_norm": 0.43843444195575226, "learning_rate": 7.535834914116396e-05, "loss": 0.9747, "step": 2642 }, { "epoch": 1.1975532396918895, "grad_norm": 0.41970142439008756, "learning_rate": 7.535243095455643e-05, "loss": 0.9719, "step": 2643 }, { "epoch": 1.1980063434526507, "grad_norm": 0.42415122628391294, "learning_rate": 7.534650923015066e-05, "loss": 0.9491, "step": 2644 }, { "epoch": 1.1984594472134118, "grad_norm": 0.4191478359486049, "learning_rate": 7.534058396853928e-05, "loss": 0.9802, "step": 2645 }, { "epoch": 1.1989125509741732, "grad_norm": 0.4385356986869799, "learning_rate": 7.533465517031521e-05, "loss": 0.9802, "step": 2646 }, { "epoch": 1.1993656547349343, "grad_norm": 0.48853060614504457, "learning_rate": 7.532872283607177e-05, "loss": 0.9872, "step": 2647 }, { "epoch": 1.1998187584956956, "grad_norm": 0.511299522238483, "learning_rate": 7.532278696640263e-05, "loss": 1.0013, "step": 2648 }, { "epoch": 1.2002718622564568, "grad_norm": 0.4846642540395019, "learning_rate": 7.531684756190179e-05, "loss": 0.9723, "step": 2649 }, { "epoch": 1.200724966017218, "grad_norm": 0.4969114087355663, "learning_rate": 7.531090462316363e-05, "loss": 0.9669, "step": 2650 }, { "epoch": 1.2011780697779793, "grad_norm": 0.5458172989115532, "learning_rate": 7.530495815078287e-05, "loss": 0.9766, "step": 2651 }, { "epoch": 1.2016311735387404, "grad_norm": 0.5793044992413822, "learning_rate": 7.52990081453546e-05, "loss": 0.9723, "step": 2652 }, { "epoch": 1.2020842772995015, "grad_norm": 0.5824615876930286, "learning_rate": 7.529305460747422e-05, "loss": 0.9658, "step": 2653 }, { "epoch": 1.2025373810602629, "grad_norm": 0.5245031362806558, "learning_rate": 7.528709753773754e-05, "loss": 1.0183, "step": 2654 }, { "epoch": 1.202990484821024, "grad_norm": 0.4830820009678708, "learning_rate": 7.528113693674069e-05, "loss": 0.9934, "step": 2655 }, { "epoch": 1.2034435885817851, "grad_norm": 0.455696039616923, "learning_rate": 7.527517280508017e-05, "loss": 0.9774, "step": 2656 }, { "epoch": 1.2038966923425465, "grad_norm": 0.4700101892233884, "learning_rate": 7.52692051433528e-05, "loss": 0.9791, "step": 2657 }, { "epoch": 1.2043497961033076, "grad_norm": 0.4879996565199193, "learning_rate": 7.52632339521558e-05, "loss": 0.9823, "step": 2658 }, { "epoch": 1.204802899864069, "grad_norm": 0.49660395701044346, "learning_rate": 7.525725923208673e-05, "loss": 0.9865, "step": 2659 }, { "epoch": 1.20525600362483, "grad_norm": 0.4899430149130708, "learning_rate": 7.525128098374345e-05, "loss": 0.9996, "step": 2660 }, { "epoch": 1.2057091073855912, "grad_norm": 0.4366269345990885, "learning_rate": 7.524529920772426e-05, "loss": 0.9522, "step": 2661 }, { "epoch": 1.2061622111463526, "grad_norm": 0.3869624589340156, "learning_rate": 7.523931390462774e-05, "loss": 0.9529, "step": 2662 }, { "epoch": 1.2066153149071137, "grad_norm": 0.36213441965015125, "learning_rate": 7.52333250750529e-05, "loss": 1.0108, "step": 2663 }, { "epoch": 1.207068418667875, "grad_norm": 0.29214224766791086, "learning_rate": 7.5227332719599e-05, "loss": 0.9695, "step": 2664 }, { "epoch": 1.2075215224286362, "grad_norm": 0.3295497908998063, "learning_rate": 7.522133683886574e-05, "loss": 0.971, "step": 2665 }, { "epoch": 1.2079746261893973, "grad_norm": 0.3902394648405915, "learning_rate": 7.52153374334531e-05, "loss": 0.9994, "step": 2666 }, { "epoch": 1.2084277299501587, "grad_norm": 0.3565510616444739, "learning_rate": 7.520933450396154e-05, "loss": 0.9719, "step": 2667 }, { "epoch": 1.2088808337109198, "grad_norm": 0.29443638864021887, "learning_rate": 7.52033280509917e-05, "loss": 0.9672, "step": 2668 }, { "epoch": 1.209333937471681, "grad_norm": 0.2933709284403178, "learning_rate": 7.51973180751447e-05, "loss": 0.9697, "step": 2669 }, { "epoch": 1.2097870412324423, "grad_norm": 0.32247586266477823, "learning_rate": 7.519130457702197e-05, "loss": 0.9964, "step": 2670 }, { "epoch": 1.2102401449932034, "grad_norm": 0.3760471939503257, "learning_rate": 7.518528755722528e-05, "loss": 0.9311, "step": 2671 }, { "epoch": 1.2106932487539646, "grad_norm": 0.42721697039899414, "learning_rate": 7.517926701635679e-05, "loss": 0.9751, "step": 2672 }, { "epoch": 1.211146352514726, "grad_norm": 0.415201331602769, "learning_rate": 7.517324295501897e-05, "loss": 0.9868, "step": 2673 }, { "epoch": 1.211599456275487, "grad_norm": 0.3934373983360785, "learning_rate": 7.516721537381467e-05, "loss": 0.9862, "step": 2674 }, { "epoch": 1.2120525600362484, "grad_norm": 0.4633576099289559, "learning_rate": 7.516118427334708e-05, "loss": 0.985, "step": 2675 }, { "epoch": 1.2125056637970095, "grad_norm": 0.5063995198358326, "learning_rate": 7.515514965421976e-05, "loss": 0.9942, "step": 2676 }, { "epoch": 1.2129587675577707, "grad_norm": 0.41467576143285595, "learning_rate": 7.514911151703657e-05, "loss": 0.973, "step": 2677 }, { "epoch": 1.213411871318532, "grad_norm": 0.3098967289390347, "learning_rate": 7.51430698624018e-05, "loss": 0.9891, "step": 2678 }, { "epoch": 1.2138649750792931, "grad_norm": 0.26177510678090615, "learning_rate": 7.513702469092004e-05, "loss": 0.9808, "step": 2679 }, { "epoch": 1.2143180788400545, "grad_norm": 0.3269490324250073, "learning_rate": 7.513097600319624e-05, "loss": 0.9725, "step": 2680 }, { "epoch": 1.2147711826008156, "grad_norm": 0.38662268360827595, "learning_rate": 7.512492379983571e-05, "loss": 0.9833, "step": 2681 }, { "epoch": 1.2152242863615768, "grad_norm": 0.4424188594754017, "learning_rate": 7.51188680814441e-05, "loss": 0.9759, "step": 2682 }, { "epoch": 1.215677390122338, "grad_norm": 0.48372495857666087, "learning_rate": 7.511280884862742e-05, "loss": 0.9705, "step": 2683 }, { "epoch": 1.2161304938830992, "grad_norm": 0.5085520938358397, "learning_rate": 7.510674610199203e-05, "loss": 0.9662, "step": 2684 }, { "epoch": 1.2165835976438604, "grad_norm": 0.6113158184372213, "learning_rate": 7.510067984214467e-05, "loss": 0.9922, "step": 2685 }, { "epoch": 1.2170367014046217, "grad_norm": 0.4906580526729448, "learning_rate": 7.509461006969238e-05, "loss": 0.9898, "step": 2686 }, { "epoch": 1.2174898051653829, "grad_norm": 0.5453411184889848, "learning_rate": 7.508853678524257e-05, "loss": 1.0033, "step": 2687 }, { "epoch": 1.217942908926144, "grad_norm": 0.6168352747335786, "learning_rate": 7.508245998940302e-05, "loss": 0.9647, "step": 2688 }, { "epoch": 1.2183960126869053, "grad_norm": 0.7080059518821038, "learning_rate": 7.507637968278185e-05, "loss": 0.9699, "step": 2689 }, { "epoch": 1.2188491164476665, "grad_norm": 0.7568789115194798, "learning_rate": 7.507029586598752e-05, "loss": 0.9626, "step": 2690 }, { "epoch": 1.2193022202084278, "grad_norm": 0.6991482365820005, "learning_rate": 7.506420853962887e-05, "loss": 0.9775, "step": 2691 }, { "epoch": 1.219755323969189, "grad_norm": 0.602559136933077, "learning_rate": 7.505811770431504e-05, "loss": 0.9666, "step": 2692 }, { "epoch": 1.22020842772995, "grad_norm": 0.5485714046380532, "learning_rate": 7.505202336065559e-05, "loss": 0.957, "step": 2693 }, { "epoch": 1.2206615314907114, "grad_norm": 0.5097553335416338, "learning_rate": 7.504592550926037e-05, "loss": 0.9937, "step": 2694 }, { "epoch": 1.2211146352514726, "grad_norm": 0.4658385109194483, "learning_rate": 7.503982415073961e-05, "loss": 0.9853, "step": 2695 }, { "epoch": 1.221567739012234, "grad_norm": 0.4922311881794485, "learning_rate": 7.50337192857039e-05, "loss": 0.9617, "step": 2696 }, { "epoch": 1.222020842772995, "grad_norm": 0.4611089343427015, "learning_rate": 7.502761091476416e-05, "loss": 0.9785, "step": 2697 }, { "epoch": 1.2224739465337562, "grad_norm": 0.37514076904878263, "learning_rate": 7.502149903853166e-05, "loss": 1.0049, "step": 2698 }, { "epoch": 1.2229270502945175, "grad_norm": 0.357210260968486, "learning_rate": 7.501538365761804e-05, "loss": 0.9746, "step": 2699 }, { "epoch": 1.2233801540552787, "grad_norm": 0.3324926585471427, "learning_rate": 7.500926477263527e-05, "loss": 0.996, "step": 2700 }, { "epoch": 1.2238332578160398, "grad_norm": 0.33093011620615514, "learning_rate": 7.50031423841957e-05, "loss": 0.9866, "step": 2701 }, { "epoch": 1.2242863615768012, "grad_norm": 0.40197919122956843, "learning_rate": 7.499701649291199e-05, "loss": 0.9922, "step": 2702 }, { "epoch": 1.2247394653375623, "grad_norm": 0.4759793807189813, "learning_rate": 7.49908870993972e-05, "loss": 0.9893, "step": 2703 }, { "epoch": 1.2251925690983234, "grad_norm": 0.5444815844365686, "learning_rate": 7.498475420426467e-05, "loss": 0.9972, "step": 2704 }, { "epoch": 1.2256456728590848, "grad_norm": 0.5594654695034622, "learning_rate": 7.497861780812816e-05, "loss": 0.9498, "step": 2705 }, { "epoch": 1.226098776619846, "grad_norm": 0.4856525513131899, "learning_rate": 7.497247791160174e-05, "loss": 0.9765, "step": 2706 }, { "epoch": 1.2265518803806073, "grad_norm": 0.42728600407050527, "learning_rate": 7.496633451529986e-05, "loss": 0.9846, "step": 2707 }, { "epoch": 1.2270049841413684, "grad_norm": 0.45150555805933895, "learning_rate": 7.496018761983731e-05, "loss": 0.9892, "step": 2708 }, { "epoch": 1.2274580879021295, "grad_norm": 0.5051150191103931, "learning_rate": 7.49540372258292e-05, "loss": 0.9872, "step": 2709 }, { "epoch": 1.2279111916628909, "grad_norm": 0.47467095045700064, "learning_rate": 7.494788333389103e-05, "loss": 0.9693, "step": 2710 }, { "epoch": 1.228364295423652, "grad_norm": 0.4704007651558209, "learning_rate": 7.494172594463863e-05, "loss": 0.9841, "step": 2711 }, { "epoch": 1.2288173991844134, "grad_norm": 0.4131405223542158, "learning_rate": 7.493556505868817e-05, "loss": 0.9978, "step": 2712 }, { "epoch": 1.2292705029451745, "grad_norm": 0.3340559084331828, "learning_rate": 7.492940067665619e-05, "loss": 0.9868, "step": 2713 }, { "epoch": 1.2297236067059356, "grad_norm": 0.3360664391024459, "learning_rate": 7.49232327991596e-05, "loss": 0.9823, "step": 2714 }, { "epoch": 1.230176710466697, "grad_norm": 0.38950018433814376, "learning_rate": 7.491706142681561e-05, "loss": 0.9788, "step": 2715 }, { "epoch": 1.230629814227458, "grad_norm": 0.40712762382293, "learning_rate": 7.491088656024179e-05, "loss": 0.9894, "step": 2716 }, { "epoch": 1.2310829179882192, "grad_norm": 0.3388088059324055, "learning_rate": 7.49047082000561e-05, "loss": 0.9553, "step": 2717 }, { "epoch": 1.2315360217489806, "grad_norm": 0.37155903914248517, "learning_rate": 7.48985263468768e-05, "loss": 0.9911, "step": 2718 }, { "epoch": 1.2319891255097417, "grad_norm": 0.42684710914819474, "learning_rate": 7.489234100132254e-05, "loss": 0.9885, "step": 2719 }, { "epoch": 1.2324422292705028, "grad_norm": 0.4718075082981923, "learning_rate": 7.488615216401231e-05, "loss": 0.984, "step": 2720 }, { "epoch": 1.2328953330312642, "grad_norm": 0.5094922302117105, "learning_rate": 7.487995983556541e-05, "loss": 0.961, "step": 2721 }, { "epoch": 1.2333484367920253, "grad_norm": 0.5291613569014072, "learning_rate": 7.487376401660154e-05, "loss": 0.9898, "step": 2722 }, { "epoch": 1.2338015405527867, "grad_norm": 0.5106408641151948, "learning_rate": 7.486756470774073e-05, "loss": 0.9718, "step": 2723 }, { "epoch": 1.2342546443135478, "grad_norm": 0.49063876183963107, "learning_rate": 7.486136190960336e-05, "loss": 0.9985, "step": 2724 }, { "epoch": 1.234707748074309, "grad_norm": 0.46730549974095414, "learning_rate": 7.485515562281014e-05, "loss": 0.9794, "step": 2725 }, { "epoch": 1.2351608518350703, "grad_norm": 0.41987195452972187, "learning_rate": 7.484894584798218e-05, "loss": 0.9956, "step": 2726 }, { "epoch": 1.2356139555958314, "grad_norm": 0.3644700528328795, "learning_rate": 7.484273258574088e-05, "loss": 0.9806, "step": 2727 }, { "epoch": 1.2360670593565928, "grad_norm": 0.3669023647802127, "learning_rate": 7.483651583670801e-05, "loss": 0.9732, "step": 2728 }, { "epoch": 1.236520163117354, "grad_norm": 0.4368837559156083, "learning_rate": 7.483029560150572e-05, "loss": 0.9694, "step": 2729 }, { "epoch": 1.236973266878115, "grad_norm": 0.44019331210835155, "learning_rate": 7.482407188075646e-05, "loss": 0.9958, "step": 2730 }, { "epoch": 1.2374263706388764, "grad_norm": 0.39101872189968534, "learning_rate": 7.481784467508308e-05, "loss": 0.9696, "step": 2731 }, { "epoch": 1.2378794743996375, "grad_norm": 0.419624476101042, "learning_rate": 7.481161398510873e-05, "loss": 0.9615, "step": 2732 }, { "epoch": 1.2383325781603987, "grad_norm": 0.4995060807136828, "learning_rate": 7.480537981145692e-05, "loss": 0.9744, "step": 2733 }, { "epoch": 1.23878568192116, "grad_norm": 0.5876968607568501, "learning_rate": 7.479914215475153e-05, "loss": 0.9816, "step": 2734 }, { "epoch": 1.2392387856819211, "grad_norm": 0.6699064930663945, "learning_rate": 7.479290101561678e-05, "loss": 0.9816, "step": 2735 }, { "epoch": 1.2396918894426823, "grad_norm": 0.7528651502901476, "learning_rate": 7.478665639467722e-05, "loss": 0.9643, "step": 2736 }, { "epoch": 1.2401449932034436, "grad_norm": 0.866327179434653, "learning_rate": 7.478040829255779e-05, "loss": 0.9741, "step": 2737 }, { "epoch": 1.2405980969642048, "grad_norm": 0.9764221210061519, "learning_rate": 7.477415670988373e-05, "loss": 0.9436, "step": 2738 }, { "epoch": 1.241051200724966, "grad_norm": 1.0052350805683492, "learning_rate": 7.476790164728065e-05, "loss": 0.9782, "step": 2739 }, { "epoch": 1.2415043044857272, "grad_norm": 0.8637055320617385, "learning_rate": 7.476164310537453e-05, "loss": 1.0072, "step": 2740 }, { "epoch": 1.2419574082464884, "grad_norm": 0.5441356318914148, "learning_rate": 7.475538108479165e-05, "loss": 0.9818, "step": 2741 }, { "epoch": 1.2424105120072497, "grad_norm": 0.3472776364062674, "learning_rate": 7.474911558615869e-05, "loss": 0.9738, "step": 2742 }, { "epoch": 1.2428636157680109, "grad_norm": 0.49523004513154245, "learning_rate": 7.474284661010262e-05, "loss": 0.9842, "step": 2743 }, { "epoch": 1.2433167195287722, "grad_norm": 0.5914480714290301, "learning_rate": 7.473657415725083e-05, "loss": 0.9692, "step": 2744 }, { "epoch": 1.2437698232895333, "grad_norm": 0.5184586461019769, "learning_rate": 7.473029822823098e-05, "loss": 0.972, "step": 2745 }, { "epoch": 1.2442229270502945, "grad_norm": 0.3611296725286947, "learning_rate": 7.472401882367114e-05, "loss": 0.9829, "step": 2746 }, { "epoch": 1.2446760308110558, "grad_norm": 0.3864791279352818, "learning_rate": 7.47177359441997e-05, "loss": 0.9755, "step": 2747 }, { "epoch": 1.245129134571817, "grad_norm": 0.3662113467163379, "learning_rate": 7.47114495904454e-05, "loss": 0.9796, "step": 2748 }, { "epoch": 1.245582238332578, "grad_norm": 0.29531381952430164, "learning_rate": 7.470515976303732e-05, "loss": 0.9732, "step": 2749 }, { "epoch": 1.2460353420933394, "grad_norm": 0.3285881655348586, "learning_rate": 7.469886646260493e-05, "loss": 0.9766, "step": 2750 }, { "epoch": 1.2464884458541006, "grad_norm": 0.3591373796760361, "learning_rate": 7.469256968977797e-05, "loss": 0.9626, "step": 2751 }, { "epoch": 1.2469415496148617, "grad_norm": 0.3255129701280045, "learning_rate": 7.468626944518662e-05, "loss": 0.9649, "step": 2752 }, { "epoch": 1.247394653375623, "grad_norm": 0.3593099352001449, "learning_rate": 7.467996572946132e-05, "loss": 0.9653, "step": 2753 }, { "epoch": 1.2478477571363842, "grad_norm": 0.3773241945159573, "learning_rate": 7.467365854323293e-05, "loss": 0.9694, "step": 2754 }, { "epoch": 1.2483008608971455, "grad_norm": 0.350708473207466, "learning_rate": 7.466734788713259e-05, "loss": 0.9718, "step": 2755 }, { "epoch": 1.2487539646579067, "grad_norm": 0.3359199384307156, "learning_rate": 7.466103376179183e-05, "loss": 0.9757, "step": 2756 }, { "epoch": 1.2492070684186678, "grad_norm": 0.3938374302287242, "learning_rate": 7.465471616784254e-05, "loss": 0.9572, "step": 2757 }, { "epoch": 1.2496601721794292, "grad_norm": 0.4696560335331136, "learning_rate": 7.464839510591693e-05, "loss": 0.991, "step": 2758 }, { "epoch": 1.2501132759401903, "grad_norm": 0.5225624524341596, "learning_rate": 7.464207057664755e-05, "loss": 0.9736, "step": 2759 }, { "epoch": 1.2505663797009516, "grad_norm": 0.506806030825011, "learning_rate": 7.46357425806673e-05, "loss": 0.9841, "step": 2760 }, { "epoch": 1.2510194834617128, "grad_norm": 0.4792455089198548, "learning_rate": 7.462941111860949e-05, "loss": 0.9653, "step": 2761 }, { "epoch": 1.251472587222474, "grad_norm": 0.4835059989444514, "learning_rate": 7.462307619110766e-05, "loss": 0.995, "step": 2762 }, { "epoch": 1.251925690983235, "grad_norm": 0.46360763874004574, "learning_rate": 7.461673779879579e-05, "loss": 0.9625, "step": 2763 }, { "epoch": 1.2523787947439964, "grad_norm": 0.40336922190973473, "learning_rate": 7.461039594230817e-05, "loss": 0.9683, "step": 2764 }, { "epoch": 1.2528318985047575, "grad_norm": 0.35194165224931395, "learning_rate": 7.460405062227947e-05, "loss": 0.9689, "step": 2765 }, { "epoch": 1.2532850022655189, "grad_norm": 0.33297875532200333, "learning_rate": 7.459770183934465e-05, "loss": 1.0007, "step": 2766 }, { "epoch": 1.25373810602628, "grad_norm": 0.3620092806120333, "learning_rate": 7.459134959413906e-05, "loss": 0.9786, "step": 2767 }, { "epoch": 1.2541912097870411, "grad_norm": 0.39543062352057523, "learning_rate": 7.458499388729837e-05, "loss": 0.9669, "step": 2768 }, { "epoch": 1.2546443135478025, "grad_norm": 0.4517188395401382, "learning_rate": 7.457863471945864e-05, "loss": 0.9435, "step": 2769 }, { "epoch": 1.2550974173085636, "grad_norm": 0.5002285160629518, "learning_rate": 7.45722720912562e-05, "loss": 0.9898, "step": 2770 }, { "epoch": 1.255550521069325, "grad_norm": 0.5221833919013155, "learning_rate": 7.456590600332782e-05, "loss": 0.9889, "step": 2771 }, { "epoch": 1.256003624830086, "grad_norm": 0.5000193973856186, "learning_rate": 7.455953645631054e-05, "loss": 0.9889, "step": 2772 }, { "epoch": 1.2564567285908472, "grad_norm": 0.45937926647587496, "learning_rate": 7.455316345084179e-05, "loss": 0.9801, "step": 2773 }, { "epoch": 1.2569098323516086, "grad_norm": 0.42653041231143995, "learning_rate": 7.454678698755932e-05, "loss": 0.9957, "step": 2774 }, { "epoch": 1.2573629361123697, "grad_norm": 0.3661340080117397, "learning_rate": 7.454040706710125e-05, "loss": 0.9889, "step": 2775 }, { "epoch": 1.257816039873131, "grad_norm": 0.3397115677953029, "learning_rate": 7.453402369010602e-05, "loss": 0.9551, "step": 2776 }, { "epoch": 1.2582691436338922, "grad_norm": 0.3798795300436346, "learning_rate": 7.452763685721243e-05, "loss": 0.9983, "step": 2777 }, { "epoch": 1.2587222473946533, "grad_norm": 0.39495794447197924, "learning_rate": 7.452124656905963e-05, "loss": 0.9912, "step": 2778 }, { "epoch": 1.2591753511554145, "grad_norm": 0.4211794222417007, "learning_rate": 7.451485282628713e-05, "loss": 0.9885, "step": 2779 }, { "epoch": 1.2596284549161758, "grad_norm": 0.46753204128552384, "learning_rate": 7.450845562953472e-05, "loss": 0.9639, "step": 2780 }, { "epoch": 1.260081558676937, "grad_norm": 0.47808807063769115, "learning_rate": 7.450205497944262e-05, "loss": 0.9708, "step": 2781 }, { "epoch": 1.2605346624376983, "grad_norm": 0.46403000358856866, "learning_rate": 7.449565087665135e-05, "loss": 0.9795, "step": 2782 }, { "epoch": 1.2609877661984594, "grad_norm": 0.43866659956473236, "learning_rate": 7.448924332180178e-05, "loss": 0.9687, "step": 2783 }, { "epoch": 1.2614408699592206, "grad_norm": 0.4340450141085792, "learning_rate": 7.44828323155351e-05, "loss": 0.9815, "step": 2784 }, { "epoch": 1.261893973719982, "grad_norm": 0.38292425870764646, "learning_rate": 7.447641785849293e-05, "loss": 0.975, "step": 2785 }, { "epoch": 1.262347077480743, "grad_norm": 0.408798863522611, "learning_rate": 7.446999995131715e-05, "loss": 0.9845, "step": 2786 }, { "epoch": 1.2628001812415044, "grad_norm": 0.4711731369727142, "learning_rate": 7.446357859465e-05, "loss": 0.9815, "step": 2787 }, { "epoch": 1.2632532850022655, "grad_norm": 0.5265129541021558, "learning_rate": 7.445715378913412e-05, "loss": 0.9915, "step": 2788 }, { "epoch": 1.2637063887630267, "grad_norm": 0.4461862598801834, "learning_rate": 7.445072553541241e-05, "loss": 0.9896, "step": 2789 }, { "epoch": 1.264159492523788, "grad_norm": 0.4318425028013618, "learning_rate": 7.444429383412818e-05, "loss": 0.9814, "step": 2790 }, { "epoch": 1.2646125962845491, "grad_norm": 0.42115995476256085, "learning_rate": 7.443785868592508e-05, "loss": 0.9732, "step": 2791 }, { "epoch": 1.2650657000453105, "grad_norm": 0.3670188266848668, "learning_rate": 7.443142009144707e-05, "loss": 1.0007, "step": 2792 }, { "epoch": 1.2655188038060716, "grad_norm": 0.3548384932867155, "learning_rate": 7.442497805133848e-05, "loss": 0.9685, "step": 2793 }, { "epoch": 1.2659719075668328, "grad_norm": 0.35234473459973703, "learning_rate": 7.441853256624397e-05, "loss": 0.9787, "step": 2794 }, { "epoch": 1.2664250113275939, "grad_norm": 0.30526929021667143, "learning_rate": 7.441208363680858e-05, "loss": 0.9518, "step": 2795 }, { "epoch": 1.2668781150883552, "grad_norm": 0.2662545316266619, "learning_rate": 7.440563126367765e-05, "loss": 0.9912, "step": 2796 }, { "epoch": 1.2673312188491164, "grad_norm": 0.3129987366868706, "learning_rate": 7.439917544749687e-05, "loss": 0.9784, "step": 2797 }, { "epoch": 1.2677843226098777, "grad_norm": 0.24398222978269152, "learning_rate": 7.439271618891232e-05, "loss": 0.9912, "step": 2798 }, { "epoch": 1.2682374263706389, "grad_norm": 0.29231274578390837, "learning_rate": 7.438625348857037e-05, "loss": 0.994, "step": 2799 }, { "epoch": 1.2686905301314, "grad_norm": 0.338513580633911, "learning_rate": 7.437978734711776e-05, "loss": 0.9812, "step": 2800 }, { "epoch": 1.2691436338921613, "grad_norm": 0.3361130613924187, "learning_rate": 7.437331776520159e-05, "loss": 0.9534, "step": 2801 }, { "epoch": 1.2695967376529225, "grad_norm": 0.3577400533082603, "learning_rate": 7.436684474346928e-05, "loss": 0.9874, "step": 2802 }, { "epoch": 1.2700498414136838, "grad_norm": 0.439751423737288, "learning_rate": 7.436036828256857e-05, "loss": 0.9807, "step": 2803 }, { "epoch": 1.270502945174445, "grad_norm": 0.5239136954982951, "learning_rate": 7.43538883831476e-05, "loss": 0.981, "step": 2804 }, { "epoch": 1.270956048935206, "grad_norm": 0.6033584363945218, "learning_rate": 7.434740504585483e-05, "loss": 0.9994, "step": 2805 }, { "epoch": 1.2714091526959674, "grad_norm": 0.5951366942512416, "learning_rate": 7.434091827133906e-05, "loss": 0.9843, "step": 2806 }, { "epoch": 1.2718622564567286, "grad_norm": 0.48780431617827696, "learning_rate": 7.433442806024943e-05, "loss": 0.9852, "step": 2807 }, { "epoch": 1.27231536021749, "grad_norm": 0.3831420761917604, "learning_rate": 7.432793441323544e-05, "loss": 0.9678, "step": 2808 }, { "epoch": 1.272768463978251, "grad_norm": 0.3402829391507362, "learning_rate": 7.432143733094691e-05, "loss": 0.9788, "step": 2809 }, { "epoch": 1.2732215677390122, "grad_norm": 0.321149820911563, "learning_rate": 7.431493681403403e-05, "loss": 0.9624, "step": 2810 }, { "epoch": 1.2736746714997733, "grad_norm": 0.3330737673094195, "learning_rate": 7.430843286314732e-05, "loss": 0.973, "step": 2811 }, { "epoch": 1.2741277752605347, "grad_norm": 0.337751230460643, "learning_rate": 7.430192547893763e-05, "loss": 0.9756, "step": 2812 }, { "epoch": 1.2745808790212958, "grad_norm": 0.3610747688839774, "learning_rate": 7.429541466205619e-05, "loss": 1.0077, "step": 2813 }, { "epoch": 1.2750339827820572, "grad_norm": 0.33637123112788814, "learning_rate": 7.428890041315454e-05, "loss": 0.9847, "step": 2814 }, { "epoch": 1.2754870865428183, "grad_norm": 0.30788702893515035, "learning_rate": 7.428238273288457e-05, "loss": 0.968, "step": 2815 }, { "epoch": 1.2759401903035794, "grad_norm": 0.3212344970659931, "learning_rate": 7.427586162189853e-05, "loss": 0.9858, "step": 2816 }, { "epoch": 1.2763932940643408, "grad_norm": 0.3861546020508266, "learning_rate": 7.426933708084902e-05, "loss": 0.9658, "step": 2817 }, { "epoch": 1.276846397825102, "grad_norm": 0.3915678224416062, "learning_rate": 7.426280911038893e-05, "loss": 0.977, "step": 2818 }, { "epoch": 1.2772995015858633, "grad_norm": 0.3951002552516464, "learning_rate": 7.425627771117154e-05, "loss": 0.9807, "step": 2819 }, { "epoch": 1.2777526053466244, "grad_norm": 0.38444733017591226, "learning_rate": 7.424974288385048e-05, "loss": 0.9674, "step": 2820 }, { "epoch": 1.2782057091073855, "grad_norm": 0.3385384804519514, "learning_rate": 7.424320462907969e-05, "loss": 1.0059, "step": 2821 }, { "epoch": 1.2786588128681469, "grad_norm": 0.2949209573016732, "learning_rate": 7.423666294751345e-05, "loss": 0.9743, "step": 2822 }, { "epoch": 1.279111916628908, "grad_norm": 0.271127772086665, "learning_rate": 7.423011783980645e-05, "loss": 0.9808, "step": 2823 }, { "epoch": 1.2795650203896693, "grad_norm": 0.28822481427025753, "learning_rate": 7.422356930661362e-05, "loss": 0.9965, "step": 2824 }, { "epoch": 1.2800181241504305, "grad_norm": 0.2583906256752606, "learning_rate": 7.421701734859032e-05, "loss": 0.9737, "step": 2825 }, { "epoch": 1.2804712279111916, "grad_norm": 0.30297858605485495, "learning_rate": 7.421046196639222e-05, "loss": 0.9678, "step": 2826 }, { "epoch": 1.2809243316719527, "grad_norm": 0.3163045989075176, "learning_rate": 7.420390316067531e-05, "loss": 0.9798, "step": 2827 }, { "epoch": 1.281377435432714, "grad_norm": 0.3234710051683798, "learning_rate": 7.419734093209597e-05, "loss": 0.9869, "step": 2828 }, { "epoch": 1.2818305391934752, "grad_norm": 0.30450679663010377, "learning_rate": 7.419077528131088e-05, "loss": 0.9791, "step": 2829 }, { "epoch": 1.2822836429542366, "grad_norm": 0.3391010033859362, "learning_rate": 7.418420620897708e-05, "loss": 0.9721, "step": 2830 }, { "epoch": 1.2827367467149977, "grad_norm": 0.3891929710644479, "learning_rate": 7.417763371575196e-05, "loss": 0.9757, "step": 2831 }, { "epoch": 1.2831898504757588, "grad_norm": 0.40879858442787337, "learning_rate": 7.417105780229323e-05, "loss": 0.9787, "step": 2832 }, { "epoch": 1.2836429542365202, "grad_norm": 0.38294728704673714, "learning_rate": 7.416447846925897e-05, "loss": 0.9794, "step": 2833 }, { "epoch": 1.2840960579972813, "grad_norm": 0.4370163392206628, "learning_rate": 7.415789571730759e-05, "loss": 0.9866, "step": 2834 }, { "epoch": 1.2845491617580427, "grad_norm": 0.5091695660684329, "learning_rate": 7.415130954709784e-05, "loss": 1.0083, "step": 2835 }, { "epoch": 1.2850022655188038, "grad_norm": 0.5969425416257151, "learning_rate": 7.41447199592888e-05, "loss": 0.9823, "step": 2836 }, { "epoch": 1.285455369279565, "grad_norm": 0.6853550754435488, "learning_rate": 7.413812695453993e-05, "loss": 0.9741, "step": 2837 }, { "epoch": 1.2859084730403263, "grad_norm": 0.7805757205273532, "learning_rate": 7.413153053351097e-05, "loss": 0.9846, "step": 2838 }, { "epoch": 1.2863615768010874, "grad_norm": 0.8408356726627825, "learning_rate": 7.412493069686209e-05, "loss": 0.9554, "step": 2839 }, { "epoch": 1.2868146805618488, "grad_norm": 0.8041899545033028, "learning_rate": 7.41183274452537e-05, "loss": 0.9839, "step": 2840 }, { "epoch": 1.28726778432261, "grad_norm": 0.6840701913931332, "learning_rate": 7.411172077934663e-05, "loss": 0.977, "step": 2841 }, { "epoch": 1.287720888083371, "grad_norm": 0.6138843368783012, "learning_rate": 7.410511069980201e-05, "loss": 0.9548, "step": 2842 }, { "epoch": 1.2881739918441322, "grad_norm": 0.5527032855874605, "learning_rate": 7.409849720728135e-05, "loss": 0.9835, "step": 2843 }, { "epoch": 1.2886270956048935, "grad_norm": 0.4638081720283939, "learning_rate": 7.409188030244645e-05, "loss": 0.9741, "step": 2844 }, { "epoch": 1.2890801993656547, "grad_norm": 0.3507405127218755, "learning_rate": 7.408525998595948e-05, "loss": 0.9805, "step": 2845 }, { "epoch": 1.289533303126416, "grad_norm": 0.34691604561037825, "learning_rate": 7.407863625848298e-05, "loss": 0.9957, "step": 2846 }, { "epoch": 1.2899864068871771, "grad_norm": 0.38827691425541017, "learning_rate": 7.407200912067978e-05, "loss": 0.9571, "step": 2847 }, { "epoch": 1.2904395106479383, "grad_norm": 0.3981741768383386, "learning_rate": 7.406537857321306e-05, "loss": 1.008, "step": 2848 }, { "epoch": 1.2908926144086996, "grad_norm": 0.4659906542999362, "learning_rate": 7.405874461674638e-05, "loss": 0.9617, "step": 2849 }, { "epoch": 1.2913457181694608, "grad_norm": 0.5453020200739536, "learning_rate": 7.40521072519436e-05, "loss": 0.9774, "step": 2850 }, { "epoch": 1.291798821930222, "grad_norm": 0.5376842936047692, "learning_rate": 7.404546647946893e-05, "loss": 0.9837, "step": 2851 }, { "epoch": 1.2922519256909832, "grad_norm": 0.46403019272738516, "learning_rate": 7.403882229998694e-05, "loss": 0.9744, "step": 2852 }, { "epoch": 1.2927050294517444, "grad_norm": 0.5013839185617786, "learning_rate": 7.403217471416254e-05, "loss": 0.979, "step": 2853 }, { "epoch": 1.2931581332125057, "grad_norm": 0.4925295881290431, "learning_rate": 7.402552372266094e-05, "loss": 0.9731, "step": 2854 }, { "epoch": 1.2936112369732669, "grad_norm": 0.4725710635881239, "learning_rate": 7.401886932614774e-05, "loss": 0.9689, "step": 2855 }, { "epoch": 1.2940643407340282, "grad_norm": 0.5165783137854884, "learning_rate": 7.401221152528884e-05, "loss": 0.9536, "step": 2856 }, { "epoch": 1.2945174444947893, "grad_norm": 0.4936540324435497, "learning_rate": 7.400555032075054e-05, "loss": 0.9679, "step": 2857 }, { "epoch": 1.2949705482555505, "grad_norm": 0.36533179460335014, "learning_rate": 7.399888571319939e-05, "loss": 0.9752, "step": 2858 }, { "epoch": 1.2954236520163116, "grad_norm": 0.29991236175099634, "learning_rate": 7.399221770330239e-05, "loss": 0.9782, "step": 2859 }, { "epoch": 1.295876755777073, "grad_norm": 0.34837326243518124, "learning_rate": 7.398554629172676e-05, "loss": 0.9767, "step": 2860 }, { "epoch": 1.296329859537834, "grad_norm": 0.3075003715981375, "learning_rate": 7.397887147914017e-05, "loss": 0.922, "step": 2861 }, { "epoch": 1.2967829632985954, "grad_norm": 0.28467622362705003, "learning_rate": 7.397219326621057e-05, "loss": 0.9877, "step": 2862 }, { "epoch": 1.2972360670593566, "grad_norm": 0.307243170015877, "learning_rate": 7.396551165360627e-05, "loss": 0.9792, "step": 2863 }, { "epoch": 1.2976891708201177, "grad_norm": 0.26319042454020086, "learning_rate": 7.39588266419959e-05, "loss": 0.9836, "step": 2864 }, { "epoch": 1.298142274580879, "grad_norm": 0.30030327850118016, "learning_rate": 7.395213823204845e-05, "loss": 0.9734, "step": 2865 }, { "epoch": 1.2985953783416402, "grad_norm": 0.3901923037677621, "learning_rate": 7.394544642443325e-05, "loss": 0.9695, "step": 2866 }, { "epoch": 1.2990484821024015, "grad_norm": 0.39149518867616445, "learning_rate": 7.393875121981997e-05, "loss": 0.9695, "step": 2867 }, { "epoch": 1.2995015858631627, "grad_norm": 0.40486820083295644, "learning_rate": 7.393205261887861e-05, "loss": 0.9699, "step": 2868 }, { "epoch": 1.2999546896239238, "grad_norm": 0.4710399499626502, "learning_rate": 7.39253506222795e-05, "loss": 0.97, "step": 2869 }, { "epoch": 1.3004077933846852, "grad_norm": 0.5261359027137591, "learning_rate": 7.391864523069333e-05, "loss": 0.9704, "step": 2870 }, { "epoch": 1.3008608971454463, "grad_norm": 0.5143419529986429, "learning_rate": 7.391193644479113e-05, "loss": 0.976, "step": 2871 }, { "epoch": 1.3013140009062076, "grad_norm": 0.41999211449533796, "learning_rate": 7.390522426524427e-05, "loss": 0.9724, "step": 2872 }, { "epoch": 1.3017671046669688, "grad_norm": 0.36342859931294585, "learning_rate": 7.389850869272442e-05, "loss": 0.9743, "step": 2873 }, { "epoch": 1.30222020842773, "grad_norm": 0.3347174886083207, "learning_rate": 7.389178972790365e-05, "loss": 1.0002, "step": 2874 }, { "epoch": 1.302673312188491, "grad_norm": 0.30008222092061676, "learning_rate": 7.388506737145436e-05, "loss": 0.983, "step": 2875 }, { "epoch": 1.3031264159492524, "grad_norm": 0.2965313343117334, "learning_rate": 7.387834162404924e-05, "loss": 0.9772, "step": 2876 }, { "epoch": 1.3035795197100135, "grad_norm": 0.29921626633063286, "learning_rate": 7.387161248636136e-05, "loss": 0.9938, "step": 2877 }, { "epoch": 1.3040326234707749, "grad_norm": 0.31000183236517553, "learning_rate": 7.386487995906413e-05, "loss": 0.9399, "step": 2878 }, { "epoch": 1.304485727231536, "grad_norm": 0.3435043180193703, "learning_rate": 7.385814404283128e-05, "loss": 0.9884, "step": 2879 }, { "epoch": 1.3049388309922971, "grad_norm": 0.3973007607743587, "learning_rate": 7.385140473833688e-05, "loss": 0.9767, "step": 2880 }, { "epoch": 1.3053919347530585, "grad_norm": 0.453173665591289, "learning_rate": 7.384466204625536e-05, "loss": 0.9851, "step": 2881 }, { "epoch": 1.3058450385138196, "grad_norm": 0.4750569157916338, "learning_rate": 7.383791596726146e-05, "loss": 0.9737, "step": 2882 }, { "epoch": 1.306298142274581, "grad_norm": 0.4296623608320374, "learning_rate": 7.383116650203031e-05, "loss": 0.9683, "step": 2883 }, { "epoch": 1.306751246035342, "grad_norm": 0.364035841195653, "learning_rate": 7.382441365123733e-05, "loss": 0.9729, "step": 2884 }, { "epoch": 1.3072043497961032, "grad_norm": 0.3657450133172182, "learning_rate": 7.381765741555827e-05, "loss": 0.9807, "step": 2885 }, { "epoch": 1.3076574535568646, "grad_norm": 0.4412993253596879, "learning_rate": 7.381089779566924e-05, "loss": 0.9931, "step": 2886 }, { "epoch": 1.3081105573176257, "grad_norm": 0.46938437778826114, "learning_rate": 7.380413479224674e-05, "loss": 0.9888, "step": 2887 }, { "epoch": 1.308563661078387, "grad_norm": 0.4250322601708508, "learning_rate": 7.379736840596751e-05, "loss": 0.9807, "step": 2888 }, { "epoch": 1.3090167648391482, "grad_norm": 0.4068073874346438, "learning_rate": 7.37905986375087e-05, "loss": 0.999, "step": 2889 }, { "epoch": 1.3094698685999093, "grad_norm": 0.4060173545290123, "learning_rate": 7.378382548754778e-05, "loss": 0.9826, "step": 2890 }, { "epoch": 1.3099229723606705, "grad_norm": 0.4298592545318606, "learning_rate": 7.377704895676253e-05, "loss": 0.9547, "step": 2891 }, { "epoch": 1.3103760761214318, "grad_norm": 0.46431438740619596, "learning_rate": 7.377026904583113e-05, "loss": 0.9648, "step": 2892 }, { "epoch": 1.310829179882193, "grad_norm": 0.48995989369031884, "learning_rate": 7.376348575543203e-05, "loss": 0.9874, "step": 2893 }, { "epoch": 1.3112822836429543, "grad_norm": 0.4388185003163677, "learning_rate": 7.375669908624406e-05, "loss": 0.9686, "step": 2894 }, { "epoch": 1.3117353874037154, "grad_norm": 0.3905827824291021, "learning_rate": 7.374990903894636e-05, "loss": 0.9774, "step": 2895 }, { "epoch": 1.3121884911644766, "grad_norm": 0.4385381203608434, "learning_rate": 7.374311561421846e-05, "loss": 0.9609, "step": 2896 }, { "epoch": 1.312641594925238, "grad_norm": 0.5427273837881273, "learning_rate": 7.373631881274018e-05, "loss": 0.9626, "step": 2897 }, { "epoch": 1.313094698685999, "grad_norm": 0.542908663123143, "learning_rate": 7.372951863519169e-05, "loss": 0.9726, "step": 2898 }, { "epoch": 1.3135478024467604, "grad_norm": 0.5118566611263031, "learning_rate": 7.372271508225348e-05, "loss": 0.9918, "step": 2899 }, { "epoch": 1.3140009062075215, "grad_norm": 0.4653544713044797, "learning_rate": 7.371590815460642e-05, "loss": 0.9823, "step": 2900 }, { "epoch": 1.3144540099682827, "grad_norm": 0.4417069912942185, "learning_rate": 7.370909785293169e-05, "loss": 0.9957, "step": 2901 }, { "epoch": 1.314907113729044, "grad_norm": 0.4087293400922698, "learning_rate": 7.370228417791081e-05, "loss": 0.9667, "step": 2902 }, { "epoch": 1.3153602174898051, "grad_norm": 0.4311665840735184, "learning_rate": 7.369546713022565e-05, "loss": 0.9913, "step": 2903 }, { "epoch": 1.3158133212505665, "grad_norm": 0.33007368657216946, "learning_rate": 7.368864671055837e-05, "loss": 0.9754, "step": 2904 }, { "epoch": 1.3162664250113276, "grad_norm": 0.24741345153193617, "learning_rate": 7.368182291959155e-05, "loss": 0.9833, "step": 2905 }, { "epoch": 1.3167195287720888, "grad_norm": 0.27714027314036094, "learning_rate": 7.367499575800806e-05, "loss": 0.995, "step": 2906 }, { "epoch": 1.3171726325328499, "grad_norm": 0.29682053751841386, "learning_rate": 7.366816522649107e-05, "loss": 0.9841, "step": 2907 }, { "epoch": 1.3176257362936112, "grad_norm": 0.2859096239538841, "learning_rate": 7.366133132572415e-05, "loss": 0.9778, "step": 2908 }, { "epoch": 1.3180788400543724, "grad_norm": 0.2865129327316132, "learning_rate": 7.365449405639121e-05, "loss": 0.9677, "step": 2909 }, { "epoch": 1.3185319438151337, "grad_norm": 0.3096556655999069, "learning_rate": 7.364765341917644e-05, "loss": 0.9832, "step": 2910 }, { "epoch": 1.3189850475758949, "grad_norm": 0.34881734719049795, "learning_rate": 7.36408094147644e-05, "loss": 0.9716, "step": 2911 }, { "epoch": 1.319438151336656, "grad_norm": 0.41899506800019976, "learning_rate": 7.363396204383998e-05, "loss": 0.9658, "step": 2912 }, { "epoch": 1.3198912550974173, "grad_norm": 0.41247100034892076, "learning_rate": 7.362711130708844e-05, "loss": 0.9719, "step": 2913 }, { "epoch": 1.3203443588581785, "grad_norm": 0.43015033231076943, "learning_rate": 7.362025720519532e-05, "loss": 0.9711, "step": 2914 }, { "epoch": 1.3207974626189398, "grad_norm": 0.458930091640701, "learning_rate": 7.361339973884654e-05, "loss": 0.9831, "step": 2915 }, { "epoch": 1.321250566379701, "grad_norm": 0.4158725366672497, "learning_rate": 7.360653890872834e-05, "loss": 0.9868, "step": 2916 }, { "epoch": 1.321703670140462, "grad_norm": 0.3370524584910825, "learning_rate": 7.359967471552729e-05, "loss": 0.9581, "step": 2917 }, { "epoch": 1.3221567739012234, "grad_norm": 0.29652898538953015, "learning_rate": 7.359280715993032e-05, "loss": 0.9773, "step": 2918 }, { "epoch": 1.3226098776619846, "grad_norm": 0.33019388086307183, "learning_rate": 7.358593624262468e-05, "loss": 0.9872, "step": 2919 }, { "epoch": 1.323062981422746, "grad_norm": 0.4242690234990251, "learning_rate": 7.357906196429795e-05, "loss": 0.9836, "step": 2920 }, { "epoch": 1.323516085183507, "grad_norm": 0.474723288245326, "learning_rate": 7.357218432563805e-05, "loss": 0.9749, "step": 2921 }, { "epoch": 1.3239691889442682, "grad_norm": 0.46654781078189095, "learning_rate": 7.356530332733325e-05, "loss": 0.9896, "step": 2922 }, { "epoch": 1.3244222927050293, "grad_norm": 0.4705467736553687, "learning_rate": 7.355841897007215e-05, "loss": 0.9749, "step": 2923 }, { "epoch": 1.3248753964657907, "grad_norm": 0.5260887175700584, "learning_rate": 7.35515312545437e-05, "loss": 0.9625, "step": 2924 }, { "epoch": 1.3253285002265518, "grad_norm": 0.6119776959758224, "learning_rate": 7.354464018143712e-05, "loss": 0.9677, "step": 2925 }, { "epoch": 1.3257816039873132, "grad_norm": 0.7097974116098831, "learning_rate": 7.353774575144204e-05, "loss": 0.9719, "step": 2926 }, { "epoch": 1.3262347077480743, "grad_norm": 0.7730080551638314, "learning_rate": 7.353084796524842e-05, "loss": 0.9794, "step": 2927 }, { "epoch": 1.3266878115088354, "grad_norm": 0.767845678210595, "learning_rate": 7.352394682354653e-05, "loss": 0.9646, "step": 2928 }, { "epoch": 1.3271409152695968, "grad_norm": 0.7039196179266085, "learning_rate": 7.351704232702697e-05, "loss": 0.9718, "step": 2929 }, { "epoch": 1.327594019030358, "grad_norm": 0.570008542652062, "learning_rate": 7.351013447638068e-05, "loss": 0.9676, "step": 2930 }, { "epoch": 1.3280471227911192, "grad_norm": 0.43749072608975265, "learning_rate": 7.350322327229897e-05, "loss": 0.9578, "step": 2931 }, { "epoch": 1.3285002265518804, "grad_norm": 0.38181214327140656, "learning_rate": 7.349630871547345e-05, "loss": 0.9684, "step": 2932 }, { "epoch": 1.3289533303126415, "grad_norm": 0.37356367339661345, "learning_rate": 7.348939080659607e-05, "loss": 0.9744, "step": 2933 }, { "epoch": 1.3294064340734029, "grad_norm": 0.40382182780269826, "learning_rate": 7.348246954635912e-05, "loss": 0.9964, "step": 2934 }, { "epoch": 1.329859537834164, "grad_norm": 0.4631854939401302, "learning_rate": 7.347554493545525e-05, "loss": 0.9656, "step": 2935 }, { "epoch": 1.3303126415949253, "grad_norm": 0.5544603539926307, "learning_rate": 7.34686169745774e-05, "loss": 0.9698, "step": 2936 }, { "epoch": 1.3307657453556865, "grad_norm": 0.6003737285181869, "learning_rate": 7.346168566441884e-05, "loss": 0.9883, "step": 2937 }, { "epoch": 1.3312188491164476, "grad_norm": 0.5900178609822805, "learning_rate": 7.345475100567326e-05, "loss": 0.9602, "step": 2938 }, { "epoch": 1.3316719528772087, "grad_norm": 0.5202938622603157, "learning_rate": 7.344781299903458e-05, "loss": 0.9844, "step": 2939 }, { "epoch": 1.33212505663797, "grad_norm": 0.4435531892659002, "learning_rate": 7.344087164519714e-05, "loss": 0.9612, "step": 2940 }, { "epoch": 1.3325781603987312, "grad_norm": 0.41482704818998684, "learning_rate": 7.343392694485555e-05, "loss": 0.9755, "step": 2941 }, { "epoch": 1.3330312641594926, "grad_norm": 0.40785382004843934, "learning_rate": 7.342697889870478e-05, "loss": 0.976, "step": 2942 }, { "epoch": 1.3334843679202537, "grad_norm": 0.4111210496579791, "learning_rate": 7.342002750744016e-05, "loss": 0.9632, "step": 2943 }, { "epoch": 1.3339374716810148, "grad_norm": 0.4552738256455185, "learning_rate": 7.341307277175731e-05, "loss": 0.9756, "step": 2944 }, { "epoch": 1.3343905754417762, "grad_norm": 0.45124633212136367, "learning_rate": 7.340611469235222e-05, "loss": 0.9407, "step": 2945 }, { "epoch": 1.3348436792025373, "grad_norm": 0.40812782557167765, "learning_rate": 7.339915326992119e-05, "loss": 0.9411, "step": 2946 }, { "epoch": 1.3352967829632987, "grad_norm": 0.3530700627952266, "learning_rate": 7.339218850516087e-05, "loss": 0.975, "step": 2947 }, { "epoch": 1.3357498867240598, "grad_norm": 0.3200823137129883, "learning_rate": 7.338522039876825e-05, "loss": 0.9722, "step": 2948 }, { "epoch": 1.336202990484821, "grad_norm": 0.3049489207409156, "learning_rate": 7.337824895144063e-05, "loss": 0.9883, "step": 2949 }, { "epoch": 1.3366560942455823, "grad_norm": 0.3152832466836355, "learning_rate": 7.337127416387566e-05, "loss": 0.9709, "step": 2950 }, { "epoch": 1.3371091980063434, "grad_norm": 0.3737886373070675, "learning_rate": 7.336429603677133e-05, "loss": 0.989, "step": 2951 }, { "epoch": 1.3375623017671048, "grad_norm": 0.3638418494007138, "learning_rate": 7.335731457082594e-05, "loss": 0.9768, "step": 2952 }, { "epoch": 1.338015405527866, "grad_norm": 0.26600149503607173, "learning_rate": 7.335032976673818e-05, "loss": 0.9839, "step": 2953 }, { "epoch": 1.338468509288627, "grad_norm": 0.2481563657797431, "learning_rate": 7.334334162520698e-05, "loss": 0.9794, "step": 2954 }, { "epoch": 1.3389216130493884, "grad_norm": 0.29345953358353777, "learning_rate": 7.33363501469317e-05, "loss": 0.9692, "step": 2955 }, { "epoch": 1.3393747168101495, "grad_norm": 0.27034943914750914, "learning_rate": 7.3329355332612e-05, "loss": 0.9631, "step": 2956 }, { "epoch": 1.3398278205709109, "grad_norm": 0.2595950563765037, "learning_rate": 7.332235718294784e-05, "loss": 0.9776, "step": 2957 }, { "epoch": 1.340280924331672, "grad_norm": 0.3287431579033178, "learning_rate": 7.331535569863953e-05, "loss": 0.9617, "step": 2958 }, { "epoch": 1.3407340280924331, "grad_norm": 0.3609943320951992, "learning_rate": 7.330835088038776e-05, "loss": 0.9467, "step": 2959 }, { "epoch": 1.3411871318531943, "grad_norm": 0.3762059758506403, "learning_rate": 7.33013427288935e-05, "loss": 0.9847, "step": 2960 }, { "epoch": 1.3416402356139556, "grad_norm": 0.4057673482849163, "learning_rate": 7.329433124485807e-05, "loss": 0.9836, "step": 2961 }, { "epoch": 1.3420933393747168, "grad_norm": 0.4270109698859757, "learning_rate": 7.328731642898315e-05, "loss": 0.9715, "step": 2962 }, { "epoch": 1.342546443135478, "grad_norm": 0.44329143016574474, "learning_rate": 7.328029828197067e-05, "loss": 0.9694, "step": 2963 }, { "epoch": 1.3429995468962392, "grad_norm": 0.5273177974957398, "learning_rate": 7.327327680452302e-05, "loss": 0.9794, "step": 2964 }, { "epoch": 1.3434526506570004, "grad_norm": 0.5474136344837968, "learning_rate": 7.32662519973428e-05, "loss": 0.9867, "step": 2965 }, { "epoch": 1.3439057544177617, "grad_norm": 0.5379533912876399, "learning_rate": 7.325922386113305e-05, "loss": 0.9991, "step": 2966 }, { "epoch": 1.3443588581785229, "grad_norm": 0.5219940310324396, "learning_rate": 7.325219239659706e-05, "loss": 0.9586, "step": 2967 }, { "epoch": 1.3448119619392842, "grad_norm": 0.5158452294702616, "learning_rate": 7.324515760443849e-05, "loss": 0.9879, "step": 2968 }, { "epoch": 1.3452650657000453, "grad_norm": 0.5743677759264811, "learning_rate": 7.323811948536131e-05, "loss": 0.9596, "step": 2969 }, { "epoch": 1.3457181694608065, "grad_norm": 0.5696992805118174, "learning_rate": 7.323107804006988e-05, "loss": 0.986, "step": 2970 }, { "epoch": 1.3461712732215678, "grad_norm": 0.49407936589158696, "learning_rate": 7.322403326926883e-05, "loss": 0.9677, "step": 2971 }, { "epoch": 1.346624376982329, "grad_norm": 0.4393373144934096, "learning_rate": 7.321698517366313e-05, "loss": 0.9716, "step": 2972 }, { "epoch": 1.3470774807430903, "grad_norm": 0.4267203460872746, "learning_rate": 7.320993375395813e-05, "loss": 0.9757, "step": 2973 }, { "epoch": 1.3475305845038514, "grad_norm": 0.40715056215293094, "learning_rate": 7.320287901085946e-05, "loss": 0.9719, "step": 2974 }, { "epoch": 1.3479836882646126, "grad_norm": 0.377466777608707, "learning_rate": 7.319582094507312e-05, "loss": 0.9771, "step": 2975 }, { "epoch": 1.3484367920253737, "grad_norm": 0.35438555723790927, "learning_rate": 7.31887595573054e-05, "loss": 0.9939, "step": 2976 }, { "epoch": 1.348889895786135, "grad_norm": 0.35537620097607314, "learning_rate": 7.3181694848263e-05, "loss": 0.9766, "step": 2977 }, { "epoch": 1.3493429995468962, "grad_norm": 0.3359312705278679, "learning_rate": 7.317462681865284e-05, "loss": 0.9577, "step": 2978 }, { "epoch": 1.3497961033076575, "grad_norm": 0.3682669110963614, "learning_rate": 7.316755546918227e-05, "loss": 0.9864, "step": 2979 }, { "epoch": 1.3502492070684187, "grad_norm": 0.4101352822584593, "learning_rate": 7.316048080055895e-05, "loss": 0.9792, "step": 2980 }, { "epoch": 1.3507023108291798, "grad_norm": 0.34513022317123304, "learning_rate": 7.315340281349082e-05, "loss": 0.9749, "step": 2981 }, { "epoch": 1.3511554145899412, "grad_norm": 0.37018263503807874, "learning_rate": 7.31463215086862e-05, "loss": 0.9513, "step": 2982 }, { "epoch": 1.3516085183507023, "grad_norm": 0.3909300177144061, "learning_rate": 7.313923688685375e-05, "loss": 0.9831, "step": 2983 }, { "epoch": 1.3520616221114636, "grad_norm": 0.39787673901482046, "learning_rate": 7.313214894870244e-05, "loss": 0.971, "step": 2984 }, { "epoch": 1.3525147258722248, "grad_norm": 0.4079917472820215, "learning_rate": 7.312505769494157e-05, "loss": 0.9701, "step": 2985 }, { "epoch": 1.352967829632986, "grad_norm": 0.4101073162867466, "learning_rate": 7.311796312628078e-05, "loss": 0.9752, "step": 2986 }, { "epoch": 1.3534209333937472, "grad_norm": 0.41496232727295557, "learning_rate": 7.311086524343003e-05, "loss": 0.9878, "step": 2987 }, { "epoch": 1.3538740371545084, "grad_norm": 0.4230397919677434, "learning_rate": 7.310376404709963e-05, "loss": 0.9775, "step": 2988 }, { "epoch": 1.3543271409152697, "grad_norm": 0.4087646880064671, "learning_rate": 7.309665953800023e-05, "loss": 0.9699, "step": 2989 }, { "epoch": 1.3547802446760309, "grad_norm": 0.36941376527772696, "learning_rate": 7.308955171684275e-05, "loss": 0.9875, "step": 2990 }, { "epoch": 1.355233348436792, "grad_norm": 0.39396070825299745, "learning_rate": 7.308244058433853e-05, "loss": 0.9593, "step": 2991 }, { "epoch": 1.3556864521975531, "grad_norm": 0.4250806095893163, "learning_rate": 7.307532614119919e-05, "loss": 0.957, "step": 2992 }, { "epoch": 1.3561395559583145, "grad_norm": 0.3999282415747276, "learning_rate": 7.306820838813666e-05, "loss": 0.9784, "step": 2993 }, { "epoch": 1.3565926597190756, "grad_norm": 0.3704762423635232, "learning_rate": 7.306108732586325e-05, "loss": 0.9816, "step": 2994 }, { "epoch": 1.357045763479837, "grad_norm": 0.3267143500674719, "learning_rate": 7.305396295509157e-05, "loss": 0.9675, "step": 2995 }, { "epoch": 1.357498867240598, "grad_norm": 0.27405527495564946, "learning_rate": 7.30468352765346e-05, "loss": 0.9639, "step": 2996 }, { "epoch": 1.3579519710013592, "grad_norm": 0.26451108045271843, "learning_rate": 7.303970429090559e-05, "loss": 0.9971, "step": 2997 }, { "epoch": 1.3584050747621206, "grad_norm": 0.2790788606742112, "learning_rate": 7.303256999891818e-05, "loss": 0.9563, "step": 2998 }, { "epoch": 1.3588581785228817, "grad_norm": 0.31889979425906734, "learning_rate": 7.30254324012863e-05, "loss": 0.976, "step": 2999 }, { "epoch": 1.359311282283643, "grad_norm": 0.30453423143796793, "learning_rate": 7.301829149872422e-05, "loss": 0.9696, "step": 3000 }, { "epoch": 1.3597643860444042, "grad_norm": 0.27437734584705226, "learning_rate": 7.301114729194655e-05, "loss": 0.9899, "step": 3001 }, { "epoch": 1.3602174898051653, "grad_norm": 0.3064739093311518, "learning_rate": 7.300399978166825e-05, "loss": 0.9656, "step": 3002 }, { "epoch": 1.3606705935659267, "grad_norm": 0.3642214648861554, "learning_rate": 7.299684896860457e-05, "loss": 0.9995, "step": 3003 }, { "epoch": 1.3611236973266878, "grad_norm": 0.39150293927881435, "learning_rate": 7.29896948534711e-05, "loss": 0.9645, "step": 3004 }, { "epoch": 1.3615768010874492, "grad_norm": 0.37092525150701555, "learning_rate": 7.298253743698376e-05, "loss": 0.9673, "step": 3005 }, { "epoch": 1.3620299048482103, "grad_norm": 0.3904693221622385, "learning_rate": 7.297537671985885e-05, "loss": 0.972, "step": 3006 }, { "epoch": 1.3624830086089714, "grad_norm": 0.4378440355462469, "learning_rate": 7.296821270281292e-05, "loss": 0.9798, "step": 3007 }, { "epoch": 1.3629361123697326, "grad_norm": 0.3860128305847876, "learning_rate": 7.296104538656291e-05, "loss": 0.9681, "step": 3008 }, { "epoch": 1.363389216130494, "grad_norm": 0.3675868286937055, "learning_rate": 7.295387477182607e-05, "loss": 0.9702, "step": 3009 }, { "epoch": 1.363842319891255, "grad_norm": 0.4105267691868475, "learning_rate": 7.294670085931996e-05, "loss": 1.0132, "step": 3010 }, { "epoch": 1.3642954236520164, "grad_norm": 0.4461506896248297, "learning_rate": 7.293952364976251e-05, "loss": 0.9759, "step": 3011 }, { "epoch": 1.3647485274127775, "grad_norm": 0.41750703885268065, "learning_rate": 7.293234314387196e-05, "loss": 0.971, "step": 3012 }, { "epoch": 1.3652016311735387, "grad_norm": 0.41244165052684956, "learning_rate": 7.292515934236687e-05, "loss": 0.9847, "step": 3013 }, { "epoch": 1.3656547349343, "grad_norm": 0.4556402454877585, "learning_rate": 7.291797224596613e-05, "loss": 0.9686, "step": 3014 }, { "epoch": 1.3661078386950611, "grad_norm": 0.4883396398334453, "learning_rate": 7.291078185538899e-05, "loss": 0.9484, "step": 3015 }, { "epoch": 1.3665609424558225, "grad_norm": 0.5530528786140053, "learning_rate": 7.2903588171355e-05, "loss": 0.9898, "step": 3016 }, { "epoch": 1.3670140462165836, "grad_norm": 0.6278994385683223, "learning_rate": 7.289639119458404e-05, "loss": 0.9618, "step": 3017 }, { "epoch": 1.3674671499773448, "grad_norm": 0.675603546457148, "learning_rate": 7.288919092579634e-05, "loss": 0.9776, "step": 3018 }, { "epoch": 1.367920253738106, "grad_norm": 0.6888667692966998, "learning_rate": 7.288198736571245e-05, "loss": 0.9665, "step": 3019 }, { "epoch": 1.3683733574988672, "grad_norm": 0.6289698353775709, "learning_rate": 7.287478051505323e-05, "loss": 0.9886, "step": 3020 }, { "epoch": 1.3688264612596286, "grad_norm": 0.6025060704848948, "learning_rate": 7.286757037453991e-05, "loss": 0.9725, "step": 3021 }, { "epoch": 1.3692795650203897, "grad_norm": 0.6896448440195843, "learning_rate": 7.286035694489399e-05, "loss": 0.976, "step": 3022 }, { "epoch": 1.3697326687811509, "grad_norm": 0.7034281530778926, "learning_rate": 7.285314022683737e-05, "loss": 0.9876, "step": 3023 }, { "epoch": 1.370185772541912, "grad_norm": 0.5973863272074154, "learning_rate": 7.284592022109222e-05, "loss": 0.9694, "step": 3024 }, { "epoch": 1.3706388763026733, "grad_norm": 0.46538953685100787, "learning_rate": 7.283869692838108e-05, "loss": 0.9751, "step": 3025 }, { "epoch": 1.3710919800634345, "grad_norm": 0.3957932467034368, "learning_rate": 7.283147034942679e-05, "loss": 0.9715, "step": 3026 }, { "epoch": 1.3715450838241958, "grad_norm": 0.40277615359536945, "learning_rate": 7.282424048495251e-05, "loss": 0.9706, "step": 3027 }, { "epoch": 1.371998187584957, "grad_norm": 0.4130456010244044, "learning_rate": 7.28170073356818e-05, "loss": 0.9802, "step": 3028 }, { "epoch": 1.372451291345718, "grad_norm": 0.36104140282581265, "learning_rate": 7.280977090233845e-05, "loss": 0.9943, "step": 3029 }, { "epoch": 1.3729043951064794, "grad_norm": 0.36492708059809026, "learning_rate": 7.280253118564667e-05, "loss": 0.9957, "step": 3030 }, { "epoch": 1.3733574988672406, "grad_norm": 0.4604033923931201, "learning_rate": 7.279528818633092e-05, "loss": 0.9731, "step": 3031 }, { "epoch": 1.373810602628002, "grad_norm": 0.5717447876358556, "learning_rate": 7.278804190511602e-05, "loss": 0.9888, "step": 3032 }, { "epoch": 1.374263706388763, "grad_norm": 0.6189173952652846, "learning_rate": 7.278079234272716e-05, "loss": 0.9815, "step": 3033 }, { "epoch": 1.3747168101495242, "grad_norm": 0.5678794478823922, "learning_rate": 7.277353949988978e-05, "loss": 0.9748, "step": 3034 }, { "epoch": 1.3751699139102855, "grad_norm": 0.49545494060096307, "learning_rate": 7.27662833773297e-05, "loss": 0.9713, "step": 3035 }, { "epoch": 1.3756230176710467, "grad_norm": 0.4827966341055809, "learning_rate": 7.275902397577307e-05, "loss": 0.9694, "step": 3036 }, { "epoch": 1.376076121431808, "grad_norm": 0.5039121857594921, "learning_rate": 7.275176129594635e-05, "loss": 0.9758, "step": 3037 }, { "epoch": 1.3765292251925692, "grad_norm": 0.5017825423126956, "learning_rate": 7.274449533857632e-05, "loss": 0.9577, "step": 3038 }, { "epoch": 1.3769823289533303, "grad_norm": 0.4260872535782057, "learning_rate": 7.273722610439011e-05, "loss": 0.9422, "step": 3039 }, { "epoch": 1.3774354327140914, "grad_norm": 0.3486920612668738, "learning_rate": 7.272995359411519e-05, "loss": 0.9932, "step": 3040 }, { "epoch": 1.3778885364748528, "grad_norm": 0.31146174180732017, "learning_rate": 7.272267780847928e-05, "loss": 0.9735, "step": 3041 }, { "epoch": 1.378341640235614, "grad_norm": 0.303609594285057, "learning_rate": 7.271539874821054e-05, "loss": 0.9742, "step": 3042 }, { "epoch": 1.3787947439963752, "grad_norm": 0.3699427880636268, "learning_rate": 7.270811641403738e-05, "loss": 0.9811, "step": 3043 }, { "epoch": 1.3792478477571364, "grad_norm": 0.41297076166458424, "learning_rate": 7.270083080668855e-05, "loss": 0.9828, "step": 3044 }, { "epoch": 1.3797009515178975, "grad_norm": 0.40467008816714034, "learning_rate": 7.269354192689316e-05, "loss": 0.96, "step": 3045 }, { "epoch": 1.3801540552786589, "grad_norm": 0.412638016736522, "learning_rate": 7.268624977538062e-05, "loss": 0.9851, "step": 3046 }, { "epoch": 1.38060715903942, "grad_norm": 0.49386153352386997, "learning_rate": 7.267895435288066e-05, "loss": 0.973, "step": 3047 }, { "epoch": 1.3810602628001813, "grad_norm": 0.531161130626168, "learning_rate": 7.267165566012334e-05, "loss": 0.9841, "step": 3048 }, { "epoch": 1.3815133665609425, "grad_norm": 0.5772800610852574, "learning_rate": 7.266435369783909e-05, "loss": 0.9722, "step": 3049 }, { "epoch": 1.3819664703217036, "grad_norm": 0.5924586589789905, "learning_rate": 7.265704846675862e-05, "loss": 1.0076, "step": 3050 }, { "epoch": 1.382419574082465, "grad_norm": 0.5851184667685062, "learning_rate": 7.264973996761297e-05, "loss": 0.9678, "step": 3051 }, { "epoch": 1.382872677843226, "grad_norm": 0.5634656821745222, "learning_rate": 7.264242820113353e-05, "loss": 0.9637, "step": 3052 }, { "epoch": 1.3833257816039874, "grad_norm": 0.499781708831375, "learning_rate": 7.2635113168052e-05, "loss": 0.9872, "step": 3053 }, { "epoch": 1.3837788853647486, "grad_norm": 0.3719856854894633, "learning_rate": 7.26277948691004e-05, "loss": 0.9689, "step": 3054 }, { "epoch": 1.3842319891255097, "grad_norm": 0.24538094098888433, "learning_rate": 7.26204733050111e-05, "loss": 0.9509, "step": 3055 }, { "epoch": 1.3846850928862708, "grad_norm": 0.2947098560066058, "learning_rate": 7.261314847651681e-05, "loss": 0.9658, "step": 3056 }, { "epoch": 1.3851381966470322, "grad_norm": 0.39637739459692395, "learning_rate": 7.260582038435051e-05, "loss": 0.9555, "step": 3057 }, { "epoch": 1.3855913004077933, "grad_norm": 0.4480339896079862, "learning_rate": 7.259848902924554e-05, "loss": 0.9678, "step": 3058 }, { "epoch": 1.3860444041685547, "grad_norm": 0.5029310154933957, "learning_rate": 7.259115441193557e-05, "loss": 0.9712, "step": 3059 }, { "epoch": 1.3864975079293158, "grad_norm": 0.5766208773504363, "learning_rate": 7.258381653315461e-05, "loss": 0.9696, "step": 3060 }, { "epoch": 1.386950611690077, "grad_norm": 0.6062947803839253, "learning_rate": 7.257647539363696e-05, "loss": 0.9757, "step": 3061 }, { "epoch": 1.3874037154508383, "grad_norm": 0.53803855431541, "learning_rate": 7.256913099411726e-05, "loss": 0.981, "step": 3062 }, { "epoch": 1.3878568192115994, "grad_norm": 0.42968648015041117, "learning_rate": 7.256178333533049e-05, "loss": 0.9747, "step": 3063 }, { "epoch": 1.3883099229723608, "grad_norm": 0.35076606049293996, "learning_rate": 7.255443241801196e-05, "loss": 0.9534, "step": 3064 }, { "epoch": 1.388763026733122, "grad_norm": 0.33098093811012663, "learning_rate": 7.254707824289726e-05, "loss": 0.9522, "step": 3065 }, { "epoch": 1.389216130493883, "grad_norm": 0.32015592043113467, "learning_rate": 7.253972081072237e-05, "loss": 0.9494, "step": 3066 }, { "epoch": 1.3896692342546444, "grad_norm": 0.41331993091574476, "learning_rate": 7.253236012222355e-05, "loss": 0.9812, "step": 3067 }, { "epoch": 1.3901223380154055, "grad_norm": 0.47026938096427057, "learning_rate": 7.25249961781374e-05, "loss": 0.9734, "step": 3068 }, { "epoch": 1.3905754417761669, "grad_norm": 0.4824356070525831, "learning_rate": 7.251762897920086e-05, "loss": 0.9577, "step": 3069 }, { "epoch": 1.391028545536928, "grad_norm": 0.4900850930562195, "learning_rate": 7.251025852615117e-05, "loss": 0.9736, "step": 3070 }, { "epoch": 1.3914816492976891, "grad_norm": 0.5222311474288329, "learning_rate": 7.25028848197259e-05, "loss": 0.9666, "step": 3071 }, { "epoch": 1.3919347530584503, "grad_norm": 0.5073153471693131, "learning_rate": 7.249550786066298e-05, "loss": 0.9735, "step": 3072 }, { "epoch": 1.3923878568192116, "grad_norm": 0.470813918290097, "learning_rate": 7.24881276497006e-05, "loss": 0.9621, "step": 3073 }, { "epoch": 1.3928409605799728, "grad_norm": 0.4158280797230897, "learning_rate": 7.248074418757736e-05, "loss": 0.9644, "step": 3074 }, { "epoch": 1.393294064340734, "grad_norm": 0.3532013907173983, "learning_rate": 7.247335747503211e-05, "loss": 0.9651, "step": 3075 }, { "epoch": 1.3937471681014952, "grad_norm": 0.25548251283822243, "learning_rate": 7.246596751280406e-05, "loss": 0.9794, "step": 3076 }, { "epoch": 1.3942002718622564, "grad_norm": 0.21975483953619251, "learning_rate": 7.245857430163276e-05, "loss": 0.9516, "step": 3077 }, { "epoch": 1.3946533756230177, "grad_norm": 0.2547181379865182, "learning_rate": 7.245117784225804e-05, "loss": 0.9832, "step": 3078 }, { "epoch": 1.3951064793837789, "grad_norm": 0.31987937909385167, "learning_rate": 7.244377813542008e-05, "loss": 0.9592, "step": 3079 }, { "epoch": 1.3955595831445402, "grad_norm": 0.411345624293116, "learning_rate": 7.24363751818594e-05, "loss": 1.0007, "step": 3080 }, { "epoch": 1.3960126869053013, "grad_norm": 0.39535833843318313, "learning_rate": 7.242896898231682e-05, "loss": 0.9563, "step": 3081 }, { "epoch": 1.3964657906660625, "grad_norm": 0.40071281197120456, "learning_rate": 7.242155953753351e-05, "loss": 0.9637, "step": 3082 }, { "epoch": 1.3969188944268238, "grad_norm": 0.4427689204842634, "learning_rate": 7.241414684825094e-05, "loss": 0.9808, "step": 3083 }, { "epoch": 1.397371998187585, "grad_norm": 0.44437160202118225, "learning_rate": 7.240673091521091e-05, "loss": 0.9736, "step": 3084 }, { "epoch": 1.3978251019483463, "grad_norm": 0.4664612244408703, "learning_rate": 7.239931173915556e-05, "loss": 0.9688, "step": 3085 }, { "epoch": 1.3982782057091074, "grad_norm": 0.4485680616915116, "learning_rate": 7.239188932082734e-05, "loss": 0.995, "step": 3086 }, { "epoch": 1.3987313094698686, "grad_norm": 0.3679692217264461, "learning_rate": 7.238446366096901e-05, "loss": 0.961, "step": 3087 }, { "epoch": 1.3991844132306297, "grad_norm": 0.39049765153438837, "learning_rate": 7.23770347603237e-05, "loss": 0.9805, "step": 3088 }, { "epoch": 1.399637516991391, "grad_norm": 0.40067536733859027, "learning_rate": 7.236960261963483e-05, "loss": 0.9722, "step": 3089 }, { "epoch": 1.4000906207521522, "grad_norm": 0.4497074508110077, "learning_rate": 7.236216723964615e-05, "loss": 0.9867, "step": 3090 }, { "epoch": 1.4005437245129135, "grad_norm": 0.5074606255296424, "learning_rate": 7.235472862110172e-05, "loss": 0.9455, "step": 3091 }, { "epoch": 1.4009968282736747, "grad_norm": 0.5305197776747952, "learning_rate": 7.234728676474596e-05, "loss": 0.9489, "step": 3092 }, { "epoch": 1.4014499320344358, "grad_norm": 0.5419151662727366, "learning_rate": 7.233984167132359e-05, "loss": 0.9758, "step": 3093 }, { "epoch": 1.4019030357951971, "grad_norm": 0.3733672763578869, "learning_rate": 7.233239334157966e-05, "loss": 0.9716, "step": 3094 }, { "epoch": 1.4023561395559583, "grad_norm": 0.38961369093769815, "learning_rate": 7.232494177625951e-05, "loss": 0.979, "step": 3095 }, { "epoch": 1.4028092433167196, "grad_norm": 0.4650426082496968, "learning_rate": 7.231748697610887e-05, "loss": 0.9781, "step": 3096 }, { "epoch": 1.4032623470774808, "grad_norm": 0.42056354800756807, "learning_rate": 7.231002894187375e-05, "loss": 0.9797, "step": 3097 }, { "epoch": 1.403715450838242, "grad_norm": 0.38861042320010547, "learning_rate": 7.23025676743005e-05, "loss": 0.9651, "step": 3098 }, { "epoch": 1.4041685545990032, "grad_norm": 0.3160870508318843, "learning_rate": 7.229510317413578e-05, "loss": 0.9932, "step": 3099 }, { "epoch": 1.4046216583597644, "grad_norm": 0.2829017227694029, "learning_rate": 7.228763544212656e-05, "loss": 0.971, "step": 3100 }, { "epoch": 1.4050747621205257, "grad_norm": 0.3101103177162467, "learning_rate": 7.228016447902017e-05, "loss": 1.0035, "step": 3101 }, { "epoch": 1.4055278658812869, "grad_norm": 0.30046674952522545, "learning_rate": 7.227269028556427e-05, "loss": 0.9724, "step": 3102 }, { "epoch": 1.405980969642048, "grad_norm": 0.2823025899757524, "learning_rate": 7.226521286250677e-05, "loss": 0.9686, "step": 3103 }, { "epoch": 1.4064340734028091, "grad_norm": 0.34906522958081004, "learning_rate": 7.225773221059598e-05, "loss": 0.9562, "step": 3104 }, { "epoch": 1.4068871771635705, "grad_norm": 0.48052279192313313, "learning_rate": 7.225024833058051e-05, "loss": 0.9697, "step": 3105 }, { "epoch": 1.4073402809243316, "grad_norm": 0.506975399114851, "learning_rate": 7.22427612232093e-05, "loss": 0.9675, "step": 3106 }, { "epoch": 1.407793384685093, "grad_norm": 0.5274742113311588, "learning_rate": 7.223527088923154e-05, "loss": 0.9777, "step": 3107 }, { "epoch": 1.408246488445854, "grad_norm": 0.5953914922508732, "learning_rate": 7.22277773293969e-05, "loss": 0.959, "step": 3108 }, { "epoch": 1.4086995922066152, "grad_norm": 0.5982382969088159, "learning_rate": 7.222028054445519e-05, "loss": 0.9811, "step": 3109 }, { "epoch": 1.4091526959673766, "grad_norm": 0.55933183651081, "learning_rate": 7.221278053515668e-05, "loss": 0.975, "step": 3110 }, { "epoch": 1.4096057997281377, "grad_norm": 0.5373625111209166, "learning_rate": 7.220527730225191e-05, "loss": 0.9786, "step": 3111 }, { "epoch": 1.410058903488899, "grad_norm": 0.46704651775414535, "learning_rate": 7.219777084649173e-05, "loss": 0.9736, "step": 3112 }, { "epoch": 1.4105120072496602, "grad_norm": 0.41738417786367715, "learning_rate": 7.219026116862733e-05, "loss": 0.9831, "step": 3113 }, { "epoch": 1.4109651110104213, "grad_norm": 0.35221651823930006, "learning_rate": 7.218274826941023e-05, "loss": 0.9925, "step": 3114 }, { "epoch": 1.4114182147711827, "grad_norm": 0.33833616552722073, "learning_rate": 7.217523214959225e-05, "loss": 0.9747, "step": 3115 }, { "epoch": 1.4118713185319438, "grad_norm": 0.2997336343000008, "learning_rate": 7.216771280992556e-05, "loss": 0.9555, "step": 3116 }, { "epoch": 1.4123244222927052, "grad_norm": 0.26583080333964737, "learning_rate": 7.216019025116263e-05, "loss": 0.9719, "step": 3117 }, { "epoch": 1.4127775260534663, "grad_norm": 0.3344397798675417, "learning_rate": 7.215266447405626e-05, "loss": 0.9613, "step": 3118 }, { "epoch": 1.4132306298142274, "grad_norm": 0.33420654271232786, "learning_rate": 7.214513547935956e-05, "loss": 0.9721, "step": 3119 }, { "epoch": 1.4136837335749886, "grad_norm": 0.37454288245136946, "learning_rate": 7.213760326782601e-05, "loss": 0.9799, "step": 3120 }, { "epoch": 1.41413683733575, "grad_norm": 0.3666198256422468, "learning_rate": 7.213006784020932e-05, "loss": 0.9469, "step": 3121 }, { "epoch": 1.414589941096511, "grad_norm": 0.36519700298444185, "learning_rate": 7.212252919726363e-05, "loss": 0.9536, "step": 3122 }, { "epoch": 1.4150430448572724, "grad_norm": 0.3475837681366694, "learning_rate": 7.211498733974331e-05, "loss": 0.9937, "step": 3123 }, { "epoch": 1.4154961486180335, "grad_norm": 0.29738262103013213, "learning_rate": 7.210744226840313e-05, "loss": 0.9539, "step": 3124 }, { "epoch": 1.4159492523787947, "grad_norm": 0.3203887695402772, "learning_rate": 7.20998939839981e-05, "loss": 0.9524, "step": 3125 }, { "epoch": 1.416402356139556, "grad_norm": 0.36518574438046325, "learning_rate": 7.209234248728364e-05, "loss": 0.9652, "step": 3126 }, { "epoch": 1.4168554599003171, "grad_norm": 0.3022157178121114, "learning_rate": 7.20847877790154e-05, "loss": 0.9674, "step": 3127 }, { "epoch": 1.4173085636610785, "grad_norm": 0.2503953393945263, "learning_rate": 7.207722985994941e-05, "loss": 0.9772, "step": 3128 }, { "epoch": 1.4177616674218396, "grad_norm": 0.3396192888607925, "learning_rate": 7.206966873084204e-05, "loss": 0.9685, "step": 3129 }, { "epoch": 1.4182147711826008, "grad_norm": 0.36114860634325274, "learning_rate": 7.206210439244991e-05, "loss": 0.9707, "step": 3130 }, { "epoch": 1.418667874943362, "grad_norm": 0.403155746675576, "learning_rate": 7.205453684553001e-05, "loss": 0.9784, "step": 3131 }, { "epoch": 1.4191209787041232, "grad_norm": 0.4775437005074305, "learning_rate": 7.204696609083966e-05, "loss": 0.9692, "step": 3132 }, { "epoch": 1.4195740824648846, "grad_norm": 0.5336764485228515, "learning_rate": 7.203939212913646e-05, "loss": 0.9614, "step": 3133 }, { "epoch": 1.4200271862256457, "grad_norm": 0.5563348847730859, "learning_rate": 7.203181496117836e-05, "loss": 0.9753, "step": 3134 }, { "epoch": 1.4204802899864069, "grad_norm": 0.5700810653618316, "learning_rate": 7.202423458772364e-05, "loss": 0.9666, "step": 3135 }, { "epoch": 1.420933393747168, "grad_norm": 0.5843311616494125, "learning_rate": 7.201665100953085e-05, "loss": 0.9932, "step": 3136 }, { "epoch": 1.4213864975079293, "grad_norm": 0.5876641599388145, "learning_rate": 7.200906422735895e-05, "loss": 0.9672, "step": 3137 }, { "epoch": 1.4218396012686905, "grad_norm": 0.5203875809748681, "learning_rate": 7.200147424196711e-05, "loss": 0.9853, "step": 3138 }, { "epoch": 1.4222927050294518, "grad_norm": 0.4595834169786946, "learning_rate": 7.19938810541149e-05, "loss": 0.9611, "step": 3139 }, { "epoch": 1.422745808790213, "grad_norm": 0.4084385759512968, "learning_rate": 7.198628466456221e-05, "loss": 0.9856, "step": 3140 }, { "epoch": 1.423198912550974, "grad_norm": 0.4196026682833703, "learning_rate": 7.19786850740692e-05, "loss": 0.9688, "step": 3141 }, { "epoch": 1.4236520163117354, "grad_norm": 0.42297073173834, "learning_rate": 7.197108228339638e-05, "loss": 0.9599, "step": 3142 }, { "epoch": 1.4241051200724966, "grad_norm": 0.4301227407387084, "learning_rate": 7.19634762933046e-05, "loss": 0.9707, "step": 3143 }, { "epoch": 1.424558223833258, "grad_norm": 0.4813977497718758, "learning_rate": 7.195586710455497e-05, "loss": 0.9699, "step": 3144 }, { "epoch": 1.425011327594019, "grad_norm": 0.5072573481579871, "learning_rate": 7.1948254717909e-05, "loss": 0.9811, "step": 3145 }, { "epoch": 1.4254644313547802, "grad_norm": 0.4847221598976344, "learning_rate": 7.194063913412845e-05, "loss": 0.9552, "step": 3146 }, { "epoch": 1.4259175351155415, "grad_norm": 0.47902342418887156, "learning_rate": 7.193302035397545e-05, "loss": 0.9698, "step": 3147 }, { "epoch": 1.4263706388763027, "grad_norm": 0.44899292149218273, "learning_rate": 7.192539837821242e-05, "loss": 0.9716, "step": 3148 }, { "epoch": 1.426823742637064, "grad_norm": 0.3989795717688475, "learning_rate": 7.19177732076021e-05, "loss": 0.9675, "step": 3149 }, { "epoch": 1.4272768463978251, "grad_norm": 0.34392229727760587, "learning_rate": 7.191014484290756e-05, "loss": 0.9846, "step": 3150 }, { "epoch": 1.4277299501585863, "grad_norm": 0.30182400062606696, "learning_rate": 7.19025132848922e-05, "loss": 0.9798, "step": 3151 }, { "epoch": 1.4281830539193474, "grad_norm": 0.2740088884877135, "learning_rate": 7.189487853431973e-05, "loss": 0.9669, "step": 3152 }, { "epoch": 1.4286361576801088, "grad_norm": 0.30732163276081415, "learning_rate": 7.188724059195415e-05, "loss": 0.9663, "step": 3153 }, { "epoch": 1.42908926144087, "grad_norm": 0.33376504204853746, "learning_rate": 7.187959945855984e-05, "loss": 0.9797, "step": 3154 }, { "epoch": 1.4295423652016312, "grad_norm": 0.35026992936728873, "learning_rate": 7.187195513490143e-05, "loss": 0.9643, "step": 3155 }, { "epoch": 1.4299954689623924, "grad_norm": 0.39567640702715534, "learning_rate": 7.186430762174393e-05, "loss": 0.9564, "step": 3156 }, { "epoch": 1.4304485727231535, "grad_norm": 0.353055754772434, "learning_rate": 7.185665691985265e-05, "loss": 0.9509, "step": 3157 }, { "epoch": 1.4309016764839149, "grad_norm": 0.3202080175957882, "learning_rate": 7.184900302999319e-05, "loss": 0.9694, "step": 3158 }, { "epoch": 1.431354780244676, "grad_norm": 0.2871809799964681, "learning_rate": 7.184134595293151e-05, "loss": 0.9667, "step": 3159 }, { "epoch": 1.4318078840054373, "grad_norm": 0.2597326810353414, "learning_rate": 7.183368568943386e-05, "loss": 0.9862, "step": 3160 }, { "epoch": 1.4322609877661985, "grad_norm": 0.25548617263932066, "learning_rate": 7.182602224026682e-05, "loss": 0.9649, "step": 3161 }, { "epoch": 1.4327140915269596, "grad_norm": 0.23405168893237155, "learning_rate": 7.18183556061973e-05, "loss": 0.9786, "step": 3162 }, { "epoch": 1.433167195287721, "grad_norm": 0.25822045680975364, "learning_rate": 7.181068578799252e-05, "loss": 0.9999, "step": 3163 }, { "epoch": 1.433620299048482, "grad_norm": 0.3511807764919181, "learning_rate": 7.180301278641998e-05, "loss": 0.9682, "step": 3164 }, { "epoch": 1.4340734028092434, "grad_norm": 0.4336950870904565, "learning_rate": 7.17953366022476e-05, "loss": 0.9525, "step": 3165 }, { "epoch": 1.4345265065700046, "grad_norm": 0.48947966616946265, "learning_rate": 7.178765723624349e-05, "loss": 0.9526, "step": 3166 }, { "epoch": 1.4349796103307657, "grad_norm": 0.5161683658644163, "learning_rate": 7.177997468917616e-05, "loss": 0.9512, "step": 3167 }, { "epoch": 1.4354327140915268, "grad_norm": 0.5217518050299769, "learning_rate": 7.177228896181444e-05, "loss": 0.9281, "step": 3168 }, { "epoch": 1.4358858178522882, "grad_norm": 0.5246503414262168, "learning_rate": 7.176460005492744e-05, "loss": 0.966, "step": 3169 }, { "epoch": 1.4363389216130493, "grad_norm": 0.5016172067081944, "learning_rate": 7.175690796928462e-05, "loss": 0.9768, "step": 3170 }, { "epoch": 1.4367920253738107, "grad_norm": 0.4877159098591854, "learning_rate": 7.174921270565572e-05, "loss": 0.9849, "step": 3171 }, { "epoch": 1.4372451291345718, "grad_norm": 0.493111551670646, "learning_rate": 7.174151426481086e-05, "loss": 0.9863, "step": 3172 }, { "epoch": 1.437698232895333, "grad_norm": 0.5144358445303852, "learning_rate": 7.17338126475204e-05, "loss": 0.9806, "step": 3173 }, { "epoch": 1.4381513366560943, "grad_norm": 0.44561308899569924, "learning_rate": 7.172610785455508e-05, "loss": 0.9681, "step": 3174 }, { "epoch": 1.4386044404168554, "grad_norm": 0.3339737633674465, "learning_rate": 7.171839988668595e-05, "loss": 0.986, "step": 3175 }, { "epoch": 1.4390575441776168, "grad_norm": 0.30951642636601356, "learning_rate": 7.171068874468432e-05, "loss": 0.9653, "step": 3176 }, { "epoch": 1.439510647938378, "grad_norm": 0.3488125766183838, "learning_rate": 7.170297442932191e-05, "loss": 0.9731, "step": 3177 }, { "epoch": 1.439963751699139, "grad_norm": 0.32356090947986776, "learning_rate": 7.169525694137068e-05, "loss": 0.9588, "step": 3178 }, { "epoch": 1.4404168554599004, "grad_norm": 0.2956716039759787, "learning_rate": 7.168753628160295e-05, "loss": 0.9655, "step": 3179 }, { "epoch": 1.4408699592206615, "grad_norm": 0.36555855698840256, "learning_rate": 7.167981245079132e-05, "loss": 0.9874, "step": 3180 }, { "epoch": 1.4413230629814229, "grad_norm": 0.46751567258416277, "learning_rate": 7.167208544970878e-05, "loss": 0.9735, "step": 3181 }, { "epoch": 1.441776166742184, "grad_norm": 0.5377553927024343, "learning_rate": 7.166435527912853e-05, "loss": 0.987, "step": 3182 }, { "epoch": 1.4422292705029451, "grad_norm": 0.5294894954273769, "learning_rate": 7.165662193982419e-05, "loss": 0.9836, "step": 3183 }, { "epoch": 1.4426823742637063, "grad_norm": 0.5210274728735977, "learning_rate": 7.164888543256963e-05, "loss": 0.9666, "step": 3184 }, { "epoch": 1.4431354780244676, "grad_norm": 0.5659465428005797, "learning_rate": 7.164114575813907e-05, "loss": 0.9785, "step": 3185 }, { "epoch": 1.4435885817852288, "grad_norm": 0.5907646717340903, "learning_rate": 7.163340291730705e-05, "loss": 0.9614, "step": 3186 }, { "epoch": 1.44404168554599, "grad_norm": 0.6058428426092892, "learning_rate": 7.162565691084837e-05, "loss": 0.9821, "step": 3187 }, { "epoch": 1.4444947893067512, "grad_norm": 0.5867564194907741, "learning_rate": 7.161790773953823e-05, "loss": 0.9594, "step": 3188 }, { "epoch": 1.4449478930675124, "grad_norm": 0.5189166448094523, "learning_rate": 7.161015540415209e-05, "loss": 0.9643, "step": 3189 }, { "epoch": 1.4454009968282737, "grad_norm": 0.3799776170023693, "learning_rate": 7.160239990546575e-05, "loss": 0.9926, "step": 3190 }, { "epoch": 1.4458541005890349, "grad_norm": 0.2591419846748774, "learning_rate": 7.159464124425533e-05, "loss": 0.9813, "step": 3191 }, { "epoch": 1.4463072043497962, "grad_norm": 0.265568087466485, "learning_rate": 7.158687942129725e-05, "loss": 0.9739, "step": 3192 }, { "epoch": 1.4467603081105573, "grad_norm": 0.32748249586718214, "learning_rate": 7.157911443736822e-05, "loss": 0.9922, "step": 3193 }, { "epoch": 1.4472134118713185, "grad_norm": 0.3400677676056585, "learning_rate": 7.157134629324535e-05, "loss": 0.9613, "step": 3194 }, { "epoch": 1.4476665156320798, "grad_norm": 0.29818368580846033, "learning_rate": 7.1563574989706e-05, "loss": 0.9736, "step": 3195 }, { "epoch": 1.448119619392841, "grad_norm": 0.2927605896978355, "learning_rate": 7.155580052752785e-05, "loss": 0.9837, "step": 3196 }, { "epoch": 1.4485727231536023, "grad_norm": 0.2982991844003055, "learning_rate": 7.154802290748892e-05, "loss": 0.9589, "step": 3197 }, { "epoch": 1.4490258269143634, "grad_norm": 0.3075310053013729, "learning_rate": 7.154024213036752e-05, "loss": 0.9691, "step": 3198 }, { "epoch": 1.4494789306751246, "grad_norm": 0.32328624951577767, "learning_rate": 7.153245819694231e-05, "loss": 0.966, "step": 3199 }, { "epoch": 1.4499320344358857, "grad_norm": 0.3799930195977658, "learning_rate": 7.152467110799222e-05, "loss": 0.9655, "step": 3200 }, { "epoch": 1.450385138196647, "grad_norm": 0.4757670824675854, "learning_rate": 7.151688086429655e-05, "loss": 0.9671, "step": 3201 }, { "epoch": 1.4508382419574082, "grad_norm": 0.533005649708345, "learning_rate": 7.150908746663487e-05, "loss": 0.9818, "step": 3202 }, { "epoch": 1.4512913457181695, "grad_norm": 0.6471380491271996, "learning_rate": 7.15012909157871e-05, "loss": 0.9837, "step": 3203 }, { "epoch": 1.4517444494789307, "grad_norm": 0.7016746803575051, "learning_rate": 7.149349121253343e-05, "loss": 0.9645, "step": 3204 }, { "epoch": 1.4521975532396918, "grad_norm": 0.6489572382697584, "learning_rate": 7.14856883576544e-05, "loss": 0.9861, "step": 3205 }, { "epoch": 1.4526506570004531, "grad_norm": 0.6013066904844911, "learning_rate": 7.147788235193089e-05, "loss": 0.9798, "step": 3206 }, { "epoch": 1.4531037607612143, "grad_norm": 0.5334516720693349, "learning_rate": 7.147007319614406e-05, "loss": 0.9739, "step": 3207 }, { "epoch": 1.4535568645219756, "grad_norm": 0.46799618981332175, "learning_rate": 7.146226089107533e-05, "loss": 0.9809, "step": 3208 }, { "epoch": 1.4540099682827368, "grad_norm": 0.43739143992318497, "learning_rate": 7.145444543750657e-05, "loss": 0.9786, "step": 3209 }, { "epoch": 1.454463072043498, "grad_norm": 0.3690378148106588, "learning_rate": 7.144662683621986e-05, "loss": 0.9551, "step": 3210 }, { "epoch": 1.4549161758042592, "grad_norm": 0.3038300788321441, "learning_rate": 7.143880508799761e-05, "loss": 0.9623, "step": 3211 }, { "epoch": 1.4553692795650204, "grad_norm": 0.5658196157485228, "learning_rate": 7.143098019362258e-05, "loss": 0.9729, "step": 3212 }, { "epoch": 1.4558223833257817, "grad_norm": 0.37542583793088546, "learning_rate": 7.142315215387783e-05, "loss": 0.9553, "step": 3213 }, { "epoch": 1.4562754870865429, "grad_norm": 0.4420943926738843, "learning_rate": 7.141532096954671e-05, "loss": 0.948, "step": 3214 }, { "epoch": 1.456728590847304, "grad_norm": 0.46395275629383237, "learning_rate": 7.140748664141292e-05, "loss": 0.9687, "step": 3215 }, { "epoch": 1.4571816946080651, "grad_norm": 0.4446988751524974, "learning_rate": 7.139964917026046e-05, "loss": 0.9638, "step": 3216 }, { "epoch": 1.4576347983688265, "grad_norm": 0.4608694782020171, "learning_rate": 7.139180855687362e-05, "loss": 0.9578, "step": 3217 }, { "epoch": 1.4580879021295876, "grad_norm": 0.5174809449919646, "learning_rate": 7.138396480203705e-05, "loss": 0.9623, "step": 3218 }, { "epoch": 1.458541005890349, "grad_norm": 0.5525794270213218, "learning_rate": 7.137611790653569e-05, "loss": 0.9671, "step": 3219 }, { "epoch": 1.45899410965111, "grad_norm": 0.6594275805990579, "learning_rate": 7.136826787115481e-05, "loss": 0.9479, "step": 3220 }, { "epoch": 1.4594472134118712, "grad_norm": 0.7092538699875235, "learning_rate": 7.136041469667993e-05, "loss": 0.9638, "step": 3221 }, { "epoch": 1.4599003171726326, "grad_norm": 0.6155822154236881, "learning_rate": 7.1352558383897e-05, "loss": 0.9814, "step": 3222 }, { "epoch": 1.4603534209333937, "grad_norm": 0.42541751787412474, "learning_rate": 7.134469893359218e-05, "loss": 0.9907, "step": 3223 }, { "epoch": 1.460806524694155, "grad_norm": 0.30831050131224746, "learning_rate": 7.133683634655199e-05, "loss": 0.9799, "step": 3224 }, { "epoch": 1.4612596284549162, "grad_norm": 0.33016479156945944, "learning_rate": 7.132897062356326e-05, "loss": 0.9776, "step": 3225 }, { "epoch": 1.4617127322156773, "grad_norm": 0.36905761491706546, "learning_rate": 7.132110176541313e-05, "loss": 0.975, "step": 3226 }, { "epoch": 1.4621658359764387, "grad_norm": 0.3715570945413755, "learning_rate": 7.131322977288906e-05, "loss": 0.9624, "step": 3227 }, { "epoch": 1.4626189397371998, "grad_norm": 0.3494851345137042, "learning_rate": 7.13053546467788e-05, "loss": 0.9977, "step": 3228 }, { "epoch": 1.4630720434979612, "grad_norm": 0.40654040706168604, "learning_rate": 7.129747638787045e-05, "loss": 1.0005, "step": 3229 }, { "epoch": 1.4635251472587223, "grad_norm": 0.4769042494245993, "learning_rate": 7.12895949969524e-05, "loss": 0.959, "step": 3230 }, { "epoch": 1.4639782510194834, "grad_norm": 0.48813347453929573, "learning_rate": 7.128171047481336e-05, "loss": 0.9691, "step": 3231 }, { "epoch": 1.4644313547802446, "grad_norm": 0.45411521728396304, "learning_rate": 7.127382282224235e-05, "loss": 0.9626, "step": 3232 }, { "epoch": 1.464884458541006, "grad_norm": 0.41223189476956645, "learning_rate": 7.126593204002872e-05, "loss": 0.9501, "step": 3233 }, { "epoch": 1.465337562301767, "grad_norm": 0.40523820300394114, "learning_rate": 7.125803812896209e-05, "loss": 0.9485, "step": 3234 }, { "epoch": 1.4657906660625284, "grad_norm": 0.38940891532427774, "learning_rate": 7.125014108983243e-05, "loss": 0.9626, "step": 3235 }, { "epoch": 1.4662437698232895, "grad_norm": 0.39896470694518466, "learning_rate": 7.124224092343003e-05, "loss": 0.9883, "step": 3236 }, { "epoch": 1.4666968735840507, "grad_norm": 0.4326083256540301, "learning_rate": 7.123433763054547e-05, "loss": 0.9667, "step": 3237 }, { "epoch": 1.467149977344812, "grad_norm": 0.42987968012886935, "learning_rate": 7.122643121196964e-05, "loss": 0.9915, "step": 3238 }, { "epoch": 1.4676030811055731, "grad_norm": 0.38724325935259585, "learning_rate": 7.121852166849378e-05, "loss": 0.9845, "step": 3239 }, { "epoch": 1.4680561848663345, "grad_norm": 0.37203151588024186, "learning_rate": 7.121060900090938e-05, "loss": 0.9956, "step": 3240 }, { "epoch": 1.4685092886270956, "grad_norm": 0.419840726296164, "learning_rate": 7.12026932100083e-05, "loss": 0.9539, "step": 3241 }, { "epoch": 1.4689623923878568, "grad_norm": 0.4387311198920657, "learning_rate": 7.11947742965827e-05, "loss": 0.9584, "step": 3242 }, { "epoch": 1.469415496148618, "grad_norm": 0.3491681286447227, "learning_rate": 7.118685226142501e-05, "loss": 0.9766, "step": 3243 }, { "epoch": 1.4698685999093792, "grad_norm": 0.3448910880640826, "learning_rate": 7.117892710532806e-05, "loss": 0.9856, "step": 3244 }, { "epoch": 1.4703217036701406, "grad_norm": 0.3951887910944432, "learning_rate": 7.117099882908488e-05, "loss": 0.9694, "step": 3245 }, { "epoch": 1.4707748074309017, "grad_norm": 0.42020321920233655, "learning_rate": 7.116306743348891e-05, "loss": 0.9725, "step": 3246 }, { "epoch": 1.4712279111916629, "grad_norm": 0.4370597034691595, "learning_rate": 7.115513291933384e-05, "loss": 0.9654, "step": 3247 }, { "epoch": 1.471681014952424, "grad_norm": 0.44845499754333196, "learning_rate": 7.114719528741371e-05, "loss": 0.9893, "step": 3248 }, { "epoch": 1.4721341187131853, "grad_norm": 0.4040280155846247, "learning_rate": 7.113925453852287e-05, "loss": 0.9593, "step": 3249 }, { "epoch": 1.4725872224739465, "grad_norm": 0.34217779543536664, "learning_rate": 7.113131067345592e-05, "loss": 0.9725, "step": 3250 }, { "epoch": 1.4730403262347078, "grad_norm": 0.33140671536325644, "learning_rate": 7.112336369300786e-05, "loss": 0.9766, "step": 3251 }, { "epoch": 1.473493429995469, "grad_norm": 0.3416492573292916, "learning_rate": 7.111541359797397e-05, "loss": 0.9463, "step": 3252 }, { "epoch": 1.47394653375623, "grad_norm": 0.34244281709978586, "learning_rate": 7.110746038914979e-05, "loss": 0.9683, "step": 3253 }, { "epoch": 1.4743996375169914, "grad_norm": 0.369231019049424, "learning_rate": 7.109950406733127e-05, "loss": 0.9627, "step": 3254 }, { "epoch": 1.4748527412777526, "grad_norm": 0.44326199633265345, "learning_rate": 7.109154463331458e-05, "loss": 0.9658, "step": 3255 }, { "epoch": 1.475305845038514, "grad_norm": 0.4623420329722257, "learning_rate": 7.108358208789625e-05, "loss": 0.9743, "step": 3256 }, { "epoch": 1.475758948799275, "grad_norm": 0.47173077650197304, "learning_rate": 7.10756164318731e-05, "loss": 0.9589, "step": 3257 }, { "epoch": 1.4762120525600362, "grad_norm": 0.5210988958609651, "learning_rate": 7.106764766604229e-05, "loss": 0.9808, "step": 3258 }, { "epoch": 1.4766651563207975, "grad_norm": 0.49348177965184664, "learning_rate": 7.105967579120125e-05, "loss": 0.98, "step": 3259 }, { "epoch": 1.4771182600815587, "grad_norm": 0.4151235898129536, "learning_rate": 7.105170080814777e-05, "loss": 0.9565, "step": 3260 }, { "epoch": 1.47757136384232, "grad_norm": 0.383617848826308, "learning_rate": 7.104372271767991e-05, "loss": 0.9564, "step": 3261 }, { "epoch": 1.4780244676030811, "grad_norm": 0.40222291177320346, "learning_rate": 7.103574152059607e-05, "loss": 0.9911, "step": 3262 }, { "epoch": 1.4784775713638423, "grad_norm": 0.378767434982895, "learning_rate": 7.102775721769491e-05, "loss": 0.9652, "step": 3263 }, { "epoch": 1.4789306751246034, "grad_norm": 0.3765459554636558, "learning_rate": 7.101976980977547e-05, "loss": 0.968, "step": 3264 }, { "epoch": 1.4793837788853648, "grad_norm": 0.3407697304733331, "learning_rate": 7.101177929763707e-05, "loss": 0.968, "step": 3265 }, { "epoch": 1.479836882646126, "grad_norm": 0.38561046525059123, "learning_rate": 7.100378568207932e-05, "loss": 0.9649, "step": 3266 }, { "epoch": 1.4802899864068872, "grad_norm": 0.3608796505866142, "learning_rate": 7.099578896390217e-05, "loss": 0.9839, "step": 3267 }, { "epoch": 1.4807430901676484, "grad_norm": 0.3191806552981592, "learning_rate": 7.098778914390585e-05, "loss": 0.9597, "step": 3268 }, { "epoch": 1.4811961939284095, "grad_norm": 0.30317888173976965, "learning_rate": 7.097978622289096e-05, "loss": 0.953, "step": 3269 }, { "epoch": 1.4816492976891709, "grad_norm": 0.29689379309357844, "learning_rate": 7.097178020165833e-05, "loss": 0.9556, "step": 3270 }, { "epoch": 1.482102401449932, "grad_norm": 0.28625817080177124, "learning_rate": 7.096377108100917e-05, "loss": 0.9656, "step": 3271 }, { "epoch": 1.4825555052106933, "grad_norm": 0.3066060218774467, "learning_rate": 7.095575886174497e-05, "loss": 0.98, "step": 3272 }, { "epoch": 1.4830086089714545, "grad_norm": 0.309498783453517, "learning_rate": 7.09477435446675e-05, "loss": 0.9717, "step": 3273 }, { "epoch": 1.4834617127322156, "grad_norm": 0.3787469722019609, "learning_rate": 7.09397251305789e-05, "loss": 0.9795, "step": 3274 }, { "epoch": 1.483914816492977, "grad_norm": 0.45290637678028917, "learning_rate": 7.093170362028158e-05, "loss": 0.9798, "step": 3275 }, { "epoch": 1.484367920253738, "grad_norm": 0.5706634585809769, "learning_rate": 7.092367901457827e-05, "loss": 0.9557, "step": 3276 }, { "epoch": 1.4848210240144994, "grad_norm": 0.6516085946570979, "learning_rate": 7.091565131427201e-05, "loss": 0.9588, "step": 3277 }, { "epoch": 1.4852741277752606, "grad_norm": 0.6611397582865328, "learning_rate": 7.090762052016617e-05, "loss": 0.9791, "step": 3278 }, { "epoch": 1.4857272315360217, "grad_norm": 0.6467414071513158, "learning_rate": 7.089958663306439e-05, "loss": 0.9705, "step": 3279 }, { "epoch": 1.4861803352967828, "grad_norm": 0.5865561075533884, "learning_rate": 7.089154965377064e-05, "loss": 0.9676, "step": 3280 }, { "epoch": 1.4866334390575442, "grad_norm": 0.4745660454997328, "learning_rate": 7.08835095830892e-05, "loss": 0.9624, "step": 3281 }, { "epoch": 1.4870865428183053, "grad_norm": 0.4026647854347219, "learning_rate": 7.087546642182465e-05, "loss": 0.9725, "step": 3282 }, { "epoch": 1.4875396465790667, "grad_norm": 0.31446148094496124, "learning_rate": 7.086742017078191e-05, "loss": 0.945, "step": 3283 }, { "epoch": 1.4879927503398278, "grad_norm": 0.2954760964712175, "learning_rate": 7.085937083076618e-05, "loss": 0.9721, "step": 3284 }, { "epoch": 1.488445854100589, "grad_norm": 0.3444078866660036, "learning_rate": 7.085131840258298e-05, "loss": 0.9836, "step": 3285 }, { "epoch": 1.4888989578613503, "grad_norm": 0.4405685447602796, "learning_rate": 7.084326288703809e-05, "loss": 0.9667, "step": 3286 }, { "epoch": 1.4893520616221114, "grad_norm": 0.521883579108938, "learning_rate": 7.08352042849377e-05, "loss": 0.9629, "step": 3287 }, { "epoch": 1.4898051653828728, "grad_norm": 0.45722481978977253, "learning_rate": 7.082714259708824e-05, "loss": 0.9845, "step": 3288 }, { "epoch": 1.490258269143634, "grad_norm": 0.35368239787539757, "learning_rate": 7.081907782429643e-05, "loss": 0.9724, "step": 3289 }, { "epoch": 1.490711372904395, "grad_norm": 0.3152037293613661, "learning_rate": 7.081100996736937e-05, "loss": 0.9563, "step": 3290 }, { "epoch": 1.4911644766651564, "grad_norm": 0.3354086459297963, "learning_rate": 7.080293902711441e-05, "loss": 0.9513, "step": 3291 }, { "epoch": 1.4916175804259175, "grad_norm": 0.36695856556550427, "learning_rate": 7.079486500433922e-05, "loss": 0.9506, "step": 3292 }, { "epoch": 1.4920706841866789, "grad_norm": 0.3403555975536698, "learning_rate": 7.07867878998518e-05, "loss": 0.9688, "step": 3293 }, { "epoch": 1.49252378794744, "grad_norm": 0.3148903295465063, "learning_rate": 7.077870771446045e-05, "loss": 0.9746, "step": 3294 }, { "epoch": 1.4929768917082011, "grad_norm": 0.28828040240166797, "learning_rate": 7.077062444897375e-05, "loss": 0.9751, "step": 3295 }, { "epoch": 1.4934299954689623, "grad_norm": 0.29686715986181916, "learning_rate": 7.076253810420063e-05, "loss": 0.953, "step": 3296 }, { "epoch": 1.4938830992297236, "grad_norm": 0.24279578748735794, "learning_rate": 7.07544486809503e-05, "loss": 0.9863, "step": 3297 }, { "epoch": 1.4943362029904848, "grad_norm": 0.23711096359527636, "learning_rate": 7.074635618003229e-05, "loss": 0.9442, "step": 3298 }, { "epoch": 1.494789306751246, "grad_norm": 0.3190672343227007, "learning_rate": 7.073826060225642e-05, "loss": 0.9731, "step": 3299 }, { "epoch": 1.4952424105120072, "grad_norm": 0.34390661345137075, "learning_rate": 7.073016194843285e-05, "loss": 0.9654, "step": 3300 }, { "epoch": 1.4956955142727684, "grad_norm": 0.40067363095846475, "learning_rate": 7.072206021937203e-05, "loss": 0.9793, "step": 3301 }, { "epoch": 1.4961486180335297, "grad_norm": 0.46681725536887725, "learning_rate": 7.071395541588472e-05, "loss": 0.9525, "step": 3302 }, { "epoch": 1.4966017217942909, "grad_norm": 0.46055708336283346, "learning_rate": 7.070584753878198e-05, "loss": 0.9677, "step": 3303 }, { "epoch": 1.4970548255550522, "grad_norm": 0.45529545117403974, "learning_rate": 7.069773658887519e-05, "loss": 0.9896, "step": 3304 }, { "epoch": 1.4975079293158133, "grad_norm": 0.4826247345923222, "learning_rate": 7.068962256697601e-05, "loss": 0.9928, "step": 3305 }, { "epoch": 1.4979610330765745, "grad_norm": 0.49054340254078177, "learning_rate": 7.068150547389646e-05, "loss": 0.9616, "step": 3306 }, { "epoch": 1.4984141368373358, "grad_norm": 0.4178700027018327, "learning_rate": 7.067338531044882e-05, "loss": 0.9653, "step": 3307 }, { "epoch": 1.498867240598097, "grad_norm": 0.3321162204638854, "learning_rate": 7.06652620774457e-05, "loss": 0.9743, "step": 3308 }, { "epoch": 1.4993203443588583, "grad_norm": 0.3474923566395956, "learning_rate": 7.065713577570001e-05, "loss": 0.9909, "step": 3309 }, { "epoch": 1.4997734481196194, "grad_norm": 0.39980456595740754, "learning_rate": 7.064900640602494e-05, "loss": 0.9699, "step": 3310 }, { "epoch": 1.5002265518803806, "grad_norm": 0.3850435687948706, "learning_rate": 7.064087396923403e-05, "loss": 0.9723, "step": 3311 }, { "epoch": 1.5006796556411417, "grad_norm": 0.3096274203580609, "learning_rate": 7.063273846614115e-05, "loss": 0.9765, "step": 3312 }, { "epoch": 1.501132759401903, "grad_norm": 0.24419502055701728, "learning_rate": 7.062459989756039e-05, "loss": 0.9631, "step": 3313 }, { "epoch": 1.5015858631626644, "grad_norm": 0.25709864464105914, "learning_rate": 7.061645826430621e-05, "loss": 0.997, "step": 3314 }, { "epoch": 1.5020389669234255, "grad_norm": 0.258429320349637, "learning_rate": 7.060831356719338e-05, "loss": 0.983, "step": 3315 }, { "epoch": 1.5024920706841867, "grad_norm": 0.286176266484077, "learning_rate": 7.060016580703693e-05, "loss": 0.9545, "step": 3316 }, { "epoch": 1.5029451744449478, "grad_norm": 0.29748455947767455, "learning_rate": 7.059201498465224e-05, "loss": 0.9664, "step": 3317 }, { "epoch": 1.5033982782057091, "grad_norm": 0.3390150426063305, "learning_rate": 7.058386110085497e-05, "loss": 0.9509, "step": 3318 }, { "epoch": 1.5038513819664703, "grad_norm": 0.3529232970450808, "learning_rate": 7.057570415646111e-05, "loss": 0.9665, "step": 3319 }, { "epoch": 1.5043044857272316, "grad_norm": 0.42719847764587054, "learning_rate": 7.056754415228693e-05, "loss": 0.995, "step": 3320 }, { "epoch": 1.5047575894879928, "grad_norm": 0.510257402544283, "learning_rate": 7.055938108914905e-05, "loss": 0.9662, "step": 3321 }, { "epoch": 1.505210693248754, "grad_norm": 0.560542678218709, "learning_rate": 7.055121496786433e-05, "loss": 0.9827, "step": 3322 }, { "epoch": 1.505663797009515, "grad_norm": 0.5381239812055382, "learning_rate": 7.054304578925e-05, "loss": 0.968, "step": 3323 }, { "epoch": 1.5061169007702764, "grad_norm": 0.49789172831310774, "learning_rate": 7.053487355412355e-05, "loss": 0.9637, "step": 3324 }, { "epoch": 1.5065700045310377, "grad_norm": 0.47401368011576583, "learning_rate": 7.05266982633028e-05, "loss": 0.9612, "step": 3325 }, { "epoch": 1.5070231082917989, "grad_norm": 0.465410192730969, "learning_rate": 7.051851991760588e-05, "loss": 0.9781, "step": 3326 }, { "epoch": 1.50747621205256, "grad_norm": 0.47791596661236724, "learning_rate": 7.05103385178512e-05, "loss": 0.9726, "step": 3327 }, { "epoch": 1.5079293158133211, "grad_norm": 0.5111833402920121, "learning_rate": 7.050215406485751e-05, "loss": 0.9567, "step": 3328 }, { "epoch": 1.5083824195740825, "grad_norm": 0.5029849887158395, "learning_rate": 7.049396655944381e-05, "loss": 0.9728, "step": 3329 }, { "epoch": 1.5088355233348438, "grad_norm": 0.4980997338470003, "learning_rate": 7.04857760024295e-05, "loss": 0.9591, "step": 3330 }, { "epoch": 1.509288627095605, "grad_norm": 0.43973216578104685, "learning_rate": 7.047758239463417e-05, "loss": 0.9675, "step": 3331 }, { "epoch": 1.509741730856366, "grad_norm": 0.3856085447975661, "learning_rate": 7.046938573687782e-05, "loss": 0.9718, "step": 3332 }, { "epoch": 1.5101948346171272, "grad_norm": 0.3443035916729789, "learning_rate": 7.046118602998066e-05, "loss": 0.9622, "step": 3333 }, { "epoch": 1.5106479383778886, "grad_norm": 0.3577552643872268, "learning_rate": 7.04529832747633e-05, "loss": 0.9608, "step": 3334 }, { "epoch": 1.5111010421386497, "grad_norm": 0.3761697642636337, "learning_rate": 7.044477747204656e-05, "loss": 0.9723, "step": 3335 }, { "epoch": 1.511554145899411, "grad_norm": 0.407601022556669, "learning_rate": 7.043656862265166e-05, "loss": 0.9487, "step": 3336 }, { "epoch": 1.5120072496601722, "grad_norm": 0.4301063038971208, "learning_rate": 7.042835672740007e-05, "loss": 0.9858, "step": 3337 }, { "epoch": 1.5124603534209333, "grad_norm": 0.41884143075644986, "learning_rate": 7.042014178711355e-05, "loss": 0.9668, "step": 3338 }, { "epoch": 1.5129134571816945, "grad_norm": 0.45550449374260044, "learning_rate": 7.041192380261421e-05, "loss": 0.9686, "step": 3339 }, { "epoch": 1.5133665609424558, "grad_norm": 0.4373497763052047, "learning_rate": 7.040370277472444e-05, "loss": 0.9971, "step": 3340 }, { "epoch": 1.5138196647032172, "grad_norm": 0.3895507173836733, "learning_rate": 7.03954787042669e-05, "loss": 0.9614, "step": 3341 }, { "epoch": 1.5142727684639783, "grad_norm": 0.4160901845897792, "learning_rate": 7.038725159206463e-05, "loss": 1.0053, "step": 3342 }, { "epoch": 1.5147258722247394, "grad_norm": 0.46292029738771046, "learning_rate": 7.037902143894094e-05, "loss": 0.9638, "step": 3343 }, { "epoch": 1.5151789759855006, "grad_norm": 0.46969774198527064, "learning_rate": 7.037078824571942e-05, "loss": 0.9692, "step": 3344 }, { "epoch": 1.515632079746262, "grad_norm": 0.4251435288471086, "learning_rate": 7.0362552013224e-05, "loss": 0.9645, "step": 3345 }, { "epoch": 1.5160851835070233, "grad_norm": 0.4047849370884844, "learning_rate": 7.035431274227888e-05, "loss": 0.9746, "step": 3346 }, { "epoch": 1.5165382872677844, "grad_norm": 0.39202689152114406, "learning_rate": 7.034607043370858e-05, "loss": 0.9718, "step": 3347 }, { "epoch": 1.5169913910285455, "grad_norm": 0.3966384135650265, "learning_rate": 7.033782508833797e-05, "loss": 0.9526, "step": 3348 }, { "epoch": 1.5174444947893067, "grad_norm": 0.3882291492560413, "learning_rate": 7.032957670699212e-05, "loss": 0.9501, "step": 3349 }, { "epoch": 1.517897598550068, "grad_norm": 0.327768213815262, "learning_rate": 7.032132529049652e-05, "loss": 0.9702, "step": 3350 }, { "epoch": 1.5183507023108291, "grad_norm": 0.2727051510326374, "learning_rate": 7.031307083967687e-05, "loss": 0.9728, "step": 3351 }, { "epoch": 1.5188038060715905, "grad_norm": 0.29330468115444525, "learning_rate": 7.030481335535923e-05, "loss": 0.9897, "step": 3352 }, { "epoch": 1.5192569098323516, "grad_norm": 0.30091465921401833, "learning_rate": 7.029655283836992e-05, "loss": 0.9629, "step": 3353 }, { "epoch": 1.5197100135931128, "grad_norm": 0.2639883020902882, "learning_rate": 7.028828928953562e-05, "loss": 0.9675, "step": 3354 }, { "epoch": 1.5201631173538739, "grad_norm": 0.24164349451583841, "learning_rate": 7.028002270968328e-05, "loss": 0.9659, "step": 3355 }, { "epoch": 1.5206162211146352, "grad_norm": 0.26226288860084246, "learning_rate": 7.027175309964014e-05, "loss": 0.9711, "step": 3356 }, { "epoch": 1.5210693248753966, "grad_norm": 0.31480445026668, "learning_rate": 7.026348046023376e-05, "loss": 0.9824, "step": 3357 }, { "epoch": 1.5215224286361577, "grad_norm": 0.4071562398761442, "learning_rate": 7.025520479229201e-05, "loss": 0.9741, "step": 3358 }, { "epoch": 1.5219755323969189, "grad_norm": 0.4826852679796623, "learning_rate": 7.024692609664303e-05, "loss": 0.9609, "step": 3359 }, { "epoch": 1.52242863615768, "grad_norm": 0.5167358164217412, "learning_rate": 7.023864437411533e-05, "loss": 0.9599, "step": 3360 }, { "epoch": 1.5228817399184413, "grad_norm": 0.5645081253897578, "learning_rate": 7.023035962553765e-05, "loss": 0.9666, "step": 3361 }, { "epoch": 1.5233348436792027, "grad_norm": 0.6026831949128877, "learning_rate": 7.022207185173908e-05, "loss": 0.9656, "step": 3362 }, { "epoch": 1.5237879474399638, "grad_norm": 0.6586164155571661, "learning_rate": 7.021378105354899e-05, "loss": 0.9634, "step": 3363 }, { "epoch": 1.524241051200725, "grad_norm": 0.6767337580652054, "learning_rate": 7.020548723179707e-05, "loss": 0.9805, "step": 3364 }, { "epoch": 1.524694154961486, "grad_norm": 0.6491193090368959, "learning_rate": 7.019719038731326e-05, "loss": 0.977, "step": 3365 }, { "epoch": 1.5251472587222474, "grad_norm": 0.545092633810732, "learning_rate": 7.01888905209279e-05, "loss": 0.9572, "step": 3366 }, { "epoch": 1.5256003624830086, "grad_norm": 0.3987878553803188, "learning_rate": 7.018058763347155e-05, "loss": 0.9509, "step": 3367 }, { "epoch": 1.52605346624377, "grad_norm": 0.29319161546327394, "learning_rate": 7.017228172577508e-05, "loss": 0.982, "step": 3368 }, { "epoch": 1.526506570004531, "grad_norm": 0.264950364029065, "learning_rate": 7.016397279866973e-05, "loss": 0.9601, "step": 3369 }, { "epoch": 1.5269596737652922, "grad_norm": 0.34003537254286087, "learning_rate": 7.015566085298693e-05, "loss": 0.9546, "step": 3370 }, { "epoch": 1.5274127775260533, "grad_norm": 0.45121029559700504, "learning_rate": 7.014734588955852e-05, "loss": 0.9615, "step": 3371 }, { "epoch": 1.5278658812868147, "grad_norm": 0.5782488767935151, "learning_rate": 7.01390279092166e-05, "loss": 0.9576, "step": 3372 }, { "epoch": 1.528318985047576, "grad_norm": 0.6282939869402177, "learning_rate": 7.013070691279356e-05, "loss": 0.9768, "step": 3373 }, { "epoch": 1.5287720888083371, "grad_norm": 0.6090110644108425, "learning_rate": 7.012238290112208e-05, "loss": 0.9576, "step": 3374 }, { "epoch": 1.5292251925690983, "grad_norm": 0.5447513266837195, "learning_rate": 7.011405587503519e-05, "loss": 0.9536, "step": 3375 }, { "epoch": 1.5296782963298594, "grad_norm": 0.474312842424407, "learning_rate": 7.010572583536617e-05, "loss": 0.9748, "step": 3376 }, { "epoch": 1.5301314000906208, "grad_norm": 0.3988265719815056, "learning_rate": 7.009739278294865e-05, "loss": 0.9795, "step": 3377 }, { "epoch": 1.5305845038513821, "grad_norm": 0.38677473640508053, "learning_rate": 7.008905671861654e-05, "loss": 0.9583, "step": 3378 }, { "epoch": 1.5310376076121432, "grad_norm": 0.38006275374116943, "learning_rate": 7.008071764320401e-05, "loss": 0.9506, "step": 3379 }, { "epoch": 1.5314907113729044, "grad_norm": 0.34586247765044403, "learning_rate": 7.007237555754561e-05, "loss": 0.9503, "step": 3380 }, { "epoch": 1.5319438151336655, "grad_norm": 0.36060482245453307, "learning_rate": 7.006403046247615e-05, "loss": 0.9551, "step": 3381 }, { "epoch": 1.5323969188944269, "grad_norm": 0.4269271824245829, "learning_rate": 7.005568235883073e-05, "loss": 0.989, "step": 3382 }, { "epoch": 1.532850022655188, "grad_norm": 0.4622553100133435, "learning_rate": 7.004733124744475e-05, "loss": 0.9628, "step": 3383 }, { "epoch": 1.5333031264159493, "grad_norm": 0.37181237780061216, "learning_rate": 7.003897712915394e-05, "loss": 0.9894, "step": 3384 }, { "epoch": 1.5337562301767105, "grad_norm": 0.2532085823078384, "learning_rate": 7.003062000479432e-05, "loss": 0.9825, "step": 3385 }, { "epoch": 1.5342093339374716, "grad_norm": 0.2261023934066838, "learning_rate": 7.00222598752022e-05, "loss": 0.972, "step": 3386 }, { "epoch": 1.5346624376982327, "grad_norm": 0.26560981553778545, "learning_rate": 7.00138967412142e-05, "loss": 0.9979, "step": 3387 }, { "epoch": 1.535115541458994, "grad_norm": 0.33170392556962536, "learning_rate": 7.000553060366724e-05, "loss": 0.9787, "step": 3388 }, { "epoch": 1.5355686452197554, "grad_norm": 0.3720720412608704, "learning_rate": 6.999716146339853e-05, "loss": 0.979, "step": 3389 }, { "epoch": 1.5360217489805166, "grad_norm": 0.38131326381592606, "learning_rate": 6.998878932124561e-05, "loss": 0.9812, "step": 3390 }, { "epoch": 1.5364748527412777, "grad_norm": 0.4074006801034606, "learning_rate": 6.998041417804626e-05, "loss": 0.9656, "step": 3391 }, { "epoch": 1.5369279565020388, "grad_norm": 0.43772630304495336, "learning_rate": 6.997203603463864e-05, "loss": 0.9825, "step": 3392 }, { "epoch": 1.5373810602628002, "grad_norm": 0.4673488003445348, "learning_rate": 6.996365489186114e-05, "loss": 0.956, "step": 3393 }, { "epoch": 1.5378341640235615, "grad_norm": 0.4622447546416373, "learning_rate": 6.99552707505525e-05, "loss": 0.9554, "step": 3394 }, { "epoch": 1.5382872677843227, "grad_norm": 0.4185605427778667, "learning_rate": 6.994688361155174e-05, "loss": 0.9691, "step": 3395 }, { "epoch": 1.5387403715450838, "grad_norm": 0.3563594841029583, "learning_rate": 6.993849347569816e-05, "loss": 0.9659, "step": 3396 }, { "epoch": 1.539193475305845, "grad_norm": 0.3270907300771423, "learning_rate": 6.99301003438314e-05, "loss": 0.972, "step": 3397 }, { "epoch": 1.5396465790666063, "grad_norm": 0.2949969640086655, "learning_rate": 6.992170421679138e-05, "loss": 0.953, "step": 3398 }, { "epoch": 1.5400996828273674, "grad_norm": 0.2757843802334836, "learning_rate": 6.99133050954183e-05, "loss": 0.9752, "step": 3399 }, { "epoch": 1.5405527865881288, "grad_norm": 0.3313891261137318, "learning_rate": 6.990490298055269e-05, "loss": 0.9885, "step": 3400 }, { "epoch": 1.54100589034889, "grad_norm": 0.3724352042434093, "learning_rate": 6.989649787303538e-05, "loss": 0.9436, "step": 3401 }, { "epoch": 1.541458994109651, "grad_norm": 0.32569182924972623, "learning_rate": 6.988808977370748e-05, "loss": 0.997, "step": 3402 }, { "epoch": 1.5419120978704122, "grad_norm": 0.2826103876967948, "learning_rate": 6.98796786834104e-05, "loss": 0.9717, "step": 3403 }, { "epoch": 1.5423652016311735, "grad_norm": 0.27364271807326174, "learning_rate": 6.987126460298587e-05, "loss": 0.9579, "step": 3404 }, { "epoch": 1.5428183053919349, "grad_norm": 0.2851184332581058, "learning_rate": 6.986284753327589e-05, "loss": 0.9821, "step": 3405 }, { "epoch": 1.543271409152696, "grad_norm": 0.2920084516874362, "learning_rate": 6.985442747512279e-05, "loss": 0.9771, "step": 3406 }, { "epoch": 1.5437245129134571, "grad_norm": 0.2912072435331744, "learning_rate": 6.984600442936918e-05, "loss": 0.9641, "step": 3407 }, { "epoch": 1.5441776166742183, "grad_norm": 0.27061359193921786, "learning_rate": 6.983757839685797e-05, "loss": 0.9734, "step": 3408 }, { "epoch": 1.5446307204349796, "grad_norm": 0.3434564524070744, "learning_rate": 6.982914937843238e-05, "loss": 0.975, "step": 3409 }, { "epoch": 1.545083824195741, "grad_norm": 0.4045640423448532, "learning_rate": 6.982071737493591e-05, "loss": 0.9612, "step": 3410 }, { "epoch": 1.545536927956502, "grad_norm": 0.4232060710901736, "learning_rate": 6.981228238721237e-05, "loss": 0.9696, "step": 3411 }, { "epoch": 1.5459900317172632, "grad_norm": 0.44607194141317213, "learning_rate": 6.980384441610588e-05, "loss": 0.965, "step": 3412 }, { "epoch": 1.5464431354780244, "grad_norm": 0.43756489107893354, "learning_rate": 6.979540346246085e-05, "loss": 0.9859, "step": 3413 }, { "epoch": 1.5468962392387857, "grad_norm": 0.3957524522586598, "learning_rate": 6.978695952712197e-05, "loss": 0.9771, "step": 3414 }, { "epoch": 1.5473493429995469, "grad_norm": 0.3407205366117797, "learning_rate": 6.977851261093425e-05, "loss": 0.9661, "step": 3415 }, { "epoch": 1.5478024467603082, "grad_norm": 0.346865262344219, "learning_rate": 6.977006271474299e-05, "loss": 0.9671, "step": 3416 }, { "epoch": 1.5482555505210693, "grad_norm": 0.37530620483999755, "learning_rate": 6.97616098393938e-05, "loss": 0.9762, "step": 3417 }, { "epoch": 1.5487086542818305, "grad_norm": 0.4360285846423763, "learning_rate": 6.975315398573257e-05, "loss": 0.9525, "step": 3418 }, { "epoch": 1.5491617580425916, "grad_norm": 0.45033865970829207, "learning_rate": 6.974469515460548e-05, "loss": 0.9755, "step": 3419 }, { "epoch": 1.549614861803353, "grad_norm": 0.44029306168005156, "learning_rate": 6.973623334685906e-05, "loss": 0.9718, "step": 3420 }, { "epoch": 1.5500679655641143, "grad_norm": 0.3351725804667881, "learning_rate": 6.972776856334008e-05, "loss": 0.9678, "step": 3421 }, { "epoch": 1.5505210693248754, "grad_norm": 0.2367471622219221, "learning_rate": 6.971930080489564e-05, "loss": 0.9767, "step": 3422 }, { "epoch": 1.5509741730856366, "grad_norm": 0.23590747483727523, "learning_rate": 6.971083007237312e-05, "loss": 0.9534, "step": 3423 }, { "epoch": 1.5514272768463977, "grad_norm": 0.2430724329209363, "learning_rate": 6.97023563666202e-05, "loss": 0.9601, "step": 3424 }, { "epoch": 1.551880380607159, "grad_norm": 0.2874274078006037, "learning_rate": 6.969387968848487e-05, "loss": 0.9901, "step": 3425 }, { "epoch": 1.5523334843679204, "grad_norm": 0.3510968518632312, "learning_rate": 6.968540003881542e-05, "loss": 0.9872, "step": 3426 }, { "epoch": 1.5527865881286815, "grad_norm": 0.4252060660316481, "learning_rate": 6.967691741846043e-05, "loss": 0.9634, "step": 3427 }, { "epoch": 1.5532396918894427, "grad_norm": 0.4848271520656821, "learning_rate": 6.966843182826875e-05, "loss": 0.9878, "step": 3428 }, { "epoch": 1.5536927956502038, "grad_norm": 0.5355200718241138, "learning_rate": 6.965994326908956e-05, "loss": 0.9944, "step": 3429 }, { "epoch": 1.5541458994109651, "grad_norm": 0.5967436410428402, "learning_rate": 6.965145174177234e-05, "loss": 0.9712, "step": 3430 }, { "epoch": 1.5545990031717263, "grad_norm": 0.6572403424506565, "learning_rate": 6.964295724716688e-05, "loss": 0.9492, "step": 3431 }, { "epoch": 1.5550521069324876, "grad_norm": 0.659777326282369, "learning_rate": 6.963445978612319e-05, "loss": 0.9742, "step": 3432 }, { "epoch": 1.5555052106932488, "grad_norm": 0.5790940695710692, "learning_rate": 6.962595935949165e-05, "loss": 0.9765, "step": 3433 }, { "epoch": 1.55595831445401, "grad_norm": 0.42196964530489933, "learning_rate": 6.961745596812294e-05, "loss": 0.9788, "step": 3434 }, { "epoch": 1.556411418214771, "grad_norm": 0.2628213793598541, "learning_rate": 6.9608949612868e-05, "loss": 0.9624, "step": 3435 }, { "epoch": 1.5568645219755324, "grad_norm": 0.2692023013954562, "learning_rate": 6.960044029457807e-05, "loss": 0.9706, "step": 3436 }, { "epoch": 1.5573176257362937, "grad_norm": 0.31159351389425016, "learning_rate": 6.95919280141047e-05, "loss": 0.9599, "step": 3437 }, { "epoch": 1.5577707294970549, "grad_norm": 0.3323764054900298, "learning_rate": 6.958341277229974e-05, "loss": 0.9665, "step": 3438 }, { "epoch": 1.558223833257816, "grad_norm": 0.3374321309886447, "learning_rate": 6.957489457001534e-05, "loss": 1.0031, "step": 3439 }, { "epoch": 1.5586769370185771, "grad_norm": 0.31792526395878934, "learning_rate": 6.956637340810391e-05, "loss": 0.9721, "step": 3440 }, { "epoch": 1.5591300407793385, "grad_norm": 0.31436629533338273, "learning_rate": 6.95578492874182e-05, "loss": 0.9679, "step": 3441 }, { "epoch": 1.5595831445400998, "grad_norm": 0.29620644949893693, "learning_rate": 6.954932220881122e-05, "loss": 0.967, "step": 3442 }, { "epoch": 1.560036248300861, "grad_norm": 0.36584455927315596, "learning_rate": 6.954079217313632e-05, "loss": 0.985, "step": 3443 }, { "epoch": 1.560489352061622, "grad_norm": 0.4619580508190666, "learning_rate": 6.953225918124711e-05, "loss": 0.9498, "step": 3444 }, { "epoch": 1.5609424558223832, "grad_norm": 0.5194932676905349, "learning_rate": 6.95237232339975e-05, "loss": 0.9593, "step": 3445 }, { "epoch": 1.5613955595831446, "grad_norm": 0.543218377662478, "learning_rate": 6.95151843322417e-05, "loss": 0.988, "step": 3446 }, { "epoch": 1.5618486633439057, "grad_norm": 0.5376425125419881, "learning_rate": 6.950664247683423e-05, "loss": 0.9624, "step": 3447 }, { "epoch": 1.562301767104667, "grad_norm": 0.5173534123603607, "learning_rate": 6.949809766862989e-05, "loss": 0.9644, "step": 3448 }, { "epoch": 1.5627548708654282, "grad_norm": 0.4950562174929473, "learning_rate": 6.948954990848376e-05, "loss": 0.969, "step": 3449 }, { "epoch": 1.5632079746261893, "grad_norm": 0.40947730579477937, "learning_rate": 6.948099919725126e-05, "loss": 0.977, "step": 3450 }, { "epoch": 1.5636610783869505, "grad_norm": 0.2989190313255066, "learning_rate": 6.947244553578806e-05, "loss": 0.9902, "step": 3451 }, { "epoch": 1.5641141821477118, "grad_norm": 0.2658136330986648, "learning_rate": 6.946388892495016e-05, "loss": 0.9629, "step": 3452 }, { "epoch": 1.5645672859084732, "grad_norm": 0.354938968066047, "learning_rate": 6.945532936559383e-05, "loss": 0.9648, "step": 3453 }, { "epoch": 1.5650203896692343, "grad_norm": 0.39135825349850856, "learning_rate": 6.944676685857564e-05, "loss": 0.9642, "step": 3454 }, { "epoch": 1.5654734934299954, "grad_norm": 0.3598910824698183, "learning_rate": 6.943820140475246e-05, "loss": 0.9708, "step": 3455 }, { "epoch": 1.5659265971907566, "grad_norm": 0.32490146238092354, "learning_rate": 6.942963300498147e-05, "loss": 0.9617, "step": 3456 }, { "epoch": 1.566379700951518, "grad_norm": 0.2501585035139084, "learning_rate": 6.942106166012011e-05, "loss": 0.9698, "step": 3457 }, { "epoch": 1.5668328047122793, "grad_norm": 0.269900321549305, "learning_rate": 6.941248737102614e-05, "loss": 0.9619, "step": 3458 }, { "epoch": 1.5672859084730404, "grad_norm": 0.25328104270249263, "learning_rate": 6.940391013855762e-05, "loss": 0.973, "step": 3459 }, { "epoch": 1.5677390122338015, "grad_norm": 0.22019481165538612, "learning_rate": 6.939532996357289e-05, "loss": 0.9694, "step": 3460 }, { "epoch": 1.5681921159945627, "grad_norm": 0.23191087769125496, "learning_rate": 6.938674684693057e-05, "loss": 0.9662, "step": 3461 }, { "epoch": 1.568645219755324, "grad_norm": 0.24383471713879823, "learning_rate": 6.93781607894896e-05, "loss": 0.9617, "step": 3462 }, { "epoch": 1.5690983235160851, "grad_norm": 0.21511937901168765, "learning_rate": 6.936957179210922e-05, "loss": 0.9639, "step": 3463 }, { "epoch": 1.5695514272768465, "grad_norm": 0.2558792227799732, "learning_rate": 6.936097985564894e-05, "loss": 0.9775, "step": 3464 }, { "epoch": 1.5700045310376076, "grad_norm": 0.29147431325019424, "learning_rate": 6.935238498096858e-05, "loss": 0.9685, "step": 3465 }, { "epoch": 1.5704576347983688, "grad_norm": 0.29206035494660854, "learning_rate": 6.934378716892824e-05, "loss": 0.9729, "step": 3466 }, { "epoch": 1.5709107385591299, "grad_norm": 0.30179430191732043, "learning_rate": 6.933518642038832e-05, "loss": 0.9539, "step": 3467 }, { "epoch": 1.5713638423198912, "grad_norm": 0.3335853338937421, "learning_rate": 6.932658273620954e-05, "loss": 0.972, "step": 3468 }, { "epoch": 1.5718169460806526, "grad_norm": 0.3548075239404393, "learning_rate": 6.931797611725285e-05, "loss": 0.9514, "step": 3469 }, { "epoch": 1.5722700498414137, "grad_norm": 0.3723999982932913, "learning_rate": 6.930936656437958e-05, "loss": 0.9799, "step": 3470 }, { "epoch": 1.5727231536021749, "grad_norm": 0.445169179537405, "learning_rate": 6.930075407845127e-05, "loss": 0.9708, "step": 3471 }, { "epoch": 1.573176257362936, "grad_norm": 0.5116014907985649, "learning_rate": 6.929213866032981e-05, "loss": 0.9549, "step": 3472 }, { "epoch": 1.5736293611236973, "grad_norm": 0.5562618341949797, "learning_rate": 6.928352031087737e-05, "loss": 0.954, "step": 3473 }, { "epoch": 1.5740824648844587, "grad_norm": 0.5653003432650864, "learning_rate": 6.92748990309564e-05, "loss": 0.9769, "step": 3474 }, { "epoch": 1.5745355686452198, "grad_norm": 0.5008387618112807, "learning_rate": 6.926627482142965e-05, "loss": 0.9729, "step": 3475 }, { "epoch": 1.574988672405981, "grad_norm": 0.379827560093161, "learning_rate": 6.925764768316015e-05, "loss": 0.9938, "step": 3476 }, { "epoch": 1.575441776166742, "grad_norm": 0.3249373946542213, "learning_rate": 6.924901761701124e-05, "loss": 0.9506, "step": 3477 }, { "epoch": 1.5758948799275034, "grad_norm": 0.2688977111677589, "learning_rate": 6.924038462384659e-05, "loss": 0.9396, "step": 3478 }, { "epoch": 1.5763479836882646, "grad_norm": 0.2787376884037513, "learning_rate": 6.923174870453008e-05, "loss": 0.9505, "step": 3479 }, { "epoch": 1.576801087449026, "grad_norm": 0.33800078237495484, "learning_rate": 6.922310985992594e-05, "loss": 0.9653, "step": 3480 }, { "epoch": 1.577254191209787, "grad_norm": 0.39192559793858767, "learning_rate": 6.921446809089869e-05, "loss": 0.975, "step": 3481 }, { "epoch": 1.5777072949705482, "grad_norm": 0.4272721477634254, "learning_rate": 6.920582339831311e-05, "loss": 0.967, "step": 3482 }, { "epoch": 1.5781603987313093, "grad_norm": 0.4891834314987272, "learning_rate": 6.919717578303432e-05, "loss": 0.957, "step": 3483 }, { "epoch": 1.5786135024920707, "grad_norm": 0.5637740838503673, "learning_rate": 6.918852524592769e-05, "loss": 0.9671, "step": 3484 }, { "epoch": 1.579066606252832, "grad_norm": 0.635876182849075, "learning_rate": 6.917987178785889e-05, "loss": 0.9759, "step": 3485 }, { "epoch": 1.5795197100135931, "grad_norm": 0.5944878711963959, "learning_rate": 6.91712154096939e-05, "loss": 0.9444, "step": 3486 }, { "epoch": 1.5799728137743543, "grad_norm": 0.49802055758493935, "learning_rate": 6.916255611229899e-05, "loss": 0.9553, "step": 3487 }, { "epoch": 1.5804259175351154, "grad_norm": 0.36803864191318025, "learning_rate": 6.915389389654072e-05, "loss": 0.9674, "step": 3488 }, { "epoch": 1.5808790212958768, "grad_norm": 0.30003548414619186, "learning_rate": 6.914522876328592e-05, "loss": 0.9806, "step": 3489 }, { "epoch": 1.5813321250566381, "grad_norm": 0.3057918077176273, "learning_rate": 6.913656071340175e-05, "loss": 0.9827, "step": 3490 }, { "epoch": 1.5817852288173992, "grad_norm": 0.33637938851640414, "learning_rate": 6.91278897477556e-05, "loss": 0.9585, "step": 3491 }, { "epoch": 1.5822383325781604, "grad_norm": 0.3407205313972132, "learning_rate": 6.911921586721526e-05, "loss": 0.9466, "step": 3492 }, { "epoch": 1.5826914363389215, "grad_norm": 0.3280965428720684, "learning_rate": 6.91105390726487e-05, "loss": 0.9626, "step": 3493 }, { "epoch": 1.5831445400996829, "grad_norm": 0.3790096982018943, "learning_rate": 6.910185936492422e-05, "loss": 0.9564, "step": 3494 }, { "epoch": 1.583597643860444, "grad_norm": 0.47704586638783275, "learning_rate": 6.909317674491044e-05, "loss": 0.9605, "step": 3495 }, { "epoch": 1.5840507476212053, "grad_norm": 0.5943355020972824, "learning_rate": 6.908449121347624e-05, "loss": 0.9608, "step": 3496 }, { "epoch": 1.5845038513819665, "grad_norm": 0.6206931777555351, "learning_rate": 6.90758027714908e-05, "loss": 0.9521, "step": 3497 }, { "epoch": 1.5849569551427276, "grad_norm": 0.5850192461933867, "learning_rate": 6.906711141982361e-05, "loss": 0.9904, "step": 3498 }, { "epoch": 1.5854100589034887, "grad_norm": 0.43801933285461286, "learning_rate": 6.905841715934441e-05, "loss": 0.963, "step": 3499 }, { "epoch": 1.58586316266425, "grad_norm": 0.3331740062780844, "learning_rate": 6.904971999092327e-05, "loss": 0.9519, "step": 3500 }, { "epoch": 1.5863162664250114, "grad_norm": 0.4171759140732559, "learning_rate": 6.904101991543052e-05, "loss": 0.9826, "step": 3501 }, { "epoch": 1.5867693701857726, "grad_norm": 0.41888848338214657, "learning_rate": 6.903231693373681e-05, "loss": 0.9764, "step": 3502 }, { "epoch": 1.5872224739465337, "grad_norm": 0.3544670646932112, "learning_rate": 6.902361104671306e-05, "loss": 0.9662, "step": 3503 }, { "epoch": 1.5876755777072948, "grad_norm": 0.4193084578916191, "learning_rate": 6.901490225523049e-05, "loss": 0.9531, "step": 3504 }, { "epoch": 1.5881286814680562, "grad_norm": 0.43832482908354803, "learning_rate": 6.900619056016061e-05, "loss": 0.9815, "step": 3505 }, { "epoch": 1.5885817852288175, "grad_norm": 0.34693690457162496, "learning_rate": 6.899747596237522e-05, "loss": 0.9635, "step": 3506 }, { "epoch": 1.5890348889895787, "grad_norm": 0.2945104705368795, "learning_rate": 6.898875846274638e-05, "loss": 0.9699, "step": 3507 }, { "epoch": 1.5894879927503398, "grad_norm": 0.3482330988932461, "learning_rate": 6.898003806214654e-05, "loss": 0.951, "step": 3508 }, { "epoch": 1.589941096511101, "grad_norm": 0.3429242693347135, "learning_rate": 6.897131476144831e-05, "loss": 0.9601, "step": 3509 }, { "epoch": 1.5903942002718623, "grad_norm": 0.28618967607463236, "learning_rate": 6.896258856152467e-05, "loss": 0.9572, "step": 3510 }, { "epoch": 1.5908473040326234, "grad_norm": 0.3121796101457461, "learning_rate": 6.895385946324886e-05, "loss": 0.975, "step": 3511 }, { "epoch": 1.5913004077933848, "grad_norm": 0.3129394908567863, "learning_rate": 6.894512746749444e-05, "loss": 0.9561, "step": 3512 }, { "epoch": 1.591753511554146, "grad_norm": 0.2831669473215101, "learning_rate": 6.893639257513523e-05, "loss": 0.9756, "step": 3513 }, { "epoch": 1.592206615314907, "grad_norm": 0.3129699091958004, "learning_rate": 6.892765478704536e-05, "loss": 0.9702, "step": 3514 }, { "epoch": 1.5926597190756682, "grad_norm": 0.38985300888179286, "learning_rate": 6.891891410409921e-05, "loss": 0.9855, "step": 3515 }, { "epoch": 1.5931128228364295, "grad_norm": 0.3569845959733536, "learning_rate": 6.891017052717151e-05, "loss": 0.9497, "step": 3516 }, { "epoch": 1.5935659265971909, "grad_norm": 0.3156573442027538, "learning_rate": 6.890142405713725e-05, "loss": 0.9808, "step": 3517 }, { "epoch": 1.594019030357952, "grad_norm": 0.289971831754571, "learning_rate": 6.88926746948717e-05, "loss": 0.9761, "step": 3518 }, { "epoch": 1.5944721341187131, "grad_norm": 0.308846798923489, "learning_rate": 6.888392244125042e-05, "loss": 0.9774, "step": 3519 }, { "epoch": 1.5949252378794743, "grad_norm": 0.2818275363715582, "learning_rate": 6.887516729714929e-05, "loss": 0.9704, "step": 3520 }, { "epoch": 1.5953783416402356, "grad_norm": 0.2727932097280708, "learning_rate": 6.886640926344442e-05, "loss": 0.9647, "step": 3521 }, { "epoch": 1.595831445400997, "grad_norm": 0.32649331407431287, "learning_rate": 6.88576483410123e-05, "loss": 0.9713, "step": 3522 }, { "epoch": 1.596284549161758, "grad_norm": 0.35368780514272186, "learning_rate": 6.88488845307296e-05, "loss": 0.958, "step": 3523 }, { "epoch": 1.5967376529225192, "grad_norm": 0.27750787754394174, "learning_rate": 6.884011783347336e-05, "loss": 0.9624, "step": 3524 }, { "epoch": 1.5971907566832804, "grad_norm": 0.2950816786871272, "learning_rate": 6.883134825012091e-05, "loss": 0.9642, "step": 3525 }, { "epoch": 1.5976438604440417, "grad_norm": 0.32617370709087345, "learning_rate": 6.88225757815498e-05, "loss": 0.9484, "step": 3526 }, { "epoch": 1.5980969642048029, "grad_norm": 0.28514517756104696, "learning_rate": 6.881380042863793e-05, "loss": 0.9809, "step": 3527 }, { "epoch": 1.5985500679655642, "grad_norm": 0.3511306254933778, "learning_rate": 6.880502219226346e-05, "loss": 0.9677, "step": 3528 }, { "epoch": 1.5990031717263253, "grad_norm": 0.37446402430982756, "learning_rate": 6.879624107330486e-05, "loss": 0.9836, "step": 3529 }, { "epoch": 1.5994562754870865, "grad_norm": 0.35246926057360545, "learning_rate": 6.878745707264088e-05, "loss": 0.961, "step": 3530 }, { "epoch": 1.5999093792478476, "grad_norm": 0.3755327044852268, "learning_rate": 6.877867019115054e-05, "loss": 0.9823, "step": 3531 }, { "epoch": 1.600362483008609, "grad_norm": 0.4144331267954858, "learning_rate": 6.876988042971315e-05, "loss": 0.9755, "step": 3532 }, { "epoch": 1.6008155867693703, "grad_norm": 0.4087901891947792, "learning_rate": 6.876108778920836e-05, "loss": 0.9802, "step": 3533 }, { "epoch": 1.6012686905301314, "grad_norm": 0.3812620450059409, "learning_rate": 6.875229227051604e-05, "loss": 0.9814, "step": 3534 }, { "epoch": 1.6017217942908926, "grad_norm": 0.3685235536714701, "learning_rate": 6.87434938745164e-05, "loss": 0.969, "step": 3535 }, { "epoch": 1.6021748980516537, "grad_norm": 0.3433894748464223, "learning_rate": 6.87346926020899e-05, "loss": 0.9767, "step": 3536 }, { "epoch": 1.602628001812415, "grad_norm": 0.35055600225273487, "learning_rate": 6.87258884541173e-05, "loss": 0.9646, "step": 3537 }, { "epoch": 1.6030811055731764, "grad_norm": 0.3719212457450887, "learning_rate": 6.871708143147966e-05, "loss": 0.963, "step": 3538 }, { "epoch": 1.6035342093339375, "grad_norm": 0.35635146436608545, "learning_rate": 6.870827153505831e-05, "loss": 0.9703, "step": 3539 }, { "epoch": 1.6039873130946987, "grad_norm": 0.355891414258369, "learning_rate": 6.869945876573489e-05, "loss": 0.9708, "step": 3540 }, { "epoch": 1.6044404168554598, "grad_norm": 0.37846147766123556, "learning_rate": 6.869064312439132e-05, "loss": 0.9768, "step": 3541 }, { "epoch": 1.6048935206162211, "grad_norm": 0.4515923317859654, "learning_rate": 6.868182461190976e-05, "loss": 0.948, "step": 3542 }, { "epoch": 1.6053466243769823, "grad_norm": 0.5441354278915811, "learning_rate": 6.867300322917275e-05, "loss": 0.9952, "step": 3543 }, { "epoch": 1.6057997281377436, "grad_norm": 0.6164972722988954, "learning_rate": 6.866417897706303e-05, "loss": 0.9722, "step": 3544 }, { "epoch": 1.6062528318985048, "grad_norm": 0.6844697740925653, "learning_rate": 6.86553518564637e-05, "loss": 0.9694, "step": 3545 }, { "epoch": 1.606705935659266, "grad_norm": 0.6872941209966537, "learning_rate": 6.864652186825806e-05, "loss": 0.986, "step": 3546 }, { "epoch": 1.607159039420027, "grad_norm": 0.6725115001726542, "learning_rate": 6.863768901332978e-05, "loss": 0.983, "step": 3547 }, { "epoch": 1.6076121431807884, "grad_norm": 0.6426737348551411, "learning_rate": 6.862885329256277e-05, "loss": 0.9474, "step": 3548 }, { "epoch": 1.6080652469415497, "grad_norm": 0.5964268176062217, "learning_rate": 6.862001470684124e-05, "loss": 0.963, "step": 3549 }, { "epoch": 1.6085183507023109, "grad_norm": 0.5044586905883737, "learning_rate": 6.861117325704971e-05, "loss": 0.9826, "step": 3550 }, { "epoch": 1.608971454463072, "grad_norm": 0.35310200453264307, "learning_rate": 6.860232894407295e-05, "loss": 0.9753, "step": 3551 }, { "epoch": 1.6094245582238331, "grad_norm": 0.20540941568350327, "learning_rate": 6.859348176879603e-05, "loss": 0.9533, "step": 3552 }, { "epoch": 1.6098776619845945, "grad_norm": 0.2790765915909539, "learning_rate": 6.85846317321043e-05, "loss": 0.9443, "step": 3553 }, { "epoch": 1.6103307657453558, "grad_norm": 0.4033073655666044, "learning_rate": 6.857577883488341e-05, "loss": 0.9712, "step": 3554 }, { "epoch": 1.610783869506117, "grad_norm": 0.4466642598127796, "learning_rate": 6.856692307801929e-05, "loss": 0.9632, "step": 3555 }, { "epoch": 1.611236973266878, "grad_norm": 0.4327256113922817, "learning_rate": 6.855806446239815e-05, "loss": 0.9633, "step": 3556 }, { "epoch": 1.6116900770276392, "grad_norm": 0.3947844506536204, "learning_rate": 6.854920298890651e-05, "loss": 0.9416, "step": 3557 }, { "epoch": 1.6121431807884006, "grad_norm": 0.34663914902323445, "learning_rate": 6.854033865843112e-05, "loss": 0.9521, "step": 3558 }, { "epoch": 1.6125962845491617, "grad_norm": 0.3492240106682421, "learning_rate": 6.85314714718591e-05, "loss": 0.9449, "step": 3559 }, { "epoch": 1.613049388309923, "grad_norm": 0.5372959070509937, "learning_rate": 6.852260143007778e-05, "loss": 0.961, "step": 3560 }, { "epoch": 1.6135024920706842, "grad_norm": 0.3965524018582907, "learning_rate": 6.851372853397483e-05, "loss": 0.9515, "step": 3561 }, { "epoch": 1.6139555958314453, "grad_norm": 0.33792270499467053, "learning_rate": 6.850485278443814e-05, "loss": 0.9536, "step": 3562 }, { "epoch": 1.6144086995922065, "grad_norm": 0.2712952356737577, "learning_rate": 6.849597418235596e-05, "loss": 0.9509, "step": 3563 }, { "epoch": 1.6148618033529678, "grad_norm": 0.22757533928688387, "learning_rate": 6.848709272861677e-05, "loss": 0.9582, "step": 3564 }, { "epoch": 1.6153149071137292, "grad_norm": 0.2294694136922304, "learning_rate": 6.847820842410938e-05, "loss": 0.9798, "step": 3565 }, { "epoch": 1.6157680108744903, "grad_norm": 0.2764020372453608, "learning_rate": 6.846932126972286e-05, "loss": 0.9701, "step": 3566 }, { "epoch": 1.6162211146352514, "grad_norm": 0.3273193539121622, "learning_rate": 6.846043126634655e-05, "loss": 0.9722, "step": 3567 }, { "epoch": 1.6166742183960126, "grad_norm": 0.3406790232701032, "learning_rate": 6.84515384148701e-05, "loss": 0.9605, "step": 3568 }, { "epoch": 1.617127322156774, "grad_norm": 0.39513611400384246, "learning_rate": 6.844264271618345e-05, "loss": 0.9649, "step": 3569 }, { "epoch": 1.6175804259175353, "grad_norm": 0.44211468518051067, "learning_rate": 6.84337441711768e-05, "loss": 0.968, "step": 3570 }, { "epoch": 1.6180335296782964, "grad_norm": 0.47745225276032827, "learning_rate": 6.842484278074064e-05, "loss": 0.9668, "step": 3571 }, { "epoch": 1.6184866334390575, "grad_norm": 0.48440126359124225, "learning_rate": 6.841593854576579e-05, "loss": 0.9752, "step": 3572 }, { "epoch": 1.6189397371998187, "grad_norm": 0.5043232426266192, "learning_rate": 6.840703146714326e-05, "loss": 0.9733, "step": 3573 }, { "epoch": 1.61939284096058, "grad_norm": 0.5027968383722287, "learning_rate": 6.839812154576445e-05, "loss": 0.963, "step": 3574 }, { "epoch": 1.6198459447213411, "grad_norm": 0.47254895171690475, "learning_rate": 6.838920878252097e-05, "loss": 0.9555, "step": 3575 }, { "epoch": 1.6202990484821025, "grad_norm": 0.43034960290367325, "learning_rate": 6.838029317830476e-05, "loss": 0.9646, "step": 3576 }, { "epoch": 1.6207521522428636, "grad_norm": 0.3834247638906782, "learning_rate": 6.8371374734008e-05, "loss": 0.9676, "step": 3577 }, { "epoch": 1.6212052560036248, "grad_norm": 0.3854217398727078, "learning_rate": 6.83624534505232e-05, "loss": 0.9687, "step": 3578 }, { "epoch": 1.6216583597643859, "grad_norm": 0.3646573719900149, "learning_rate": 6.835352932874312e-05, "loss": 0.9586, "step": 3579 }, { "epoch": 1.6221114635251472, "grad_norm": 0.35112411237653385, "learning_rate": 6.834460236956084e-05, "loss": 0.9699, "step": 3580 }, { "epoch": 1.6225645672859086, "grad_norm": 0.30828728055030136, "learning_rate": 6.833567257386967e-05, "loss": 0.9616, "step": 3581 }, { "epoch": 1.6230176710466697, "grad_norm": 0.26177685928593813, "learning_rate": 6.832673994256326e-05, "loss": 0.9764, "step": 3582 }, { "epoch": 1.6234707748074308, "grad_norm": 0.22359893410682205, "learning_rate": 6.83178044765355e-05, "loss": 0.9575, "step": 3583 }, { "epoch": 1.623923878568192, "grad_norm": 0.2387232326708191, "learning_rate": 6.830886617668058e-05, "loss": 0.9775, "step": 3584 }, { "epoch": 1.6243769823289533, "grad_norm": 0.26474962368814325, "learning_rate": 6.8299925043893e-05, "loss": 0.9446, "step": 3585 }, { "epoch": 1.6248300860897147, "grad_norm": 0.2640186389258179, "learning_rate": 6.82909810790675e-05, "loss": 0.9973, "step": 3586 }, { "epoch": 1.6252831898504758, "grad_norm": 0.29966743855530037, "learning_rate": 6.828203428309913e-05, "loss": 0.9623, "step": 3587 }, { "epoch": 1.625736293611237, "grad_norm": 0.31813698524144157, "learning_rate": 6.82730846568832e-05, "loss": 0.95, "step": 3588 }, { "epoch": 1.626189397371998, "grad_norm": 0.34551721960799925, "learning_rate": 6.826413220131536e-05, "loss": 0.9547, "step": 3589 }, { "epoch": 1.6266425011327594, "grad_norm": 0.4231109465402758, "learning_rate": 6.825517691729148e-05, "loss": 0.9764, "step": 3590 }, { "epoch": 1.6270956048935206, "grad_norm": 0.4335224609171798, "learning_rate": 6.824621880570771e-05, "loss": 0.9583, "step": 3591 }, { "epoch": 1.627548708654282, "grad_norm": 0.42643079088921787, "learning_rate": 6.823725786746053e-05, "loss": 0.9749, "step": 3592 }, { "epoch": 1.628001812415043, "grad_norm": 0.5049252727878745, "learning_rate": 6.82282941034467e-05, "loss": 0.9661, "step": 3593 }, { "epoch": 1.6284549161758042, "grad_norm": 0.53156885839464, "learning_rate": 6.821932751456322e-05, "loss": 0.9583, "step": 3594 }, { "epoch": 1.6289080199365653, "grad_norm": 0.5979282678394178, "learning_rate": 6.82103581017074e-05, "loss": 0.955, "step": 3595 }, { "epoch": 1.6293611236973267, "grad_norm": 0.5421023066899305, "learning_rate": 6.820138586577685e-05, "loss": 0.9791, "step": 3596 }, { "epoch": 1.629814227458088, "grad_norm": 0.4967807451916777, "learning_rate": 6.81924108076694e-05, "loss": 0.9677, "step": 3597 }, { "epoch": 1.6302673312188491, "grad_norm": 0.5273085564310105, "learning_rate": 6.818343292828325e-05, "loss": 0.9668, "step": 3598 }, { "epoch": 1.6307204349796103, "grad_norm": 0.5286254325519292, "learning_rate": 6.817445222851682e-05, "loss": 0.9531, "step": 3599 }, { "epoch": 1.6311735387403714, "grad_norm": 0.5135567171165907, "learning_rate": 6.816546870926881e-05, "loss": 0.9764, "step": 3600 }, { "epoch": 1.6316266425011328, "grad_norm": 0.4514212402771094, "learning_rate": 6.815648237143825e-05, "loss": 0.9431, "step": 3601 }, { "epoch": 1.6320797462618941, "grad_norm": 0.374750205588834, "learning_rate": 6.81474932159244e-05, "loss": 0.9601, "step": 3602 }, { "epoch": 1.6325328500226552, "grad_norm": 0.37063496588018563, "learning_rate": 6.813850124362684e-05, "loss": 0.9861, "step": 3603 }, { "epoch": 1.6329859537834164, "grad_norm": 0.24728338125280835, "learning_rate": 6.812950645544541e-05, "loss": 0.9753, "step": 3604 }, { "epoch": 1.6334390575441775, "grad_norm": 0.28887094943116004, "learning_rate": 6.812050885228024e-05, "loss": 0.9607, "step": 3605 }, { "epoch": 1.6338921613049389, "grad_norm": 0.35117588369915786, "learning_rate": 6.811150843503175e-05, "loss": 0.9842, "step": 3606 }, { "epoch": 1.6343452650657, "grad_norm": 0.40709561557292273, "learning_rate": 6.810250520460063e-05, "loss": 0.9881, "step": 3607 }, { "epoch": 1.6347983688264613, "grad_norm": 0.43203697638438204, "learning_rate": 6.809349916188784e-05, "loss": 0.9534, "step": 3608 }, { "epoch": 1.6352514725872225, "grad_norm": 0.3870247841220019, "learning_rate": 6.808449030779464e-05, "loss": 0.9476, "step": 3609 }, { "epoch": 1.6357045763479836, "grad_norm": 0.3285664919956267, "learning_rate": 6.807547864322257e-05, "loss": 0.9523, "step": 3610 }, { "epoch": 1.6361576801087447, "grad_norm": 0.30085048406408393, "learning_rate": 6.806646416907345e-05, "loss": 0.9573, "step": 3611 }, { "epoch": 1.636610783869506, "grad_norm": 0.30433686636665225, "learning_rate": 6.805744688624938e-05, "loss": 0.9593, "step": 3612 }, { "epoch": 1.6370638876302674, "grad_norm": 0.32603876225620165, "learning_rate": 6.804842679565273e-05, "loss": 0.9798, "step": 3613 }, { "epoch": 1.6375169913910286, "grad_norm": 0.36191974545803557, "learning_rate": 6.803940389818618e-05, "loss": 0.9675, "step": 3614 }, { "epoch": 1.6379700951517897, "grad_norm": 0.3530439559456027, "learning_rate": 6.803037819475265e-05, "loss": 0.9828, "step": 3615 }, { "epoch": 1.6384231989125508, "grad_norm": 0.33767549001664937, "learning_rate": 6.802134968625538e-05, "loss": 0.9643, "step": 3616 }, { "epoch": 1.6388763026733122, "grad_norm": 0.33060481294585065, "learning_rate": 6.801231837359786e-05, "loss": 0.9683, "step": 3617 }, { "epoch": 1.6393294064340735, "grad_norm": 0.35245006031992737, "learning_rate": 6.800328425768388e-05, "loss": 0.9591, "step": 3618 }, { "epoch": 1.6397825101948347, "grad_norm": 0.4249613266030163, "learning_rate": 6.79942473394175e-05, "loss": 1.0042, "step": 3619 }, { "epoch": 1.6402356139555958, "grad_norm": 0.5018716811160896, "learning_rate": 6.798520761970308e-05, "loss": 0.9826, "step": 3620 }, { "epoch": 1.640688717716357, "grad_norm": 0.5262053551304675, "learning_rate": 6.797616509944524e-05, "loss": 0.9683, "step": 3621 }, { "epoch": 1.6411418214771183, "grad_norm": 0.48812155816039576, "learning_rate": 6.796711977954886e-05, "loss": 0.9541, "step": 3622 }, { "epoch": 1.6415949252378794, "grad_norm": 0.396664105860478, "learning_rate": 6.795807166091916e-05, "loss": 0.9698, "step": 3623 }, { "epoch": 1.6420480289986408, "grad_norm": 0.3518996330672198, "learning_rate": 6.794902074446159e-05, "loss": 0.9737, "step": 3624 }, { "epoch": 1.642501132759402, "grad_norm": 0.31960663726978106, "learning_rate": 6.79399670310819e-05, "loss": 0.977, "step": 3625 }, { "epoch": 1.642954236520163, "grad_norm": 0.30018153300598227, "learning_rate": 6.79309105216861e-05, "loss": 0.9648, "step": 3626 }, { "epoch": 1.6434073402809242, "grad_norm": 0.272051974282387, "learning_rate": 6.792185121718053e-05, "loss": 0.9701, "step": 3627 }, { "epoch": 1.6438604440416855, "grad_norm": 0.25401848087147055, "learning_rate": 6.791278911847172e-05, "loss": 0.9588, "step": 3628 }, { "epoch": 1.6443135478024469, "grad_norm": 0.3185702081638906, "learning_rate": 6.790372422646662e-05, "loss": 0.9765, "step": 3629 }, { "epoch": 1.644766651563208, "grad_norm": 0.3712349679311133, "learning_rate": 6.789465654207228e-05, "loss": 0.9883, "step": 3630 }, { "epoch": 1.6452197553239691, "grad_norm": 0.3111778481715736, "learning_rate": 6.788558606619619e-05, "loss": 0.9717, "step": 3631 }, { "epoch": 1.6456728590847303, "grad_norm": 2.743067578219481, "learning_rate": 6.787651279974602e-05, "loss": 0.9812, "step": 3632 }, { "epoch": 1.6461259628454916, "grad_norm": 0.40307493968699654, "learning_rate": 6.786743674362977e-05, "loss": 0.9511, "step": 3633 }, { "epoch": 1.646579066606253, "grad_norm": 0.7363707711081947, "learning_rate": 6.785835789875569e-05, "loss": 0.9689, "step": 3634 }, { "epoch": 1.647032170367014, "grad_norm": 0.8588970472262045, "learning_rate": 6.784927626603232e-05, "loss": 0.9824, "step": 3635 }, { "epoch": 1.6474852741277752, "grad_norm": 0.812589155748232, "learning_rate": 6.784019184636849e-05, "loss": 0.9641, "step": 3636 }, { "epoch": 1.6479383778885364, "grad_norm": 0.6405211709059534, "learning_rate": 6.783110464067329e-05, "loss": 0.9856, "step": 3637 }, { "epoch": 1.6483914816492977, "grad_norm": 0.5955738558454104, "learning_rate": 6.78220146498561e-05, "loss": 0.9912, "step": 3638 }, { "epoch": 1.6488445854100588, "grad_norm": 0.597694825119125, "learning_rate": 6.781292187482658e-05, "loss": 0.9656, "step": 3639 }, { "epoch": 1.6492976891708202, "grad_norm": 0.48770433391332024, "learning_rate": 6.780382631649467e-05, "loss": 1.0081, "step": 3640 }, { "epoch": 1.6497507929315813, "grad_norm": 0.42192995085434537, "learning_rate": 6.779472797577056e-05, "loss": 0.9873, "step": 3641 }, { "epoch": 1.6502038966923425, "grad_norm": 0.532358303930385, "learning_rate": 6.778562685356477e-05, "loss": 0.9697, "step": 3642 }, { "epoch": 1.6506570004531036, "grad_norm": 0.5145219818671386, "learning_rate": 6.777652295078804e-05, "loss": 0.9505, "step": 3643 }, { "epoch": 1.651110104213865, "grad_norm": 0.42842652595015157, "learning_rate": 6.776741626835144e-05, "loss": 0.9568, "step": 3644 }, { "epoch": 1.6515632079746263, "grad_norm": 0.5059684765224437, "learning_rate": 6.77583068071663e-05, "loss": 0.9892, "step": 3645 }, { "epoch": 1.6520163117353874, "grad_norm": 0.47770734552056615, "learning_rate": 6.774919456814421e-05, "loss": 0.9553, "step": 3646 }, { "epoch": 1.6524694154961486, "grad_norm": 0.41105436921382027, "learning_rate": 6.774007955219706e-05, "loss": 0.9733, "step": 3647 }, { "epoch": 1.6529225192569097, "grad_norm": 0.3995018834604289, "learning_rate": 6.7730961760237e-05, "loss": 0.9523, "step": 3648 }, { "epoch": 1.653375623017671, "grad_norm": 0.36798053582747964, "learning_rate": 6.772184119317649e-05, "loss": 0.9723, "step": 3649 }, { "epoch": 1.6538287267784324, "grad_norm": 0.31411154157959625, "learning_rate": 6.771271785192821e-05, "loss": 0.9518, "step": 3650 }, { "epoch": 1.6542818305391935, "grad_norm": 0.29166395978169635, "learning_rate": 6.770359173740521e-05, "loss": 0.9728, "step": 3651 }, { "epoch": 1.6547349342999547, "grad_norm": 0.3024561226815327, "learning_rate": 6.76944628505207e-05, "loss": 0.9705, "step": 3652 }, { "epoch": 1.6551880380607158, "grad_norm": 0.3103368182402752, "learning_rate": 6.768533119218827e-05, "loss": 0.98, "step": 3653 }, { "epoch": 1.6556411418214771, "grad_norm": 0.40183596563140656, "learning_rate": 6.767619676332173e-05, "loss": 0.9771, "step": 3654 }, { "epoch": 1.6560942455822383, "grad_norm": 0.467388756183635, "learning_rate": 6.766705956483518e-05, "loss": 0.9802, "step": 3655 }, { "epoch": 1.6565473493429996, "grad_norm": 0.4303708806403964, "learning_rate": 6.7657919597643e-05, "loss": 0.9725, "step": 3656 }, { "epoch": 1.6570004531037608, "grad_norm": 0.3539279289081673, "learning_rate": 6.764877686265985e-05, "loss": 0.9833, "step": 3657 }, { "epoch": 1.657453556864522, "grad_norm": 0.30651354763661554, "learning_rate": 6.763963136080067e-05, "loss": 0.9614, "step": 3658 }, { "epoch": 1.657906660625283, "grad_norm": 0.2785004624077476, "learning_rate": 6.763048309298069e-05, "loss": 0.9828, "step": 3659 }, { "epoch": 1.6583597643860444, "grad_norm": 0.248689684244206, "learning_rate": 6.762133206011535e-05, "loss": 0.9803, "step": 3660 }, { "epoch": 1.6588128681468057, "grad_norm": 0.2645503923822664, "learning_rate": 6.761217826312044e-05, "loss": 0.9579, "step": 3661 }, { "epoch": 1.6592659719075669, "grad_norm": 0.36069695265373075, "learning_rate": 6.760302170291202e-05, "loss": 0.9533, "step": 3662 }, { "epoch": 1.659719075668328, "grad_norm": 0.36316977150720675, "learning_rate": 6.759386238040637e-05, "loss": 0.9667, "step": 3663 }, { "epoch": 1.6601721794290891, "grad_norm": 0.3277583683645094, "learning_rate": 6.758470029652011e-05, "loss": 0.9773, "step": 3664 }, { "epoch": 1.6606252831898505, "grad_norm": 0.3164485738531524, "learning_rate": 6.757553545217011e-05, "loss": 0.9692, "step": 3665 }, { "epoch": 1.6610783869506118, "grad_norm": 0.3019267460762152, "learning_rate": 6.756636784827351e-05, "loss": 0.9472, "step": 3666 }, { "epoch": 1.661531490711373, "grad_norm": 0.2750559602165625, "learning_rate": 6.755719748574773e-05, "loss": 0.9604, "step": 3667 }, { "epoch": 1.661984594472134, "grad_norm": 0.23843928059363265, "learning_rate": 6.754802436551046e-05, "loss": 0.9558, "step": 3668 }, { "epoch": 1.6624376982328952, "grad_norm": 0.24075335832927955, "learning_rate": 6.753884848847972e-05, "loss": 0.9697, "step": 3669 }, { "epoch": 1.6628908019936566, "grad_norm": 0.2451491555741636, "learning_rate": 6.75296698555737e-05, "loss": 0.964, "step": 3670 }, { "epoch": 1.6633439057544177, "grad_norm": 0.2923190852487624, "learning_rate": 6.752048846771097e-05, "loss": 0.9665, "step": 3671 }, { "epoch": 1.663797009515179, "grad_norm": 0.29067639624541886, "learning_rate": 6.75113043258103e-05, "loss": 0.9641, "step": 3672 }, { "epoch": 1.6642501132759402, "grad_norm": 0.28169408226328024, "learning_rate": 6.75021174307908e-05, "loss": 0.9588, "step": 3673 }, { "epoch": 1.6647032170367013, "grad_norm": 0.2995328313096785, "learning_rate": 6.749292778357182e-05, "loss": 0.9524, "step": 3674 }, { "epoch": 1.6651563207974625, "grad_norm": 0.30149051129149657, "learning_rate": 6.748373538507298e-05, "loss": 0.9637, "step": 3675 }, { "epoch": 1.6656094245582238, "grad_norm": 0.2932152860584444, "learning_rate": 6.747454023621416e-05, "loss": 0.9584, "step": 3676 }, { "epoch": 1.6660625283189852, "grad_norm": 0.2885082916257768, "learning_rate": 6.746534233791558e-05, "loss": 0.9416, "step": 3677 }, { "epoch": 1.6665156320797463, "grad_norm": 0.28152819787254735, "learning_rate": 6.745614169109767e-05, "loss": 0.948, "step": 3678 }, { "epoch": 1.6669687358405074, "grad_norm": 0.2554505665468379, "learning_rate": 6.744693829668116e-05, "loss": 0.9583, "step": 3679 }, { "epoch": 1.6674218396012686, "grad_norm": 0.23180441953867875, "learning_rate": 6.743773215558706e-05, "loss": 0.9679, "step": 3680 }, { "epoch": 1.66787494336203, "grad_norm": 0.2697522663761293, "learning_rate": 6.742852326873667e-05, "loss": 0.9795, "step": 3681 }, { "epoch": 1.6683280471227913, "grad_norm": 0.2859099305152244, "learning_rate": 6.74193116370515e-05, "loss": 0.9662, "step": 3682 }, { "epoch": 1.6687811508835524, "grad_norm": 0.3118434888352911, "learning_rate": 6.741009726145343e-05, "loss": 0.9523, "step": 3683 }, { "epoch": 1.6692342546443135, "grad_norm": 0.3267671252349535, "learning_rate": 6.740088014286451e-05, "loss": 0.9748, "step": 3684 }, { "epoch": 1.6696873584050747, "grad_norm": 0.36887489110564525, "learning_rate": 6.739166028220716e-05, "loss": 0.9731, "step": 3685 }, { "epoch": 1.670140462165836, "grad_norm": 0.41769938925626554, "learning_rate": 6.738243768040402e-05, "loss": 0.9706, "step": 3686 }, { "epoch": 1.6705935659265974, "grad_norm": 0.42400379286731543, "learning_rate": 6.737321233837802e-05, "loss": 0.9936, "step": 3687 }, { "epoch": 1.6710466696873585, "grad_norm": 0.42091319705348573, "learning_rate": 6.736398425705235e-05, "loss": 0.9661, "step": 3688 }, { "epoch": 1.6714997734481196, "grad_norm": 0.39926419163427934, "learning_rate": 6.735475343735049e-05, "loss": 0.9716, "step": 3689 }, { "epoch": 1.6719528772088808, "grad_norm": 0.37365128268901227, "learning_rate": 6.734551988019618e-05, "loss": 0.966, "step": 3690 }, { "epoch": 1.672405980969642, "grad_norm": 0.40282783051182747, "learning_rate": 6.733628358651346e-05, "loss": 0.9349, "step": 3691 }, { "epoch": 1.6728590847304032, "grad_norm": 0.4530253821131322, "learning_rate": 6.732704455722662e-05, "loss": 0.9645, "step": 3692 }, { "epoch": 1.6733121884911646, "grad_norm": 0.4202738689978344, "learning_rate": 6.731780279326023e-05, "loss": 0.9738, "step": 3693 }, { "epoch": 1.6737652922519257, "grad_norm": 0.35364826288963463, "learning_rate": 6.730855829553914e-05, "loss": 0.9736, "step": 3694 }, { "epoch": 1.6742183960126868, "grad_norm": 0.3399277592151065, "learning_rate": 6.729931106498846e-05, "loss": 0.9702, "step": 3695 }, { "epoch": 1.674671499773448, "grad_norm": 0.3174362633702431, "learning_rate": 6.729006110253358e-05, "loss": 0.9534, "step": 3696 }, { "epoch": 1.6751246035342093, "grad_norm": 0.3267597964941867, "learning_rate": 6.728080840910018e-05, "loss": 0.9768, "step": 3697 }, { "epoch": 1.6755777072949707, "grad_norm": 0.2824454732749402, "learning_rate": 6.727155298561418e-05, "loss": 0.9607, "step": 3698 }, { "epoch": 1.6760308110557318, "grad_norm": 0.2556591800547431, "learning_rate": 6.72622948330018e-05, "loss": 0.9799, "step": 3699 }, { "epoch": 1.676483914816493, "grad_norm": 0.2440451834088967, "learning_rate": 6.72530339521895e-05, "loss": 0.9742, "step": 3700 }, { "epoch": 1.676937018577254, "grad_norm": 0.23325115510463604, "learning_rate": 6.724377034410409e-05, "loss": 0.9457, "step": 3701 }, { "epoch": 1.6773901223380154, "grad_norm": 0.26967074031856203, "learning_rate": 6.723450400967257e-05, "loss": 0.9971, "step": 3702 }, { "epoch": 1.6778432260987768, "grad_norm": 0.25785406535222377, "learning_rate": 6.722523494982223e-05, "loss": 0.9586, "step": 3703 }, { "epoch": 1.678296329859538, "grad_norm": 0.23510446163103493, "learning_rate": 6.721596316548067e-05, "loss": 0.9669, "step": 3704 }, { "epoch": 1.678749433620299, "grad_norm": 0.28662213849456025, "learning_rate": 6.720668865757571e-05, "loss": 0.9507, "step": 3705 }, { "epoch": 1.6792025373810602, "grad_norm": 0.33443033794322125, "learning_rate": 6.71974114270355e-05, "loss": 0.9626, "step": 3706 }, { "epoch": 1.6796556411418215, "grad_norm": 0.381015581110621, "learning_rate": 6.718813147478843e-05, "loss": 0.9653, "step": 3707 }, { "epoch": 1.6801087449025827, "grad_norm": 0.3772018358716622, "learning_rate": 6.717884880176316e-05, "loss": 0.9544, "step": 3708 }, { "epoch": 1.680561848663344, "grad_norm": 0.3341263110426139, "learning_rate": 6.716956340888861e-05, "loss": 0.9754, "step": 3709 }, { "epoch": 1.6810149524241051, "grad_norm": 0.33602608559912417, "learning_rate": 6.716027529709403e-05, "loss": 0.9659, "step": 3710 }, { "epoch": 1.6814680561848663, "grad_norm": 0.33457982312933837, "learning_rate": 6.715098446730885e-05, "loss": 0.9631, "step": 3711 }, { "epoch": 1.6819211599456274, "grad_norm": 0.35322238803689465, "learning_rate": 6.714169092046288e-05, "loss": 0.9677, "step": 3712 }, { "epoch": 1.6823742637063888, "grad_norm": 0.4086379226231597, "learning_rate": 6.713239465748611e-05, "loss": 0.9697, "step": 3713 }, { "epoch": 1.6828273674671501, "grad_norm": 0.4466930640575917, "learning_rate": 6.712309567930886e-05, "loss": 0.9581, "step": 3714 }, { "epoch": 1.6832804712279112, "grad_norm": 0.7245697601181732, "learning_rate": 6.711379398686168e-05, "loss": 0.9586, "step": 3715 }, { "epoch": 1.6837335749886724, "grad_norm": 0.48331535642996776, "learning_rate": 6.710448958107541e-05, "loss": 0.9502, "step": 3716 }, { "epoch": 1.6841866787494335, "grad_norm": 0.4340761775441182, "learning_rate": 6.70951824628812e-05, "loss": 0.9692, "step": 3717 }, { "epoch": 1.6846397825101949, "grad_norm": 0.382773579526921, "learning_rate": 6.708587263321039e-05, "loss": 0.9466, "step": 3718 }, { "epoch": 1.6850928862709562, "grad_norm": 0.3343697462147229, "learning_rate": 6.707656009299465e-05, "loss": 0.9529, "step": 3719 }, { "epoch": 1.6855459900317173, "grad_norm": 0.2697256084581556, "learning_rate": 6.706724484316592e-05, "loss": 0.9614, "step": 3720 }, { "epoch": 1.6859990937924785, "grad_norm": 0.2926012978168552, "learning_rate": 6.70579268846564e-05, "loss": 0.9486, "step": 3721 }, { "epoch": 1.6864521975532396, "grad_norm": 0.36250630497670405, "learning_rate": 6.704860621839853e-05, "loss": 0.9588, "step": 3722 }, { "epoch": 1.686905301314001, "grad_norm": 0.38980923360303626, "learning_rate": 6.703928284532508e-05, "loss": 0.9749, "step": 3723 }, { "epoch": 1.687358405074762, "grad_norm": 0.37681306918681184, "learning_rate": 6.702995676636904e-05, "loss": 0.9598, "step": 3724 }, { "epoch": 1.6878115088355234, "grad_norm": 0.33412163309325793, "learning_rate": 6.702062798246371e-05, "loss": 0.9776, "step": 3725 }, { "epoch": 1.6882646125962846, "grad_norm": 0.3018737931709297, "learning_rate": 6.701129649454262e-05, "loss": 0.9497, "step": 3726 }, { "epoch": 1.6887177163570457, "grad_norm": 0.30012362582293994, "learning_rate": 6.700196230353962e-05, "loss": 0.9382, "step": 3727 }, { "epoch": 1.6891708201178068, "grad_norm": 0.34880350551352535, "learning_rate": 6.699262541038879e-05, "loss": 0.956, "step": 3728 }, { "epoch": 1.6896239238785682, "grad_norm": 0.33841920557167654, "learning_rate": 6.69832858160245e-05, "loss": 0.9481, "step": 3729 }, { "epoch": 1.6900770276393295, "grad_norm": 0.286521132106846, "learning_rate": 6.697394352138137e-05, "loss": 0.9727, "step": 3730 }, { "epoch": 1.6905301314000907, "grad_norm": 0.26710431153416325, "learning_rate": 6.696459852739432e-05, "loss": 0.97, "step": 3731 }, { "epoch": 1.6909832351608518, "grad_norm": 0.25670452536770255, "learning_rate": 6.695525083499852e-05, "loss": 0.9518, "step": 3732 }, { "epoch": 1.691436338921613, "grad_norm": 0.24787329390455892, "learning_rate": 6.694590044512943e-05, "loss": 0.9687, "step": 3733 }, { "epoch": 1.6918894426823743, "grad_norm": 0.28501301410039903, "learning_rate": 6.693654735872273e-05, "loss": 0.9567, "step": 3734 }, { "epoch": 1.6923425464431356, "grad_norm": 0.2956095877966132, "learning_rate": 6.692719157671443e-05, "loss": 0.9629, "step": 3735 }, { "epoch": 1.6927956502038968, "grad_norm": 0.25362663675328845, "learning_rate": 6.691783310004077e-05, "loss": 0.9747, "step": 3736 }, { "epoch": 1.693248753964658, "grad_norm": 0.22038613874406285, "learning_rate": 6.69084719296383e-05, "loss": 0.9631, "step": 3737 }, { "epoch": 1.693701857725419, "grad_norm": 0.2539932013128234, "learning_rate": 6.689910806644378e-05, "loss": 0.9492, "step": 3738 }, { "epoch": 1.6941549614861804, "grad_norm": 0.3561645268900019, "learning_rate": 6.68897415113943e-05, "loss": 0.9877, "step": 3739 }, { "epoch": 1.6946080652469415, "grad_norm": 0.3902005803686216, "learning_rate": 6.688037226542717e-05, "loss": 0.9393, "step": 3740 }, { "epoch": 1.6950611690077029, "grad_norm": 0.3946279232620356, "learning_rate": 6.687100032948002e-05, "loss": 0.9694, "step": 3741 }, { "epoch": 1.695514272768464, "grad_norm": 0.37227054000919113, "learning_rate": 6.686162570449068e-05, "loss": 0.9579, "step": 3742 }, { "epoch": 1.6959673765292251, "grad_norm": 0.28755614378443983, "learning_rate": 6.685224839139734e-05, "loss": 0.9796, "step": 3743 }, { "epoch": 1.6964204802899863, "grad_norm": 0.2663015927833727, "learning_rate": 6.684286839113837e-05, "loss": 0.942, "step": 3744 }, { "epoch": 1.6968735840507476, "grad_norm": 0.3079501084948899, "learning_rate": 6.683348570465247e-05, "loss": 0.939, "step": 3745 }, { "epoch": 1.697326687811509, "grad_norm": 0.3572788908692385, "learning_rate": 6.682410033287857e-05, "loss": 0.9426, "step": 3746 }, { "epoch": 1.69777979157227, "grad_norm": 0.4259962271217581, "learning_rate": 6.681471227675589e-05, "loss": 0.9703, "step": 3747 }, { "epoch": 1.6982328953330312, "grad_norm": 0.5128050609158199, "learning_rate": 6.680532153722392e-05, "loss": 0.9529, "step": 3748 }, { "epoch": 1.6986859990937924, "grad_norm": 0.5595150496838288, "learning_rate": 6.679592811522243e-05, "loss": 0.9856, "step": 3749 }, { "epoch": 1.6991391028545537, "grad_norm": 0.56168664842818, "learning_rate": 6.67865320116914e-05, "loss": 0.9568, "step": 3750 }, { "epoch": 1.699592206615315, "grad_norm": 0.5498700187740913, "learning_rate": 6.677713322757115e-05, "loss": 0.9375, "step": 3751 }, { "epoch": 1.7000453103760762, "grad_norm": 0.5623222709483945, "learning_rate": 6.676773176380222e-05, "loss": 0.9842, "step": 3752 }, { "epoch": 1.7004984141368373, "grad_norm": 0.5188383870796908, "learning_rate": 6.675832762132545e-05, "loss": 0.9593, "step": 3753 }, { "epoch": 1.7009515178975985, "grad_norm": 0.45119176410951983, "learning_rate": 6.674892080108193e-05, "loss": 0.9595, "step": 3754 }, { "epoch": 1.7014046216583598, "grad_norm": 0.39064560969026596, "learning_rate": 6.673951130401303e-05, "loss": 0.9666, "step": 3755 }, { "epoch": 1.701857725419121, "grad_norm": 0.358993841649825, "learning_rate": 6.673009913106035e-05, "loss": 0.9705, "step": 3756 }, { "epoch": 1.7023108291798823, "grad_norm": 0.2983837284429146, "learning_rate": 6.672068428316583e-05, "loss": 0.958, "step": 3757 }, { "epoch": 1.7027639329406434, "grad_norm": 0.28421276101392157, "learning_rate": 6.671126676127159e-05, "loss": 0.9683, "step": 3758 }, { "epoch": 1.7032170367014046, "grad_norm": 0.39655413680200435, "learning_rate": 6.670184656632011e-05, "loss": 0.9592, "step": 3759 }, { "epoch": 1.7036701404621657, "grad_norm": 0.4659249437618335, "learning_rate": 6.669242369925406e-05, "loss": 0.9625, "step": 3760 }, { "epoch": 1.704123244222927, "grad_norm": 0.43215203373162964, "learning_rate": 6.668299816101639e-05, "loss": 0.9843, "step": 3761 }, { "epoch": 1.7045763479836884, "grad_norm": 0.4179349529005325, "learning_rate": 6.667356995255038e-05, "loss": 0.964, "step": 3762 }, { "epoch": 1.7050294517444495, "grad_norm": 0.45028063864325485, "learning_rate": 6.666413907479951e-05, "loss": 0.9743, "step": 3763 }, { "epoch": 1.7054825555052107, "grad_norm": 0.4506703087225789, "learning_rate": 6.665470552870755e-05, "loss": 0.9568, "step": 3764 }, { "epoch": 1.7059356592659718, "grad_norm": 0.42635786642669526, "learning_rate": 6.664526931521855e-05, "loss": 0.9541, "step": 3765 }, { "epoch": 1.7063887630267331, "grad_norm": 0.35431850220702055, "learning_rate": 6.663583043527679e-05, "loss": 0.9711, "step": 3766 }, { "epoch": 1.7068418667874945, "grad_norm": 0.2781912697420968, "learning_rate": 6.662638888982684e-05, "loss": 0.9689, "step": 3767 }, { "epoch": 1.7072949705482556, "grad_norm": 0.28001463552586814, "learning_rate": 6.661694467981356e-05, "loss": 0.9683, "step": 3768 }, { "epoch": 1.7077480743090168, "grad_norm": 0.29109865205390695, "learning_rate": 6.660749780618204e-05, "loss": 0.9683, "step": 3769 }, { "epoch": 1.708201178069778, "grad_norm": 0.34935668973311984, "learning_rate": 6.659804826987765e-05, "loss": 0.9504, "step": 3770 }, { "epoch": 1.7086542818305392, "grad_norm": 0.38542660258883815, "learning_rate": 6.658859607184604e-05, "loss": 0.966, "step": 3771 }, { "epoch": 1.7091073855913004, "grad_norm": 0.37858268961971686, "learning_rate": 6.657914121303309e-05, "loss": 0.9759, "step": 3772 }, { "epoch": 1.7095604893520617, "grad_norm": 0.37931237492961495, "learning_rate": 6.656968369438498e-05, "loss": 0.9681, "step": 3773 }, { "epoch": 1.7100135931128229, "grad_norm": 0.3915650282158177, "learning_rate": 6.656022351684815e-05, "loss": 0.9775, "step": 3774 }, { "epoch": 1.710466696873584, "grad_norm": 0.3601896095259817, "learning_rate": 6.655076068136929e-05, "loss": 0.9413, "step": 3775 }, { "epoch": 1.7109198006343451, "grad_norm": 0.3753938918851317, "learning_rate": 6.65412951888954e-05, "loss": 0.9739, "step": 3776 }, { "epoch": 1.7113729043951065, "grad_norm": 0.43401733289524796, "learning_rate": 6.653182704037365e-05, "loss": 0.9838, "step": 3777 }, { "epoch": 1.7118260081558678, "grad_norm": 0.43123697925006155, "learning_rate": 6.652235623675159e-05, "loss": 0.968, "step": 3778 }, { "epoch": 1.712279111916629, "grad_norm": 0.40250580301419486, "learning_rate": 6.651288277897698e-05, "loss": 0.9654, "step": 3779 }, { "epoch": 1.71273221567739, "grad_norm": 0.36017876182876535, "learning_rate": 6.650340666799784e-05, "loss": 0.9614, "step": 3780 }, { "epoch": 1.7131853194381512, "grad_norm": 0.33453404287316485, "learning_rate": 6.649392790476247e-05, "loss": 0.9586, "step": 3781 }, { "epoch": 1.7136384231989126, "grad_norm": 0.34948492207409615, "learning_rate": 6.648444649021944e-05, "loss": 0.9521, "step": 3782 }, { "epoch": 1.714091526959674, "grad_norm": 0.42332854923332097, "learning_rate": 6.647496242531756e-05, "loss": 0.9687, "step": 3783 }, { "epoch": 1.714544630720435, "grad_norm": 0.5437025573763942, "learning_rate": 6.646547571100591e-05, "loss": 0.9529, "step": 3784 }, { "epoch": 1.7149977344811962, "grad_norm": 0.5965156462604102, "learning_rate": 6.645598634823388e-05, "loss": 0.963, "step": 3785 }, { "epoch": 1.7154508382419573, "grad_norm": 0.5492852776019393, "learning_rate": 6.644649433795109e-05, "loss": 0.9758, "step": 3786 }, { "epoch": 1.7159039420027187, "grad_norm": 0.4283493059928893, "learning_rate": 6.64369996811074e-05, "loss": 0.9345, "step": 3787 }, { "epoch": 1.7163570457634798, "grad_norm": 0.3194901040879756, "learning_rate": 6.6427502378653e-05, "loss": 0.9571, "step": 3788 }, { "epoch": 1.7168101495242412, "grad_norm": 0.2736657020674601, "learning_rate": 6.641800243153825e-05, "loss": 0.966, "step": 3789 }, { "epoch": 1.7172632532850023, "grad_norm": 0.2792564864664222, "learning_rate": 6.640849984071388e-05, "loss": 0.9644, "step": 3790 }, { "epoch": 1.7177163570457634, "grad_norm": 0.278149083371104, "learning_rate": 6.639899460713082e-05, "loss": 0.9593, "step": 3791 }, { "epoch": 1.7181694608065246, "grad_norm": 0.28981539629008357, "learning_rate": 6.638948673174028e-05, "loss": 0.9779, "step": 3792 }, { "epoch": 1.718622564567286, "grad_norm": 0.3070393511941138, "learning_rate": 6.637997621549375e-05, "loss": 0.9884, "step": 3793 }, { "epoch": 1.7190756683280473, "grad_norm": 0.3437143419204042, "learning_rate": 6.637046305934295e-05, "loss": 0.9488, "step": 3794 }, { "epoch": 1.7195287720888084, "grad_norm": 0.3269027795080655, "learning_rate": 6.636094726423988e-05, "loss": 0.9519, "step": 3795 }, { "epoch": 1.7199818758495695, "grad_norm": 0.3367295505464125, "learning_rate": 6.635142883113681e-05, "loss": 0.9647, "step": 3796 }, { "epoch": 1.7204349796103307, "grad_norm": 0.38790010327123936, "learning_rate": 6.634190776098629e-05, "loss": 0.9705, "step": 3797 }, { "epoch": 1.720888083371092, "grad_norm": 0.3648332138193338, "learning_rate": 6.63323840547411e-05, "loss": 0.9555, "step": 3798 }, { "epoch": 1.7213411871318534, "grad_norm": 0.3574657475406491, "learning_rate": 6.63228577133543e-05, "loss": 0.9637, "step": 3799 }, { "epoch": 1.7217942908926145, "grad_norm": 0.3543334100104129, "learning_rate": 6.63133287377792e-05, "loss": 0.9684, "step": 3800 }, { "epoch": 1.7222473946533756, "grad_norm": 0.3445905869652699, "learning_rate": 6.630379712896942e-05, "loss": 0.9533, "step": 3801 }, { "epoch": 1.7227004984141367, "grad_norm": 0.35641020387762506, "learning_rate": 6.629426288787878e-05, "loss": 0.9626, "step": 3802 }, { "epoch": 1.723153602174898, "grad_norm": 0.3948213726094587, "learning_rate": 6.62847260154614e-05, "loss": 0.9493, "step": 3803 }, { "epoch": 1.7236067059356592, "grad_norm": 0.4301764284230644, "learning_rate": 6.627518651267167e-05, "loss": 0.952, "step": 3804 }, { "epoch": 1.7240598096964206, "grad_norm": 0.4051097228231305, "learning_rate": 6.62656443804642e-05, "loss": 0.9682, "step": 3805 }, { "epoch": 1.7245129134571817, "grad_norm": 0.3747516178058588, "learning_rate": 6.625609961979394e-05, "loss": 0.9876, "step": 3806 }, { "epoch": 1.7249660172179428, "grad_norm": 0.34398578455487616, "learning_rate": 6.624655223161602e-05, "loss": 0.967, "step": 3807 }, { "epoch": 1.725419120978704, "grad_norm": 0.2597329174305455, "learning_rate": 6.623700221688585e-05, "loss": 0.9598, "step": 3808 }, { "epoch": 1.7258722247394653, "grad_norm": 0.23101861817932118, "learning_rate": 6.622744957655918e-05, "loss": 0.9487, "step": 3809 }, { "epoch": 1.7263253285002267, "grad_norm": 0.2178815338145204, "learning_rate": 6.621789431159192e-05, "loss": 0.9743, "step": 3810 }, { "epoch": 1.7267784322609878, "grad_norm": 0.2543424653215342, "learning_rate": 6.620833642294031e-05, "loss": 0.9418, "step": 3811 }, { "epoch": 1.727231536021749, "grad_norm": 0.2681773516239126, "learning_rate": 6.619877591156081e-05, "loss": 0.9976, "step": 3812 }, { "epoch": 1.72768463978251, "grad_norm": 0.2977310552533092, "learning_rate": 6.618921277841017e-05, "loss": 0.9683, "step": 3813 }, { "epoch": 1.7281377435432714, "grad_norm": 0.36132403307980004, "learning_rate": 6.61796470244454e-05, "loss": 0.957, "step": 3814 }, { "epoch": 1.7285908473040328, "grad_norm": 0.37735714663384695, "learning_rate": 6.617007865062377e-05, "loss": 0.9717, "step": 3815 }, { "epoch": 1.729043951064794, "grad_norm": 0.36370894139019766, "learning_rate": 6.616050765790279e-05, "loss": 0.9747, "step": 3816 }, { "epoch": 1.729497054825555, "grad_norm": 0.35206062593178095, "learning_rate": 6.615093404724028e-05, "loss": 0.9578, "step": 3817 }, { "epoch": 1.7299501585863162, "grad_norm": 0.3453038537378844, "learning_rate": 6.614135781959426e-05, "loss": 0.9535, "step": 3818 }, { "epoch": 1.7304032623470775, "grad_norm": 0.2701259902802974, "learning_rate": 6.613177897592307e-05, "loss": 0.9721, "step": 3819 }, { "epoch": 1.7308563661078387, "grad_norm": 0.2226625356461987, "learning_rate": 6.612219751718527e-05, "loss": 0.9728, "step": 3820 }, { "epoch": 1.7313094698686, "grad_norm": 0.2645537602598002, "learning_rate": 6.611261344433971e-05, "loss": 0.9678, "step": 3821 }, { "epoch": 1.7317625736293611, "grad_norm": 0.27973479424623854, "learning_rate": 6.610302675834549e-05, "loss": 0.9679, "step": 3822 }, { "epoch": 1.7322156773901223, "grad_norm": 0.2862044880061425, "learning_rate": 6.609343746016196e-05, "loss": 0.9663, "step": 3823 }, { "epoch": 1.7326687811508834, "grad_norm": 0.3280120642841347, "learning_rate": 6.608384555074875e-05, "loss": 0.9552, "step": 3824 }, { "epoch": 1.7331218849116448, "grad_norm": 0.34423681627375596, "learning_rate": 6.607425103106576e-05, "loss": 0.9633, "step": 3825 }, { "epoch": 1.7335749886724061, "grad_norm": 0.34939463911307367, "learning_rate": 6.606465390207311e-05, "loss": 0.9449, "step": 3826 }, { "epoch": 1.7340280924331672, "grad_norm": 0.38197108440181826, "learning_rate": 6.60550541647312e-05, "loss": 0.9503, "step": 3827 }, { "epoch": 1.7344811961939284, "grad_norm": 0.4014746374817315, "learning_rate": 6.604545182000075e-05, "loss": 0.951, "step": 3828 }, { "epoch": 1.7349342999546895, "grad_norm": 0.4120766058105364, "learning_rate": 6.603584686884262e-05, "loss": 0.9721, "step": 3829 }, { "epoch": 1.7353874037154509, "grad_norm": 0.4123328219978293, "learning_rate": 6.602623931221805e-05, "loss": 0.9561, "step": 3830 }, { "epoch": 1.7358405074762122, "grad_norm": 0.4471672038252659, "learning_rate": 6.601662915108847e-05, "loss": 0.9597, "step": 3831 }, { "epoch": 1.7362936112369733, "grad_norm": 0.5275090548139921, "learning_rate": 6.600701638641559e-05, "loss": 0.9949, "step": 3832 }, { "epoch": 1.7367467149977345, "grad_norm": 0.6149791143730325, "learning_rate": 6.599740101916138e-05, "loss": 0.9597, "step": 3833 }, { "epoch": 1.7371998187584956, "grad_norm": 0.6589109797474393, "learning_rate": 6.59877830502881e-05, "loss": 0.9594, "step": 3834 }, { "epoch": 1.737652922519257, "grad_norm": 0.6684481094582725, "learning_rate": 6.597816248075819e-05, "loss": 0.9527, "step": 3835 }, { "epoch": 1.738106026280018, "grad_norm": 0.5760669360974684, "learning_rate": 6.596853931153444e-05, "loss": 0.9641, "step": 3836 }, { "epoch": 1.7385591300407794, "grad_norm": 0.4128940427436101, "learning_rate": 6.595891354357985e-05, "loss": 0.9639, "step": 3837 }, { "epoch": 1.7390122338015406, "grad_norm": 0.30401823304228137, "learning_rate": 6.59492851778577e-05, "loss": 0.957, "step": 3838 }, { "epoch": 1.7394653375623017, "grad_norm": 0.28848557390971763, "learning_rate": 6.59396542153315e-05, "loss": 0.9671, "step": 3839 }, { "epoch": 1.7399184413230628, "grad_norm": 0.30596057627853135, "learning_rate": 6.593002065696508e-05, "loss": 0.9491, "step": 3840 }, { "epoch": 1.7403715450838242, "grad_norm": 0.29529633110569675, "learning_rate": 6.592038450372247e-05, "loss": 0.9695, "step": 3841 }, { "epoch": 1.7408246488445855, "grad_norm": 0.24329759323572545, "learning_rate": 6.591074575656799e-05, "loss": 0.9544, "step": 3842 }, { "epoch": 1.7412777526053467, "grad_norm": 0.22739239636488345, "learning_rate": 6.590110441646619e-05, "loss": 0.9377, "step": 3843 }, { "epoch": 1.7417308563661078, "grad_norm": 0.2539840514790894, "learning_rate": 6.589146048438192e-05, "loss": 0.9615, "step": 3844 }, { "epoch": 1.742183960126869, "grad_norm": 0.2992963483240323, "learning_rate": 6.588181396128027e-05, "loss": 0.9689, "step": 3845 }, { "epoch": 1.7426370638876303, "grad_norm": 0.32918145495540124, "learning_rate": 6.587216484812657e-05, "loss": 0.9574, "step": 3846 }, { "epoch": 1.7430901676483916, "grad_norm": 0.35858288419920753, "learning_rate": 6.586251314588646e-05, "loss": 0.978, "step": 3847 }, { "epoch": 1.7435432714091528, "grad_norm": 0.37781863065260196, "learning_rate": 6.585285885552581e-05, "loss": 0.9615, "step": 3848 }, { "epoch": 1.743996375169914, "grad_norm": 0.3877087645826915, "learning_rate": 6.58432019780107e-05, "loss": 0.9682, "step": 3849 }, { "epoch": 1.744449478930675, "grad_norm": 0.4025898869487086, "learning_rate": 6.583354251430754e-05, "loss": 0.9611, "step": 3850 }, { "epoch": 1.7449025826914364, "grad_norm": 0.419226911467587, "learning_rate": 6.582388046538301e-05, "loss": 0.9659, "step": 3851 }, { "epoch": 1.7453556864521975, "grad_norm": 0.38277381865005017, "learning_rate": 6.581421583220396e-05, "loss": 0.9869, "step": 3852 }, { "epoch": 1.7458087902129589, "grad_norm": 0.3516191759353198, "learning_rate": 6.580454861573758e-05, "loss": 0.9738, "step": 3853 }, { "epoch": 1.74626189397372, "grad_norm": 0.31859512564787057, "learning_rate": 6.579487881695129e-05, "loss": 0.9537, "step": 3854 }, { "epoch": 1.7467149977344811, "grad_norm": 0.27131354184331563, "learning_rate": 6.578520643681275e-05, "loss": 0.9623, "step": 3855 }, { "epoch": 1.7471681014952423, "grad_norm": 0.2472044053068463, "learning_rate": 6.577553147628993e-05, "loss": 0.9715, "step": 3856 }, { "epoch": 1.7476212052560036, "grad_norm": 0.26464853330561944, "learning_rate": 6.5765853936351e-05, "loss": 0.9485, "step": 3857 }, { "epoch": 1.748074309016765, "grad_norm": 0.28034039795065896, "learning_rate": 6.575617381796441e-05, "loss": 0.9444, "step": 3858 }, { "epoch": 1.748527412777526, "grad_norm": 0.24126323832270433, "learning_rate": 6.574649112209888e-05, "loss": 0.9623, "step": 3859 }, { "epoch": 1.7489805165382872, "grad_norm": 0.21182639069339568, "learning_rate": 6.573680584972339e-05, "loss": 0.9456, "step": 3860 }, { "epoch": 1.7494336202990484, "grad_norm": 0.23850872135823664, "learning_rate": 6.572711800180715e-05, "loss": 0.9603, "step": 3861 }, { "epoch": 1.7498867240598097, "grad_norm": 0.2396544494995407, "learning_rate": 6.571742757931966e-05, "loss": 0.9618, "step": 3862 }, { "epoch": 1.750339827820571, "grad_norm": 0.23779331850047727, "learning_rate": 6.570773458323066e-05, "loss": 0.9522, "step": 3863 }, { "epoch": 1.7507929315813322, "grad_norm": 0.25040590707937177, "learning_rate": 6.569803901451014e-05, "loss": 0.9673, "step": 3864 }, { "epoch": 1.7512460353420933, "grad_norm": 0.25310625152488125, "learning_rate": 6.568834087412836e-05, "loss": 0.9709, "step": 3865 }, { "epoch": 1.7516991391028545, "grad_norm": 0.2608318721526977, "learning_rate": 6.567864016305583e-05, "loss": 0.9771, "step": 3866 }, { "epoch": 1.7521522428636158, "grad_norm": 0.3526163512857203, "learning_rate": 6.566893688226333e-05, "loss": 0.9591, "step": 3867 }, { "epoch": 1.752605346624377, "grad_norm": 0.3872392627932795, "learning_rate": 6.565923103272191e-05, "loss": 0.9668, "step": 3868 }, { "epoch": 1.7530584503851383, "grad_norm": 0.3273149021257026, "learning_rate": 6.564952261540283e-05, "loss": 0.9572, "step": 3869 }, { "epoch": 1.7535115541458994, "grad_norm": 0.30403874721001584, "learning_rate": 6.563981163127762e-05, "loss": 0.939, "step": 3870 }, { "epoch": 1.7539646579066606, "grad_norm": 0.30894032099202745, "learning_rate": 6.563009808131811e-05, "loss": 0.9414, "step": 3871 }, { "epoch": 1.7544177616674217, "grad_norm": 0.3110855642233549, "learning_rate": 6.562038196649635e-05, "loss": 0.9608, "step": 3872 }, { "epoch": 1.754870865428183, "grad_norm": 0.3916649734252422, "learning_rate": 6.561066328778467e-05, "loss": 0.9634, "step": 3873 }, { "epoch": 1.7553239691889444, "grad_norm": 0.5086721493333778, "learning_rate": 6.56009420461556e-05, "loss": 0.9664, "step": 3874 }, { "epoch": 1.7557770729497055, "grad_norm": 0.57811299433077, "learning_rate": 6.559121824258198e-05, "loss": 0.9609, "step": 3875 }, { "epoch": 1.7562301767104667, "grad_norm": 0.6027159813169624, "learning_rate": 6.558149187803691e-05, "loss": 0.9751, "step": 3876 }, { "epoch": 1.7566832804712278, "grad_norm": 0.5906070559120317, "learning_rate": 6.557176295349372e-05, "loss": 0.9735, "step": 3877 }, { "epoch": 1.7571363842319891, "grad_norm": 0.5329259455388659, "learning_rate": 6.556203146992601e-05, "loss": 0.9835, "step": 3878 }, { "epoch": 1.7575894879927505, "grad_norm": 0.4924997118631332, "learning_rate": 6.555229742830762e-05, "loss": 0.9698, "step": 3879 }, { "epoch": 1.7580425917535116, "grad_norm": 0.41339469510142035, "learning_rate": 6.554256082961268e-05, "loss": 0.9565, "step": 3880 }, { "epoch": 1.7584956955142728, "grad_norm": 0.3094257390451732, "learning_rate": 6.553282167481552e-05, "loss": 0.9535, "step": 3881 }, { "epoch": 1.758948799275034, "grad_norm": 0.25499236760451727, "learning_rate": 6.552307996489079e-05, "loss": 0.9718, "step": 3882 }, { "epoch": 1.7594019030357952, "grad_norm": 0.24471112309356208, "learning_rate": 6.551333570081335e-05, "loss": 0.9699, "step": 3883 }, { "epoch": 1.7598550067965564, "grad_norm": 0.27510157882225833, "learning_rate": 6.550358888355833e-05, "loss": 0.956, "step": 3884 }, { "epoch": 1.7603081105573177, "grad_norm": 0.27485738086867156, "learning_rate": 6.549383951410113e-05, "loss": 0.9496, "step": 3885 }, { "epoch": 1.7607612143180789, "grad_norm": 0.3082648861328713, "learning_rate": 6.548408759341737e-05, "loss": 0.9758, "step": 3886 }, { "epoch": 1.76121431807884, "grad_norm": 0.24549448098804863, "learning_rate": 6.547433312248297e-05, "loss": 0.9728, "step": 3887 }, { "epoch": 1.7616674218396011, "grad_norm": 0.23959465592456797, "learning_rate": 6.546457610227408e-05, "loss": 0.9678, "step": 3888 }, { "epoch": 1.7621205256003625, "grad_norm": 0.330601866317788, "learning_rate": 6.545481653376707e-05, "loss": 0.9805, "step": 3889 }, { "epoch": 1.7625736293611238, "grad_norm": 0.384372769521644, "learning_rate": 6.544505441793866e-05, "loss": 0.9569, "step": 3890 }, { "epoch": 1.763026733121885, "grad_norm": 1.2717496905638443, "learning_rate": 6.54352897557657e-05, "loss": 0.9604, "step": 3891 }, { "epoch": 1.763479836882646, "grad_norm": 0.21620505203704013, "learning_rate": 6.542552254822542e-05, "loss": 0.9495, "step": 3892 }, { "epoch": 1.7639329406434072, "grad_norm": 0.33030542186684053, "learning_rate": 6.541575279629524e-05, "loss": 0.9729, "step": 3893 }, { "epoch": 1.7643860444041686, "grad_norm": 0.4626523708031906, "learning_rate": 6.540598050095281e-05, "loss": 0.9529, "step": 3894 }, { "epoch": 1.76483914816493, "grad_norm": 0.4563849958765944, "learning_rate": 6.539620566317609e-05, "loss": 0.9723, "step": 3895 }, { "epoch": 1.765292251925691, "grad_norm": 0.4618596051853344, "learning_rate": 6.538642828394325e-05, "loss": 0.9537, "step": 3896 }, { "epoch": 1.7657453556864522, "grad_norm": 0.48604774445155424, "learning_rate": 6.537664836423276e-05, "loss": 0.9811, "step": 3897 }, { "epoch": 1.7661984594472133, "grad_norm": 0.49565140977319566, "learning_rate": 6.536686590502331e-05, "loss": 0.9764, "step": 3898 }, { "epoch": 1.7666515632079747, "grad_norm": 0.5221073888797431, "learning_rate": 6.535708090729385e-05, "loss": 0.9769, "step": 3899 }, { "epoch": 1.7671046669687358, "grad_norm": 0.5437473328753282, "learning_rate": 6.534729337202358e-05, "loss": 0.9887, "step": 3900 }, { "epoch": 1.7675577707294972, "grad_norm": 0.5119533641583568, "learning_rate": 6.533750330019197e-05, "loss": 0.9685, "step": 3901 }, { "epoch": 1.7680108744902583, "grad_norm": 0.48934156141782714, "learning_rate": 6.532771069277874e-05, "loss": 0.957, "step": 3902 }, { "epoch": 1.7684639782510194, "grad_norm": 0.48019426896127554, "learning_rate": 6.531791555076384e-05, "loss": 0.987, "step": 3903 }, { "epoch": 1.7689170820117806, "grad_norm": 0.41462146262005517, "learning_rate": 6.530811787512751e-05, "loss": 0.9639, "step": 3904 }, { "epoch": 1.769370185772542, "grad_norm": 0.32757983997605306, "learning_rate": 6.529831766685022e-05, "loss": 0.9781, "step": 3905 }, { "epoch": 1.7698232895333033, "grad_norm": 0.29101101901816323, "learning_rate": 6.528851492691269e-05, "loss": 0.9702, "step": 3906 }, { "epoch": 1.7702763932940644, "grad_norm": 0.4375362213355063, "learning_rate": 6.527870965629591e-05, "loss": 1.0127, "step": 3907 }, { "epoch": 1.7707294970548255, "grad_norm": 0.349312091636619, "learning_rate": 6.526890185598113e-05, "loss": 0.9604, "step": 3908 }, { "epoch": 1.7711826008155866, "grad_norm": 0.36320952882100616, "learning_rate": 6.525909152694982e-05, "loss": 0.9588, "step": 3909 }, { "epoch": 1.771635704576348, "grad_norm": 0.3280906538255698, "learning_rate": 6.524927867018373e-05, "loss": 0.9454, "step": 3910 }, { "epoch": 1.7720888083371094, "grad_norm": 0.30526039051529236, "learning_rate": 6.523946328666484e-05, "loss": 0.9824, "step": 3911 }, { "epoch": 1.7725419120978705, "grad_norm": 0.3548094736238382, "learning_rate": 6.52296453773754e-05, "loss": 0.9533, "step": 3912 }, { "epoch": 1.7729950158586316, "grad_norm": 0.281269656470077, "learning_rate": 6.521982494329794e-05, "loss": 0.9418, "step": 3913 }, { "epoch": 1.7734481196193927, "grad_norm": 0.2818603807116819, "learning_rate": 6.521000198541519e-05, "loss": 0.9601, "step": 3914 }, { "epoch": 1.773901223380154, "grad_norm": 0.3311888796664892, "learning_rate": 6.520017650471018e-05, "loss": 0.9667, "step": 3915 }, { "epoch": 1.7743543271409152, "grad_norm": 0.31279883610690734, "learning_rate": 6.519034850216612e-05, "loss": 0.9691, "step": 3916 }, { "epoch": 1.7748074309016766, "grad_norm": 0.25881720398512714, "learning_rate": 6.518051797876656e-05, "loss": 0.9534, "step": 3917 }, { "epoch": 1.7752605346624377, "grad_norm": 0.275042217665405, "learning_rate": 6.517068493549523e-05, "loss": 0.9587, "step": 3918 }, { "epoch": 1.7757136384231988, "grad_norm": 0.3147225255112322, "learning_rate": 6.516084937333618e-05, "loss": 0.9783, "step": 3919 }, { "epoch": 1.77616674218396, "grad_norm": 0.31482725323658367, "learning_rate": 6.515101129327369e-05, "loss": 0.9858, "step": 3920 }, { "epoch": 1.7766198459447213, "grad_norm": 0.29968345553463416, "learning_rate": 6.514117069629222e-05, "loss": 0.9545, "step": 3921 }, { "epoch": 1.7770729497054827, "grad_norm": 0.33693639122631647, "learning_rate": 6.513132758337657e-05, "loss": 0.9482, "step": 3922 }, { "epoch": 1.7775260534662438, "grad_norm": 0.40935623033755825, "learning_rate": 6.512148195551178e-05, "loss": 0.9727, "step": 3923 }, { "epoch": 1.777979157227005, "grad_norm": 0.4893918183491706, "learning_rate": 6.511163381368312e-05, "loss": 0.9592, "step": 3924 }, { "epoch": 1.778432260987766, "grad_norm": 0.5592008428385468, "learning_rate": 6.510178315887607e-05, "loss": 0.9681, "step": 3925 }, { "epoch": 1.7788853647485274, "grad_norm": 0.5410260484399495, "learning_rate": 6.509192999207649e-05, "loss": 0.9516, "step": 3926 }, { "epoch": 1.7793384685092888, "grad_norm": 0.43842286178650786, "learning_rate": 6.508207431427033e-05, "loss": 0.9517, "step": 3927 }, { "epoch": 1.77979157227005, "grad_norm": 0.3502449269411218, "learning_rate": 6.507221612644392e-05, "loss": 0.9847, "step": 3928 }, { "epoch": 1.780244676030811, "grad_norm": 0.262165833998237, "learning_rate": 6.506235542958376e-05, "loss": 0.993, "step": 3929 }, { "epoch": 1.7806977797915722, "grad_norm": 0.2619309424136776, "learning_rate": 6.505249222467665e-05, "loss": 0.9609, "step": 3930 }, { "epoch": 1.7811508835523335, "grad_norm": 0.3019625576743058, "learning_rate": 6.504262651270962e-05, "loss": 0.9383, "step": 3931 }, { "epoch": 1.7816039873130947, "grad_norm": 0.4048629395493005, "learning_rate": 6.503275829466996e-05, "loss": 0.9699, "step": 3932 }, { "epoch": 1.782057091073856, "grad_norm": 0.3484199945809545, "learning_rate": 6.502288757154519e-05, "loss": 0.9673, "step": 3933 }, { "epoch": 1.7825101948346171, "grad_norm": 0.4455580176367527, "learning_rate": 6.501301434432312e-05, "loss": 0.9796, "step": 3934 }, { "epoch": 1.7829632985953783, "grad_norm": 0.4677792448367538, "learning_rate": 6.500313861399175e-05, "loss": 0.9553, "step": 3935 }, { "epoch": 1.7834164023561394, "grad_norm": 0.4521444551641366, "learning_rate": 6.499326038153942e-05, "loss": 0.977, "step": 3936 }, { "epoch": 1.7838695061169008, "grad_norm": 0.4623068839952339, "learning_rate": 6.498337964795459e-05, "loss": 0.9825, "step": 3937 }, { "epoch": 1.7843226098776621, "grad_norm": 0.4477623517076979, "learning_rate": 6.497349641422612e-05, "loss": 0.9707, "step": 3938 }, { "epoch": 1.7847757136384232, "grad_norm": 0.39163128748747766, "learning_rate": 6.496361068134301e-05, "loss": 0.9462, "step": 3939 }, { "epoch": 1.7852288173991844, "grad_norm": 0.3001134318021097, "learning_rate": 6.495372245029457e-05, "loss": 0.957, "step": 3940 }, { "epoch": 1.7856819211599455, "grad_norm": 0.29299879771655424, "learning_rate": 6.494383172207032e-05, "loss": 0.984, "step": 3941 }, { "epoch": 1.7861350249207069, "grad_norm": 0.34073503468466193, "learning_rate": 6.493393849766005e-05, "loss": 1.0007, "step": 3942 }, { "epoch": 1.7865881286814682, "grad_norm": 0.37128058886567694, "learning_rate": 6.492404277805382e-05, "loss": 0.9683, "step": 3943 }, { "epoch": 1.7870412324422293, "grad_norm": 0.38343378935516265, "learning_rate": 6.491414456424188e-05, "loss": 0.9554, "step": 3944 }, { "epoch": 1.7874943362029905, "grad_norm": 0.4254968298844659, "learning_rate": 6.490424385721479e-05, "loss": 0.972, "step": 3945 }, { "epoch": 1.7879474399637516, "grad_norm": 0.31581978258682036, "learning_rate": 6.489434065796332e-05, "loss": 0.9693, "step": 3946 }, { "epoch": 1.788400543724513, "grad_norm": 0.29047714439260164, "learning_rate": 6.488443496747852e-05, "loss": 0.9516, "step": 3947 }, { "epoch": 1.788853647485274, "grad_norm": 0.28970529421953634, "learning_rate": 6.487452678675169e-05, "loss": 0.9765, "step": 3948 }, { "epoch": 1.7893067512460354, "grad_norm": 0.2510534841150238, "learning_rate": 6.486461611677433e-05, "loss": 0.9654, "step": 3949 }, { "epoch": 1.7897598550067966, "grad_norm": 0.2435267329526528, "learning_rate": 6.485470295853825e-05, "loss": 0.9466, "step": 3950 }, { "epoch": 1.7902129587675577, "grad_norm": 0.2179259130643031, "learning_rate": 6.484478731303545e-05, "loss": 0.9647, "step": 3951 }, { "epoch": 1.7906660625283188, "grad_norm": 0.21359640734045235, "learning_rate": 6.483486918125825e-05, "loss": 0.9597, "step": 3952 }, { "epoch": 1.7911191662890802, "grad_norm": 0.255011415350603, "learning_rate": 6.482494856419915e-05, "loss": 0.9668, "step": 3953 }, { "epoch": 1.7915722700498415, "grad_norm": 0.35812606606400593, "learning_rate": 6.481502546285097e-05, "loss": 0.9691, "step": 3954 }, { "epoch": 1.7920253738106027, "grad_norm": 0.4562384786468138, "learning_rate": 6.480509987820668e-05, "loss": 0.9563, "step": 3955 }, { "epoch": 1.7924784775713638, "grad_norm": 0.5306729469342375, "learning_rate": 6.47951718112596e-05, "loss": 0.9762, "step": 3956 }, { "epoch": 1.792931581332125, "grad_norm": 0.6087495480657512, "learning_rate": 6.478524126300324e-05, "loss": 0.9572, "step": 3957 }, { "epoch": 1.7933846850928863, "grad_norm": 0.6363745167601166, "learning_rate": 6.477530823443138e-05, "loss": 0.951, "step": 3958 }, { "epoch": 1.7938377888536476, "grad_norm": 0.5324063817941025, "learning_rate": 6.476537272653801e-05, "loss": 0.9666, "step": 3959 }, { "epoch": 1.7942908926144088, "grad_norm": 0.38232877238894936, "learning_rate": 6.475543474031742e-05, "loss": 0.9528, "step": 3960 }, { "epoch": 1.79474399637517, "grad_norm": 0.27103899728352315, "learning_rate": 6.474549427676415e-05, "loss": 0.956, "step": 3961 }, { "epoch": 1.795197100135931, "grad_norm": 0.22002456149252494, "learning_rate": 6.473555133687292e-05, "loss": 0.969, "step": 3962 }, { "epoch": 1.7956502038966924, "grad_norm": 0.27994106043486117, "learning_rate": 6.472560592163879e-05, "loss": 0.9889, "step": 3963 }, { "epoch": 1.7961033076574535, "grad_norm": 0.4187763135991805, "learning_rate": 6.471565803205698e-05, "loss": 0.9676, "step": 3964 }, { "epoch": 1.7965564114182149, "grad_norm": 0.46906391095100464, "learning_rate": 6.470570766912302e-05, "loss": 0.9558, "step": 3965 }, { "epoch": 1.797009515178976, "grad_norm": 0.47457052932966265, "learning_rate": 6.469575483383265e-05, "loss": 0.9917, "step": 3966 }, { "epoch": 1.7974626189397371, "grad_norm": 0.4628310951947401, "learning_rate": 6.46857995271819e-05, "loss": 0.967, "step": 3967 }, { "epoch": 1.7979157227004983, "grad_norm": 0.41698775450374115, "learning_rate": 6.467584175016699e-05, "loss": 0.9622, "step": 3968 }, { "epoch": 1.7983688264612596, "grad_norm": 0.36146755801831165, "learning_rate": 6.466588150378443e-05, "loss": 0.9797, "step": 3969 }, { "epoch": 1.798821930222021, "grad_norm": 0.3154730755303909, "learning_rate": 6.465591878903096e-05, "loss": 0.9514, "step": 3970 }, { "epoch": 1.799275033982782, "grad_norm": 0.30391892681422483, "learning_rate": 6.46459536069036e-05, "loss": 0.9649, "step": 3971 }, { "epoch": 1.7997281377435432, "grad_norm": 0.2755539779180359, "learning_rate": 6.463598595839953e-05, "loss": 0.9402, "step": 3972 }, { "epoch": 1.8001812415043044, "grad_norm": 0.2394058589110086, "learning_rate": 6.46260158445163e-05, "loss": 0.9587, "step": 3973 }, { "epoch": 1.8006343452650657, "grad_norm": 0.3261778653999901, "learning_rate": 6.46160432662516e-05, "loss": 0.9784, "step": 3974 }, { "epoch": 1.801087449025827, "grad_norm": 0.41085668910925705, "learning_rate": 6.460606822460341e-05, "loss": 0.9611, "step": 3975 }, { "epoch": 1.8015405527865882, "grad_norm": 0.43651725535211183, "learning_rate": 6.459609072056998e-05, "loss": 0.9518, "step": 3976 }, { "epoch": 1.8019936565473493, "grad_norm": 0.3934238907079356, "learning_rate": 6.458611075514977e-05, "loss": 0.9752, "step": 3977 }, { "epoch": 1.8024467603081105, "grad_norm": 0.3294680296843117, "learning_rate": 6.457612832934148e-05, "loss": 0.9652, "step": 3978 }, { "epoch": 1.8028998640688718, "grad_norm": 0.32795565095268897, "learning_rate": 6.456614344414409e-05, "loss": 0.9631, "step": 3979 }, { "epoch": 1.803352967829633, "grad_norm": 0.34860628849494496, "learning_rate": 6.455615610055681e-05, "loss": 0.9772, "step": 3980 }, { "epoch": 1.8038060715903943, "grad_norm": 0.3352052265293342, "learning_rate": 6.45461662995791e-05, "loss": 0.9712, "step": 3981 }, { "epoch": 1.8042591753511554, "grad_norm": 0.3420224356592389, "learning_rate": 6.453617404221066e-05, "loss": 0.9722, "step": 3982 }, { "epoch": 1.8047122791119166, "grad_norm": 0.31590581822219865, "learning_rate": 6.452617932945143e-05, "loss": 0.9902, "step": 3983 }, { "epoch": 1.8051653828726777, "grad_norm": 0.29358682417324905, "learning_rate": 6.451618216230162e-05, "loss": 0.9456, "step": 3984 }, { "epoch": 1.805618486633439, "grad_norm": 0.28561605407156687, "learning_rate": 6.450618254176165e-05, "loss": 0.95, "step": 3985 }, { "epoch": 1.8060715903942004, "grad_norm": 0.3133462323276506, "learning_rate": 6.449618046883222e-05, "loss": 0.9689, "step": 3986 }, { "epoch": 1.8065246941549615, "grad_norm": 0.29294387823813156, "learning_rate": 6.448617594451424e-05, "loss": 0.9643, "step": 3987 }, { "epoch": 1.8069777979157227, "grad_norm": 0.25076520617540393, "learning_rate": 6.44761689698089e-05, "loss": 0.9584, "step": 3988 }, { "epoch": 1.8074309016764838, "grad_norm": 0.2600477173849382, "learning_rate": 6.446615954571764e-05, "loss": 0.962, "step": 3989 }, { "epoch": 1.8078840054372451, "grad_norm": 0.2807829993220264, "learning_rate": 6.445614767324208e-05, "loss": 0.9578, "step": 3990 }, { "epoch": 1.8083371091980065, "grad_norm": 0.34925354200471953, "learning_rate": 6.444613335338416e-05, "loss": 0.9608, "step": 3991 }, { "epoch": 1.8087902129587676, "grad_norm": 0.3457310655988422, "learning_rate": 6.443611658714604e-05, "loss": 0.9482, "step": 3992 }, { "epoch": 1.8092433167195288, "grad_norm": 0.3626501907636255, "learning_rate": 6.442609737553012e-05, "loss": 0.9534, "step": 3993 }, { "epoch": 1.80969642048029, "grad_norm": 0.40882685826533754, "learning_rate": 6.441607571953902e-05, "loss": 0.965, "step": 3994 }, { "epoch": 1.8101495242410512, "grad_norm": 0.4425476402882528, "learning_rate": 6.440605162017566e-05, "loss": 0.9525, "step": 3995 }, { "epoch": 1.8106026280018124, "grad_norm": 0.46281854412529283, "learning_rate": 6.439602507844316e-05, "loss": 0.9694, "step": 3996 }, { "epoch": 1.8110557317625737, "grad_norm": 0.48503354427802814, "learning_rate": 6.438599609534491e-05, "loss": 0.9622, "step": 3997 }, { "epoch": 1.8115088355233349, "grad_norm": 0.4516693086285219, "learning_rate": 6.437596467188451e-05, "loss": 0.9662, "step": 3998 }, { "epoch": 1.811961939284096, "grad_norm": 0.41211142065843104, "learning_rate": 6.436593080906585e-05, "loss": 0.9652, "step": 3999 }, { "epoch": 1.8124150430448571, "grad_norm": 0.3643628700802737, "learning_rate": 6.435589450789301e-05, "loss": 0.953, "step": 4000 }, { "epoch": 1.8128681468056185, "grad_norm": 0.31941372394680884, "learning_rate": 6.434585576937041e-05, "loss": 0.9764, "step": 4001 }, { "epoch": 1.8133212505663798, "grad_norm": 0.2411075547369455, "learning_rate": 6.433581459450258e-05, "loss": 0.9865, "step": 4002 }, { "epoch": 1.813774354327141, "grad_norm": 0.234258846140491, "learning_rate": 6.432577098429441e-05, "loss": 0.9468, "step": 4003 }, { "epoch": 1.814227458087902, "grad_norm": 0.2768036357841629, "learning_rate": 6.431572493975095e-05, "loss": 0.9529, "step": 4004 }, { "epoch": 1.8146805618486632, "grad_norm": 0.30820874919826885, "learning_rate": 6.430567646187755e-05, "loss": 0.9484, "step": 4005 }, { "epoch": 1.8151336656094246, "grad_norm": 0.29176501426065543, "learning_rate": 6.429562555167979e-05, "loss": 0.9524, "step": 4006 }, { "epoch": 1.815586769370186, "grad_norm": 0.31739329186031223, "learning_rate": 6.428557221016349e-05, "loss": 0.9497, "step": 4007 }, { "epoch": 1.816039873130947, "grad_norm": 0.3460663331138575, "learning_rate": 6.427551643833469e-05, "loss": 0.946, "step": 4008 }, { "epoch": 1.8164929768917082, "grad_norm": 0.34544316688134774, "learning_rate": 6.42654582371997e-05, "loss": 0.9767, "step": 4009 }, { "epoch": 1.8169460806524693, "grad_norm": 0.3055796481259138, "learning_rate": 6.425539760776507e-05, "loss": 0.9583, "step": 4010 }, { "epoch": 1.8173991844132307, "grad_norm": 0.29398495989668405, "learning_rate": 6.42453345510376e-05, "loss": 0.9672, "step": 4011 }, { "epoch": 1.8178522881739918, "grad_norm": 0.29267014877091135, "learning_rate": 6.423526906802431e-05, "loss": 0.9521, "step": 4012 }, { "epoch": 1.8183053919347532, "grad_norm": 0.3075827759348859, "learning_rate": 6.42252011597325e-05, "loss": 0.9637, "step": 4013 }, { "epoch": 1.8187584956955143, "grad_norm": 0.3194936361207015, "learning_rate": 6.421513082716965e-05, "loss": 0.9791, "step": 4014 }, { "epoch": 1.8192115994562754, "grad_norm": 0.2727961703277067, "learning_rate": 6.420505807134355e-05, "loss": 0.9595, "step": 4015 }, { "epoch": 1.8196647032170365, "grad_norm": 0.2769083808394489, "learning_rate": 6.419498289326218e-05, "loss": 0.9655, "step": 4016 }, { "epoch": 1.820117806977798, "grad_norm": 0.3122932890020639, "learning_rate": 6.418490529393382e-05, "loss": 0.9452, "step": 4017 }, { "epoch": 1.8205709107385593, "grad_norm": 0.33388904967089106, "learning_rate": 6.417482527436692e-05, "loss": 0.9671, "step": 4018 }, { "epoch": 1.8210240144993204, "grad_norm": 0.3273565934145807, "learning_rate": 6.416474283557024e-05, "loss": 0.9667, "step": 4019 }, { "epoch": 1.8214771182600815, "grad_norm": 0.292221971638999, "learning_rate": 6.415465797855274e-05, "loss": 0.9506, "step": 4020 }, { "epoch": 1.8219302220208426, "grad_norm": 0.24172925712142987, "learning_rate": 6.414457070432363e-05, "loss": 0.9463, "step": 4021 }, { "epoch": 1.822383325781604, "grad_norm": 0.24949091641101737, "learning_rate": 6.413448101389239e-05, "loss": 0.9788, "step": 4022 }, { "epoch": 1.8228364295423654, "grad_norm": 0.24870601617171406, "learning_rate": 6.412438890826869e-05, "loss": 0.9683, "step": 4023 }, { "epoch": 1.8232895333031265, "grad_norm": 0.2702350959485261, "learning_rate": 6.411429438846248e-05, "loss": 0.9702, "step": 4024 }, { "epoch": 1.8237426370638876, "grad_norm": 0.31256756705758215, "learning_rate": 6.410419745548394e-05, "loss": 0.9427, "step": 4025 }, { "epoch": 1.8241957408246487, "grad_norm": 0.3849042218361986, "learning_rate": 6.40940981103435e-05, "loss": 0.9628, "step": 4026 }, { "epoch": 1.82464884458541, "grad_norm": 0.5083809552043927, "learning_rate": 6.408399635405182e-05, "loss": 0.9464, "step": 4027 }, { "epoch": 1.8251019483461712, "grad_norm": 0.5558418433595159, "learning_rate": 6.407389218761981e-05, "loss": 0.9581, "step": 4028 }, { "epoch": 1.8255550521069326, "grad_norm": 0.5351707231945753, "learning_rate": 6.40637856120586e-05, "loss": 0.9539, "step": 4029 }, { "epoch": 1.8260081558676937, "grad_norm": 0.4737735220345536, "learning_rate": 6.40536766283796e-05, "loss": 0.9629, "step": 4030 }, { "epoch": 1.8264612596284548, "grad_norm": 0.4140650186347061, "learning_rate": 6.404356523759444e-05, "loss": 0.9827, "step": 4031 }, { "epoch": 1.826914363389216, "grad_norm": 0.32311547986657785, "learning_rate": 6.403345144071496e-05, "loss": 0.9754, "step": 4032 }, { "epoch": 1.8273674671499773, "grad_norm": 0.25703755228237396, "learning_rate": 6.402333523875331e-05, "loss": 0.9359, "step": 4033 }, { "epoch": 1.8278205709107387, "grad_norm": 0.2712237308179196, "learning_rate": 6.40132166327218e-05, "loss": 0.9716, "step": 4034 }, { "epoch": 1.8282736746714998, "grad_norm": 0.34132222479632746, "learning_rate": 6.400309562363305e-05, "loss": 0.9514, "step": 4035 }, { "epoch": 1.828726778432261, "grad_norm": 0.44429832873031555, "learning_rate": 6.39929722124999e-05, "loss": 0.9675, "step": 4036 }, { "epoch": 1.829179882193022, "grad_norm": 0.4299981312904889, "learning_rate": 6.39828464003354e-05, "loss": 0.9565, "step": 4037 }, { "epoch": 1.8296329859537834, "grad_norm": 0.3806753427539694, "learning_rate": 6.397271818815287e-05, "loss": 0.9821, "step": 4038 }, { "epoch": 1.8300860897145448, "grad_norm": 0.38990095582489304, "learning_rate": 6.396258757696586e-05, "loss": 0.9785, "step": 4039 }, { "epoch": 1.830539193475306, "grad_norm": 0.34761951816200537, "learning_rate": 6.395245456778818e-05, "loss": 0.9361, "step": 4040 }, { "epoch": 1.830992297236067, "grad_norm": 0.2619088203807847, "learning_rate": 6.394231916163385e-05, "loss": 0.9672, "step": 4041 }, { "epoch": 1.8314454009968282, "grad_norm": 0.24638892806378984, "learning_rate": 6.393218135951714e-05, "loss": 0.9706, "step": 4042 }, { "epoch": 1.8318985047575895, "grad_norm": 0.23455909523803878, "learning_rate": 6.392204116245257e-05, "loss": 0.9505, "step": 4043 }, { "epoch": 1.8323516085183507, "grad_norm": 0.18633778619842983, "learning_rate": 6.391189857145489e-05, "loss": 0.9518, "step": 4044 }, { "epoch": 1.832804712279112, "grad_norm": 0.27059455325479703, "learning_rate": 6.390175358753909e-05, "loss": 0.9709, "step": 4045 }, { "epoch": 1.8332578160398731, "grad_norm": 0.3282736538427965, "learning_rate": 6.38916062117204e-05, "loss": 0.9695, "step": 4046 }, { "epoch": 1.8337109198006343, "grad_norm": 0.33715792730921273, "learning_rate": 6.388145644501432e-05, "loss": 0.9727, "step": 4047 }, { "epoch": 1.8341640235613954, "grad_norm": 0.2844167681541638, "learning_rate": 6.387130428843651e-05, "loss": 0.9833, "step": 4048 }, { "epoch": 1.8346171273221568, "grad_norm": 0.24071179214768482, "learning_rate": 6.386114974300295e-05, "loss": 0.9512, "step": 4049 }, { "epoch": 1.8350702310829181, "grad_norm": 0.23591273253911343, "learning_rate": 6.385099280972983e-05, "loss": 0.9544, "step": 4050 }, { "epoch": 1.8355233348436792, "grad_norm": 0.22421089483998774, "learning_rate": 6.384083348963357e-05, "loss": 0.9584, "step": 4051 }, { "epoch": 1.8359764386044404, "grad_norm": 0.21958944512614506, "learning_rate": 6.383067178373083e-05, "loss": 0.9369, "step": 4052 }, { "epoch": 1.8364295423652015, "grad_norm": 0.23891879827377138, "learning_rate": 6.382050769303852e-05, "loss": 0.9681, "step": 4053 }, { "epoch": 1.8368826461259629, "grad_norm": 0.2680864666026457, "learning_rate": 6.381034121857381e-05, "loss": 0.934, "step": 4054 }, { "epoch": 1.8373357498867242, "grad_norm": 0.27152806418238334, "learning_rate": 6.380017236135404e-05, "loss": 0.9591, "step": 4055 }, { "epoch": 1.8377888536474853, "grad_norm": 0.24749495431644286, "learning_rate": 6.379000112239684e-05, "loss": 0.9594, "step": 4056 }, { "epoch": 1.8382419574082465, "grad_norm": 0.2537050873555254, "learning_rate": 6.37798275027201e-05, "loss": 0.9625, "step": 4057 }, { "epoch": 1.8386950611690076, "grad_norm": 0.2841809643430826, "learning_rate": 6.376965150334188e-05, "loss": 0.9793, "step": 4058 }, { "epoch": 1.839148164929769, "grad_norm": 0.29274736860566786, "learning_rate": 6.375947312528054e-05, "loss": 0.9656, "step": 4059 }, { "epoch": 1.83960126869053, "grad_norm": 0.28937576236959384, "learning_rate": 6.374929236955463e-05, "loss": 0.9461, "step": 4060 }, { "epoch": 1.8400543724512914, "grad_norm": 0.29362696867466437, "learning_rate": 6.373910923718299e-05, "loss": 0.9472, "step": 4061 }, { "epoch": 1.8405074762120526, "grad_norm": 0.315054786733498, "learning_rate": 6.372892372918465e-05, "loss": 0.9763, "step": 4062 }, { "epoch": 1.8409605799728137, "grad_norm": 0.3468465833213505, "learning_rate": 6.37187358465789e-05, "loss": 0.9539, "step": 4063 }, { "epoch": 1.8414136837335748, "grad_norm": 0.3507855352708534, "learning_rate": 6.370854559038526e-05, "loss": 0.9654, "step": 4064 }, { "epoch": 1.8418667874943362, "grad_norm": 0.3133581418283567, "learning_rate": 6.369835296162351e-05, "loss": 0.9758, "step": 4065 }, { "epoch": 1.8423198912550975, "grad_norm": 0.27204216492764544, "learning_rate": 6.368815796131363e-05, "loss": 0.9398, "step": 4066 }, { "epoch": 1.8427729950158587, "grad_norm": 0.221883862683992, "learning_rate": 6.367796059047586e-05, "loss": 0.959, "step": 4067 }, { "epoch": 1.8432260987766198, "grad_norm": 0.24518729023350525, "learning_rate": 6.366776085013069e-05, "loss": 0.9675, "step": 4068 }, { "epoch": 1.843679202537381, "grad_norm": 0.3579025633110214, "learning_rate": 6.365755874129882e-05, "loss": 0.9552, "step": 4069 }, { "epoch": 1.8441323062981423, "grad_norm": 0.37066134266443207, "learning_rate": 6.364735426500119e-05, "loss": 0.9544, "step": 4070 }, { "epoch": 1.8445854100589036, "grad_norm": 0.3062084460904077, "learning_rate": 6.3637147422259e-05, "loss": 0.9633, "step": 4071 }, { "epoch": 1.8450385138196648, "grad_norm": 0.3263973107769448, "learning_rate": 6.362693821409366e-05, "loss": 0.9522, "step": 4072 }, { "epoch": 1.845491617580426, "grad_norm": 0.363674666002364, "learning_rate": 6.361672664152684e-05, "loss": 0.9686, "step": 4073 }, { "epoch": 1.845944721341187, "grad_norm": 0.37524096444335836, "learning_rate": 6.360651270558042e-05, "loss": 0.972, "step": 4074 }, { "epoch": 1.8463978251019484, "grad_norm": 0.3598814492919252, "learning_rate": 6.359629640727655e-05, "loss": 0.9675, "step": 4075 }, { "epoch": 1.8468509288627095, "grad_norm": 0.3478699003341858, "learning_rate": 6.358607774763759e-05, "loss": 0.9433, "step": 4076 }, { "epoch": 1.8473040326234709, "grad_norm": 0.3103241536833286, "learning_rate": 6.357585672768615e-05, "loss": 0.9739, "step": 4077 }, { "epoch": 1.847757136384232, "grad_norm": 0.33126912386748614, "learning_rate": 6.356563334844507e-05, "loss": 0.9612, "step": 4078 }, { "epoch": 1.8482102401449931, "grad_norm": 0.3787447210136553, "learning_rate": 6.355540761093739e-05, "loss": 0.9518, "step": 4079 }, { "epoch": 1.8486633439057543, "grad_norm": 0.3635326939325441, "learning_rate": 6.354517951618648e-05, "loss": 0.963, "step": 4080 }, { "epoch": 1.8491164476665156, "grad_norm": 0.28877819906623375, "learning_rate": 6.353494906521585e-05, "loss": 0.9697, "step": 4081 }, { "epoch": 1.849569551427277, "grad_norm": 0.29007627885567194, "learning_rate": 6.352471625904932e-05, "loss": 0.9712, "step": 4082 }, { "epoch": 1.850022655188038, "grad_norm": 0.34696236194484087, "learning_rate": 6.351448109871087e-05, "loss": 0.9495, "step": 4083 }, { "epoch": 1.8504757589487992, "grad_norm": 0.3352347529184095, "learning_rate": 6.350424358522478e-05, "loss": 0.99, "step": 4084 }, { "epoch": 1.8509288627095604, "grad_norm": 0.26709482365708664, "learning_rate": 6.349400371961553e-05, "loss": 0.9632, "step": 4085 }, { "epoch": 1.8513819664703217, "grad_norm": 0.26798719622612993, "learning_rate": 6.348376150290785e-05, "loss": 0.9425, "step": 4086 }, { "epoch": 1.851835070231083, "grad_norm": 0.303016326616463, "learning_rate": 6.347351693612671e-05, "loss": 0.9723, "step": 4087 }, { "epoch": 1.8522881739918442, "grad_norm": 0.3179703705683223, "learning_rate": 6.34632700202973e-05, "loss": 0.9715, "step": 4088 }, { "epoch": 1.8527412777526053, "grad_norm": 0.3332220538163646, "learning_rate": 6.345302075644504e-05, "loss": 0.9551, "step": 4089 }, { "epoch": 1.8531943815133665, "grad_norm": 0.3966891162573196, "learning_rate": 6.344276914559564e-05, "loss": 0.9729, "step": 4090 }, { "epoch": 1.8536474852741278, "grad_norm": 0.4629457952551116, "learning_rate": 6.343251518877495e-05, "loss": 0.9534, "step": 4091 }, { "epoch": 1.854100589034889, "grad_norm": 0.47199020343118964, "learning_rate": 6.342225888700912e-05, "loss": 0.9754, "step": 4092 }, { "epoch": 1.8545536927956503, "grad_norm": 0.42895129408725874, "learning_rate": 6.341200024132455e-05, "loss": 0.962, "step": 4093 }, { "epoch": 1.8550067965564114, "grad_norm": 0.33775684892908536, "learning_rate": 6.340173925274781e-05, "loss": 0.9674, "step": 4094 }, { "epoch": 1.8554599003171726, "grad_norm": 0.32520109384383183, "learning_rate": 6.339147592230577e-05, "loss": 0.9619, "step": 4095 }, { "epoch": 1.8559130040779337, "grad_norm": 0.3140676827109771, "learning_rate": 6.338121025102549e-05, "loss": 0.9535, "step": 4096 }, { "epoch": 1.856366107838695, "grad_norm": 0.25170743626958325, "learning_rate": 6.337094223993427e-05, "loss": 0.9659, "step": 4097 }, { "epoch": 1.8568192115994564, "grad_norm": 0.271231402704049, "learning_rate": 6.336067189005969e-05, "loss": 0.954, "step": 4098 }, { "epoch": 1.8572723153602175, "grad_norm": 0.3806325973161579, "learning_rate": 6.335039920242948e-05, "loss": 0.9812, "step": 4099 }, { "epoch": 1.8577254191209787, "grad_norm": 0.4598041388100438, "learning_rate": 6.334012417807166e-05, "loss": 0.9543, "step": 4100 }, { "epoch": 1.8581785228817398, "grad_norm": 0.4979790963198932, "learning_rate": 6.332984681801452e-05, "loss": 0.9632, "step": 4101 }, { "epoch": 1.8586316266425011, "grad_norm": 0.5223539448719391, "learning_rate": 6.331956712328649e-05, "loss": 0.9589, "step": 4102 }, { "epoch": 1.8590847304032625, "grad_norm": 0.510470317668037, "learning_rate": 6.33092850949163e-05, "loss": 0.9433, "step": 4103 }, { "epoch": 1.8595378341640236, "grad_norm": 0.43825399909307333, "learning_rate": 6.329900073393288e-05, "loss": 0.9536, "step": 4104 }, { "epoch": 1.8599909379247848, "grad_norm": 0.40688334540798377, "learning_rate": 6.328871404136544e-05, "loss": 0.962, "step": 4105 }, { "epoch": 1.860444041685546, "grad_norm": 0.4356776389476663, "learning_rate": 6.327842501824339e-05, "loss": 0.9636, "step": 4106 }, { "epoch": 1.8608971454463072, "grad_norm": 0.4128117956661725, "learning_rate": 6.326813366559634e-05, "loss": 0.9859, "step": 4107 }, { "epoch": 1.8613502492070684, "grad_norm": 0.393733397013343, "learning_rate": 6.32578399844542e-05, "loss": 0.9637, "step": 4108 }, { "epoch": 1.8618033529678297, "grad_norm": 0.37976505321631376, "learning_rate": 6.324754397584709e-05, "loss": 0.953, "step": 4109 }, { "epoch": 1.8622564567285909, "grad_norm": 0.38006641542126923, "learning_rate": 6.323724564080532e-05, "loss": 0.9677, "step": 4110 }, { "epoch": 1.862709560489352, "grad_norm": 0.37411856241973146, "learning_rate": 6.322694498035949e-05, "loss": 0.961, "step": 4111 }, { "epoch": 1.8631626642501131, "grad_norm": 0.38693671917600087, "learning_rate": 6.321664199554041e-05, "loss": 0.9437, "step": 4112 }, { "epoch": 1.8636157680108745, "grad_norm": 0.3758789023180022, "learning_rate": 6.320633668737911e-05, "loss": 0.9425, "step": 4113 }, { "epoch": 1.8640688717716358, "grad_norm": 0.34788084912876266, "learning_rate": 6.319602905690689e-05, "loss": 0.9878, "step": 4114 }, { "epoch": 1.864521975532397, "grad_norm": 0.33952259204668256, "learning_rate": 6.318571910515525e-05, "loss": 0.9404, "step": 4115 }, { "epoch": 1.864975079293158, "grad_norm": 0.2946701002190047, "learning_rate": 6.317540683315593e-05, "loss": 0.9677, "step": 4116 }, { "epoch": 1.8654281830539192, "grad_norm": 0.2939316458888824, "learning_rate": 6.316509224194087e-05, "loss": 0.9748, "step": 4117 }, { "epoch": 1.8658812868146806, "grad_norm": 0.32911504219194604, "learning_rate": 6.315477533254232e-05, "loss": 0.9495, "step": 4118 }, { "epoch": 1.866334390575442, "grad_norm": 0.34765113089361976, "learning_rate": 6.314445610599269e-05, "loss": 0.9509, "step": 4119 }, { "epoch": 1.866787494336203, "grad_norm": 0.34297855515948417, "learning_rate": 6.313413456332466e-05, "loss": 0.9603, "step": 4120 }, { "epoch": 1.8672405980969642, "grad_norm": 0.32657621274963733, "learning_rate": 6.312381070557113e-05, "loss": 0.9941, "step": 4121 }, { "epoch": 1.8676937018577253, "grad_norm": 0.33313002437209, "learning_rate": 6.311348453376523e-05, "loss": 0.9502, "step": 4122 }, { "epoch": 1.8681468056184867, "grad_norm": 0.3103268023490667, "learning_rate": 6.310315604894031e-05, "loss": 0.9667, "step": 4123 }, { "epoch": 1.8685999093792478, "grad_norm": 0.2644366283842374, "learning_rate": 6.309282525212999e-05, "loss": 0.9692, "step": 4124 }, { "epoch": 1.8690530131400092, "grad_norm": 0.23851698657139167, "learning_rate": 6.308249214436808e-05, "loss": 0.9474, "step": 4125 }, { "epoch": 1.8695061169007703, "grad_norm": 0.30486005655196385, "learning_rate": 6.307215672668866e-05, "loss": 0.9561, "step": 4126 }, { "epoch": 1.8699592206615314, "grad_norm": 0.3234588442420982, "learning_rate": 6.306181900012598e-05, "loss": 0.9728, "step": 4127 }, { "epoch": 1.8704123244222925, "grad_norm": 0.30917512046741963, "learning_rate": 6.305147896571457e-05, "loss": 0.9644, "step": 4128 }, { "epoch": 1.870865428183054, "grad_norm": 0.33590141533319945, "learning_rate": 6.304113662448922e-05, "loss": 0.9611, "step": 4129 }, { "epoch": 1.8713185319438153, "grad_norm": 0.4274015790306838, "learning_rate": 6.303079197748487e-05, "loss": 0.9636, "step": 4130 }, { "epoch": 1.8717716357045764, "grad_norm": 0.49162756214776293, "learning_rate": 6.302044502573674e-05, "loss": 0.944, "step": 4131 }, { "epoch": 1.8722247394653375, "grad_norm": 0.5042936853001874, "learning_rate": 6.301009577028028e-05, "loss": 0.955, "step": 4132 }, { "epoch": 1.8726778432260986, "grad_norm": 0.45674644771595646, "learning_rate": 6.299974421215116e-05, "loss": 0.9539, "step": 4133 }, { "epoch": 1.87313094698686, "grad_norm": 0.3674894392499377, "learning_rate": 6.298939035238529e-05, "loss": 0.9736, "step": 4134 }, { "epoch": 1.8735840507476214, "grad_norm": 0.33104428398516494, "learning_rate": 6.297903419201881e-05, "loss": 0.9653, "step": 4135 }, { "epoch": 1.8740371545083825, "grad_norm": 0.2790266751981636, "learning_rate": 6.296867573208807e-05, "loss": 0.9817, "step": 4136 }, { "epoch": 1.8744902582691436, "grad_norm": 0.24096424336271285, "learning_rate": 6.295831497362965e-05, "loss": 0.956, "step": 4137 }, { "epoch": 1.8749433620299047, "grad_norm": 0.3022403397306582, "learning_rate": 6.29479519176804e-05, "loss": 0.9723, "step": 4138 }, { "epoch": 1.875396465790666, "grad_norm": 0.350414916295371, "learning_rate": 6.293758656527739e-05, "loss": 0.9538, "step": 4139 }, { "epoch": 1.8758495695514272, "grad_norm": 0.4018260989610819, "learning_rate": 6.292721891745785e-05, "loss": 0.979, "step": 4140 }, { "epoch": 1.8763026733121886, "grad_norm": 0.43204694655818604, "learning_rate": 6.291684897525935e-05, "loss": 0.9766, "step": 4141 }, { "epoch": 1.8767557770729497, "grad_norm": 0.44893979910967996, "learning_rate": 6.290647673971961e-05, "loss": 0.9641, "step": 4142 }, { "epoch": 1.8772088808337108, "grad_norm": 0.42504980770279105, "learning_rate": 6.289610221187659e-05, "loss": 0.9592, "step": 4143 }, { "epoch": 1.877661984594472, "grad_norm": 0.379052327510647, "learning_rate": 6.288572539276854e-05, "loss": 0.9879, "step": 4144 }, { "epoch": 1.8781150883552333, "grad_norm": 0.31908579120159425, "learning_rate": 6.287534628343382e-05, "loss": 0.9694, "step": 4145 }, { "epoch": 1.8785681921159947, "grad_norm": 0.26688174299001544, "learning_rate": 6.286496488491115e-05, "loss": 0.969, "step": 4146 }, { "epoch": 1.8790212958767558, "grad_norm": 0.25046883792912555, "learning_rate": 6.285458119823939e-05, "loss": 0.96, "step": 4147 }, { "epoch": 1.879474399637517, "grad_norm": 0.24538824480625443, "learning_rate": 6.284419522445766e-05, "loss": 0.9756, "step": 4148 }, { "epoch": 1.879927503398278, "grad_norm": 0.3597344076736837, "learning_rate": 6.283380696460534e-05, "loss": 0.9548, "step": 4149 }, { "epoch": 1.8803806071590394, "grad_norm": 0.43819805345963375, "learning_rate": 6.282341641972198e-05, "loss": 0.94, "step": 4150 }, { "epoch": 1.8808337109198008, "grad_norm": 0.34460178909174793, "learning_rate": 6.281302359084737e-05, "loss": 0.9627, "step": 4151 }, { "epoch": 1.881286814680562, "grad_norm": 0.28833993631699606, "learning_rate": 6.280262847902158e-05, "loss": 0.9577, "step": 4152 }, { "epoch": 1.881739918441323, "grad_norm": 0.30286776530447196, "learning_rate": 6.279223108528485e-05, "loss": 0.9572, "step": 4153 }, { "epoch": 1.8821930222020842, "grad_norm": 0.2549192136302779, "learning_rate": 6.278183141067768e-05, "loss": 0.9719, "step": 4154 }, { "epoch": 1.8826461259628455, "grad_norm": 0.28415952412755563, "learning_rate": 6.277142945624079e-05, "loss": 0.9602, "step": 4155 }, { "epoch": 1.8830992297236067, "grad_norm": 0.2922247937493228, "learning_rate": 6.276102522301511e-05, "loss": 0.9432, "step": 4156 }, { "epoch": 1.883552333484368, "grad_norm": 0.3132373736881197, "learning_rate": 6.275061871204183e-05, "loss": 0.9719, "step": 4157 }, { "epoch": 1.8840054372451291, "grad_norm": 0.31296978964788497, "learning_rate": 6.274020992436236e-05, "loss": 0.9889, "step": 4158 }, { "epoch": 1.8844585410058903, "grad_norm": 0.25720666090709937, "learning_rate": 6.272979886101831e-05, "loss": 0.9447, "step": 4159 }, { "epoch": 1.8849116447666514, "grad_norm": 0.21254393397383556, "learning_rate": 6.271938552305157e-05, "loss": 0.9681, "step": 4160 }, { "epoch": 1.8853647485274128, "grad_norm": 0.22625313009179981, "learning_rate": 6.270896991150419e-05, "loss": 0.9572, "step": 4161 }, { "epoch": 1.8858178522881741, "grad_norm": 0.23055047444591034, "learning_rate": 6.269855202741852e-05, "loss": 0.9643, "step": 4162 }, { "epoch": 1.8862709560489352, "grad_norm": 0.27567448103541653, "learning_rate": 6.268813187183707e-05, "loss": 0.9573, "step": 4163 }, { "epoch": 1.8867240598096964, "grad_norm": 0.3049934543413265, "learning_rate": 6.267770944580262e-05, "loss": 0.9463, "step": 4164 }, { "epoch": 1.8871771635704575, "grad_norm": 0.36556629284601144, "learning_rate": 6.266728475035818e-05, "loss": 0.9635, "step": 4165 }, { "epoch": 1.8876302673312189, "grad_norm": 0.3822426870861703, "learning_rate": 6.265685778654694e-05, "loss": 0.9634, "step": 4166 }, { "epoch": 1.8880833710919802, "grad_norm": 0.33827681661546866, "learning_rate": 6.264642855541239e-05, "loss": 0.9663, "step": 4167 }, { "epoch": 1.8885364748527413, "grad_norm": 0.25350915455004236, "learning_rate": 6.263599705799819e-05, "loss": 0.9482, "step": 4168 }, { "epoch": 1.8889895786135025, "grad_norm": 0.2239635960095019, "learning_rate": 6.262556329534822e-05, "loss": 0.9717, "step": 4169 }, { "epoch": 1.8894426823742636, "grad_norm": 0.27883848412218837, "learning_rate": 6.261512726850666e-05, "loss": 0.9553, "step": 4170 }, { "epoch": 1.889895786135025, "grad_norm": 0.3172943948909215, "learning_rate": 6.26046889785178e-05, "loss": 0.9628, "step": 4171 }, { "epoch": 1.890348889895786, "grad_norm": 0.3617943865208761, "learning_rate": 6.25942484264263e-05, "loss": 0.9691, "step": 4172 }, { "epoch": 1.8908019936565474, "grad_norm": 0.40160324469546843, "learning_rate": 6.258380561327691e-05, "loss": 0.9434, "step": 4173 }, { "epoch": 1.8912550974173086, "grad_norm": 0.41169680176494305, "learning_rate": 6.25733605401147e-05, "loss": 0.9462, "step": 4174 }, { "epoch": 1.8917082011780697, "grad_norm": 0.4360698250102241, "learning_rate": 6.256291320798491e-05, "loss": 0.9497, "step": 4175 }, { "epoch": 1.8921613049388308, "grad_norm": 0.47263943907886297, "learning_rate": 6.255246361793305e-05, "loss": 0.9654, "step": 4176 }, { "epoch": 1.8926144086995922, "grad_norm": 0.492888585594115, "learning_rate": 6.254201177100482e-05, "loss": 0.9592, "step": 4177 }, { "epoch": 1.8930675124603535, "grad_norm": 0.4727263111322721, "learning_rate": 6.253155766824616e-05, "loss": 0.9713, "step": 4178 }, { "epoch": 1.8935206162211147, "grad_norm": 0.4297063500421653, "learning_rate": 6.252110131070323e-05, "loss": 0.9655, "step": 4179 }, { "epoch": 1.8939737199818758, "grad_norm": 0.3903025932582141, "learning_rate": 6.251064269942245e-05, "loss": 0.9653, "step": 4180 }, { "epoch": 1.894426823742637, "grad_norm": 0.35976644216222425, "learning_rate": 6.25001818354504e-05, "loss": 0.9663, "step": 4181 }, { "epoch": 1.8948799275033983, "grad_norm": 0.37963169303468897, "learning_rate": 6.248971871983395e-05, "loss": 0.9897, "step": 4182 }, { "epoch": 1.8953330312641596, "grad_norm": 0.3995463921366461, "learning_rate": 6.247925335362014e-05, "loss": 0.9693, "step": 4183 }, { "epoch": 1.8957861350249208, "grad_norm": 0.34191182149535376, "learning_rate": 6.24687857378563e-05, "loss": 0.9672, "step": 4184 }, { "epoch": 1.896239238785682, "grad_norm": 0.28769269226739386, "learning_rate": 6.245831587358992e-05, "loss": 0.9692, "step": 4185 }, { "epoch": 1.896692342546443, "grad_norm": 0.3461952574411583, "learning_rate": 6.244784376186876e-05, "loss": 0.9742, "step": 4186 }, { "epoch": 1.8971454463072044, "grad_norm": 0.3786063391714815, "learning_rate": 6.243736940374078e-05, "loss": 0.956, "step": 4187 }, { "epoch": 1.8975985500679655, "grad_norm": 0.397404971969139, "learning_rate": 6.242689280025415e-05, "loss": 0.9777, "step": 4188 }, { "epoch": 1.8980516538287269, "grad_norm": 0.41546087601276194, "learning_rate": 6.241641395245733e-05, "loss": 0.9416, "step": 4189 }, { "epoch": 1.898504757589488, "grad_norm": 0.3903823255755893, "learning_rate": 6.240593286139895e-05, "loss": 0.9503, "step": 4190 }, { "epoch": 1.8989578613502491, "grad_norm": 0.3560045133042323, "learning_rate": 6.239544952812786e-05, "loss": 0.9627, "step": 4191 }, { "epoch": 1.8994109651110103, "grad_norm": 0.3131428241198897, "learning_rate": 6.238496395369316e-05, "loss": 0.9626, "step": 4192 }, { "epoch": 1.8998640688717716, "grad_norm": 0.3183476226476776, "learning_rate": 6.237447613914415e-05, "loss": 0.9719, "step": 4193 }, { "epoch": 1.900317172632533, "grad_norm": 0.2764331635854591, "learning_rate": 6.236398608553041e-05, "loss": 0.9373, "step": 4194 }, { "epoch": 1.900770276393294, "grad_norm": 0.3016488182763755, "learning_rate": 6.235349379390166e-05, "loss": 0.9499, "step": 4195 }, { "epoch": 1.9012233801540552, "grad_norm": 0.3394918786493137, "learning_rate": 6.234299926530792e-05, "loss": 0.9565, "step": 4196 }, { "epoch": 1.9016764839148164, "grad_norm": 0.2932553318573757, "learning_rate": 6.233250250079938e-05, "loss": 0.9595, "step": 4197 }, { "epoch": 1.9021295876755777, "grad_norm": 0.27765309118394127, "learning_rate": 6.232200350142649e-05, "loss": 0.9562, "step": 4198 }, { "epoch": 1.902582691436339, "grad_norm": 0.28592329245131604, "learning_rate": 6.231150226823989e-05, "loss": 0.9309, "step": 4199 }, { "epoch": 1.9030357951971002, "grad_norm": 0.32265755089303005, "learning_rate": 6.230099880229046e-05, "loss": 0.9457, "step": 4200 }, { "epoch": 1.9034888989578613, "grad_norm": 0.26733961814297225, "learning_rate": 6.229049310462936e-05, "loss": 0.9751, "step": 4201 }, { "epoch": 1.9039420027186225, "grad_norm": 0.2839062131489708, "learning_rate": 6.227998517630785e-05, "loss": 0.9474, "step": 4202 }, { "epoch": 1.9043951064793838, "grad_norm": 0.29450558294415285, "learning_rate": 6.226947501837752e-05, "loss": 0.9308, "step": 4203 }, { "epoch": 1.904848210240145, "grad_norm": 0.3349032843779522, "learning_rate": 6.225896263189013e-05, "loss": 0.9529, "step": 4204 }, { "epoch": 1.9053013140009063, "grad_norm": 0.38791020904506174, "learning_rate": 6.224844801789769e-05, "loss": 0.9809, "step": 4205 }, { "epoch": 1.9057544177616674, "grad_norm": 0.4005470291237818, "learning_rate": 6.22379311774524e-05, "loss": 0.9512, "step": 4206 }, { "epoch": 1.9062075215224286, "grad_norm": 0.3749160562976337, "learning_rate": 6.222741211160675e-05, "loss": 0.9601, "step": 4207 }, { "epoch": 1.9066606252831897, "grad_norm": 0.3456291611524224, "learning_rate": 6.221689082141335e-05, "loss": 0.9677, "step": 4208 }, { "epoch": 1.907113729043951, "grad_norm": 0.3735211650552589, "learning_rate": 6.220636730792512e-05, "loss": 0.9697, "step": 4209 }, { "epoch": 1.9075668328047124, "grad_norm": 0.3534055352826561, "learning_rate": 6.219584157219517e-05, "loss": 0.9778, "step": 4210 }, { "epoch": 1.9080199365654735, "grad_norm": 0.3505175130662083, "learning_rate": 6.218531361527686e-05, "loss": 0.9544, "step": 4211 }, { "epoch": 1.9084730403262347, "grad_norm": 0.32187255056207903, "learning_rate": 6.217478343822368e-05, "loss": 0.9642, "step": 4212 }, { "epoch": 1.9089261440869958, "grad_norm": 0.28590593985349966, "learning_rate": 6.216425104208947e-05, "loss": 0.9744, "step": 4213 }, { "epoch": 1.9093792478477571, "grad_norm": 0.29246307922547804, "learning_rate": 6.21537164279282e-05, "loss": 0.9548, "step": 4214 }, { "epoch": 1.9098323516085185, "grad_norm": 0.2880260267308533, "learning_rate": 6.21431795967941e-05, "loss": 0.9704, "step": 4215 }, { "epoch": 1.9102854553692796, "grad_norm": 0.26513764629643155, "learning_rate": 6.213264054974163e-05, "loss": 0.9686, "step": 4216 }, { "epoch": 1.9107385591300408, "grad_norm": 0.8232841410324683, "learning_rate": 6.212209928782544e-05, "loss": 0.9705, "step": 4217 }, { "epoch": 1.911191662890802, "grad_norm": 0.2842007723262045, "learning_rate": 6.211155581210041e-05, "loss": 0.9693, "step": 4218 }, { "epoch": 1.9116447666515632, "grad_norm": 0.2947923825273287, "learning_rate": 6.210101012362166e-05, "loss": 0.9631, "step": 4219 }, { "epoch": 1.9120978704123244, "grad_norm": 0.27252628840090287, "learning_rate": 6.209046222344453e-05, "loss": 0.9599, "step": 4220 }, { "epoch": 1.9125509741730857, "grad_norm": 0.2973478831779515, "learning_rate": 6.207991211262456e-05, "loss": 0.9646, "step": 4221 }, { "epoch": 1.9130040779338469, "grad_norm": 0.2795680577224374, "learning_rate": 6.206935979221752e-05, "loss": 0.9503, "step": 4222 }, { "epoch": 1.913457181694608, "grad_norm": 0.23752599345199776, "learning_rate": 6.205880526327942e-05, "loss": 0.9505, "step": 4223 }, { "epoch": 1.9139102854553691, "grad_norm": 0.22606246085326326, "learning_rate": 6.204824852686647e-05, "loss": 0.9741, "step": 4224 }, { "epoch": 1.9143633892161305, "grad_norm": 0.2534446087482585, "learning_rate": 6.203768958403511e-05, "loss": 0.9471, "step": 4225 }, { "epoch": 1.9148164929768918, "grad_norm": 0.3131202727929122, "learning_rate": 6.202712843584198e-05, "loss": 0.9674, "step": 4226 }, { "epoch": 1.915269596737653, "grad_norm": 0.25870507440521423, "learning_rate": 6.201656508334398e-05, "loss": 0.9727, "step": 4227 }, { "epoch": 1.915722700498414, "grad_norm": 0.2599147185437996, "learning_rate": 6.200599952759821e-05, "loss": 0.9746, "step": 4228 }, { "epoch": 1.9161758042591752, "grad_norm": 0.2874650018013163, "learning_rate": 6.199543176966195e-05, "loss": 0.9692, "step": 4229 }, { "epoch": 1.9166289080199366, "grad_norm": 0.34964748488512803, "learning_rate": 6.198486181059279e-05, "loss": 0.9504, "step": 4230 }, { "epoch": 1.917082011780698, "grad_norm": 0.3771694047549155, "learning_rate": 6.197428965144846e-05, "loss": 0.9713, "step": 4231 }, { "epoch": 1.917535115541459, "grad_norm": 0.42178944377613276, "learning_rate": 6.196371529328694e-05, "loss": 0.9828, "step": 4232 }, { "epoch": 1.9179882193022202, "grad_norm": 0.40916708773498667, "learning_rate": 6.195313873716644e-05, "loss": 0.9557, "step": 4233 }, { "epoch": 1.9184413230629813, "grad_norm": 0.4147266070507474, "learning_rate": 6.19425599841454e-05, "loss": 0.956, "step": 4234 }, { "epoch": 1.9188944268237427, "grad_norm": 0.46253878058450487, "learning_rate": 6.193197903528241e-05, "loss": 0.9734, "step": 4235 }, { "epoch": 1.9193475305845038, "grad_norm": 0.5065541225435718, "learning_rate": 6.192139589163636e-05, "loss": 0.9631, "step": 4236 }, { "epoch": 1.9198006343452652, "grad_norm": 0.4888489036496714, "learning_rate": 6.191081055426633e-05, "loss": 0.9786, "step": 4237 }, { "epoch": 1.9202537381060263, "grad_norm": 0.443390682381111, "learning_rate": 6.190022302423161e-05, "loss": 0.964, "step": 4238 }, { "epoch": 1.9207068418667874, "grad_norm": 0.47007023408881005, "learning_rate": 6.188963330259174e-05, "loss": 0.9691, "step": 4239 }, { "epoch": 1.9211599456275485, "grad_norm": 0.4820244629566285, "learning_rate": 6.187904139040644e-05, "loss": 0.9644, "step": 4240 }, { "epoch": 1.92161304938831, "grad_norm": 0.4455813511331769, "learning_rate": 6.186844728873566e-05, "loss": 0.9613, "step": 4241 }, { "epoch": 1.9220661531490713, "grad_norm": 0.44065800209816197, "learning_rate": 6.185785099863959e-05, "loss": 0.9585, "step": 4242 }, { "epoch": 1.9225192569098324, "grad_norm": 0.4215733302380273, "learning_rate": 6.18472525211786e-05, "loss": 0.97, "step": 4243 }, { "epoch": 1.9229723606705935, "grad_norm": 0.353880253449188, "learning_rate": 6.183665185741334e-05, "loss": 0.9478, "step": 4244 }, { "epoch": 1.9234254644313546, "grad_norm": 0.30048328582941497, "learning_rate": 6.182604900840462e-05, "loss": 0.9729, "step": 4245 }, { "epoch": 1.923878568192116, "grad_norm": 0.2745621442992756, "learning_rate": 6.181544397521349e-05, "loss": 0.9331, "step": 4246 }, { "epoch": 1.9243316719528774, "grad_norm": 0.24378289199511533, "learning_rate": 6.180483675890124e-05, "loss": 0.95, "step": 4247 }, { "epoch": 1.9247847757136385, "grad_norm": 0.22152605980562837, "learning_rate": 6.179422736052934e-05, "loss": 0.9546, "step": 4248 }, { "epoch": 1.9252378794743996, "grad_norm": 0.25933332022315225, "learning_rate": 6.178361578115949e-05, "loss": 0.9811, "step": 4249 }, { "epoch": 1.9256909832351607, "grad_norm": 0.30855428254700484, "learning_rate": 6.177300202185362e-05, "loss": 0.9766, "step": 4250 }, { "epoch": 1.926144086995922, "grad_norm": 0.3014862479644237, "learning_rate": 6.176238608367391e-05, "loss": 0.9424, "step": 4251 }, { "epoch": 1.9265971907566832, "grad_norm": 0.2772713668332886, "learning_rate": 6.175176796768266e-05, "loss": 0.939, "step": 4252 }, { "epoch": 1.9270502945174446, "grad_norm": 0.25475752118456924, "learning_rate": 6.174114767494251e-05, "loss": 0.9489, "step": 4253 }, { "epoch": 1.9275033982782057, "grad_norm": 0.2641328572923247, "learning_rate": 6.17305252065162e-05, "loss": 0.934, "step": 4254 }, { "epoch": 1.9279565020389668, "grad_norm": 0.24500568410720927, "learning_rate": 6.171990056346679e-05, "loss": 0.9747, "step": 4255 }, { "epoch": 1.928409605799728, "grad_norm": 0.2112507549667263, "learning_rate": 6.170927374685747e-05, "loss": 0.9634, "step": 4256 }, { "epoch": 1.9288627095604893, "grad_norm": 0.24713666974391313, "learning_rate": 6.169864475775172e-05, "loss": 0.9509, "step": 4257 }, { "epoch": 1.9293158133212507, "grad_norm": 0.2810689462961619, "learning_rate": 6.168801359721321e-05, "loss": 0.9571, "step": 4258 }, { "epoch": 1.9297689170820118, "grad_norm": 0.2866481524430757, "learning_rate": 6.16773802663058e-05, "loss": 0.9571, "step": 4259 }, { "epoch": 1.930222020842773, "grad_norm": 0.2911608772989753, "learning_rate": 6.166674476609362e-05, "loss": 0.9555, "step": 4260 }, { "epoch": 1.930675124603534, "grad_norm": 0.33976872639371636, "learning_rate": 6.165610709764096e-05, "loss": 0.9573, "step": 4261 }, { "epoch": 1.9311282283642954, "grad_norm": 0.3635835630732887, "learning_rate": 6.164546726201238e-05, "loss": 0.9786, "step": 4262 }, { "epoch": 1.9315813321250568, "grad_norm": 0.34018723645366306, "learning_rate": 6.163482526027262e-05, "loss": 0.9811, "step": 4263 }, { "epoch": 1.932034435885818, "grad_norm": 0.33908496461832854, "learning_rate": 6.162418109348665e-05, "loss": 0.9531, "step": 4264 }, { "epoch": 1.932487539646579, "grad_norm": 0.3241378722156642, "learning_rate": 6.161353476271967e-05, "loss": 0.9622, "step": 4265 }, { "epoch": 1.9329406434073402, "grad_norm": 0.26178564952881056, "learning_rate": 6.160288626903705e-05, "loss": 0.9406, "step": 4266 }, { "epoch": 1.9333937471681015, "grad_norm": 0.2473330980716556, "learning_rate": 6.159223561350444e-05, "loss": 0.9537, "step": 4267 }, { "epoch": 1.9338468509288627, "grad_norm": 0.30789668657357355, "learning_rate": 6.158158279718766e-05, "loss": 0.9593, "step": 4268 }, { "epoch": 1.934299954689624, "grad_norm": 0.3046624034543418, "learning_rate": 6.157092782115276e-05, "loss": 0.964, "step": 4269 }, { "epoch": 1.9347530584503851, "grad_norm": 0.2693591378376364, "learning_rate": 6.156027068646604e-05, "loss": 0.9488, "step": 4270 }, { "epoch": 1.9352061622111463, "grad_norm": 0.2278363532252313, "learning_rate": 6.154961139419394e-05, "loss": 0.9426, "step": 4271 }, { "epoch": 1.9356592659719074, "grad_norm": 0.3094330316130483, "learning_rate": 6.153894994540319e-05, "loss": 0.9568, "step": 4272 }, { "epoch": 1.9361123697326688, "grad_norm": 0.3547059781273133, "learning_rate": 6.152828634116067e-05, "loss": 0.9739, "step": 4273 }, { "epoch": 1.9365654734934301, "grad_norm": 0.33594226745058847, "learning_rate": 6.151762058253355e-05, "loss": 0.9633, "step": 4274 }, { "epoch": 1.9370185772541912, "grad_norm": 0.29668144873478736, "learning_rate": 6.150695267058916e-05, "loss": 0.9662, "step": 4275 }, { "epoch": 1.9374716810149524, "grad_norm": 0.3284502413632245, "learning_rate": 6.149628260639505e-05, "loss": 0.9656, "step": 4276 }, { "epoch": 1.9379247847757135, "grad_norm": 0.42652402667687855, "learning_rate": 6.148561039101903e-05, "loss": 0.9731, "step": 4277 }, { "epoch": 1.9383778885364749, "grad_norm": 0.48196631572061127, "learning_rate": 6.147493602552906e-05, "loss": 0.9849, "step": 4278 }, { "epoch": 1.9388309922972362, "grad_norm": 0.5292125904653294, "learning_rate": 6.146425951099338e-05, "loss": 0.9759, "step": 4279 }, { "epoch": 1.9392840960579973, "grad_norm": 0.547691119880986, "learning_rate": 6.145358084848038e-05, "loss": 0.9551, "step": 4280 }, { "epoch": 1.9397371998187585, "grad_norm": 0.5488511362879088, "learning_rate": 6.14429000390587e-05, "loss": 0.937, "step": 4281 }, { "epoch": 1.9401903035795196, "grad_norm": 0.5344880775481058, "learning_rate": 6.143221708379722e-05, "loss": 0.9477, "step": 4282 }, { "epoch": 1.940643407340281, "grad_norm": 0.545003498058616, "learning_rate": 6.1421531983765e-05, "loss": 0.9558, "step": 4283 }, { "epoch": 1.941096511101042, "grad_norm": 0.5311364850371998, "learning_rate": 6.14108447400313e-05, "loss": 0.9431, "step": 4284 }, { "epoch": 1.9415496148618034, "grad_norm": 0.48011177802610705, "learning_rate": 6.140015535366564e-05, "loss": 0.9571, "step": 4285 }, { "epoch": 1.9420027186225646, "grad_norm": 0.4165576563728937, "learning_rate": 6.138946382573772e-05, "loss": 0.982, "step": 4286 }, { "epoch": 1.9424558223833257, "grad_norm": 0.36021275630510685, "learning_rate": 6.137877015731749e-05, "loss": 0.963, "step": 4287 }, { "epoch": 1.9429089261440868, "grad_norm": 0.26938336442454236, "learning_rate": 6.136807434947505e-05, "loss": 0.9589, "step": 4288 }, { "epoch": 1.9433620299048482, "grad_norm": 0.1979736165641151, "learning_rate": 6.135737640328078e-05, "loss": 0.9653, "step": 4289 }, { "epoch": 1.9438151336656095, "grad_norm": 0.23416227994342126, "learning_rate": 6.134667631980523e-05, "loss": 0.9498, "step": 4290 }, { "epoch": 1.9442682374263707, "grad_norm": 0.281588787723853, "learning_rate": 6.13359741001192e-05, "loss": 0.961, "step": 4291 }, { "epoch": 1.9447213411871318, "grad_norm": 0.28705336496277567, "learning_rate": 6.132526974529367e-05, "loss": 0.9731, "step": 4292 }, { "epoch": 1.945174444947893, "grad_norm": 0.25299232421792994, "learning_rate": 6.131456325639987e-05, "loss": 0.9447, "step": 4293 }, { "epoch": 1.9456275487086543, "grad_norm": 0.22133948151261768, "learning_rate": 6.130385463450919e-05, "loss": 0.9567, "step": 4294 }, { "epoch": 1.9460806524694156, "grad_norm": 0.25030610413216775, "learning_rate": 6.129314388069331e-05, "loss": 0.9607, "step": 4295 }, { "epoch": 1.9465337562301768, "grad_norm": 0.26055743741805415, "learning_rate": 6.128243099602404e-05, "loss": 0.9563, "step": 4296 }, { "epoch": 1.946986859990938, "grad_norm": 0.24879829814261045, "learning_rate": 6.127171598157347e-05, "loss": 0.9698, "step": 4297 }, { "epoch": 1.947439963751699, "grad_norm": 0.2543501205554824, "learning_rate": 6.126099883841384e-05, "loss": 0.9618, "step": 4298 }, { "epoch": 1.9478930675124604, "grad_norm": 0.2539502562742908, "learning_rate": 6.125027956761768e-05, "loss": 0.975, "step": 4299 }, { "epoch": 1.9483461712732215, "grad_norm": 0.27843684741349795, "learning_rate": 6.123955817025768e-05, "loss": 0.9635, "step": 4300 }, { "epoch": 1.9487992750339829, "grad_norm": 0.34872362526817724, "learning_rate": 6.122883464740673e-05, "loss": 0.9484, "step": 4301 }, { "epoch": 1.949252378794744, "grad_norm": 0.35375107512629833, "learning_rate": 6.1218109000138e-05, "loss": 0.975, "step": 4302 }, { "epoch": 1.9497054825555051, "grad_norm": 0.35128545354857493, "learning_rate": 6.12073812295248e-05, "loss": 0.9579, "step": 4303 }, { "epoch": 1.9501585863162663, "grad_norm": 0.3170492404538022, "learning_rate": 6.119665133664068e-05, "loss": 0.9694, "step": 4304 }, { "epoch": 1.9506116900770276, "grad_norm": 0.25216494073222645, "learning_rate": 6.118591932255942e-05, "loss": 0.9482, "step": 4305 }, { "epoch": 1.951064793837789, "grad_norm": 0.22596734917460393, "learning_rate": 6.117518518835499e-05, "loss": 0.9611, "step": 4306 }, { "epoch": 1.95151789759855, "grad_norm": 0.2254702987043959, "learning_rate": 6.116444893510159e-05, "loss": 0.949, "step": 4307 }, { "epoch": 1.9519710013593112, "grad_norm": 0.19596582533301143, "learning_rate": 6.115371056387358e-05, "loss": 0.9506, "step": 4308 }, { "epoch": 1.9524241051200724, "grad_norm": 0.17832510605286506, "learning_rate": 6.114297007574564e-05, "loss": 0.9401, "step": 4309 }, { "epoch": 1.9528772088808337, "grad_norm": 0.1870104210273291, "learning_rate": 6.113222747179254e-05, "loss": 0.9436, "step": 4310 }, { "epoch": 1.953330312641595, "grad_norm": 0.22921711172582857, "learning_rate": 6.112148275308933e-05, "loss": 0.9499, "step": 4311 }, { "epoch": 1.9537834164023562, "grad_norm": 0.2853665193289077, "learning_rate": 6.111073592071128e-05, "loss": 0.9618, "step": 4312 }, { "epoch": 1.9542365201631173, "grad_norm": 0.34089461963679246, "learning_rate": 6.109998697573384e-05, "loss": 0.9593, "step": 4313 }, { "epoch": 1.9546896239238785, "grad_norm": 0.37778975555286687, "learning_rate": 6.108923591923267e-05, "loss": 0.9451, "step": 4314 }, { "epoch": 1.9551427276846398, "grad_norm": 0.35590810320957933, "learning_rate": 6.107848275228365e-05, "loss": 0.9792, "step": 4315 }, { "epoch": 1.955595831445401, "grad_norm": 0.3288239791170118, "learning_rate": 6.106772747596291e-05, "loss": 0.9477, "step": 4316 }, { "epoch": 1.9560489352061623, "grad_norm": 0.34725573628393575, "learning_rate": 6.10569700913467e-05, "loss": 0.9624, "step": 4317 }, { "epoch": 1.9565020389669234, "grad_norm": 0.3625475919951475, "learning_rate": 6.104621059951157e-05, "loss": 0.9558, "step": 4318 }, { "epoch": 1.9569551427276846, "grad_norm": 0.3399226924343531, "learning_rate": 6.103544900153425e-05, "loss": 0.9508, "step": 4319 }, { "epoch": 1.9574082464884457, "grad_norm": 0.30572486469554977, "learning_rate": 6.102468529849166e-05, "loss": 0.9732, "step": 4320 }, { "epoch": 1.957861350249207, "grad_norm": 0.2556808064756185, "learning_rate": 6.101391949146095e-05, "loss": 0.9547, "step": 4321 }, { "epoch": 1.9583144540099684, "grad_norm": 0.2112417945477102, "learning_rate": 6.100315158151949e-05, "loss": 0.9775, "step": 4322 }, { "epoch": 1.9587675577707295, "grad_norm": 0.20990544941579203, "learning_rate": 6.0992381569744864e-05, "loss": 0.9476, "step": 4323 }, { "epoch": 1.9592206615314907, "grad_norm": 0.2259521394634219, "learning_rate": 6.098160945721481e-05, "loss": 0.9728, "step": 4324 }, { "epoch": 1.9596737652922518, "grad_norm": 0.30634407679413134, "learning_rate": 6.097083524500735e-05, "loss": 0.9614, "step": 4325 }, { "epoch": 1.9601268690530131, "grad_norm": 0.37677077677916404, "learning_rate": 6.096005893420067e-05, "loss": 0.955, "step": 4326 }, { "epoch": 1.9605799728137745, "grad_norm": 0.3757934176830056, "learning_rate": 6.094928052587318e-05, "loss": 0.9574, "step": 4327 }, { "epoch": 1.9610330765745356, "grad_norm": 0.3702925713060765, "learning_rate": 6.0938500021103514e-05, "loss": 0.9347, "step": 4328 }, { "epoch": 1.9614861803352968, "grad_norm": 0.32936227102691573, "learning_rate": 6.0927717420970485e-05, "loss": 0.9575, "step": 4329 }, { "epoch": 1.961939284096058, "grad_norm": 0.2704498236420659, "learning_rate": 6.0916932726553136e-05, "loss": 0.9552, "step": 4330 }, { "epoch": 1.9623923878568192, "grad_norm": 0.24847010983492362, "learning_rate": 6.090614593893072e-05, "loss": 0.9642, "step": 4331 }, { "epoch": 1.9628454916175804, "grad_norm": 0.27369891869637253, "learning_rate": 6.0895357059182716e-05, "loss": 0.9515, "step": 4332 }, { "epoch": 1.9632985953783417, "grad_norm": 0.320996932842233, "learning_rate": 6.088456608838874e-05, "loss": 0.9725, "step": 4333 }, { "epoch": 1.9637516991391029, "grad_norm": 0.32152378818637306, "learning_rate": 6.0873773027628716e-05, "loss": 0.9514, "step": 4334 }, { "epoch": 1.964204802899864, "grad_norm": 0.3023936653919213, "learning_rate": 6.0862977877982705e-05, "loss": 0.9502, "step": 4335 }, { "epoch": 1.9646579066606251, "grad_norm": 0.23709142544193246, "learning_rate": 6.0852180640531016e-05, "loss": 0.9631, "step": 4336 }, { "epoch": 1.9651110104213865, "grad_norm": 0.29974657294943224, "learning_rate": 6.084138131635414e-05, "loss": 0.9438, "step": 4337 }, { "epoch": 1.9655641141821478, "grad_norm": 0.29397759449892896, "learning_rate": 6.08305799065328e-05, "loss": 0.9668, "step": 4338 }, { "epoch": 1.966017217942909, "grad_norm": 0.2586742781432724, "learning_rate": 6.081977641214792e-05, "loss": 0.9273, "step": 4339 }, { "epoch": 1.96647032170367, "grad_norm": 0.25582440470819595, "learning_rate": 6.080897083428063e-05, "loss": 0.9588, "step": 4340 }, { "epoch": 1.9669234254644312, "grad_norm": 0.25475656981497335, "learning_rate": 6.079816317401225e-05, "loss": 0.9661, "step": 4341 }, { "epoch": 1.9673765292251926, "grad_norm": 0.22892705027719473, "learning_rate": 6.078735343242435e-05, "loss": 0.9716, "step": 4342 }, { "epoch": 1.967829632985954, "grad_norm": 0.18976753426052523, "learning_rate": 6.077654161059868e-05, "loss": 0.9674, "step": 4343 }, { "epoch": 1.968282736746715, "grad_norm": 0.24924081745104568, "learning_rate": 6.0765727709617194e-05, "loss": 0.9448, "step": 4344 }, { "epoch": 1.9687358405074762, "grad_norm": 0.27826882163568584, "learning_rate": 6.0754911730562075e-05, "loss": 0.9602, "step": 4345 }, { "epoch": 1.9691889442682373, "grad_norm": 0.3114653477055089, "learning_rate": 6.0744093674515695e-05, "loss": 0.961, "step": 4346 }, { "epoch": 1.9696420480289987, "grad_norm": 0.3097803981628157, "learning_rate": 6.073327354256064e-05, "loss": 0.9561, "step": 4347 }, { "epoch": 1.9700951517897598, "grad_norm": 0.37487530728837654, "learning_rate": 6.072245133577971e-05, "loss": 0.9533, "step": 4348 }, { "epoch": 1.9705482555505212, "grad_norm": 0.4507074851134849, "learning_rate": 6.07116270552559e-05, "loss": 0.9462, "step": 4349 }, { "epoch": 1.9710013593112823, "grad_norm": 0.47905192310180894, "learning_rate": 6.070080070207245e-05, "loss": 0.9459, "step": 4350 }, { "epoch": 1.9714544630720434, "grad_norm": 0.48269486730977806, "learning_rate": 6.068997227731274e-05, "loss": 0.9253, "step": 4351 }, { "epoch": 1.9719075668328045, "grad_norm": 0.42613622725936207, "learning_rate": 6.067914178206041e-05, "loss": 0.9622, "step": 4352 }, { "epoch": 1.972360670593566, "grad_norm": 0.34408148248369175, "learning_rate": 6.0668309217399294e-05, "loss": 0.9412, "step": 4353 }, { "epoch": 1.9728137743543273, "grad_norm": 0.3101371133784038, "learning_rate": 6.0657474584413433e-05, "loss": 0.9496, "step": 4354 }, { "epoch": 1.9732668781150884, "grad_norm": 0.2235945367870132, "learning_rate": 6.064663788418706e-05, "loss": 0.9523, "step": 4355 }, { "epoch": 1.9737199818758495, "grad_norm": 0.22906060040134943, "learning_rate": 6.0635799117804655e-05, "loss": 0.9415, "step": 4356 }, { "epoch": 1.9741730856366106, "grad_norm": 0.24396996789281736, "learning_rate": 6.062495828635086e-05, "loss": 0.9534, "step": 4357 }, { "epoch": 1.974626189397372, "grad_norm": 0.29363110730584685, "learning_rate": 6.0614115390910544e-05, "loss": 0.9782, "step": 4358 }, { "epoch": 1.9750792931581334, "grad_norm": 0.384284084704633, "learning_rate": 6.060327043256877e-05, "loss": 0.9719, "step": 4359 }, { "epoch": 1.9755323969188945, "grad_norm": 0.37382178915005865, "learning_rate": 6.059242341241083e-05, "loss": 0.9433, "step": 4360 }, { "epoch": 1.9759855006796556, "grad_norm": 0.3967229145594574, "learning_rate": 6.0581574331522216e-05, "loss": 0.9612, "step": 4361 }, { "epoch": 1.9764386044404167, "grad_norm": 0.41528869757356635, "learning_rate": 6.0570723190988605e-05, "loss": 0.9597, "step": 4362 }, { "epoch": 1.976891708201178, "grad_norm": 0.42563364668884157, "learning_rate": 6.05598699918959e-05, "loss": 0.9619, "step": 4363 }, { "epoch": 1.9773448119619392, "grad_norm": 0.4246115983606182, "learning_rate": 6.0549014735330204e-05, "loss": 0.9541, "step": 4364 }, { "epoch": 1.9777979157227006, "grad_norm": 0.42388042352168664, "learning_rate": 6.053815742237785e-05, "loss": 0.9659, "step": 4365 }, { "epoch": 1.9782510194834617, "grad_norm": 0.4600936904122245, "learning_rate": 6.052729805412532e-05, "loss": 0.9645, "step": 4366 }, { "epoch": 1.9787041232442228, "grad_norm": 0.47094808211133277, "learning_rate": 6.051643663165936e-05, "loss": 0.9592, "step": 4367 }, { "epoch": 1.979157227004984, "grad_norm": 0.40981128164349306, "learning_rate": 6.0505573156066876e-05, "loss": 0.9544, "step": 4368 }, { "epoch": 1.9796103307657453, "grad_norm": 0.30314257742886624, "learning_rate": 6.0494707628435e-05, "loss": 0.9504, "step": 4369 }, { "epoch": 1.9800634345265067, "grad_norm": 0.3032929681801511, "learning_rate": 6.04838400498511e-05, "loss": 0.9447, "step": 4370 }, { "epoch": 1.9805165382872678, "grad_norm": 0.3631724279273831, "learning_rate": 6.0472970421402706e-05, "loss": 0.9616, "step": 4371 }, { "epoch": 1.980969642048029, "grad_norm": 0.36510874236880886, "learning_rate": 6.046209874417756e-05, "loss": 0.9586, "step": 4372 }, { "epoch": 1.98142274580879, "grad_norm": 1.0035653194486145, "learning_rate": 6.045122501926361e-05, "loss": 0.9682, "step": 4373 }, { "epoch": 1.9818758495695514, "grad_norm": 0.41343788193053815, "learning_rate": 6.044034924774903e-05, "loss": 0.959, "step": 4374 }, { "epoch": 1.9823289533303128, "grad_norm": 0.3431404099088022, "learning_rate": 6.042947143072218e-05, "loss": 0.9558, "step": 4375 }, { "epoch": 1.982782057091074, "grad_norm": 0.3001353008739645, "learning_rate": 6.041859156927161e-05, "loss": 0.9532, "step": 4376 }, { "epoch": 1.983235160851835, "grad_norm": 0.35639126439356317, "learning_rate": 6.0407709664486104e-05, "loss": 0.9556, "step": 4377 }, { "epoch": 1.9836882646125962, "grad_norm": 0.31789935952580656, "learning_rate": 6.039682571745466e-05, "loss": 0.9386, "step": 4378 }, { "epoch": 1.9841413683733575, "grad_norm": 0.22866872144385672, "learning_rate": 6.038593972926641e-05, "loss": 0.9631, "step": 4379 }, { "epoch": 1.9845944721341187, "grad_norm": 0.3101495400180509, "learning_rate": 6.03750517010108e-05, "loss": 0.9549, "step": 4380 }, { "epoch": 1.98504757589488, "grad_norm": 0.41937030749559695, "learning_rate": 6.0364161633777374e-05, "loss": 0.9717, "step": 4381 }, { "epoch": 1.9855006796556411, "grad_norm": 0.4216651483332542, "learning_rate": 6.035326952865594e-05, "loss": 0.9746, "step": 4382 }, { "epoch": 1.9859537834164023, "grad_norm": 0.4013160726043846, "learning_rate": 6.0342375386736506e-05, "loss": 0.9499, "step": 4383 }, { "epoch": 1.9864068871771634, "grad_norm": 0.45953577513410604, "learning_rate": 6.0331479209109266e-05, "loss": 0.9448, "step": 4384 }, { "epoch": 1.9868599909379248, "grad_norm": 0.47774935425117054, "learning_rate": 6.032058099686462e-05, "loss": 0.9482, "step": 4385 }, { "epoch": 1.9873130946986861, "grad_norm": 0.43251144333484104, "learning_rate": 6.030968075109317e-05, "loss": 0.9575, "step": 4386 }, { "epoch": 1.9877661984594472, "grad_norm": 0.3792471810511788, "learning_rate": 6.029877847288575e-05, "loss": 0.9475, "step": 4387 }, { "epoch": 1.9882193022202084, "grad_norm": 0.29989941071258586, "learning_rate": 6.0287874163333355e-05, "loss": 0.9514, "step": 4388 }, { "epoch": 1.9886724059809695, "grad_norm": 0.6065102698498833, "learning_rate": 6.027696782352722e-05, "loss": 0.9855, "step": 4389 }, { "epoch": 1.9891255097417309, "grad_norm": 0.2352201705473711, "learning_rate": 6.026605945455876e-05, "loss": 0.983, "step": 4390 }, { "epoch": 1.9895786135024922, "grad_norm": 0.2700003904272411, "learning_rate": 6.02551490575196e-05, "loss": 0.9549, "step": 4391 }, { "epoch": 1.9900317172632533, "grad_norm": 0.3155246494855787, "learning_rate": 6.024423663350158e-05, "loss": 0.9571, "step": 4392 }, { "epoch": 1.9904848210240145, "grad_norm": 0.3343435557614182, "learning_rate": 6.023332218359672e-05, "loss": 0.9539, "step": 4393 }, { "epoch": 1.9909379247847756, "grad_norm": 0.33647318606160803, "learning_rate": 6.022240570889725e-05, "loss": 0.9683, "step": 4394 }, { "epoch": 1.991391028545537, "grad_norm": 0.32738102100654864, "learning_rate": 6.021148721049561e-05, "loss": 0.9562, "step": 4395 }, { "epoch": 1.991844132306298, "grad_norm": 0.35215049291590417, "learning_rate": 6.020056668948444e-05, "loss": 0.9775, "step": 4396 }, { "epoch": 1.9922972360670594, "grad_norm": 0.3491811050436323, "learning_rate": 6.018964414695658e-05, "loss": 0.9521, "step": 4397 }, { "epoch": 1.9927503398278206, "grad_norm": 0.29033155424686535, "learning_rate": 6.0178719584005086e-05, "loss": 0.965, "step": 4398 }, { "epoch": 1.9932034435885817, "grad_norm": 0.19820371282753219, "learning_rate": 6.016779300172318e-05, "loss": 0.9654, "step": 4399 }, { "epoch": 1.9936565473493428, "grad_norm": 0.2058552662511048, "learning_rate": 6.015686440120433e-05, "loss": 0.9582, "step": 4400 }, { "epoch": 1.9941096511101042, "grad_norm": 0.2555121183570235, "learning_rate": 6.014593378354219e-05, "loss": 0.9663, "step": 4401 }, { "epoch": 1.9945627548708655, "grad_norm": 0.3079197299630231, "learning_rate": 6.013500114983059e-05, "loss": 0.9354, "step": 4402 }, { "epoch": 1.9950158586316267, "grad_norm": 0.3417214580174099, "learning_rate": 6.01240665011636e-05, "loss": 0.9319, "step": 4403 }, { "epoch": 1.9954689623923878, "grad_norm": 0.3433392646128172, "learning_rate": 6.011312983863546e-05, "loss": 0.9431, "step": 4404 }, { "epoch": 1.995922066153149, "grad_norm": 0.3167223728535196, "learning_rate": 6.010219116334064e-05, "loss": 0.9427, "step": 4405 }, { "epoch": 1.9963751699139103, "grad_norm": 0.28494279977372927, "learning_rate": 6.009125047637379e-05, "loss": 0.9786, "step": 4406 }, { "epoch": 1.9968282736746716, "grad_norm": 0.27113132822269514, "learning_rate": 6.008030777882978e-05, "loss": 0.9677, "step": 4407 }, { "epoch": 1.9972813774354328, "grad_norm": 0.29848445433417187, "learning_rate": 6.0069363071803664e-05, "loss": 0.9423, "step": 4408 }, { "epoch": 1.997734481196194, "grad_norm": 0.31850397989440454, "learning_rate": 6.005841635639069e-05, "loss": 1.0, "step": 4409 }, { "epoch": 1.998187584956955, "grad_norm": 0.3894915964163607, "learning_rate": 6.0047467633686355e-05, "loss": 0.9683, "step": 4410 }, { "epoch": 1.9986406887177164, "grad_norm": 0.4500816175493467, "learning_rate": 6.003651690478631e-05, "loss": 0.9313, "step": 4411 }, { "epoch": 1.9990937924784775, "grad_norm": 0.46529295996780295, "learning_rate": 6.002556417078639e-05, "loss": 0.9634, "step": 4412 }, { "epoch": 1.9995468962392389, "grad_norm": 0.4439544556035283, "learning_rate": 6.001460943278269e-05, "loss": 0.9762, "step": 4413 }, { "epoch": 2.0, "grad_norm": 0.4061223230442913, "learning_rate": 6.0003652691871465e-05, "loss": 0.969, "step": 4414 }, { "epoch": 2.000453103760761, "grad_norm": 0.3192367835402206, "learning_rate": 5.999269394914919e-05, "loss": 0.9594, "step": 4415 }, { "epoch": 2.0009062075215223, "grad_norm": 0.2711170413014288, "learning_rate": 5.998173320571252e-05, "loss": 0.9448, "step": 4416 }, { "epoch": 2.001359311282284, "grad_norm": 0.7297329499003941, "learning_rate": 5.997077046265832e-05, "loss": 0.9524, "step": 4417 }, { "epoch": 2.001812415043045, "grad_norm": 0.3556205700371329, "learning_rate": 5.9959805721083676e-05, "loss": 0.9513, "step": 4418 }, { "epoch": 2.002265518803806, "grad_norm": 0.42002763998372794, "learning_rate": 5.994883898208584e-05, "loss": 0.9588, "step": 4419 }, { "epoch": 2.0027186225645672, "grad_norm": 0.43423496576416704, "learning_rate": 5.993787024676228e-05, "loss": 0.9646, "step": 4420 }, { "epoch": 2.0031717263253284, "grad_norm": 0.3428998356878248, "learning_rate": 5.9926899516210665e-05, "loss": 0.9286, "step": 4421 }, { "epoch": 2.00362483008609, "grad_norm": 0.24150486639582344, "learning_rate": 5.991592679152884e-05, "loss": 0.9624, "step": 4422 }, { "epoch": 2.004077933846851, "grad_norm": 0.2232468415871376, "learning_rate": 5.9904952073814915e-05, "loss": 0.9595, "step": 4423 }, { "epoch": 2.004531037607612, "grad_norm": 0.30175017327262477, "learning_rate": 5.989397536416711e-05, "loss": 0.9601, "step": 4424 }, { "epoch": 2.0049841413683733, "grad_norm": 0.3563225798431963, "learning_rate": 5.9882996663683916e-05, "loss": 0.9408, "step": 4425 }, { "epoch": 2.0054372451291345, "grad_norm": 0.32089739488273045, "learning_rate": 5.987201597346399e-05, "loss": 0.9294, "step": 4426 }, { "epoch": 2.0058903488898956, "grad_norm": 0.26463501326041733, "learning_rate": 5.98610332946062e-05, "loss": 0.9595, "step": 4427 }, { "epoch": 2.006343452650657, "grad_norm": 0.2670013901106552, "learning_rate": 5.985004862820959e-05, "loss": 0.9427, "step": 4428 }, { "epoch": 2.0067965564114183, "grad_norm": 0.28801347896101326, "learning_rate": 5.983906197537344e-05, "loss": 0.9607, "step": 4429 }, { "epoch": 2.0072496601721794, "grad_norm": 0.34447674262059513, "learning_rate": 5.98280733371972e-05, "loss": 0.9657, "step": 4430 }, { "epoch": 2.0077027639329406, "grad_norm": 0.34433384080003776, "learning_rate": 5.981708271478052e-05, "loss": 0.9435, "step": 4431 }, { "epoch": 2.0081558676937017, "grad_norm": 0.3256584913524052, "learning_rate": 5.9806090109223275e-05, "loss": 0.9506, "step": 4432 }, { "epoch": 2.0086089714544633, "grad_norm": 0.2938166441445394, "learning_rate": 5.979509552162551e-05, "loss": 0.963, "step": 4433 }, { "epoch": 2.0090620752152244, "grad_norm": 0.24723194970356135, "learning_rate": 5.978409895308748e-05, "loss": 0.9536, "step": 4434 }, { "epoch": 2.0095151789759855, "grad_norm": 0.24377394067772068, "learning_rate": 5.977310040470963e-05, "loss": 0.9539, "step": 4435 }, { "epoch": 2.0099682827367467, "grad_norm": 0.2524177078870052, "learning_rate": 5.9762099877592614e-05, "loss": 0.9399, "step": 4436 }, { "epoch": 2.010421386497508, "grad_norm": 0.23269105602146514, "learning_rate": 5.9751097372837296e-05, "loss": 0.9811, "step": 4437 }, { "epoch": 2.0108744902582694, "grad_norm": 0.24766712871911492, "learning_rate": 5.9740092891544704e-05, "loss": 0.9631, "step": 4438 }, { "epoch": 2.0113275940190305, "grad_norm": 0.30413108610720513, "learning_rate": 5.9729086434816075e-05, "loss": 0.9416, "step": 4439 }, { "epoch": 2.0117806977797916, "grad_norm": 0.33874593989274154, "learning_rate": 5.971807800375286e-05, "loss": 0.9462, "step": 4440 }, { "epoch": 2.0122338015405528, "grad_norm": 0.3674169887099697, "learning_rate": 5.97070675994567e-05, "loss": 0.9384, "step": 4441 }, { "epoch": 2.012686905301314, "grad_norm": 0.30710181861190566, "learning_rate": 5.969605522302943e-05, "loss": 0.9556, "step": 4442 }, { "epoch": 2.013140009062075, "grad_norm": 0.2836483588538309, "learning_rate": 5.968504087557309e-05, "loss": 0.967, "step": 4443 }, { "epoch": 2.0135931128228366, "grad_norm": 0.2835825806381495, "learning_rate": 5.967402455818989e-05, "loss": 0.9601, "step": 4444 }, { "epoch": 2.0140462165835977, "grad_norm": 0.29571999173979774, "learning_rate": 5.9663006271982284e-05, "loss": 0.9376, "step": 4445 }, { "epoch": 2.014499320344359, "grad_norm": 0.26951870532450156, "learning_rate": 5.965198601805288e-05, "loss": 0.9528, "step": 4446 }, { "epoch": 2.01495242410512, "grad_norm": 0.22922746858308554, "learning_rate": 5.96409637975045e-05, "loss": 0.9469, "step": 4447 }, { "epoch": 2.015405527865881, "grad_norm": 0.2574570582844164, "learning_rate": 5.962993961144016e-05, "loss": 0.939, "step": 4448 }, { "epoch": 2.0158586316266427, "grad_norm": 0.2718749023912252, "learning_rate": 5.96189134609631e-05, "loss": 0.9414, "step": 4449 }, { "epoch": 2.016311735387404, "grad_norm": 0.3021347774541089, "learning_rate": 5.960788534717671e-05, "loss": 0.9685, "step": 4450 }, { "epoch": 2.016764839148165, "grad_norm": 0.3222708697511324, "learning_rate": 5.959685527118459e-05, "loss": 0.9676, "step": 4451 }, { "epoch": 2.017217942908926, "grad_norm": 0.3174658325455308, "learning_rate": 5.958582323409056e-05, "loss": 0.9345, "step": 4452 }, { "epoch": 2.017671046669687, "grad_norm": 0.3214387364267551, "learning_rate": 5.957478923699864e-05, "loss": 0.9625, "step": 4453 }, { "epoch": 2.018124150430449, "grad_norm": 0.3577619020254807, "learning_rate": 5.9563753281012984e-05, "loss": 0.9523, "step": 4454 }, { "epoch": 2.01857725419121, "grad_norm": 0.3713486493653708, "learning_rate": 5.9552715367238017e-05, "loss": 0.959, "step": 4455 }, { "epoch": 2.019030357951971, "grad_norm": 0.3962373301171569, "learning_rate": 5.954167549677831e-05, "loss": 0.9692, "step": 4456 }, { "epoch": 2.019483461712732, "grad_norm": 0.4145169284694434, "learning_rate": 5.953063367073865e-05, "loss": 0.9375, "step": 4457 }, { "epoch": 2.0199365654734933, "grad_norm": 0.3694247199404924, "learning_rate": 5.9519589890224053e-05, "loss": 0.9627, "step": 4458 }, { "epoch": 2.0203896692342544, "grad_norm": 0.30694596872448443, "learning_rate": 5.950854415633964e-05, "loss": 0.9422, "step": 4459 }, { "epoch": 2.020842772995016, "grad_norm": 0.2728617041983484, "learning_rate": 5.9497496470190815e-05, "loss": 0.9373, "step": 4460 }, { "epoch": 2.021295876755777, "grad_norm": 0.2681319177541482, "learning_rate": 5.948644683288313e-05, "loss": 0.9336, "step": 4461 }, { "epoch": 2.0217489805165383, "grad_norm": 0.2876219784163782, "learning_rate": 5.9475395245522376e-05, "loss": 0.9518, "step": 4462 }, { "epoch": 2.0222020842772994, "grad_norm": 0.2531115753852645, "learning_rate": 5.9464341709214484e-05, "loss": 0.9647, "step": 4463 }, { "epoch": 2.0226551880380605, "grad_norm": 0.22063485675534547, "learning_rate": 5.94532862250656e-05, "loss": 0.9449, "step": 4464 }, { "epoch": 2.023108291798822, "grad_norm": 0.21316472709461023, "learning_rate": 5.9442228794182104e-05, "loss": 0.9329, "step": 4465 }, { "epoch": 2.0235613955595833, "grad_norm": 0.2367988807889223, "learning_rate": 5.9431169417670506e-05, "loss": 0.9417, "step": 4466 }, { "epoch": 2.0240144993203444, "grad_norm": 0.23053086874404383, "learning_rate": 5.942010809663756e-05, "loss": 0.9411, "step": 4467 }, { "epoch": 2.0244676030811055, "grad_norm": 0.2606263808970464, "learning_rate": 5.9409044832190186e-05, "loss": 0.9397, "step": 4468 }, { "epoch": 2.0249207068418666, "grad_norm": 0.3007330644467128, "learning_rate": 5.9397979625435534e-05, "loss": 0.9517, "step": 4469 }, { "epoch": 2.025373810602628, "grad_norm": 0.3107058629811465, "learning_rate": 5.9386912477480894e-05, "loss": 0.9624, "step": 4470 }, { "epoch": 2.0258269143633894, "grad_norm": 0.3644696944579455, "learning_rate": 5.93758433894338e-05, "loss": 0.9651, "step": 4471 }, { "epoch": 2.0262800181241505, "grad_norm": 0.3905785123320811, "learning_rate": 5.936477236240196e-05, "loss": 0.9548, "step": 4472 }, { "epoch": 2.0267331218849116, "grad_norm": 0.43779465563044234, "learning_rate": 5.935369939749327e-05, "loss": 0.9423, "step": 4473 }, { "epoch": 2.0271862256456727, "grad_norm": 0.42050111576623794, "learning_rate": 5.934262449581582e-05, "loss": 0.9347, "step": 4474 }, { "epoch": 2.027639329406434, "grad_norm": 0.3205454123406077, "learning_rate": 5.933154765847792e-05, "loss": 0.9518, "step": 4475 }, { "epoch": 2.0280924331671955, "grad_norm": 0.2109766246864531, "learning_rate": 5.932046888658804e-05, "loss": 0.9541, "step": 4476 }, { "epoch": 2.0285455369279566, "grad_norm": 0.18755822533563987, "learning_rate": 5.930938818125485e-05, "loss": 0.967, "step": 4477 }, { "epoch": 2.0289986406887177, "grad_norm": 0.20304904943548258, "learning_rate": 5.929830554358726e-05, "loss": 0.9393, "step": 4478 }, { "epoch": 2.029451744449479, "grad_norm": 0.2461658067523231, "learning_rate": 5.928722097469429e-05, "loss": 0.9601, "step": 4479 }, { "epoch": 2.02990484821024, "grad_norm": 0.2578343741784966, "learning_rate": 5.9276134475685226e-05, "loss": 0.9563, "step": 4480 }, { "epoch": 2.0303579519710016, "grad_norm": 0.2393746384844887, "learning_rate": 5.9265046047669505e-05, "loss": 0.9267, "step": 4481 }, { "epoch": 2.0308110557317627, "grad_norm": 0.21930559699896532, "learning_rate": 5.925395569175677e-05, "loss": 0.9481, "step": 4482 }, { "epoch": 2.031264159492524, "grad_norm": 0.22094515141766174, "learning_rate": 5.924286340905687e-05, "loss": 0.958, "step": 4483 }, { "epoch": 2.031717263253285, "grad_norm": 0.20973418723782972, "learning_rate": 5.923176920067983e-05, "loss": 0.9559, "step": 4484 }, { "epoch": 2.032170367014046, "grad_norm": 0.24688603656167868, "learning_rate": 5.922067306773587e-05, "loss": 0.944, "step": 4485 }, { "epoch": 2.0326234707748077, "grad_norm": 0.2886842863595173, "learning_rate": 5.92095750113354e-05, "loss": 0.9446, "step": 4486 }, { "epoch": 2.033076574535569, "grad_norm": 0.3159865879796814, "learning_rate": 5.919847503258904e-05, "loss": 0.9602, "step": 4487 }, { "epoch": 2.03352967829633, "grad_norm": 0.32310575288374677, "learning_rate": 5.918737313260759e-05, "loss": 0.9546, "step": 4488 }, { "epoch": 2.033982782057091, "grad_norm": 0.3188939777977131, "learning_rate": 5.917626931250203e-05, "loss": 0.9537, "step": 4489 }, { "epoch": 2.034435885817852, "grad_norm": 0.2622937989343297, "learning_rate": 5.9165163573383564e-05, "loss": 0.9532, "step": 4490 }, { "epoch": 2.0348889895786133, "grad_norm": 0.25746501281551226, "learning_rate": 5.915405591636354e-05, "loss": 0.9355, "step": 4491 }, { "epoch": 2.035342093339375, "grad_norm": 0.2719112556423754, "learning_rate": 5.914294634255355e-05, "loss": 0.9712, "step": 4492 }, { "epoch": 2.035795197100136, "grad_norm": 0.3281004159513575, "learning_rate": 5.913183485306535e-05, "loss": 0.9476, "step": 4493 }, { "epoch": 2.036248300860897, "grad_norm": 0.36380990389587653, "learning_rate": 5.912072144901088e-05, "loss": 0.9412, "step": 4494 }, { "epoch": 2.0367014046216583, "grad_norm": 0.3233586427047116, "learning_rate": 5.91096061315023e-05, "loss": 0.9707, "step": 4495 }, { "epoch": 2.0371545083824194, "grad_norm": 0.2525047644408415, "learning_rate": 5.909848890165194e-05, "loss": 0.9447, "step": 4496 }, { "epoch": 2.037607612143181, "grad_norm": 0.25248191453569824, "learning_rate": 5.908736976057231e-05, "loss": 0.9412, "step": 4497 }, { "epoch": 2.038060715903942, "grad_norm": 0.2720974812328481, "learning_rate": 5.907624870937616e-05, "loss": 0.937, "step": 4498 }, { "epoch": 2.0385138196647032, "grad_norm": 0.24355100591210754, "learning_rate": 5.9065125749176376e-05, "loss": 0.9363, "step": 4499 }, { "epoch": 2.0389669234254644, "grad_norm": 0.22266189096650346, "learning_rate": 5.905400088108606e-05, "loss": 0.9343, "step": 4500 }, { "epoch": 2.0394200271862255, "grad_norm": 0.212670002760969, "learning_rate": 5.9042874106218507e-05, "loss": 0.9554, "step": 4501 }, { "epoch": 2.0398731309469866, "grad_norm": 0.22335325539234838, "learning_rate": 5.903174542568719e-05, "loss": 0.9469, "step": 4502 }, { "epoch": 2.040326234707748, "grad_norm": 0.2796157953856338, "learning_rate": 5.9020614840605804e-05, "loss": 0.9295, "step": 4503 }, { "epoch": 2.0407793384685093, "grad_norm": 0.3265306115692817, "learning_rate": 5.9009482352088186e-05, "loss": 0.9678, "step": 4504 }, { "epoch": 2.0412324422292705, "grad_norm": 0.3325853213790503, "learning_rate": 5.8998347961248404e-05, "loss": 0.964, "step": 4505 }, { "epoch": 2.0416855459900316, "grad_norm": 0.32402483285494144, "learning_rate": 5.8987211669200704e-05, "loss": 0.9339, "step": 4506 }, { "epoch": 2.0421386497507927, "grad_norm": 0.3449992659615639, "learning_rate": 5.8976073477059505e-05, "loss": 0.9586, "step": 4507 }, { "epoch": 2.0425917535115543, "grad_norm": 0.3538766537507339, "learning_rate": 5.896493338593945e-05, "loss": 0.965, "step": 4508 }, { "epoch": 2.0430448572723154, "grad_norm": 0.31304120349954034, "learning_rate": 5.895379139695533e-05, "loss": 0.9399, "step": 4509 }, { "epoch": 2.0434979610330766, "grad_norm": 0.22886224738830419, "learning_rate": 5.894264751122217e-05, "loss": 0.9535, "step": 4510 }, { "epoch": 2.0439510647938377, "grad_norm": 0.23314975148354755, "learning_rate": 5.893150172985517e-05, "loss": 0.9291, "step": 4511 }, { "epoch": 2.044404168554599, "grad_norm": 0.20118502703496804, "learning_rate": 5.892035405396968e-05, "loss": 0.9487, "step": 4512 }, { "epoch": 2.0448572723153604, "grad_norm": 0.18124057200796498, "learning_rate": 5.8909204484681305e-05, "loss": 0.9356, "step": 4513 }, { "epoch": 2.0453103760761215, "grad_norm": 0.1973557775423821, "learning_rate": 5.8898053023105794e-05, "loss": 0.9643, "step": 4514 }, { "epoch": 2.0457634798368827, "grad_norm": 0.20447721184650638, "learning_rate": 5.888689967035911e-05, "loss": 0.946, "step": 4515 }, { "epoch": 2.046216583597644, "grad_norm": 0.2109271508795064, "learning_rate": 5.887574442755738e-05, "loss": 0.9407, "step": 4516 }, { "epoch": 2.046669687358405, "grad_norm": 0.2454108784473607, "learning_rate": 5.886458729581694e-05, "loss": 0.9416, "step": 4517 }, { "epoch": 2.0471227911191665, "grad_norm": 0.2955125093101998, "learning_rate": 5.885342827625431e-05, "loss": 0.9412, "step": 4518 }, { "epoch": 2.0475758948799276, "grad_norm": 0.3177969568044585, "learning_rate": 5.884226736998621e-05, "loss": 0.9578, "step": 4519 }, { "epoch": 2.0480289986406888, "grad_norm": 0.28494655915098893, "learning_rate": 5.883110457812951e-05, "loss": 0.9494, "step": 4520 }, { "epoch": 2.04848210240145, "grad_norm": 0.2621109481909598, "learning_rate": 5.8819939901801314e-05, "loss": 0.9701, "step": 4521 }, { "epoch": 2.048935206162211, "grad_norm": 0.2801516676576227, "learning_rate": 5.88087733421189e-05, "loss": 0.9668, "step": 4522 }, { "epoch": 2.049388309922972, "grad_norm": 0.280544697395742, "learning_rate": 5.879760490019973e-05, "loss": 0.9632, "step": 4523 }, { "epoch": 2.0498414136837337, "grad_norm": 0.347776667638672, "learning_rate": 5.878643457716145e-05, "loss": 0.9447, "step": 4524 }, { "epoch": 2.050294517444495, "grad_norm": 0.3425145830442405, "learning_rate": 5.87752623741219e-05, "loss": 0.9633, "step": 4525 }, { "epoch": 2.050747621205256, "grad_norm": 0.2874285752140883, "learning_rate": 5.876408829219911e-05, "loss": 0.9506, "step": 4526 }, { "epoch": 2.051200724966017, "grad_norm": 0.2763458206216656, "learning_rate": 5.875291233251128e-05, "loss": 0.9395, "step": 4527 }, { "epoch": 2.0516538287267783, "grad_norm": 0.2511482766498809, "learning_rate": 5.874173449617684e-05, "loss": 0.9413, "step": 4528 }, { "epoch": 2.05210693248754, "grad_norm": 0.2620318533446036, "learning_rate": 5.873055478431437e-05, "loss": 0.9743, "step": 4529 }, { "epoch": 2.052560036248301, "grad_norm": 0.2708365002582713, "learning_rate": 5.8719373198042644e-05, "loss": 0.9611, "step": 4530 }, { "epoch": 2.053013140009062, "grad_norm": 0.258510292842463, "learning_rate": 5.870818973848064e-05, "loss": 0.9572, "step": 4531 }, { "epoch": 2.0534662437698232, "grad_norm": 0.1965541275592126, "learning_rate": 5.8697004406747506e-05, "loss": 0.9518, "step": 4532 }, { "epoch": 2.0539193475305844, "grad_norm": 0.25038151956155785, "learning_rate": 5.8685817203962586e-05, "loss": 0.9616, "step": 4533 }, { "epoch": 2.054372451291346, "grad_norm": 0.3508522163434966, "learning_rate": 5.86746281312454e-05, "loss": 0.976, "step": 4534 }, { "epoch": 2.054825555052107, "grad_norm": 0.422454826930881, "learning_rate": 5.8663437189715674e-05, "loss": 0.9512, "step": 4535 }, { "epoch": 2.055278658812868, "grad_norm": 0.3950495909182895, "learning_rate": 5.86522443804933e-05, "loss": 0.9494, "step": 4536 }, { "epoch": 2.0557317625736293, "grad_norm": 0.36841495181014267, "learning_rate": 5.864104970469838e-05, "loss": 0.941, "step": 4537 }, { "epoch": 2.0561848663343905, "grad_norm": 0.39661660933090986, "learning_rate": 5.862985316345119e-05, "loss": 0.9372, "step": 4538 }, { "epoch": 2.0566379700951516, "grad_norm": 0.4222396824803306, "learning_rate": 5.861865475787219e-05, "loss": 0.9348, "step": 4539 }, { "epoch": 2.057091073855913, "grad_norm": 0.4135954798631861, "learning_rate": 5.8607454489082026e-05, "loss": 0.9754, "step": 4540 }, { "epoch": 2.0575441776166743, "grad_norm": 0.38157997525107046, "learning_rate": 5.859625235820154e-05, "loss": 0.9632, "step": 4541 }, { "epoch": 2.0579972813774354, "grad_norm": 0.34562655407734183, "learning_rate": 5.858504836635176e-05, "loss": 0.9553, "step": 4542 }, { "epoch": 2.0584503851381966, "grad_norm": 0.30980017337214616, "learning_rate": 5.857384251465387e-05, "loss": 0.9801, "step": 4543 }, { "epoch": 2.0589034888989577, "grad_norm": 0.30299121342479485, "learning_rate": 5.8562634804229296e-05, "loss": 0.9413, "step": 4544 }, { "epoch": 2.0593565926597193, "grad_norm": 0.2805330365521605, "learning_rate": 5.85514252361996e-05, "loss": 0.9337, "step": 4545 }, { "epoch": 2.0598096964204804, "grad_norm": 0.3722465759729258, "learning_rate": 5.8540213811686555e-05, "loss": 0.9511, "step": 4546 }, { "epoch": 2.0602628001812415, "grad_norm": 0.4948182165110687, "learning_rate": 5.852900053181211e-05, "loss": 0.9747, "step": 4547 }, { "epoch": 2.0607159039420027, "grad_norm": 0.6330996252791355, "learning_rate": 5.851778539769841e-05, "loss": 0.9658, "step": 4548 }, { "epoch": 2.061169007702764, "grad_norm": 0.7113687629816958, "learning_rate": 5.850656841046778e-05, "loss": 0.9169, "step": 4549 }, { "epoch": 2.0616221114635254, "grad_norm": 0.7284415982984113, "learning_rate": 5.849534957124272e-05, "loss": 0.9511, "step": 4550 }, { "epoch": 2.0620752152242865, "grad_norm": 0.6361742719325323, "learning_rate": 5.848412888114593e-05, "loss": 0.9654, "step": 4551 }, { "epoch": 2.0625283189850476, "grad_norm": 0.5135961833656362, "learning_rate": 5.84729063413003e-05, "loss": 0.9514, "step": 4552 }, { "epoch": 2.0629814227458088, "grad_norm": 0.40320901777689483, "learning_rate": 5.846168195282888e-05, "loss": 0.9343, "step": 4553 }, { "epoch": 2.06343452650657, "grad_norm": 0.3114288668299737, "learning_rate": 5.845045571685492e-05, "loss": 0.9511, "step": 4554 }, { "epoch": 2.063887630267331, "grad_norm": 0.26428390166922017, "learning_rate": 5.8439227634501866e-05, "loss": 0.9454, "step": 4555 }, { "epoch": 2.0643407340280926, "grad_norm": 0.3042534402774912, "learning_rate": 5.842799770689333e-05, "loss": 0.9408, "step": 4556 }, { "epoch": 2.0647938377888537, "grad_norm": 0.32315040937497036, "learning_rate": 5.841676593515311e-05, "loss": 0.9483, "step": 4557 }, { "epoch": 2.065246941549615, "grad_norm": 0.6282850467404921, "learning_rate": 5.8405532320405204e-05, "loss": 0.9579, "step": 4558 }, { "epoch": 2.065700045310376, "grad_norm": 0.2720173446396792, "learning_rate": 5.83942968637738e-05, "loss": 0.9557, "step": 4559 }, { "epoch": 2.066153149071137, "grad_norm": 0.2003365774723633, "learning_rate": 5.838305956638323e-05, "loss": 0.9378, "step": 4560 }, { "epoch": 2.0666062528318987, "grad_norm": 0.19916125538165433, "learning_rate": 5.837182042935805e-05, "loss": 0.9525, "step": 4561 }, { "epoch": 2.06705935659266, "grad_norm": 0.23006142833603602, "learning_rate": 5.8360579453822974e-05, "loss": 0.943, "step": 4562 }, { "epoch": 2.067512460353421, "grad_norm": 0.25389676090742624, "learning_rate": 5.834933664090293e-05, "loss": 0.972, "step": 4563 }, { "epoch": 2.067965564114182, "grad_norm": 0.2722474875089904, "learning_rate": 5.8338091991723e-05, "loss": 0.9563, "step": 4564 }, { "epoch": 2.068418667874943, "grad_norm": 0.26216773794577947, "learning_rate": 5.832684550740846e-05, "loss": 0.9538, "step": 4565 }, { "epoch": 2.0688717716357043, "grad_norm": 0.22947172107730726, "learning_rate": 5.8315597189084784e-05, "loss": 0.9546, "step": 4566 }, { "epoch": 2.069324875396466, "grad_norm": 0.2016783074001718, "learning_rate": 5.8304347037877604e-05, "loss": 0.9398, "step": 4567 }, { "epoch": 2.069777979157227, "grad_norm": 0.21170656537535174, "learning_rate": 5.829309505491275e-05, "loss": 0.986, "step": 4568 }, { "epoch": 2.070231082917988, "grad_norm": 0.20470205349958828, "learning_rate": 5.828184124131624e-05, "loss": 0.933, "step": 4569 }, { "epoch": 2.0706841866787493, "grad_norm": 0.2362712489137601, "learning_rate": 5.8270585598214264e-05, "loss": 0.9375, "step": 4570 }, { "epoch": 2.0711372904395104, "grad_norm": 0.30270246914031873, "learning_rate": 5.8259328126733195e-05, "loss": 0.9603, "step": 4571 }, { "epoch": 2.071590394200272, "grad_norm": 0.21251699294603987, "learning_rate": 5.82480688279996e-05, "loss": 0.9339, "step": 4572 }, { "epoch": 2.072043497961033, "grad_norm": 0.2046139077319895, "learning_rate": 5.8236807703140234e-05, "loss": 0.9563, "step": 4573 }, { "epoch": 2.0724966017217943, "grad_norm": 0.22404110520823647, "learning_rate": 5.8225544753282e-05, "loss": 0.9353, "step": 4574 }, { "epoch": 2.0729497054825554, "grad_norm": 0.24966426348327014, "learning_rate": 5.8214279979552016e-05, "loss": 0.9493, "step": 4575 }, { "epoch": 2.0734028092433165, "grad_norm": 0.23705972928626343, "learning_rate": 5.820301338307758e-05, "loss": 0.9531, "step": 4576 }, { "epoch": 2.073855913004078, "grad_norm": 0.24116388727610838, "learning_rate": 5.819174496498617e-05, "loss": 0.9549, "step": 4577 }, { "epoch": 2.0743090167648393, "grad_norm": 0.21345876751882448, "learning_rate": 5.8180474726405425e-05, "loss": 0.9474, "step": 4578 }, { "epoch": 2.0747621205256004, "grad_norm": 0.2171923934257038, "learning_rate": 5.8169202668463193e-05, "loss": 0.9372, "step": 4579 }, { "epoch": 2.0752152242863615, "grad_norm": 0.241384968168043, "learning_rate": 5.815792879228749e-05, "loss": 0.9297, "step": 4580 }, { "epoch": 2.0756683280471226, "grad_norm": 0.22008943967352623, "learning_rate": 5.814665309900653e-05, "loss": 0.9546, "step": 4581 }, { "epoch": 2.076121431807884, "grad_norm": 0.25685015372621123, "learning_rate": 5.8135375589748686e-05, "loss": 0.9519, "step": 4582 }, { "epoch": 2.0765745355686454, "grad_norm": 0.2912131891442711, "learning_rate": 5.812409626564252e-05, "loss": 0.931, "step": 4583 }, { "epoch": 2.0770276393294065, "grad_norm": 0.27195991001365266, "learning_rate": 5.81128151278168e-05, "loss": 0.955, "step": 4584 }, { "epoch": 2.0774807430901676, "grad_norm": 0.272640567943391, "learning_rate": 5.8101532177400446e-05, "loss": 0.9552, "step": 4585 }, { "epoch": 2.0779338468509287, "grad_norm": 0.2531142717882277, "learning_rate": 5.809024741552255e-05, "loss": 0.9399, "step": 4586 }, { "epoch": 2.07838695061169, "grad_norm": 0.2163224955456695, "learning_rate": 5.8078960843312434e-05, "loss": 0.9504, "step": 4587 }, { "epoch": 2.0788400543724515, "grad_norm": 0.26026719640483026, "learning_rate": 5.806767246189954e-05, "loss": 0.9594, "step": 4588 }, { "epoch": 2.0792931581332126, "grad_norm": 0.3259871247551642, "learning_rate": 5.8056382272413554e-05, "loss": 0.966, "step": 4589 }, { "epoch": 2.0797462618939737, "grad_norm": 0.35154708017009356, "learning_rate": 5.804509027598428e-05, "loss": 0.9482, "step": 4590 }, { "epoch": 2.080199365654735, "grad_norm": 0.3694075072400204, "learning_rate": 5.8033796473741755e-05, "loss": 0.9601, "step": 4591 }, { "epoch": 2.080652469415496, "grad_norm": 0.42100250330231676, "learning_rate": 5.802250086681617e-05, "loss": 0.9495, "step": 4592 }, { "epoch": 2.0811055731762576, "grad_norm": 0.42878965023247817, "learning_rate": 5.801120345633791e-05, "loss": 0.9385, "step": 4593 }, { "epoch": 2.0815586769370187, "grad_norm": 0.3936099706823543, "learning_rate": 5.799990424343751e-05, "loss": 0.9335, "step": 4594 }, { "epoch": 2.08201178069778, "grad_norm": 0.3397676970866049, "learning_rate": 5.7988603229245724e-05, "loss": 0.9465, "step": 4595 }, { "epoch": 2.082464884458541, "grad_norm": 0.2845295386571889, "learning_rate": 5.797730041489347e-05, "loss": 0.9673, "step": 4596 }, { "epoch": 2.082917988219302, "grad_norm": 0.22757057944081707, "learning_rate": 5.796599580151184e-05, "loss": 0.9428, "step": 4597 }, { "epoch": 2.0833710919800636, "grad_norm": 0.24511426724691532, "learning_rate": 5.7954689390232123e-05, "loss": 0.9412, "step": 4598 }, { "epoch": 2.083824195740825, "grad_norm": 0.24702041488979717, "learning_rate": 5.794338118218577e-05, "loss": 0.9245, "step": 4599 }, { "epoch": 2.084277299501586, "grad_norm": 0.19664246179038583, "learning_rate": 5.793207117850442e-05, "loss": 0.9621, "step": 4600 }, { "epoch": 2.084730403262347, "grad_norm": 0.17665033721077472, "learning_rate": 5.792075938031989e-05, "loss": 0.9549, "step": 4601 }, { "epoch": 2.085183507023108, "grad_norm": 0.23782842973872287, "learning_rate": 5.790944578876418e-05, "loss": 0.9686, "step": 4602 }, { "epoch": 2.0856366107838693, "grad_norm": 0.28376959467380525, "learning_rate": 5.789813040496947e-05, "loss": 0.9637, "step": 4603 }, { "epoch": 2.086089714544631, "grad_norm": 0.3233198031634023, "learning_rate": 5.7886813230068106e-05, "loss": 0.96, "step": 4604 }, { "epoch": 2.086542818305392, "grad_norm": 0.3181125597631431, "learning_rate": 5.787549426519262e-05, "loss": 0.9491, "step": 4605 }, { "epoch": 2.086995922066153, "grad_norm": 0.31256480558084626, "learning_rate": 5.786417351147575e-05, "loss": 0.9538, "step": 4606 }, { "epoch": 2.0874490258269143, "grad_norm": 0.3220170484625561, "learning_rate": 5.785285097005037e-05, "loss": 0.9356, "step": 4607 }, { "epoch": 2.0879021295876754, "grad_norm": 0.302282863735246, "learning_rate": 5.7841526642049554e-05, "loss": 0.9418, "step": 4608 }, { "epoch": 2.088355233348437, "grad_norm": 0.23718370516165305, "learning_rate": 5.783020052860656e-05, "loss": 0.9504, "step": 4609 }, { "epoch": 2.088808337109198, "grad_norm": 0.2181276190972891, "learning_rate": 5.7818872630854804e-05, "loss": 0.9337, "step": 4610 }, { "epoch": 2.0892614408699592, "grad_norm": 0.27129959382646424, "learning_rate": 5.7807542949927914e-05, "loss": 0.9493, "step": 4611 }, { "epoch": 2.0897145446307204, "grad_norm": 0.35455437892220243, "learning_rate": 5.779621148695968e-05, "loss": 0.9348, "step": 4612 }, { "epoch": 2.0901676483914815, "grad_norm": 0.44032468233289634, "learning_rate": 5.7784878243084046e-05, "loss": 0.9764, "step": 4613 }, { "epoch": 2.090620752152243, "grad_norm": 0.5053422025293348, "learning_rate": 5.777354321943516e-05, "loss": 0.939, "step": 4614 }, { "epoch": 2.091073855913004, "grad_norm": 0.5147282602755698, "learning_rate": 5.776220641714735e-05, "loss": 0.9526, "step": 4615 }, { "epoch": 2.0915269596737653, "grad_norm": 0.4737172214087899, "learning_rate": 5.7750867837355114e-05, "loss": 0.9491, "step": 4616 }, { "epoch": 2.0919800634345265, "grad_norm": 0.3840297692832029, "learning_rate": 5.7739527481193135e-05, "loss": 0.9612, "step": 4617 }, { "epoch": 2.0924331671952876, "grad_norm": 0.28826669688787115, "learning_rate": 5.7728185349796255e-05, "loss": 0.9391, "step": 4618 }, { "epoch": 2.0928862709560487, "grad_norm": 0.23022807233258782, "learning_rate": 5.771684144429952e-05, "loss": 0.942, "step": 4619 }, { "epoch": 2.0933393747168103, "grad_norm": 0.24746592495519698, "learning_rate": 5.770549576583814e-05, "loss": 0.9669, "step": 4620 }, { "epoch": 2.0937924784775714, "grad_norm": 0.353083732917116, "learning_rate": 5.769414831554749e-05, "loss": 0.9467, "step": 4621 }, { "epoch": 2.0942455822383326, "grad_norm": 0.4140055126608578, "learning_rate": 5.768279909456315e-05, "loss": 0.9447, "step": 4622 }, { "epoch": 2.0946986859990937, "grad_norm": 0.3987778297890464, "learning_rate": 5.767144810402085e-05, "loss": 0.9651, "step": 4623 }, { "epoch": 2.095151789759855, "grad_norm": 0.31546327156777704, "learning_rate": 5.766009534505652e-05, "loss": 0.9458, "step": 4624 }, { "epoch": 2.0956048935206164, "grad_norm": 0.26091916448912006, "learning_rate": 5.764874081880625e-05, "loss": 0.9768, "step": 4625 }, { "epoch": 2.0960579972813775, "grad_norm": 0.28488887953036224, "learning_rate": 5.763738452640631e-05, "loss": 0.9424, "step": 4626 }, { "epoch": 2.0965111010421387, "grad_norm": 0.2925507852674397, "learning_rate": 5.762602646899316e-05, "loss": 0.9585, "step": 4627 }, { "epoch": 2.0969642048029, "grad_norm": 0.2603879552441674, "learning_rate": 5.761466664770343e-05, "loss": 0.9545, "step": 4628 }, { "epoch": 2.097417308563661, "grad_norm": 0.27592982118238213, "learning_rate": 5.760330506367391e-05, "loss": 0.9582, "step": 4629 }, { "epoch": 2.0978704123244225, "grad_norm": 0.26962864428961814, "learning_rate": 5.759194171804158e-05, "loss": 0.9708, "step": 4630 }, { "epoch": 2.0983235160851836, "grad_norm": 0.238228392091756, "learning_rate": 5.7580576611943605e-05, "loss": 0.9523, "step": 4631 }, { "epoch": 2.0987766198459448, "grad_norm": 0.20293596239937045, "learning_rate": 5.7569209746517305e-05, "loss": 0.959, "step": 4632 }, { "epoch": 2.099229723606706, "grad_norm": 0.19402269870429262, "learning_rate": 5.7557841122900197e-05, "loss": 0.9327, "step": 4633 }, { "epoch": 2.099682827367467, "grad_norm": 0.22183652293692022, "learning_rate": 5.754647074222997e-05, "loss": 0.9348, "step": 4634 }, { "epoch": 2.100135931128228, "grad_norm": 0.24119025273129638, "learning_rate": 5.753509860564446e-05, "loss": 0.9592, "step": 4635 }, { "epoch": 2.1005890348889897, "grad_norm": 0.24201562147453312, "learning_rate": 5.7523724714281734e-05, "loss": 0.9377, "step": 4636 }, { "epoch": 2.101042138649751, "grad_norm": 0.2779763753287297, "learning_rate": 5.751234906927998e-05, "loss": 0.9465, "step": 4637 }, { "epoch": 2.101495242410512, "grad_norm": 0.26634400933663466, "learning_rate": 5.7500971671777613e-05, "loss": 0.9325, "step": 4638 }, { "epoch": 2.101948346171273, "grad_norm": 0.2154703492752708, "learning_rate": 5.7489592522913165e-05, "loss": 0.9607, "step": 4639 }, { "epoch": 2.1024014499320343, "grad_norm": 0.3265355710123383, "learning_rate": 5.7478211623825384e-05, "loss": 0.9657, "step": 4640 }, { "epoch": 2.102854553692796, "grad_norm": 0.18986609383992323, "learning_rate": 5.746682897565317e-05, "loss": 0.9779, "step": 4641 }, { "epoch": 2.103307657453557, "grad_norm": 0.2185144977436657, "learning_rate": 5.745544457953564e-05, "loss": 0.9409, "step": 4642 }, { "epoch": 2.103760761214318, "grad_norm": 0.2489005571879461, "learning_rate": 5.744405843661203e-05, "loss": 0.9458, "step": 4643 }, { "epoch": 2.1042138649750792, "grad_norm": 0.2886363757239785, "learning_rate": 5.7432670548021793e-05, "loss": 0.9491, "step": 4644 }, { "epoch": 2.1046669687358404, "grad_norm": 0.31787072881854805, "learning_rate": 5.742128091490453e-05, "loss": 0.9552, "step": 4645 }, { "epoch": 2.105120072496602, "grad_norm": 0.3654678250465455, "learning_rate": 5.7409889538400044e-05, "loss": 0.9607, "step": 4646 }, { "epoch": 2.105573176257363, "grad_norm": 0.36519269241499486, "learning_rate": 5.739849641964828e-05, "loss": 0.9413, "step": 4647 }, { "epoch": 2.106026280018124, "grad_norm": 0.35537853109123374, "learning_rate": 5.7387101559789364e-05, "loss": 0.9458, "step": 4648 }, { "epoch": 2.1064793837788853, "grad_norm": 0.370431789533665, "learning_rate": 5.7375704959963626e-05, "loss": 0.9614, "step": 4649 }, { "epoch": 2.1069324875396465, "grad_norm": 0.3643422667693138, "learning_rate": 5.736430662131154e-05, "loss": 0.9506, "step": 4650 }, { "epoch": 2.1073855913004076, "grad_norm": 0.33353746948858004, "learning_rate": 5.7352906544973776e-05, "loss": 0.952, "step": 4651 }, { "epoch": 2.107838695061169, "grad_norm": 0.3818830919264095, "learning_rate": 5.734150473209115e-05, "loss": 0.9445, "step": 4652 }, { "epoch": 2.1082917988219303, "grad_norm": 0.43863679772003705, "learning_rate": 5.733010118380467e-05, "loss": 0.9214, "step": 4653 }, { "epoch": 2.1087449025826914, "grad_norm": 0.4539332382858592, "learning_rate": 5.731869590125554e-05, "loss": 0.9597, "step": 4654 }, { "epoch": 2.1091980063434526, "grad_norm": 0.4311766785226558, "learning_rate": 5.730728888558506e-05, "loss": 0.9671, "step": 4655 }, { "epoch": 2.1096511101042137, "grad_norm": 0.3571747828016946, "learning_rate": 5.729588013793481e-05, "loss": 0.9639, "step": 4656 }, { "epoch": 2.1101042138649753, "grad_norm": 0.26651390409703013, "learning_rate": 5.728446965944646e-05, "loss": 0.9467, "step": 4657 }, { "epoch": 2.1105573176257364, "grad_norm": 0.19274176968927706, "learning_rate": 5.727305745126189e-05, "loss": 0.9398, "step": 4658 }, { "epoch": 2.1110104213864975, "grad_norm": 0.18444696681957032, "learning_rate": 5.7261643514523135e-05, "loss": 0.9399, "step": 4659 }, { "epoch": 2.1114635251472587, "grad_norm": 0.21006037435433111, "learning_rate": 5.7250227850372436e-05, "loss": 0.94, "step": 4660 }, { "epoch": 2.11191662890802, "grad_norm": 0.273449200425047, "learning_rate": 5.723881045995216e-05, "loss": 0.9509, "step": 4661 }, { "epoch": 2.1123697326687814, "grad_norm": 0.28505956262663196, "learning_rate": 5.722739134440489e-05, "loss": 0.9458, "step": 4662 }, { "epoch": 2.1128228364295425, "grad_norm": 0.2953835536870418, "learning_rate": 5.721597050487335e-05, "loss": 0.9471, "step": 4663 }, { "epoch": 2.1132759401903036, "grad_norm": 0.29254993008808156, "learning_rate": 5.7204547942500465e-05, "loss": 0.944, "step": 4664 }, { "epoch": 2.1137290439510648, "grad_norm": 0.2913553482305547, "learning_rate": 5.71931236584293e-05, "loss": 0.9498, "step": 4665 }, { "epoch": 2.114182147711826, "grad_norm": 0.26516526092973847, "learning_rate": 5.718169765380312e-05, "loss": 0.9441, "step": 4666 }, { "epoch": 2.114635251472587, "grad_norm": 0.26051768550767285, "learning_rate": 5.717026992976535e-05, "loss": 0.9335, "step": 4667 }, { "epoch": 2.1150883552333486, "grad_norm": 0.2916891264052074, "learning_rate": 5.715884048745958e-05, "loss": 0.9516, "step": 4668 }, { "epoch": 2.1155414589941097, "grad_norm": 0.32061350116552156, "learning_rate": 5.714740932802958e-05, "loss": 0.9476, "step": 4669 }, { "epoch": 2.115994562754871, "grad_norm": 0.32539626569670993, "learning_rate": 5.713597645261932e-05, "loss": 0.9675, "step": 4670 }, { "epoch": 2.116447666515632, "grad_norm": 0.289029421021565, "learning_rate": 5.712454186237287e-05, "loss": 0.9376, "step": 4671 }, { "epoch": 2.116900770276393, "grad_norm": 0.2454667265914951, "learning_rate": 5.7113105558434545e-05, "loss": 0.9565, "step": 4672 }, { "epoch": 2.1173538740371547, "grad_norm": 0.24911392090708587, "learning_rate": 5.71016675419488e-05, "loss": 0.948, "step": 4673 }, { "epoch": 2.117806977797916, "grad_norm": 0.2586905597676481, "learning_rate": 5.709022781406026e-05, "loss": 0.9524, "step": 4674 }, { "epoch": 2.118260081558677, "grad_norm": 0.2884664126889799, "learning_rate": 5.707878637591371e-05, "loss": 0.9365, "step": 4675 }, { "epoch": 2.118713185319438, "grad_norm": 0.3020472089281962, "learning_rate": 5.706734322865414e-05, "loss": 0.9532, "step": 4676 }, { "epoch": 2.119166289080199, "grad_norm": 0.3236378773085727, "learning_rate": 5.7055898373426684e-05, "loss": 0.9632, "step": 4677 }, { "epoch": 2.119619392840961, "grad_norm": 0.33756721298546055, "learning_rate": 5.704445181137665e-05, "loss": 0.9319, "step": 4678 }, { "epoch": 2.120072496601722, "grad_norm": 0.3133168598573443, "learning_rate": 5.703300354364953e-05, "loss": 0.9564, "step": 4679 }, { "epoch": 2.120525600362483, "grad_norm": 0.2506457026208356, "learning_rate": 5.7021553571390985e-05, "loss": 0.9776, "step": 4680 }, { "epoch": 2.120978704123244, "grad_norm": 0.23414369755650052, "learning_rate": 5.701010189574683e-05, "loss": 0.9329, "step": 4681 }, { "epoch": 2.1214318078840053, "grad_norm": 0.21653570133106445, "learning_rate": 5.699864851786306e-05, "loss": 0.9509, "step": 4682 }, { "epoch": 2.1218849116447664, "grad_norm": 0.22780240265949533, "learning_rate": 5.698719343888583e-05, "loss": 0.9706, "step": 4683 }, { "epoch": 2.122338015405528, "grad_norm": 0.2500555991756223, "learning_rate": 5.6975736659961495e-05, "loss": 0.9558, "step": 4684 }, { "epoch": 2.122791119166289, "grad_norm": 0.29972419044942145, "learning_rate": 5.696427818223654e-05, "loss": 0.9487, "step": 4685 }, { "epoch": 2.1232442229270503, "grad_norm": 0.31333648309793255, "learning_rate": 5.6952818006857684e-05, "loss": 0.9493, "step": 4686 }, { "epoch": 2.1236973266878114, "grad_norm": 0.2812266570485444, "learning_rate": 5.694135613497172e-05, "loss": 0.9483, "step": 4687 }, { "epoch": 2.1241504304485725, "grad_norm": 0.23486344171952023, "learning_rate": 5.6929892567725686e-05, "loss": 0.9469, "step": 4688 }, { "epoch": 2.124603534209334, "grad_norm": 0.1982667241555384, "learning_rate": 5.691842730626677e-05, "loss": 0.9787, "step": 4689 }, { "epoch": 2.1250566379700953, "grad_norm": 0.18536405609604012, "learning_rate": 5.690696035174233e-05, "loss": 0.9143, "step": 4690 }, { "epoch": 2.1255097417308564, "grad_norm": 0.1993662908784922, "learning_rate": 5.689549170529989e-05, "loss": 0.9402, "step": 4691 }, { "epoch": 2.1259628454916175, "grad_norm": 0.2119725947180386, "learning_rate": 5.688402136808712e-05, "loss": 0.9547, "step": 4692 }, { "epoch": 2.1264159492523786, "grad_norm": 0.5225288740893262, "learning_rate": 5.6872549341251924e-05, "loss": 0.9564, "step": 4693 }, { "epoch": 2.1268690530131398, "grad_norm": 0.21100926289165028, "learning_rate": 5.686107562594229e-05, "loss": 0.9383, "step": 4694 }, { "epoch": 2.1273221567739014, "grad_norm": 0.2095084086901534, "learning_rate": 5.6849600223306445e-05, "loss": 0.9513, "step": 4695 }, { "epoch": 2.1277752605346625, "grad_norm": 0.18915125814481043, "learning_rate": 5.6838123134492755e-05, "loss": 0.9598, "step": 4696 }, { "epoch": 2.1282283642954236, "grad_norm": 0.18423731862099074, "learning_rate": 5.682664436064976e-05, "loss": 0.9693, "step": 4697 }, { "epoch": 2.1286814680561847, "grad_norm": 0.1870235124228652, "learning_rate": 5.681516390292616e-05, "loss": 0.9729, "step": 4698 }, { "epoch": 2.129134571816946, "grad_norm": 0.19178792263010955, "learning_rate": 5.6803681762470844e-05, "loss": 0.9522, "step": 4699 }, { "epoch": 2.1295876755777075, "grad_norm": 0.2035621645384094, "learning_rate": 5.679219794043283e-05, "loss": 0.947, "step": 4700 }, { "epoch": 2.1300407793384686, "grad_norm": 0.24552993192770436, "learning_rate": 5.6780712437961356e-05, "loss": 0.9446, "step": 4701 }, { "epoch": 2.1304938830992297, "grad_norm": 0.26183816061660947, "learning_rate": 5.676922525620579e-05, "loss": 0.9523, "step": 4702 }, { "epoch": 2.130946986859991, "grad_norm": 0.2693467392678051, "learning_rate": 5.675773639631568e-05, "loss": 0.9542, "step": 4703 }, { "epoch": 2.131400090620752, "grad_norm": 0.286875934303222, "learning_rate": 5.674624585944076e-05, "loss": 0.937, "step": 4704 }, { "epoch": 2.1318531943815135, "grad_norm": 0.3237303390649252, "learning_rate": 5.673475364673089e-05, "loss": 0.9465, "step": 4705 }, { "epoch": 2.1323062981422747, "grad_norm": 0.3350982151368362, "learning_rate": 5.6723259759336124e-05, "loss": 0.9806, "step": 4706 }, { "epoch": 2.132759401903036, "grad_norm": 0.3660349477893495, "learning_rate": 5.67117641984067e-05, "loss": 0.9435, "step": 4707 }, { "epoch": 2.133212505663797, "grad_norm": 0.34428574654031024, "learning_rate": 5.6700266965092986e-05, "loss": 0.9419, "step": 4708 }, { "epoch": 2.133665609424558, "grad_norm": 0.38061672315238243, "learning_rate": 5.6688768060545553e-05, "loss": 0.9305, "step": 4709 }, { "epoch": 2.1341187131853196, "grad_norm": 0.4287753028495363, "learning_rate": 5.6677267485915104e-05, "loss": 0.9553, "step": 4710 }, { "epoch": 2.134571816946081, "grad_norm": 0.4535181819619256, "learning_rate": 5.666576524235254e-05, "loss": 0.9427, "step": 4711 }, { "epoch": 2.135024920706842, "grad_norm": 0.37676054371483786, "learning_rate": 5.665426133100892e-05, "loss": 0.9613, "step": 4712 }, { "epoch": 2.135478024467603, "grad_norm": 0.2844855779195375, "learning_rate": 5.6642755753035454e-05, "loss": 0.9439, "step": 4713 }, { "epoch": 2.135931128228364, "grad_norm": 0.23721141654476863, "learning_rate": 5.6631248509583544e-05, "loss": 0.9646, "step": 4714 }, { "epoch": 2.1363842319891253, "grad_norm": 0.21527290670940125, "learning_rate": 5.661973960180474e-05, "loss": 0.9476, "step": 4715 }, { "epoch": 2.136837335749887, "grad_norm": 0.221706350275383, "learning_rate": 5.660822903085077e-05, "loss": 0.9336, "step": 4716 }, { "epoch": 2.137290439510648, "grad_norm": 0.2399149749900436, "learning_rate": 5.659671679787352e-05, "loss": 0.9492, "step": 4717 }, { "epoch": 2.137743543271409, "grad_norm": 0.3451067368352172, "learning_rate": 5.6585202904025024e-05, "loss": 0.9437, "step": 4718 }, { "epoch": 2.1381966470321703, "grad_norm": 0.41808280290206806, "learning_rate": 5.657368735045754e-05, "loss": 0.9601, "step": 4719 }, { "epoch": 2.1386497507929314, "grad_norm": 0.42668333328864566, "learning_rate": 5.6562170138323424e-05, "loss": 0.9744, "step": 4720 }, { "epoch": 2.139102854553693, "grad_norm": 0.3405915194901683, "learning_rate": 5.6550651268775257e-05, "loss": 0.9592, "step": 4721 }, { "epoch": 2.139555958314454, "grad_norm": 0.30302855999924666, "learning_rate": 5.653913074296574e-05, "loss": 0.9528, "step": 4722 }, { "epoch": 2.1400090620752152, "grad_norm": 0.26575213428257, "learning_rate": 5.6527608562047756e-05, "loss": 0.967, "step": 4723 }, { "epoch": 2.1404621658359764, "grad_norm": 0.2519495875703834, "learning_rate": 5.651608472717438e-05, "loss": 0.979, "step": 4724 }, { "epoch": 2.1409152695967375, "grad_norm": 0.17858027199136048, "learning_rate": 5.650455923949881e-05, "loss": 0.9469, "step": 4725 }, { "epoch": 2.1413683733574986, "grad_norm": 0.1996807740005248, "learning_rate": 5.6493032100174416e-05, "loss": 0.9869, "step": 4726 }, { "epoch": 2.14182147711826, "grad_norm": 0.1818163600766244, "learning_rate": 5.648150331035478e-05, "loss": 0.9545, "step": 4727 }, { "epoch": 2.1422745808790213, "grad_norm": 0.2226471876351401, "learning_rate": 5.646997287119357e-05, "loss": 0.9382, "step": 4728 }, { "epoch": 2.1427276846397825, "grad_norm": 0.23855426812423827, "learning_rate": 5.6458440783844693e-05, "loss": 0.9726, "step": 4729 }, { "epoch": 2.1431807884005436, "grad_norm": 0.2627274251577347, "learning_rate": 5.64469070494622e-05, "loss": 0.9441, "step": 4730 }, { "epoch": 2.1436338921613047, "grad_norm": 0.33762830470557365, "learning_rate": 5.643537166920027e-05, "loss": 0.9368, "step": 4731 }, { "epoch": 2.1440869959220663, "grad_norm": 0.3419043077215317, "learning_rate": 5.642383464421328e-05, "loss": 0.9369, "step": 4732 }, { "epoch": 2.1445400996828274, "grad_norm": 0.32037350492017164, "learning_rate": 5.6412295975655796e-05, "loss": 0.9512, "step": 4733 }, { "epoch": 2.1449932034435886, "grad_norm": 0.33249183004706095, "learning_rate": 5.640075566468248e-05, "loss": 0.9672, "step": 4734 }, { "epoch": 2.1454463072043497, "grad_norm": 0.32706192061469985, "learning_rate": 5.6389213712448215e-05, "loss": 0.935, "step": 4735 }, { "epoch": 2.145899410965111, "grad_norm": 0.31528650042438183, "learning_rate": 5.637767012010804e-05, "loss": 0.9645, "step": 4736 }, { "epoch": 2.1463525147258724, "grad_norm": 0.30372770019362916, "learning_rate": 5.636612488881712e-05, "loss": 0.9335, "step": 4737 }, { "epoch": 2.1468056184866335, "grad_norm": 0.28064044384070486, "learning_rate": 5.6354578019730844e-05, "loss": 0.9384, "step": 4738 }, { "epoch": 2.1472587222473947, "grad_norm": 0.26646139949086634, "learning_rate": 5.634302951400472e-05, "loss": 0.9723, "step": 4739 }, { "epoch": 2.147711826008156, "grad_norm": 0.2799278761041008, "learning_rate": 5.633147937279443e-05, "loss": 0.9445, "step": 4740 }, { "epoch": 2.148164929768917, "grad_norm": 0.2973473050620337, "learning_rate": 5.6319927597255816e-05, "loss": 0.9499, "step": 4741 }, { "epoch": 2.1486180335296785, "grad_norm": 0.2588303498360831, "learning_rate": 5.6308374188544926e-05, "loss": 0.9493, "step": 4742 }, { "epoch": 2.1490711372904396, "grad_norm": 0.23090683564226733, "learning_rate": 5.62968191478179e-05, "loss": 0.9142, "step": 4743 }, { "epoch": 2.1495242410512008, "grad_norm": 0.22151968160289254, "learning_rate": 5.628526247623108e-05, "loss": 0.9601, "step": 4744 }, { "epoch": 2.149977344811962, "grad_norm": 0.21121567880134223, "learning_rate": 5.6273704174940975e-05, "loss": 0.9421, "step": 4745 }, { "epoch": 2.150430448572723, "grad_norm": 0.23607459544284892, "learning_rate": 5.626214424510426e-05, "loss": 0.9593, "step": 4746 }, { "epoch": 2.150883552333484, "grad_norm": 0.2791118048997649, "learning_rate": 5.625058268787775e-05, "loss": 0.9609, "step": 4747 }, { "epoch": 2.1513366560942457, "grad_norm": 0.27632494874880875, "learning_rate": 5.623901950441844e-05, "loss": 0.9338, "step": 4748 }, { "epoch": 2.151789759855007, "grad_norm": 0.24196462549322056, "learning_rate": 5.622745469588349e-05, "loss": 0.936, "step": 4749 }, { "epoch": 2.152242863615768, "grad_norm": 0.21593336412931657, "learning_rate": 5.62158882634302e-05, "loss": 0.9467, "step": 4750 }, { "epoch": 2.152695967376529, "grad_norm": 0.19735746823369285, "learning_rate": 5.620432020821609e-05, "loss": 0.9416, "step": 4751 }, { "epoch": 2.1531490711372903, "grad_norm": 0.2194279123718849, "learning_rate": 5.619275053139875e-05, "loss": 0.9532, "step": 4752 }, { "epoch": 2.153602174898052, "grad_norm": 0.2227044628409108, "learning_rate": 5.6181179234136006e-05, "loss": 0.9515, "step": 4753 }, { "epoch": 2.154055278658813, "grad_norm": 0.18373498527752855, "learning_rate": 5.616960631758584e-05, "loss": 0.9515, "step": 4754 }, { "epoch": 2.154508382419574, "grad_norm": 0.21283662701408934, "learning_rate": 5.615803178290636e-05, "loss": 0.9351, "step": 4755 }, { "epoch": 2.1549614861803352, "grad_norm": 0.28016584959986746, "learning_rate": 5.614645563125585e-05, "loss": 0.9696, "step": 4756 }, { "epoch": 2.1554145899410964, "grad_norm": 0.28785022552356937, "learning_rate": 5.6134877863792775e-05, "loss": 0.9566, "step": 4757 }, { "epoch": 2.1558676937018575, "grad_norm": 0.22177846346257857, "learning_rate": 5.6123298481675754e-05, "loss": 0.9358, "step": 4758 }, { "epoch": 2.156320797462619, "grad_norm": 0.22367723302882644, "learning_rate": 5.6111717486063555e-05, "loss": 0.9519, "step": 4759 }, { "epoch": 2.15677390122338, "grad_norm": 0.23955813016617425, "learning_rate": 5.610013487811512e-05, "loss": 0.9367, "step": 4760 }, { "epoch": 2.1572270049841413, "grad_norm": 0.2978111500727643, "learning_rate": 5.608855065898953e-05, "loss": 0.9537, "step": 4761 }, { "epoch": 2.1576801087449025, "grad_norm": 0.3669750364783911, "learning_rate": 5.6076964829846066e-05, "loss": 0.9688, "step": 4762 }, { "epoch": 2.1581332125056636, "grad_norm": 0.43906267719229636, "learning_rate": 5.6065377391844134e-05, "loss": 0.9536, "step": 4763 }, { "epoch": 2.158586316266425, "grad_norm": 0.43986953270414897, "learning_rate": 5.605378834614331e-05, "loss": 0.9411, "step": 4764 }, { "epoch": 2.1590394200271863, "grad_norm": 0.3926575106138099, "learning_rate": 5.6042197693903345e-05, "loss": 0.9564, "step": 4765 }, { "epoch": 2.1594925237879474, "grad_norm": 0.333090827686041, "learning_rate": 5.6030605436284146e-05, "loss": 0.935, "step": 4766 }, { "epoch": 2.1599456275487086, "grad_norm": 0.2962839683801935, "learning_rate": 5.6019011574445775e-05, "loss": 0.9471, "step": 4767 }, { "epoch": 2.1603987313094697, "grad_norm": 0.2977346516791384, "learning_rate": 5.6007416109548464e-05, "loss": 0.9418, "step": 4768 }, { "epoch": 2.1608518350702313, "grad_norm": 0.31344055781249325, "learning_rate": 5.599581904275258e-05, "loss": 0.9784, "step": 4769 }, { "epoch": 2.1613049388309924, "grad_norm": 0.3036847606308525, "learning_rate": 5.5984220375218675e-05, "loss": 0.9409, "step": 4770 }, { "epoch": 2.1617580425917535, "grad_norm": 0.27903263395336386, "learning_rate": 5.5972620108107453e-05, "loss": 0.9433, "step": 4771 }, { "epoch": 2.1622111463525147, "grad_norm": 0.22020067473538965, "learning_rate": 5.5961018242579785e-05, "loss": 0.9497, "step": 4772 }, { "epoch": 2.162664250113276, "grad_norm": 0.20617714471191573, "learning_rate": 5.594941477979668e-05, "loss": 0.9668, "step": 4773 }, { "epoch": 2.1631173538740374, "grad_norm": 0.229521442763799, "learning_rate": 5.593780972091934e-05, "loss": 0.9526, "step": 4774 }, { "epoch": 2.1635704576347985, "grad_norm": 0.24673283734665397, "learning_rate": 5.5926203067109114e-05, "loss": 0.9666, "step": 4775 }, { "epoch": 2.1640235613955596, "grad_norm": 0.26406164146142513, "learning_rate": 5.591459481952749e-05, "loss": 0.9385, "step": 4776 }, { "epoch": 2.1644766651563208, "grad_norm": 0.2570495031667013, "learning_rate": 5.5902984979336125e-05, "loss": 0.9642, "step": 4777 }, { "epoch": 2.164929768917082, "grad_norm": 0.21573858336570717, "learning_rate": 5.589137354769688e-05, "loss": 0.9651, "step": 4778 }, { "epoch": 2.165382872677843, "grad_norm": 0.25880767895122664, "learning_rate": 5.58797605257717e-05, "loss": 0.9521, "step": 4779 }, { "epoch": 2.1658359764386046, "grad_norm": 0.3533377910343337, "learning_rate": 5.5868145914722736e-05, "loss": 0.9557, "step": 4780 }, { "epoch": 2.1662890801993657, "grad_norm": 0.4321257184339892, "learning_rate": 5.5856529715712295e-05, "loss": 0.9614, "step": 4781 }, { "epoch": 2.166742183960127, "grad_norm": 0.4408066334521399, "learning_rate": 5.584491192990284e-05, "loss": 0.9501, "step": 4782 }, { "epoch": 2.167195287720888, "grad_norm": 0.39720444999542126, "learning_rate": 5.583329255845697e-05, "loss": 0.9424, "step": 4783 }, { "epoch": 2.167648391481649, "grad_norm": 0.37785828502122626, "learning_rate": 5.582167160253748e-05, "loss": 0.9372, "step": 4784 }, { "epoch": 2.1681014952424107, "grad_norm": 0.3580196316150505, "learning_rate": 5.58100490633073e-05, "loss": 0.9517, "step": 4785 }, { "epoch": 2.168554599003172, "grad_norm": 0.32679825253463296, "learning_rate": 5.579842494192953e-05, "loss": 0.9697, "step": 4786 }, { "epoch": 2.169007702763933, "grad_norm": 0.3225909015106907, "learning_rate": 5.578679923956741e-05, "loss": 0.9508, "step": 4787 }, { "epoch": 2.169460806524694, "grad_norm": 0.28645615105587385, "learning_rate": 5.577517195738436e-05, "loss": 0.9581, "step": 4788 }, { "epoch": 2.169913910285455, "grad_norm": 0.26361579878780295, "learning_rate": 5.576354309654394e-05, "loss": 0.9494, "step": 4789 }, { "epoch": 2.170367014046217, "grad_norm": 0.2631601563204481, "learning_rate": 5.575191265820989e-05, "loss": 0.9473, "step": 4790 }, { "epoch": 2.170820117806978, "grad_norm": 0.2459620207157422, "learning_rate": 5.574028064354608e-05, "loss": 0.9715, "step": 4791 }, { "epoch": 2.171273221567739, "grad_norm": 0.22611336824736664, "learning_rate": 5.5728647053716555e-05, "loss": 0.9794, "step": 4792 }, { "epoch": 2.1717263253285, "grad_norm": 0.24476094098561574, "learning_rate": 5.571701188988552e-05, "loss": 0.9608, "step": 4793 }, { "epoch": 2.1721794290892613, "grad_norm": 0.2504439975335515, "learning_rate": 5.570537515321735e-05, "loss": 0.9464, "step": 4794 }, { "epoch": 2.172632532850023, "grad_norm": 0.27002295242126095, "learning_rate": 5.569373684487653e-05, "loss": 0.9433, "step": 4795 }, { "epoch": 2.173085636610784, "grad_norm": 0.25985668284580543, "learning_rate": 5.5682096966027736e-05, "loss": 0.9602, "step": 4796 }, { "epoch": 2.173538740371545, "grad_norm": 0.24864234994576023, "learning_rate": 5.567045551783579e-05, "loss": 0.9513, "step": 4797 }, { "epoch": 2.1739918441323063, "grad_norm": 0.277962164081054, "learning_rate": 5.565881250146571e-05, "loss": 0.9477, "step": 4798 }, { "epoch": 2.1744449478930674, "grad_norm": 0.2946023878922937, "learning_rate": 5.564716791808263e-05, "loss": 0.978, "step": 4799 }, { "epoch": 2.1748980516538285, "grad_norm": 0.28976968224934635, "learning_rate": 5.563552176885183e-05, "loss": 0.9542, "step": 4800 }, { "epoch": 2.17535115541459, "grad_norm": 0.28128107513304973, "learning_rate": 5.5623874054938784e-05, "loss": 0.9367, "step": 4801 }, { "epoch": 2.1758042591753513, "grad_norm": 0.28685950561293055, "learning_rate": 5.561222477750909e-05, "loss": 0.9401, "step": 4802 }, { "epoch": 2.1762573629361124, "grad_norm": 0.27893443821663094, "learning_rate": 5.560057393772852e-05, "loss": 0.9527, "step": 4803 }, { "epoch": 2.1767104666968735, "grad_norm": 0.27036437259354096, "learning_rate": 5.558892153676303e-05, "loss": 0.9465, "step": 4804 }, { "epoch": 2.1771635704576346, "grad_norm": 0.2566323202014866, "learning_rate": 5.5577267575778666e-05, "loss": 0.9422, "step": 4805 }, { "epoch": 2.177616674218396, "grad_norm": 0.2382847880768328, "learning_rate": 5.5565612055941676e-05, "loss": 0.9545, "step": 4806 }, { "epoch": 2.1780697779791574, "grad_norm": 0.2668408557138005, "learning_rate": 5.5553954978418455e-05, "loss": 0.9512, "step": 4807 }, { "epoch": 2.1785228817399185, "grad_norm": 0.22852347094878311, "learning_rate": 5.554229634437556e-05, "loss": 0.942, "step": 4808 }, { "epoch": 2.1789759855006796, "grad_norm": 0.21066995236430722, "learning_rate": 5.553063615497969e-05, "loss": 0.9372, "step": 4809 }, { "epoch": 2.1794290892614407, "grad_norm": 0.23598033113945763, "learning_rate": 5.551897441139772e-05, "loss": 0.9394, "step": 4810 }, { "epoch": 2.179882193022202, "grad_norm": 0.2449392024373374, "learning_rate": 5.550731111479664e-05, "loss": 0.9529, "step": 4811 }, { "epoch": 2.1803352967829635, "grad_norm": 0.251720714649895, "learning_rate": 5.5495646266343666e-05, "loss": 0.9563, "step": 4812 }, { "epoch": 2.1807884005437246, "grad_norm": 0.25488782435169133, "learning_rate": 5.548397986720608e-05, "loss": 0.9413, "step": 4813 }, { "epoch": 2.1812415043044857, "grad_norm": 0.1851768668764369, "learning_rate": 5.5472311918551386e-05, "loss": 0.9757, "step": 4814 }, { "epoch": 2.181694608065247, "grad_norm": 0.2185099672106625, "learning_rate": 5.546064242154723e-05, "loss": 0.9769, "step": 4815 }, { "epoch": 2.182147711826008, "grad_norm": 0.26458991292815687, "learning_rate": 5.544897137736138e-05, "loss": 0.9428, "step": 4816 }, { "epoch": 2.1826008155867695, "grad_norm": 0.23074183365647613, "learning_rate": 5.543729878716181e-05, "loss": 0.9411, "step": 4817 }, { "epoch": 2.1830539193475307, "grad_norm": 0.23946129995857762, "learning_rate": 5.542562465211662e-05, "loss": 0.9522, "step": 4818 }, { "epoch": 2.183507023108292, "grad_norm": 0.2941093337144373, "learning_rate": 5.541394897339405e-05, "loss": 0.9622, "step": 4819 }, { "epoch": 2.183960126869053, "grad_norm": 0.27762084368147827, "learning_rate": 5.540227175216252e-05, "loss": 0.9491, "step": 4820 }, { "epoch": 2.184413230629814, "grad_norm": 0.2657108733560484, "learning_rate": 5.539059298959061e-05, "loss": 0.9372, "step": 4821 }, { "epoch": 2.1848663343905756, "grad_norm": 0.25613610507763, "learning_rate": 5.5378912686847015e-05, "loss": 0.9637, "step": 4822 }, { "epoch": 2.185319438151337, "grad_norm": 0.21279184111194796, "learning_rate": 5.536723084510063e-05, "loss": 0.9345, "step": 4823 }, { "epoch": 2.185772541912098, "grad_norm": 0.21004837612799956, "learning_rate": 5.535554746552047e-05, "loss": 0.9526, "step": 4824 }, { "epoch": 2.186225645672859, "grad_norm": 0.2140741631136418, "learning_rate": 5.5343862549275726e-05, "loss": 0.9613, "step": 4825 }, { "epoch": 2.18667874943362, "grad_norm": 0.22910127187899362, "learning_rate": 5.533217609753573e-05, "loss": 0.9406, "step": 4826 }, { "epoch": 2.1871318531943817, "grad_norm": 0.286039225376932, "learning_rate": 5.532048811146998e-05, "loss": 0.9593, "step": 4827 }, { "epoch": 2.187584956955143, "grad_norm": 0.3401356910382545, "learning_rate": 5.53087985922481e-05, "loss": 0.9607, "step": 4828 }, { "epoch": 2.188038060715904, "grad_norm": 0.3726354214088426, "learning_rate": 5.529710754103992e-05, "loss": 0.9613, "step": 4829 }, { "epoch": 2.188491164476665, "grad_norm": 0.43099495007934346, "learning_rate": 5.528541495901536e-05, "loss": 0.9525, "step": 4830 }, { "epoch": 2.1889442682374263, "grad_norm": 0.46127959925849493, "learning_rate": 5.5273720847344534e-05, "loss": 0.9431, "step": 4831 }, { "epoch": 2.1893973719981874, "grad_norm": 0.3958831455763917, "learning_rate": 5.52620252071977e-05, "loss": 0.9387, "step": 4832 }, { "epoch": 2.189850475758949, "grad_norm": 0.31154027901166353, "learning_rate": 5.525032803974525e-05, "loss": 0.9647, "step": 4833 }, { "epoch": 2.19030357951971, "grad_norm": 0.24593904638388994, "learning_rate": 5.523862934615778e-05, "loss": 0.9573, "step": 4834 }, { "epoch": 2.1907566832804712, "grad_norm": 0.22469507811502482, "learning_rate": 5.5226929127605965e-05, "loss": 0.9525, "step": 4835 }, { "epoch": 2.1912097870412324, "grad_norm": 0.2474998360125758, "learning_rate": 5.52152273852607e-05, "loss": 0.9579, "step": 4836 }, { "epoch": 2.1916628908019935, "grad_norm": 0.21124507737613987, "learning_rate": 5.520352412029301e-05, "loss": 0.9346, "step": 4837 }, { "epoch": 2.192115994562755, "grad_norm": 0.25527913496353855, "learning_rate": 5.519181933387405e-05, "loss": 0.947, "step": 4838 }, { "epoch": 2.192569098323516, "grad_norm": 0.3331875332956557, "learning_rate": 5.518011302717515e-05, "loss": 0.9675, "step": 4839 }, { "epoch": 2.1930222020842773, "grad_norm": 0.32629919701094945, "learning_rate": 5.5168405201367784e-05, "loss": 0.9472, "step": 4840 }, { "epoch": 2.1934753058450385, "grad_norm": 0.25652632634867595, "learning_rate": 5.515669585762359e-05, "loss": 0.9367, "step": 4841 }, { "epoch": 2.1939284096057996, "grad_norm": 0.22895463728974122, "learning_rate": 5.5144984997114335e-05, "loss": 0.9449, "step": 4842 }, { "epoch": 2.1943815133665607, "grad_norm": 0.20519814986056448, "learning_rate": 5.5133272621011956e-05, "loss": 0.9672, "step": 4843 }, { "epoch": 2.1948346171273223, "grad_norm": 0.20411916707037608, "learning_rate": 5.512155873048855e-05, "loss": 0.9467, "step": 4844 }, { "epoch": 2.1952877208880834, "grad_norm": 0.20078292658560318, "learning_rate": 5.510984332671633e-05, "loss": 0.9346, "step": 4845 }, { "epoch": 2.1957408246488446, "grad_norm": 0.22302809664362835, "learning_rate": 5.509812641086771e-05, "loss": 0.9364, "step": 4846 }, { "epoch": 2.1961939284096057, "grad_norm": 0.2331891528952829, "learning_rate": 5.508640798411522e-05, "loss": 0.9269, "step": 4847 }, { "epoch": 2.196647032170367, "grad_norm": 0.22181164235458, "learning_rate": 5.5074688047631526e-05, "loss": 0.9361, "step": 4848 }, { "epoch": 2.1971001359311284, "grad_norm": 0.2017219861032075, "learning_rate": 5.506296660258949e-05, "loss": 0.9578, "step": 4849 }, { "epoch": 2.1975532396918895, "grad_norm": 0.19729526039370326, "learning_rate": 5.505124365016211e-05, "loss": 0.9374, "step": 4850 }, { "epoch": 2.1980063434526507, "grad_norm": 0.22598191912815666, "learning_rate": 5.503951919152252e-05, "loss": 0.9445, "step": 4851 }, { "epoch": 2.198459447213412, "grad_norm": 0.26599093415461456, "learning_rate": 5.502779322784399e-05, "loss": 0.9583, "step": 4852 }, { "epoch": 2.198912550974173, "grad_norm": 0.27696845034485573, "learning_rate": 5.501606576030001e-05, "loss": 0.9362, "step": 4853 }, { "epoch": 2.1993656547349345, "grad_norm": 0.30810787815216145, "learning_rate": 5.500433679006414e-05, "loss": 0.9485, "step": 4854 }, { "epoch": 2.1998187584956956, "grad_norm": 0.3510854217800639, "learning_rate": 5.499260631831014e-05, "loss": 0.9391, "step": 4855 }, { "epoch": 2.2002718622564568, "grad_norm": 0.35069293485005293, "learning_rate": 5.498087434621189e-05, "loss": 0.9689, "step": 4856 }, { "epoch": 2.200724966017218, "grad_norm": 0.2793168274922381, "learning_rate": 5.496914087494347e-05, "loss": 0.9215, "step": 4857 }, { "epoch": 2.201178069777979, "grad_norm": 0.23311507403826856, "learning_rate": 5.495740590567903e-05, "loss": 0.9474, "step": 4858 }, { "epoch": 2.2016311735387406, "grad_norm": 0.20456554103152366, "learning_rate": 5.494566943959294e-05, "loss": 0.9404, "step": 4859 }, { "epoch": 2.2020842772995017, "grad_norm": 0.22813396621461413, "learning_rate": 5.493393147785968e-05, "loss": 0.9639, "step": 4860 }, { "epoch": 2.202537381060263, "grad_norm": 0.23939872942272933, "learning_rate": 5.492219202165392e-05, "loss": 0.9345, "step": 4861 }, { "epoch": 2.202990484821024, "grad_norm": 0.24028532180603504, "learning_rate": 5.491045107215043e-05, "loss": 0.9437, "step": 4862 }, { "epoch": 2.203443588581785, "grad_norm": 0.2856999890499398, "learning_rate": 5.489870863052416e-05, "loss": 0.9449, "step": 4863 }, { "epoch": 2.2038966923425463, "grad_norm": 0.3149322494092849, "learning_rate": 5.488696469795022e-05, "loss": 0.9704, "step": 4864 }, { "epoch": 2.204349796103308, "grad_norm": 0.3133781091101693, "learning_rate": 5.487521927560384e-05, "loss": 0.953, "step": 4865 }, { "epoch": 2.204802899864069, "grad_norm": 0.33417741067815726, "learning_rate": 5.486347236466039e-05, "loss": 0.95, "step": 4866 }, { "epoch": 2.20525600362483, "grad_norm": 0.35366616337660334, "learning_rate": 5.4851723966295435e-05, "loss": 0.9737, "step": 4867 }, { "epoch": 2.2057091073855912, "grad_norm": 0.3400047532728363, "learning_rate": 5.4839974081684665e-05, "loss": 0.9472, "step": 4868 }, { "epoch": 2.2061622111463524, "grad_norm": 0.3063552496428732, "learning_rate": 5.482822271200389e-05, "loss": 0.9424, "step": 4869 }, { "epoch": 2.206615314907114, "grad_norm": 0.27764200486993107, "learning_rate": 5.481646985842914e-05, "loss": 0.9565, "step": 4870 }, { "epoch": 2.207068418667875, "grad_norm": 0.268951009578104, "learning_rate": 5.4804715522136515e-05, "loss": 0.9456, "step": 4871 }, { "epoch": 2.207521522428636, "grad_norm": 0.19583199254860648, "learning_rate": 5.4792959704302315e-05, "loss": 0.948, "step": 4872 }, { "epoch": 2.2079746261893973, "grad_norm": 0.21732552125620097, "learning_rate": 5.478120240610297e-05, "loss": 0.9332, "step": 4873 }, { "epoch": 2.2084277299501585, "grad_norm": 0.2446584794151823, "learning_rate": 5.476944362871505e-05, "loss": 0.9772, "step": 4874 }, { "epoch": 2.2088808337109196, "grad_norm": 0.22984787175791396, "learning_rate": 5.475768337331529e-05, "loss": 0.9452, "step": 4875 }, { "epoch": 2.209333937471681, "grad_norm": 0.20389082273181114, "learning_rate": 5.474592164108056e-05, "loss": 0.9212, "step": 4876 }, { "epoch": 2.2097870412324423, "grad_norm": 0.20813344678581552, "learning_rate": 5.473415843318789e-05, "loss": 0.934, "step": 4877 }, { "epoch": 2.2102401449932034, "grad_norm": 0.24202110551536885, "learning_rate": 5.472239375081445e-05, "loss": 0.9433, "step": 4878 }, { "epoch": 2.2106932487539646, "grad_norm": 0.2425627375508994, "learning_rate": 5.471062759513755e-05, "loss": 0.9542, "step": 4879 }, { "epoch": 2.2111463525147257, "grad_norm": 0.25308730442117205, "learning_rate": 5.469885996733467e-05, "loss": 0.9414, "step": 4880 }, { "epoch": 2.2115994562754873, "grad_norm": 0.23117045223577545, "learning_rate": 5.4687090868583417e-05, "loss": 0.9347, "step": 4881 }, { "epoch": 2.2120525600362484, "grad_norm": 0.23670736320769506, "learning_rate": 5.4675320300061546e-05, "loss": 0.9374, "step": 4882 }, { "epoch": 2.2125056637970095, "grad_norm": 0.25690139942012485, "learning_rate": 5.466354826294698e-05, "loss": 0.9186, "step": 4883 }, { "epoch": 2.2129587675577707, "grad_norm": 0.24037275707915273, "learning_rate": 5.465177475841775e-05, "loss": 0.9494, "step": 4884 }, { "epoch": 2.213411871318532, "grad_norm": 0.25141050427659883, "learning_rate": 5.463999978765207e-05, "loss": 0.9623, "step": 4885 }, { "epoch": 2.2138649750792934, "grad_norm": 0.29764015828028106, "learning_rate": 5.4628223351828304e-05, "loss": 0.9347, "step": 4886 }, { "epoch": 2.2143180788400545, "grad_norm": 0.38083303237032146, "learning_rate": 5.461644545212492e-05, "loss": 0.9462, "step": 4887 }, { "epoch": 2.2147711826008156, "grad_norm": 0.4008278184741505, "learning_rate": 5.4604666089720565e-05, "loss": 0.9361, "step": 4888 }, { "epoch": 2.2152242863615768, "grad_norm": 0.3532494957630219, "learning_rate": 5.4592885265794036e-05, "loss": 0.9421, "step": 4889 }, { "epoch": 2.215677390122338, "grad_norm": 0.2946317419051612, "learning_rate": 5.4581102981524266e-05, "loss": 0.9576, "step": 4890 }, { "epoch": 2.2161304938830995, "grad_norm": 0.23217090176438246, "learning_rate": 5.4569319238090334e-05, "loss": 0.9639, "step": 4891 }, { "epoch": 2.2165835976438606, "grad_norm": 0.20424333913731604, "learning_rate": 5.455753403667146e-05, "loss": 0.9538, "step": 4892 }, { "epoch": 2.2170367014046217, "grad_norm": 0.22933830110176728, "learning_rate": 5.454574737844701e-05, "loss": 0.9423, "step": 4893 }, { "epoch": 2.217489805165383, "grad_norm": 0.22530275126343383, "learning_rate": 5.453395926459653e-05, "loss": 0.9587, "step": 4894 }, { "epoch": 2.217942908926144, "grad_norm": 0.23755975515635916, "learning_rate": 5.452216969629965e-05, "loss": 0.9529, "step": 4895 }, { "epoch": 2.218396012686905, "grad_norm": 0.22261922670715684, "learning_rate": 5.45103786747362e-05, "loss": 0.9484, "step": 4896 }, { "epoch": 2.2188491164476667, "grad_norm": 0.2428339902470979, "learning_rate": 5.4498586201086125e-05, "loss": 0.9392, "step": 4897 }, { "epoch": 2.219302220208428, "grad_norm": 0.3244711263973161, "learning_rate": 5.448679227652953e-05, "loss": 0.9559, "step": 4898 }, { "epoch": 2.219755323969189, "grad_norm": 0.3502837236632482, "learning_rate": 5.447499690224665e-05, "loss": 0.942, "step": 4899 }, { "epoch": 2.22020842772995, "grad_norm": 0.3768818106258067, "learning_rate": 5.44632000794179e-05, "loss": 0.9545, "step": 4900 }, { "epoch": 2.220661531490711, "grad_norm": 0.3491420965137377, "learning_rate": 5.445140180922378e-05, "loss": 0.9552, "step": 4901 }, { "epoch": 2.221114635251473, "grad_norm": 0.2841317534074558, "learning_rate": 5.4439602092845e-05, "loss": 0.9581, "step": 4902 }, { "epoch": 2.221567739012234, "grad_norm": 0.21854110621884473, "learning_rate": 5.442780093146237e-05, "loss": 0.9428, "step": 4903 }, { "epoch": 2.222020842772995, "grad_norm": 0.20423470983918382, "learning_rate": 5.4415998326256854e-05, "loss": 0.9739, "step": 4904 }, { "epoch": 2.222473946533756, "grad_norm": 0.2242941901649575, "learning_rate": 5.440419427840957e-05, "loss": 0.9464, "step": 4905 }, { "epoch": 2.2229270502945173, "grad_norm": 0.23094186998827912, "learning_rate": 5.439238878910179e-05, "loss": 0.9433, "step": 4906 }, { "epoch": 2.2233801540552784, "grad_norm": 0.2274389291393358, "learning_rate": 5.438058185951489e-05, "loss": 0.9387, "step": 4907 }, { "epoch": 2.22383325781604, "grad_norm": 0.24812459037238924, "learning_rate": 5.4368773490830446e-05, "loss": 0.9316, "step": 4908 }, { "epoch": 2.224286361576801, "grad_norm": 0.23932098843052746, "learning_rate": 5.435696368423013e-05, "loss": 0.934, "step": 4909 }, { "epoch": 2.2247394653375623, "grad_norm": 0.24897195447240236, "learning_rate": 5.434515244089578e-05, "loss": 0.9617, "step": 4910 }, { "epoch": 2.2251925690983234, "grad_norm": 0.2821766335340834, "learning_rate": 5.433333976200937e-05, "loss": 0.9725, "step": 4911 }, { "epoch": 2.2256456728590845, "grad_norm": 0.29974035630391765, "learning_rate": 5.432152564875302e-05, "loss": 0.9689, "step": 4912 }, { "epoch": 2.226098776619846, "grad_norm": 0.2820658438435682, "learning_rate": 5.4309710102309003e-05, "loss": 0.9504, "step": 4913 }, { "epoch": 2.2265518803806073, "grad_norm": 0.2899590250582844, "learning_rate": 5.429789312385973e-05, "loss": 0.9486, "step": 4914 }, { "epoch": 2.2270049841413684, "grad_norm": 0.2897500728702314, "learning_rate": 5.428607471458774e-05, "loss": 0.9446, "step": 4915 }, { "epoch": 2.2274580879021295, "grad_norm": 0.29516523652540955, "learning_rate": 5.427425487567575e-05, "loss": 0.9365, "step": 4916 }, { "epoch": 2.2279111916628906, "grad_norm": 0.3324790364560805, "learning_rate": 5.426243360830657e-05, "loss": 0.9298, "step": 4917 }, { "epoch": 2.228364295423652, "grad_norm": 0.34344460453068587, "learning_rate": 5.425061091366321e-05, "loss": 0.9494, "step": 4918 }, { "epoch": 2.2288173991844134, "grad_norm": 0.26075273548889505, "learning_rate": 5.423878679292877e-05, "loss": 0.9443, "step": 4919 }, { "epoch": 2.2292705029451745, "grad_norm": 0.20874856808001258, "learning_rate": 5.422696124728652e-05, "loss": 0.938, "step": 4920 }, { "epoch": 2.2297236067059356, "grad_norm": 0.2440703209550651, "learning_rate": 5.421513427791988e-05, "loss": 0.967, "step": 4921 }, { "epoch": 2.2301767104666967, "grad_norm": 0.30088490937313184, "learning_rate": 5.420330588601239e-05, "loss": 0.957, "step": 4922 }, { "epoch": 2.2306298142274583, "grad_norm": 0.326741955588532, "learning_rate": 5.419147607274776e-05, "loss": 0.9524, "step": 4923 }, { "epoch": 2.2310829179882194, "grad_norm": 0.2967575507190523, "learning_rate": 5.4179644839309815e-05, "loss": 0.9254, "step": 4924 }, { "epoch": 2.2315360217489806, "grad_norm": 0.31374365679101235, "learning_rate": 5.416781218688254e-05, "loss": 0.9677, "step": 4925 }, { "epoch": 2.2319891255097417, "grad_norm": 0.3337307784665082, "learning_rate": 5.415597811665005e-05, "loss": 0.9619, "step": 4926 }, { "epoch": 2.232442229270503, "grad_norm": 0.3056064625997942, "learning_rate": 5.4144142629796607e-05, "loss": 0.9479, "step": 4927 }, { "epoch": 2.232895333031264, "grad_norm": 0.22933310416809807, "learning_rate": 5.413230572750662e-05, "loss": 0.9547, "step": 4928 }, { "epoch": 2.2333484367920255, "grad_norm": 0.19221711114895174, "learning_rate": 5.412046741096462e-05, "loss": 0.9525, "step": 4929 }, { "epoch": 2.2338015405527867, "grad_norm": 0.2735933795431002, "learning_rate": 5.4108627681355315e-05, "loss": 0.9529, "step": 4930 }, { "epoch": 2.234254644313548, "grad_norm": 0.31523778146514625, "learning_rate": 5.409678653986352e-05, "loss": 0.9554, "step": 4931 }, { "epoch": 2.234707748074309, "grad_norm": 0.29856593658034275, "learning_rate": 5.4084943987674215e-05, "loss": 0.9378, "step": 4932 }, { "epoch": 2.23516085183507, "grad_norm": 0.32685588374125096, "learning_rate": 5.40731000259725e-05, "loss": 0.9625, "step": 4933 }, { "epoch": 2.2356139555958316, "grad_norm": 0.3517922668348938, "learning_rate": 5.406125465594365e-05, "loss": 0.972, "step": 4934 }, { "epoch": 2.236067059356593, "grad_norm": 0.35837333598557675, "learning_rate": 5.404940787877302e-05, "loss": 0.9552, "step": 4935 }, { "epoch": 2.236520163117354, "grad_norm": 0.3706592232942297, "learning_rate": 5.403755969564617e-05, "loss": 0.9545, "step": 4936 }, { "epoch": 2.236973266878115, "grad_norm": 0.3986719044619524, "learning_rate": 5.4025710107748765e-05, "loss": 0.9344, "step": 4937 }, { "epoch": 2.237426370638876, "grad_norm": 0.381208178555527, "learning_rate": 5.401385911626663e-05, "loss": 0.9669, "step": 4938 }, { "epoch": 2.2378794743996373, "grad_norm": 0.3029237833963452, "learning_rate": 5.400200672238572e-05, "loss": 0.9473, "step": 4939 }, { "epoch": 2.238332578160399, "grad_norm": 0.2648042747349612, "learning_rate": 5.399015292729212e-05, "loss": 0.9617, "step": 4940 }, { "epoch": 2.23878568192116, "grad_norm": 0.25486412800392894, "learning_rate": 5.397829773217208e-05, "loss": 0.9299, "step": 4941 }, { "epoch": 2.239238785681921, "grad_norm": 0.2587659092249238, "learning_rate": 5.396644113821197e-05, "loss": 0.9575, "step": 4942 }, { "epoch": 2.2396918894426823, "grad_norm": 0.2368898607181471, "learning_rate": 5.3954583146598296e-05, "loss": 0.9424, "step": 4943 }, { "epoch": 2.2401449932034434, "grad_norm": 0.23291868336269073, "learning_rate": 5.394272375851773e-05, "loss": 0.9579, "step": 4944 }, { "epoch": 2.240598096964205, "grad_norm": 0.23905000192350007, "learning_rate": 5.393086297515706e-05, "loss": 0.9653, "step": 4945 }, { "epoch": 2.241051200724966, "grad_norm": 0.21702572719980898, "learning_rate": 5.391900079770322e-05, "loss": 0.9439, "step": 4946 }, { "epoch": 2.2415043044857272, "grad_norm": 0.16907341667919604, "learning_rate": 5.3907137227343296e-05, "loss": 0.9656, "step": 4947 }, { "epoch": 2.2419574082464884, "grad_norm": 0.20031675475094213, "learning_rate": 5.389527226526449e-05, "loss": 0.9303, "step": 4948 }, { "epoch": 2.2424105120072495, "grad_norm": 0.26543714579823724, "learning_rate": 5.3883405912654155e-05, "loss": 0.9491, "step": 4949 }, { "epoch": 2.242863615768011, "grad_norm": 0.30564488120087113, "learning_rate": 5.38715381706998e-05, "loss": 0.951, "step": 4950 }, { "epoch": 2.243316719528772, "grad_norm": 0.2935448368266327, "learning_rate": 5.385966904058904e-05, "loss": 0.9519, "step": 4951 }, { "epoch": 2.2437698232895333, "grad_norm": 0.29394445837922245, "learning_rate": 5.384779852350965e-05, "loss": 0.9376, "step": 4952 }, { "epoch": 2.2442229270502945, "grad_norm": 0.3393682717677905, "learning_rate": 5.383592662064956e-05, "loss": 0.9467, "step": 4953 }, { "epoch": 2.2446760308110556, "grad_norm": 0.35461455944926457, "learning_rate": 5.382405333319678e-05, "loss": 0.9412, "step": 4954 }, { "epoch": 2.245129134571817, "grad_norm": 0.3204778333162892, "learning_rate": 5.381217866233952e-05, "loss": 0.9541, "step": 4955 }, { "epoch": 2.2455822383325783, "grad_norm": 0.2902094160406287, "learning_rate": 5.38003026092661e-05, "loss": 0.9398, "step": 4956 }, { "epoch": 2.2460353420933394, "grad_norm": 0.29409192634407055, "learning_rate": 5.3788425175165e-05, "loss": 0.9672, "step": 4957 }, { "epoch": 2.2464884458541006, "grad_norm": 0.26330390840903406, "learning_rate": 5.377654636122479e-05, "loss": 0.9303, "step": 4958 }, { "epoch": 2.2469415496148617, "grad_norm": 0.2037994303870044, "learning_rate": 5.376466616863425e-05, "loss": 0.9396, "step": 4959 }, { "epoch": 2.247394653375623, "grad_norm": 0.1736872084156965, "learning_rate": 5.375278459858221e-05, "loss": 0.9507, "step": 4960 }, { "epoch": 2.2478477571363844, "grad_norm": 0.17020959027629876, "learning_rate": 5.3740901652257736e-05, "loss": 0.941, "step": 4961 }, { "epoch": 2.2483008608971455, "grad_norm": 0.20739608034576107, "learning_rate": 5.3729017330849944e-05, "loss": 0.9414, "step": 4962 }, { "epoch": 2.2487539646579067, "grad_norm": 0.1933499460110801, "learning_rate": 5.3717131635548145e-05, "loss": 0.9392, "step": 4963 }, { "epoch": 2.249207068418668, "grad_norm": 0.19344970022979088, "learning_rate": 5.370524456754176e-05, "loss": 0.9442, "step": 4964 }, { "epoch": 2.249660172179429, "grad_norm": 0.2197008755171914, "learning_rate": 5.3693356128020346e-05, "loss": 0.9385, "step": 4965 }, { "epoch": 2.2501132759401905, "grad_norm": 0.2692319860996176, "learning_rate": 5.3681466318173627e-05, "loss": 0.9564, "step": 4966 }, { "epoch": 2.2505663797009516, "grad_norm": 0.27123440344621846, "learning_rate": 5.3669575139191423e-05, "loss": 0.9738, "step": 4967 }, { "epoch": 2.2510194834617128, "grad_norm": 0.26594174823624955, "learning_rate": 5.365768259226372e-05, "loss": 0.9508, "step": 4968 }, { "epoch": 2.251472587222474, "grad_norm": 0.23720469682549525, "learning_rate": 5.3645788678580647e-05, "loss": 0.9324, "step": 4969 }, { "epoch": 2.251925690983235, "grad_norm": 0.2393509120903701, "learning_rate": 5.363389339933243e-05, "loss": 0.9508, "step": 4970 }, { "epoch": 2.252378794743996, "grad_norm": 0.22846275035601393, "learning_rate": 5.3621996755709485e-05, "loss": 0.9419, "step": 4971 }, { "epoch": 2.2528318985047577, "grad_norm": 0.21374020435719923, "learning_rate": 5.36100987489023e-05, "loss": 0.9462, "step": 4972 }, { "epoch": 2.253285002265519, "grad_norm": 0.25884347651170453, "learning_rate": 5.359819938010156e-05, "loss": 0.9397, "step": 4973 }, { "epoch": 2.25373810602628, "grad_norm": 0.23804541300892887, "learning_rate": 5.3586298650498054e-05, "loss": 0.9359, "step": 4974 }, { "epoch": 2.254191209787041, "grad_norm": 0.23122478069491317, "learning_rate": 5.3574396561282726e-05, "loss": 0.9556, "step": 4975 }, { "epoch": 2.2546443135478023, "grad_norm": 0.243701798743004, "learning_rate": 5.356249311364662e-05, "loss": 0.9293, "step": 4976 }, { "epoch": 2.255097417308564, "grad_norm": 0.2309273477041399, "learning_rate": 5.355058830878097e-05, "loss": 0.9591, "step": 4977 }, { "epoch": 2.255550521069325, "grad_norm": 0.23661978971559836, "learning_rate": 5.35386821478771e-05, "loss": 0.9545, "step": 4978 }, { "epoch": 2.256003624830086, "grad_norm": 0.2172549572151555, "learning_rate": 5.35267746321265e-05, "loss": 0.9447, "step": 4979 }, { "epoch": 2.2564567285908472, "grad_norm": 0.2271499951558067, "learning_rate": 5.3514865762720774e-05, "loss": 0.9707, "step": 4980 }, { "epoch": 2.2569098323516084, "grad_norm": 0.24058004622121676, "learning_rate": 5.350295554085165e-05, "loss": 0.9562, "step": 4981 }, { "epoch": 2.2573629361123695, "grad_norm": 0.2876639227268448, "learning_rate": 5.3491043967711043e-05, "loss": 0.9655, "step": 4982 }, { "epoch": 2.257816039873131, "grad_norm": 0.3535652118103494, "learning_rate": 5.347913104449096e-05, "loss": 0.9678, "step": 4983 }, { "epoch": 2.258269143633892, "grad_norm": 0.388236443115576, "learning_rate": 5.346721677238355e-05, "loss": 0.9412, "step": 4984 }, { "epoch": 2.2587222473946533, "grad_norm": 0.3977807670097191, "learning_rate": 5.3455301152581106e-05, "loss": 0.9612, "step": 4985 }, { "epoch": 2.2591753511554145, "grad_norm": 0.40661591179272105, "learning_rate": 5.3443384186276044e-05, "loss": 0.9594, "step": 4986 }, { "epoch": 2.259628454916176, "grad_norm": 0.3264943466459946, "learning_rate": 5.343146587466093e-05, "loss": 0.936, "step": 4987 }, { "epoch": 2.260081558676937, "grad_norm": 0.25880328686363446, "learning_rate": 5.3419546218928464e-05, "loss": 0.9426, "step": 4988 }, { "epoch": 2.2605346624376983, "grad_norm": 0.20339963968801267, "learning_rate": 5.3407625220271455e-05, "loss": 0.9446, "step": 4989 }, { "epoch": 2.2609877661984594, "grad_norm": 0.2187048300413603, "learning_rate": 5.3395702879882863e-05, "loss": 0.9511, "step": 4990 }, { "epoch": 2.2614408699592206, "grad_norm": 0.2423832842779897, "learning_rate": 5.33837791989558e-05, "loss": 0.9431, "step": 4991 }, { "epoch": 2.2618939737199817, "grad_norm": 0.24814094410045368, "learning_rate": 5.337185417868349e-05, "loss": 0.9347, "step": 4992 }, { "epoch": 2.2623470774807433, "grad_norm": 0.24114490117625975, "learning_rate": 5.335992782025929e-05, "loss": 0.933, "step": 4993 }, { "epoch": 2.2628001812415044, "grad_norm": 0.23048304860824215, "learning_rate": 5.334800012487671e-05, "loss": 0.9356, "step": 4994 }, { "epoch": 2.2632532850022655, "grad_norm": 0.24023839600030591, "learning_rate": 5.3336071093729386e-05, "loss": 0.9417, "step": 4995 }, { "epoch": 2.2637063887630267, "grad_norm": 0.28318295096660884, "learning_rate": 5.332414072801106e-05, "loss": 0.9408, "step": 4996 }, { "epoch": 2.264159492523788, "grad_norm": 0.29623422931948207, "learning_rate": 5.3312209028915644e-05, "loss": 0.9494, "step": 4997 }, { "epoch": 2.2646125962845494, "grad_norm": 0.32492825555726207, "learning_rate": 5.330027599763717e-05, "loss": 0.9388, "step": 4998 }, { "epoch": 2.2650657000453105, "grad_norm": 0.3525885679896209, "learning_rate": 5.32883416353698e-05, "loss": 0.9425, "step": 4999 }, { "epoch": 2.2655188038060716, "grad_norm": 0.4128505119193087, "learning_rate": 5.327640594330785e-05, "loss": 0.944, "step": 5000 }, { "epoch": 2.2659719075668328, "grad_norm": 0.44455323969108596, "learning_rate": 5.3264468922645726e-05, "loss": 0.9278, "step": 5001 }, { "epoch": 2.266425011327594, "grad_norm": 0.4569200617894551, "learning_rate": 5.325253057457801e-05, "loss": 0.9632, "step": 5002 }, { "epoch": 2.266878115088355, "grad_norm": 0.4598105093025227, "learning_rate": 5.3240590900299386e-05, "loss": 0.9606, "step": 5003 }, { "epoch": 2.2673312188491166, "grad_norm": 0.4301735853617183, "learning_rate": 5.3228649901004695e-05, "loss": 0.9479, "step": 5004 }, { "epoch": 2.2677843226098777, "grad_norm": 0.37181204909221155, "learning_rate": 5.321670757788891e-05, "loss": 0.9383, "step": 5005 }, { "epoch": 2.268237426370639, "grad_norm": 0.3441078227345139, "learning_rate": 5.32047639321471e-05, "loss": 0.9299, "step": 5006 }, { "epoch": 2.2686905301314, "grad_norm": 0.31171183736641944, "learning_rate": 5.319281896497452e-05, "loss": 0.9636, "step": 5007 }, { "epoch": 2.269143633892161, "grad_norm": 0.2666510659790638, "learning_rate": 5.318087267756651e-05, "loss": 0.96, "step": 5008 }, { "epoch": 2.2695967376529227, "grad_norm": 0.2168321693058084, "learning_rate": 5.316892507111856e-05, "loss": 0.9457, "step": 5009 }, { "epoch": 2.270049841413684, "grad_norm": 0.16822027668631703, "learning_rate": 5.3156976146826325e-05, "loss": 0.9575, "step": 5010 }, { "epoch": 2.270502945174445, "grad_norm": 0.1840738810260452, "learning_rate": 5.314502590588552e-05, "loss": 0.9394, "step": 5011 }, { "epoch": 2.270956048935206, "grad_norm": 0.22169945693470924, "learning_rate": 5.313307434949206e-05, "loss": 0.9498, "step": 5012 }, { "epoch": 2.271409152695967, "grad_norm": 0.19375506731994152, "learning_rate": 5.312112147884197e-05, "loss": 0.9387, "step": 5013 }, { "epoch": 2.2718622564567283, "grad_norm": 0.22342442794797843, "learning_rate": 5.310916729513137e-05, "loss": 0.9327, "step": 5014 }, { "epoch": 2.27231536021749, "grad_norm": 0.45330102956113116, "learning_rate": 5.309721179955657e-05, "loss": 0.9558, "step": 5015 }, { "epoch": 2.272768463978251, "grad_norm": 0.32985450254411774, "learning_rate": 5.3085254993313964e-05, "loss": 0.9666, "step": 5016 }, { "epoch": 2.273221567739012, "grad_norm": 0.32339146480065073, "learning_rate": 5.307329687760011e-05, "loss": 0.935, "step": 5017 }, { "epoch": 2.2736746714997733, "grad_norm": 0.30700178698478703, "learning_rate": 5.306133745361168e-05, "loss": 0.9474, "step": 5018 }, { "epoch": 2.274127775260535, "grad_norm": 0.29137856941040347, "learning_rate": 5.3049376722545486e-05, "loss": 0.9654, "step": 5019 }, { "epoch": 2.274580879021296, "grad_norm": 0.29454194976834513, "learning_rate": 5.303741468559846e-05, "loss": 0.9545, "step": 5020 }, { "epoch": 2.275033982782057, "grad_norm": 0.23039737113839265, "learning_rate": 5.3025451343967674e-05, "loss": 0.9517, "step": 5021 }, { "epoch": 2.2754870865428183, "grad_norm": 0.24358372913874848, "learning_rate": 5.3013486698850325e-05, "loss": 0.9419, "step": 5022 }, { "epoch": 2.2759401903035794, "grad_norm": 0.28725264764877634, "learning_rate": 5.300152075144373e-05, "loss": 0.9463, "step": 5023 }, { "epoch": 2.2763932940643405, "grad_norm": 0.2589527965827422, "learning_rate": 5.2989553502945365e-05, "loss": 0.9405, "step": 5024 }, { "epoch": 2.276846397825102, "grad_norm": 0.2446767528118155, "learning_rate": 5.297758495455282e-05, "loss": 0.9475, "step": 5025 }, { "epoch": 2.2772995015858633, "grad_norm": 0.22203278055234293, "learning_rate": 5.29656151074638e-05, "loss": 0.9479, "step": 5026 }, { "epoch": 2.2777526053466244, "grad_norm": 0.2811723456404688, "learning_rate": 5.2953643962876166e-05, "loss": 0.9437, "step": 5027 }, { "epoch": 2.2782057091073855, "grad_norm": 0.3366341975725217, "learning_rate": 5.2941671521987905e-05, "loss": 0.9466, "step": 5028 }, { "epoch": 2.2786588128681466, "grad_norm": 0.29002644389644844, "learning_rate": 5.2929697785997116e-05, "loss": 0.947, "step": 5029 }, { "epoch": 2.279111916628908, "grad_norm": 0.25165218626009295, "learning_rate": 5.291772275610203e-05, "loss": 0.935, "step": 5030 }, { "epoch": 2.2795650203896693, "grad_norm": 0.22460756158993933, "learning_rate": 5.2905746433501044e-05, "loss": 0.9313, "step": 5031 }, { "epoch": 2.2800181241504305, "grad_norm": 0.2128780597245496, "learning_rate": 5.2893768819392627e-05, "loss": 0.9376, "step": 5032 }, { "epoch": 2.2804712279111916, "grad_norm": 0.2636529274046485, "learning_rate": 5.288178991497541e-05, "loss": 0.9494, "step": 5033 }, { "epoch": 2.2809243316719527, "grad_norm": 0.2856715398762098, "learning_rate": 5.2869809721448154e-05, "loss": 0.9379, "step": 5034 }, { "epoch": 2.281377435432714, "grad_norm": 0.337492187824611, "learning_rate": 5.285782824000976e-05, "loss": 0.9444, "step": 5035 }, { "epoch": 2.2818305391934754, "grad_norm": 0.4157553446037521, "learning_rate": 5.284584547185921e-05, "loss": 0.9753, "step": 5036 }, { "epoch": 2.2822836429542366, "grad_norm": 0.445835530117183, "learning_rate": 5.2833861418195666e-05, "loss": 0.9529, "step": 5037 }, { "epoch": 2.2827367467149977, "grad_norm": 0.4369189327347783, "learning_rate": 5.282187608021841e-05, "loss": 0.9678, "step": 5038 }, { "epoch": 2.283189850475759, "grad_norm": 0.4015914725242856, "learning_rate": 5.280988945912682e-05, "loss": 0.9488, "step": 5039 }, { "epoch": 2.28364295423652, "grad_norm": 0.3988158246386835, "learning_rate": 5.2797901556120445e-05, "loss": 0.9216, "step": 5040 }, { "epoch": 2.2840960579972815, "grad_norm": 0.3792513002966966, "learning_rate": 5.278591237239893e-05, "loss": 0.9459, "step": 5041 }, { "epoch": 2.2845491617580427, "grad_norm": 0.3150035519921156, "learning_rate": 5.277392190916205e-05, "loss": 0.9439, "step": 5042 }, { "epoch": 2.285002265518804, "grad_norm": 0.23842746453722938, "learning_rate": 5.276193016760973e-05, "loss": 0.9258, "step": 5043 }, { "epoch": 2.285455369279565, "grad_norm": 0.17876347673815257, "learning_rate": 5.274993714894201e-05, "loss": 0.9344, "step": 5044 }, { "epoch": 2.285908473040326, "grad_norm": 0.2528727259336184, "learning_rate": 5.273794285435907e-05, "loss": 0.941, "step": 5045 }, { "epoch": 2.286361576801087, "grad_norm": 0.31507959806124464, "learning_rate": 5.2725947285061184e-05, "loss": 0.9562, "step": 5046 }, { "epoch": 2.286814680561849, "grad_norm": 0.336736710488476, "learning_rate": 5.271395044224879e-05, "loss": 0.9466, "step": 5047 }, { "epoch": 2.28726778432261, "grad_norm": 0.345617417745331, "learning_rate": 5.270195232712245e-05, "loss": 0.9519, "step": 5048 }, { "epoch": 2.287720888083371, "grad_norm": 0.29191813668630356, "learning_rate": 5.2689952940882816e-05, "loss": 0.9274, "step": 5049 }, { "epoch": 2.288173991844132, "grad_norm": 0.2567365742991662, "learning_rate": 5.2677952284730696e-05, "loss": 0.9368, "step": 5050 }, { "epoch": 2.2886270956048937, "grad_norm": 0.25602963804256723, "learning_rate": 5.2665950359867054e-05, "loss": 0.9736, "step": 5051 }, { "epoch": 2.289080199365655, "grad_norm": 0.23896045761750317, "learning_rate": 5.265394716749292e-05, "loss": 0.9425, "step": 5052 }, { "epoch": 2.289533303126416, "grad_norm": 0.21195124796481993, "learning_rate": 5.264194270880949e-05, "loss": 0.9428, "step": 5053 }, { "epoch": 2.289986406887177, "grad_norm": 0.1871002490548745, "learning_rate": 5.262993698501808e-05, "loss": 0.9505, "step": 5054 }, { "epoch": 2.2904395106479383, "grad_norm": 0.1756725807851567, "learning_rate": 5.261792999732012e-05, "loss": 0.9417, "step": 5055 }, { "epoch": 2.2908926144086994, "grad_norm": 0.1978334829119919, "learning_rate": 5.26059217469172e-05, "loss": 0.9585, "step": 5056 }, { "epoch": 2.291345718169461, "grad_norm": 0.23883886738180768, "learning_rate": 5.2593912235010996e-05, "loss": 0.9524, "step": 5057 }, { "epoch": 2.291798821930222, "grad_norm": 0.2878432031702921, "learning_rate": 5.2581901462803334e-05, "loss": 0.9491, "step": 5058 }, { "epoch": 2.2922519256909832, "grad_norm": 0.2998458793484414, "learning_rate": 5.256988943149614e-05, "loss": 0.9668, "step": 5059 }, { "epoch": 2.2927050294517444, "grad_norm": 0.2535501970279442, "learning_rate": 5.2557876142291494e-05, "loss": 0.9295, "step": 5060 }, { "epoch": 2.2931581332125055, "grad_norm": 0.23657451518403336, "learning_rate": 5.254586159639161e-05, "loss": 0.9444, "step": 5061 }, { "epoch": 2.293611236973267, "grad_norm": 0.20379099688135727, "learning_rate": 5.2533845794998803e-05, "loss": 0.9528, "step": 5062 }, { "epoch": 2.294064340734028, "grad_norm": 0.19838330440242533, "learning_rate": 5.252182873931551e-05, "loss": 0.9754, "step": 5063 }, { "epoch": 2.2945174444947893, "grad_norm": 0.262977255424258, "learning_rate": 5.250981043054431e-05, "loss": 0.9303, "step": 5064 }, { "epoch": 2.2949705482555505, "grad_norm": 0.3297620958023646, "learning_rate": 5.249779086988791e-05, "loss": 0.9564, "step": 5065 }, { "epoch": 2.2954236520163116, "grad_norm": 0.3860783612816631, "learning_rate": 5.248577005854914e-05, "loss": 0.9555, "step": 5066 }, { "epoch": 2.2958767557770727, "grad_norm": 0.39184716495074606, "learning_rate": 5.247374799773093e-05, "loss": 0.9443, "step": 5067 }, { "epoch": 2.2963298595378343, "grad_norm": 0.4343884623531036, "learning_rate": 5.2461724688636375e-05, "loss": 0.9662, "step": 5068 }, { "epoch": 2.2967829632985954, "grad_norm": 0.4589676457571387, "learning_rate": 5.2449700132468655e-05, "loss": 0.9525, "step": 5069 }, { "epoch": 2.2972360670593566, "grad_norm": 0.457210774888664, "learning_rate": 5.2437674330431114e-05, "loss": 0.9341, "step": 5070 }, { "epoch": 2.2976891708201177, "grad_norm": 0.4123310899390243, "learning_rate": 5.242564728372719e-05, "loss": 0.9594, "step": 5071 }, { "epoch": 2.298142274580879, "grad_norm": 0.33900023123386785, "learning_rate": 5.241361899356046e-05, "loss": 0.9581, "step": 5072 }, { "epoch": 2.2985953783416404, "grad_norm": 0.2821314934525407, "learning_rate": 5.240158946113461e-05, "loss": 0.9739, "step": 5073 }, { "epoch": 2.2990484821024015, "grad_norm": 0.2644303894865783, "learning_rate": 5.238955868765349e-05, "loss": 0.9516, "step": 5074 }, { "epoch": 2.2995015858631627, "grad_norm": 0.24503582197600482, "learning_rate": 5.2377526674321034e-05, "loss": 0.9504, "step": 5075 }, { "epoch": 2.299954689623924, "grad_norm": 0.24596387569394212, "learning_rate": 5.236549342234129e-05, "loss": 0.9673, "step": 5076 }, { "epoch": 2.300407793384685, "grad_norm": 0.28380370772302416, "learning_rate": 5.235345893291849e-05, "loss": 0.9281, "step": 5077 }, { "epoch": 2.300860897145446, "grad_norm": 0.28843276629818115, "learning_rate": 5.234142320725693e-05, "loss": 0.9594, "step": 5078 }, { "epoch": 2.3013140009062076, "grad_norm": 0.2787605191332223, "learning_rate": 5.232938624656106e-05, "loss": 0.9465, "step": 5079 }, { "epoch": 2.3017671046669688, "grad_norm": 0.23809804042804006, "learning_rate": 5.231734805203545e-05, "loss": 0.9478, "step": 5080 }, { "epoch": 2.30222020842773, "grad_norm": 0.275298607561226, "learning_rate": 5.230530862488477e-05, "loss": 0.9584, "step": 5081 }, { "epoch": 2.302673312188491, "grad_norm": 0.3051893637913646, "learning_rate": 5.2293267966313874e-05, "loss": 0.9536, "step": 5082 }, { "epoch": 2.3031264159492526, "grad_norm": 0.29006044039640716, "learning_rate": 5.2281226077527646e-05, "loss": 0.9575, "step": 5083 }, { "epoch": 2.3035795197100137, "grad_norm": 0.25380794205918045, "learning_rate": 5.226918295973118e-05, "loss": 0.9537, "step": 5084 }, { "epoch": 2.304032623470775, "grad_norm": 0.24239405710788595, "learning_rate": 5.225713861412965e-05, "loss": 0.9364, "step": 5085 }, { "epoch": 2.304485727231536, "grad_norm": 0.24021769769366458, "learning_rate": 5.224509304192836e-05, "loss": 0.9425, "step": 5086 }, { "epoch": 2.304938830992297, "grad_norm": 0.25898744248612415, "learning_rate": 5.223304624433273e-05, "loss": 0.9393, "step": 5087 }, { "epoch": 2.3053919347530583, "grad_norm": 0.26695557870998626, "learning_rate": 5.222099822254833e-05, "loss": 0.9542, "step": 5088 }, { "epoch": 2.30584503851382, "grad_norm": 0.23610454390567479, "learning_rate": 5.220894897778081e-05, "loss": 0.9627, "step": 5089 }, { "epoch": 2.306298142274581, "grad_norm": 0.21994057379587081, "learning_rate": 5.2196898511235974e-05, "loss": 0.9441, "step": 5090 }, { "epoch": 2.306751246035342, "grad_norm": 0.2815258507516335, "learning_rate": 5.2184846824119745e-05, "loss": 0.9591, "step": 5091 }, { "epoch": 2.3072043497961032, "grad_norm": 0.2781887123108384, "learning_rate": 5.2172793917638175e-05, "loss": 0.9496, "step": 5092 }, { "epoch": 2.3076574535568644, "grad_norm": 0.21804806307375926, "learning_rate": 5.21607397929974e-05, "loss": 0.9457, "step": 5093 }, { "epoch": 2.308110557317626, "grad_norm": 0.2561616370693486, "learning_rate": 5.2148684451403706e-05, "loss": 0.9477, "step": 5094 }, { "epoch": 2.308563661078387, "grad_norm": 0.23946990042215255, "learning_rate": 5.2136627894063515e-05, "loss": 0.9498, "step": 5095 }, { "epoch": 2.309016764839148, "grad_norm": 0.232693841047863, "learning_rate": 5.212457012218334e-05, "loss": 0.9576, "step": 5096 }, { "epoch": 2.3094698685999093, "grad_norm": 0.2590271444153375, "learning_rate": 5.211251113696984e-05, "loss": 0.9407, "step": 5097 }, { "epoch": 2.3099229723606705, "grad_norm": 0.22368428621999664, "learning_rate": 5.210045093962977e-05, "loss": 0.9349, "step": 5098 }, { "epoch": 2.3103760761214316, "grad_norm": 0.2060457793652719, "learning_rate": 5.2088389531370044e-05, "loss": 0.9282, "step": 5099 }, { "epoch": 2.310829179882193, "grad_norm": 0.20221859275934098, "learning_rate": 5.2076326913397666e-05, "loss": 0.9463, "step": 5100 }, { "epoch": 2.3112822836429543, "grad_norm": 0.25801450669376086, "learning_rate": 5.206426308691976e-05, "loss": 0.9576, "step": 5101 }, { "epoch": 2.3117353874037154, "grad_norm": 0.2649298006668564, "learning_rate": 5.205219805314359e-05, "loss": 0.9458, "step": 5102 }, { "epoch": 2.3121884911644766, "grad_norm": 0.23813038925796168, "learning_rate": 5.204013181327652e-05, "loss": 0.9371, "step": 5103 }, { "epoch": 2.3126415949252377, "grad_norm": 0.2219326459395096, "learning_rate": 5.202806436852606e-05, "loss": 0.9554, "step": 5104 }, { "epoch": 2.3130946986859993, "grad_norm": 0.2112951640011073, "learning_rate": 5.201599572009982e-05, "loss": 0.9437, "step": 5105 }, { "epoch": 2.3135478024467604, "grad_norm": 0.2144873227347361, "learning_rate": 5.200392586920553e-05, "loss": 0.9389, "step": 5106 }, { "epoch": 2.3140009062075215, "grad_norm": 0.2186283379888152, "learning_rate": 5.1991854817051064e-05, "loss": 0.9511, "step": 5107 }, { "epoch": 2.3144540099682827, "grad_norm": 0.24145620637851611, "learning_rate": 5.197978256484439e-05, "loss": 0.9424, "step": 5108 }, { "epoch": 2.314907113729044, "grad_norm": 0.20196355211634176, "learning_rate": 5.1967709113793617e-05, "loss": 0.956, "step": 5109 }, { "epoch": 2.315360217489805, "grad_norm": 0.16429621298214375, "learning_rate": 5.195563446510694e-05, "loss": 0.9738, "step": 5110 }, { "epoch": 2.3158133212505665, "grad_norm": 0.18495557642285781, "learning_rate": 5.194355861999271e-05, "loss": 0.9525, "step": 5111 }, { "epoch": 2.3162664250113276, "grad_norm": 0.17807724253818047, "learning_rate": 5.1931481579659385e-05, "loss": 0.9544, "step": 5112 }, { "epoch": 2.3167195287720888, "grad_norm": 0.1737294251341527, "learning_rate": 5.191940334531554e-05, "loss": 0.9717, "step": 5113 }, { "epoch": 2.31717263253285, "grad_norm": 0.15654105045817257, "learning_rate": 5.190732391816988e-05, "loss": 0.9777, "step": 5114 }, { "epoch": 2.3176257362936115, "grad_norm": 0.17228733829342474, "learning_rate": 5.189524329943121e-05, "loss": 0.9773, "step": 5115 }, { "epoch": 2.3180788400543726, "grad_norm": 0.18685912831892734, "learning_rate": 5.1883161490308455e-05, "loss": 0.957, "step": 5116 }, { "epoch": 2.3185319438151337, "grad_norm": 0.2149575671232496, "learning_rate": 5.1871078492010694e-05, "loss": 0.9396, "step": 5117 }, { "epoch": 2.318985047575895, "grad_norm": 0.26055013003265604, "learning_rate": 5.185899430574708e-05, "loss": 0.9351, "step": 5118 }, { "epoch": 2.319438151336656, "grad_norm": 0.26689115551702824, "learning_rate": 5.184690893272693e-05, "loss": 0.9267, "step": 5119 }, { "epoch": 2.319891255097417, "grad_norm": 0.3057574312636539, "learning_rate": 5.1834822374159614e-05, "loss": 0.9722, "step": 5120 }, { "epoch": 2.3203443588581787, "grad_norm": 0.3149926585727292, "learning_rate": 5.182273463125469e-05, "loss": 0.9636, "step": 5121 }, { "epoch": 2.32079746261894, "grad_norm": 0.3031368249618565, "learning_rate": 5.181064570522181e-05, "loss": 0.9565, "step": 5122 }, { "epoch": 2.321250566379701, "grad_norm": 0.301655305670319, "learning_rate": 5.179855559727072e-05, "loss": 0.9425, "step": 5123 }, { "epoch": 2.321703670140462, "grad_norm": 0.24328663609387602, "learning_rate": 5.178646430861133e-05, "loss": 0.9489, "step": 5124 }, { "epoch": 2.322156773901223, "grad_norm": 0.19605219439677057, "learning_rate": 5.177437184045362e-05, "loss": 0.9371, "step": 5125 }, { "epoch": 2.322609877661985, "grad_norm": 0.22054817484239467, "learning_rate": 5.1762278194007723e-05, "loss": 0.9243, "step": 5126 }, { "epoch": 2.323062981422746, "grad_norm": 0.21841109152667473, "learning_rate": 5.1750183370483874e-05, "loss": 0.9601, "step": 5127 }, { "epoch": 2.323516085183507, "grad_norm": 0.20272217943487747, "learning_rate": 5.173808737109243e-05, "loss": 0.9293, "step": 5128 }, { "epoch": 2.323969188944268, "grad_norm": 0.2169733264279019, "learning_rate": 5.1725990197043863e-05, "loss": 0.9489, "step": 5129 }, { "epoch": 2.3244222927050293, "grad_norm": 0.27356341269463424, "learning_rate": 5.1713891849548766e-05, "loss": 0.9454, "step": 5130 }, { "epoch": 2.3248753964657904, "grad_norm": 0.35331511287163325, "learning_rate": 5.170179232981786e-05, "loss": 0.9432, "step": 5131 }, { "epoch": 2.325328500226552, "grad_norm": 0.3187398142487692, "learning_rate": 5.168969163906195e-05, "loss": 0.9343, "step": 5132 }, { "epoch": 2.325781603987313, "grad_norm": 0.33923264771091133, "learning_rate": 5.167758977849199e-05, "loss": 0.9483, "step": 5133 }, { "epoch": 2.3262347077480743, "grad_norm": 0.34435407700418247, "learning_rate": 5.1665486749319034e-05, "loss": 0.9312, "step": 5134 }, { "epoch": 2.3266878115088354, "grad_norm": 0.3178553206470677, "learning_rate": 5.165338255275428e-05, "loss": 0.9651, "step": 5135 }, { "epoch": 2.3271409152695965, "grad_norm": 0.28481261035405064, "learning_rate": 5.1641277190009006e-05, "loss": 0.9445, "step": 5136 }, { "epoch": 2.327594019030358, "grad_norm": 0.19171262586355728, "learning_rate": 5.1629170662294616e-05, "loss": 0.9527, "step": 5137 }, { "epoch": 2.3280471227911192, "grad_norm": 0.16440655269891008, "learning_rate": 5.161706297082266e-05, "loss": 0.9271, "step": 5138 }, { "epoch": 2.3285002265518804, "grad_norm": 0.17758015170492936, "learning_rate": 5.160495411680476e-05, "loss": 0.9689, "step": 5139 }, { "epoch": 2.3289533303126415, "grad_norm": 0.2037745495286077, "learning_rate": 5.15928441014527e-05, "loss": 0.9334, "step": 5140 }, { "epoch": 2.3294064340734026, "grad_norm": 0.19595654528654097, "learning_rate": 5.1580732925978335e-05, "loss": 0.9307, "step": 5141 }, { "epoch": 2.3298595378341638, "grad_norm": 0.18753631765237536, "learning_rate": 5.156862059159366e-05, "loss": 0.923, "step": 5142 }, { "epoch": 2.3303126415949253, "grad_norm": 0.16706399568128427, "learning_rate": 5.155650709951081e-05, "loss": 0.9542, "step": 5143 }, { "epoch": 2.3307657453556865, "grad_norm": 0.1914417251994051, "learning_rate": 5.1544392450941976e-05, "loss": 0.9406, "step": 5144 }, { "epoch": 2.3312188491164476, "grad_norm": 0.20171073953511695, "learning_rate": 5.153227664709952e-05, "loss": 0.9284, "step": 5145 }, { "epoch": 2.3316719528772087, "grad_norm": 0.24223460951912268, "learning_rate": 5.152015968919589e-05, "loss": 0.9677, "step": 5146 }, { "epoch": 2.3321250566379703, "grad_norm": 0.33015720055337267, "learning_rate": 5.1508041578443646e-05, "loss": 0.9331, "step": 5147 }, { "epoch": 2.3325781603987314, "grad_norm": 0.3859563387698335, "learning_rate": 5.1495922316055494e-05, "loss": 0.956, "step": 5148 }, { "epoch": 2.3330312641594926, "grad_norm": 0.3535132609920976, "learning_rate": 5.1483801903244235e-05, "loss": 0.9733, "step": 5149 }, { "epoch": 2.3334843679202537, "grad_norm": 0.37088017191893624, "learning_rate": 5.147168034122277e-05, "loss": 0.9418, "step": 5150 }, { "epoch": 2.333937471681015, "grad_norm": 0.35018866999660914, "learning_rate": 5.145955763120415e-05, "loss": 0.9522, "step": 5151 }, { "epoch": 2.334390575441776, "grad_norm": 0.3128668575216464, "learning_rate": 5.144743377440151e-05, "loss": 0.9704, "step": 5152 }, { "epoch": 2.3348436792025375, "grad_norm": 0.25631042513869995, "learning_rate": 5.1435308772028126e-05, "loss": 0.9363, "step": 5153 }, { "epoch": 2.3352967829632987, "grad_norm": 0.2521422372482623, "learning_rate": 5.142318262529736e-05, "loss": 0.9483, "step": 5154 }, { "epoch": 2.33574988672406, "grad_norm": 0.28083021732414787, "learning_rate": 5.14110553354227e-05, "loss": 0.9561, "step": 5155 }, { "epoch": 2.336202990484821, "grad_norm": 0.2641616129095994, "learning_rate": 5.139892690361777e-05, "loss": 0.9553, "step": 5156 }, { "epoch": 2.336656094245582, "grad_norm": 0.31330375093653856, "learning_rate": 5.138679733109627e-05, "loss": 0.945, "step": 5157 }, { "epoch": 2.3371091980063436, "grad_norm": 0.35995649904124016, "learning_rate": 5.137466661907204e-05, "loss": 0.954, "step": 5158 }, { "epoch": 2.3375623017671048, "grad_norm": 0.3873869442201812, "learning_rate": 5.1362534768759034e-05, "loss": 0.9499, "step": 5159 }, { "epoch": 2.338015405527866, "grad_norm": 0.3160427415626671, "learning_rate": 5.135040178137131e-05, "loss": 0.9468, "step": 5160 }, { "epoch": 2.338468509288627, "grad_norm": 0.2282837633449734, "learning_rate": 5.133826765812306e-05, "loss": 0.9588, "step": 5161 }, { "epoch": 2.338921613049388, "grad_norm": 0.23815487988703024, "learning_rate": 5.132613240022855e-05, "loss": 0.9656, "step": 5162 }, { "epoch": 2.3393747168101493, "grad_norm": 0.26428370851454625, "learning_rate": 5.131399600890218e-05, "loss": 0.9712, "step": 5163 }, { "epoch": 2.339827820570911, "grad_norm": 0.284263612218575, "learning_rate": 5.1301858485358474e-05, "loss": 0.952, "step": 5164 }, { "epoch": 2.340280924331672, "grad_norm": 0.27578686332339036, "learning_rate": 5.128971983081208e-05, "loss": 0.9481, "step": 5165 }, { "epoch": 2.340734028092433, "grad_norm": 0.23912627027384825, "learning_rate": 5.127758004647772e-05, "loss": 0.9359, "step": 5166 }, { "epoch": 2.3411871318531943, "grad_norm": 0.20774685223419623, "learning_rate": 5.126543913357025e-05, "loss": 0.9547, "step": 5167 }, { "epoch": 2.341640235613956, "grad_norm": 0.22105203682722657, "learning_rate": 5.125329709330464e-05, "loss": 0.9517, "step": 5168 }, { "epoch": 2.342093339374717, "grad_norm": 0.2266695282278306, "learning_rate": 5.124115392689598e-05, "loss": 0.9373, "step": 5169 }, { "epoch": 2.342546443135478, "grad_norm": 0.2469748354531203, "learning_rate": 5.122900963555948e-05, "loss": 0.9415, "step": 5170 }, { "epoch": 2.3429995468962392, "grad_norm": 0.2954203778207343, "learning_rate": 5.121686422051042e-05, "loss": 0.9328, "step": 5171 }, { "epoch": 2.3434526506570004, "grad_norm": 0.28403969821987823, "learning_rate": 5.120471768296421e-05, "loss": 0.9549, "step": 5172 }, { "epoch": 2.3439057544177615, "grad_norm": 0.22362845574732299, "learning_rate": 5.119257002413641e-05, "loss": 0.9691, "step": 5173 }, { "epoch": 2.344358858178523, "grad_norm": 0.22116781809901212, "learning_rate": 5.118042124524264e-05, "loss": 0.9449, "step": 5174 }, { "epoch": 2.344811961939284, "grad_norm": 0.2999881229276112, "learning_rate": 5.116827134749867e-05, "loss": 0.936, "step": 5175 }, { "epoch": 2.3452650657000453, "grad_norm": 0.38841942621600145, "learning_rate": 5.115612033212037e-05, "loss": 0.9561, "step": 5176 }, { "epoch": 2.3457181694608065, "grad_norm": 0.4174358295482972, "learning_rate": 5.1143968200323715e-05, "loss": 0.9316, "step": 5177 }, { "epoch": 2.3461712732215676, "grad_norm": 0.40548490041291335, "learning_rate": 5.1131814953324797e-05, "loss": 0.9407, "step": 5178 }, { "epoch": 2.346624376982329, "grad_norm": 0.34746447163603505, "learning_rate": 5.1119660592339825e-05, "loss": 0.9463, "step": 5179 }, { "epoch": 2.3470774807430903, "grad_norm": 0.2845894861241961, "learning_rate": 5.110750511858509e-05, "loss": 0.964, "step": 5180 }, { "epoch": 2.3475305845038514, "grad_norm": 0.21273291519476692, "learning_rate": 5.109534853327705e-05, "loss": 0.9394, "step": 5181 }, { "epoch": 2.3479836882646126, "grad_norm": 0.16627630930419415, "learning_rate": 5.108319083763222e-05, "loss": 0.9417, "step": 5182 }, { "epoch": 2.3484367920253737, "grad_norm": 0.2196531442395595, "learning_rate": 5.107103203286726e-05, "loss": 0.9469, "step": 5183 }, { "epoch": 2.348889895786135, "grad_norm": 0.2998804988875275, "learning_rate": 5.105887212019892e-05, "loss": 0.9517, "step": 5184 }, { "epoch": 2.3493429995468964, "grad_norm": 0.4054074825759648, "learning_rate": 5.1046711100844074e-05, "loss": 0.9373, "step": 5185 }, { "epoch": 2.3497961033076575, "grad_norm": 0.49467424217563977, "learning_rate": 5.103454897601971e-05, "loss": 0.9595, "step": 5186 }, { "epoch": 2.3502492070684187, "grad_norm": 0.4985664903451441, "learning_rate": 5.102238574694291e-05, "loss": 0.9272, "step": 5187 }, { "epoch": 2.35070231082918, "grad_norm": 0.43179537106617766, "learning_rate": 5.101022141483089e-05, "loss": 0.9542, "step": 5188 }, { "epoch": 2.351155414589941, "grad_norm": 0.35414792530273514, "learning_rate": 5.0998055980900954e-05, "loss": 0.945, "step": 5189 }, { "epoch": 2.3516085183507025, "grad_norm": 0.2761704591282389, "learning_rate": 5.098588944637051e-05, "loss": 0.9501, "step": 5190 }, { "epoch": 2.3520616221114636, "grad_norm": 0.229045414642189, "learning_rate": 5.09737218124571e-05, "loss": 0.9461, "step": 5191 }, { "epoch": 2.3525147258722248, "grad_norm": 0.2104899820986544, "learning_rate": 5.096155308037838e-05, "loss": 0.9381, "step": 5192 }, { "epoch": 2.352967829632986, "grad_norm": 0.2093052149728301, "learning_rate": 5.094938325135209e-05, "loss": 0.94, "step": 5193 }, { "epoch": 2.353420933393747, "grad_norm": 0.2799927264301506, "learning_rate": 5.093721232659609e-05, "loss": 0.9576, "step": 5194 }, { "epoch": 2.353874037154508, "grad_norm": 0.3083622651067421, "learning_rate": 5.092504030732836e-05, "loss": 0.9399, "step": 5195 }, { "epoch": 2.3543271409152697, "grad_norm": 0.27764323728548335, "learning_rate": 5.0912867194766995e-05, "loss": 0.9567, "step": 5196 }, { "epoch": 2.354780244676031, "grad_norm": 0.2306232129398617, "learning_rate": 5.0900692990130155e-05, "loss": 0.9595, "step": 5197 }, { "epoch": 2.355233348436792, "grad_norm": 0.20697161495165162, "learning_rate": 5.088851769463616e-05, "loss": 0.9636, "step": 5198 }, { "epoch": 2.355686452197553, "grad_norm": 0.212927104604284, "learning_rate": 5.087634130950341e-05, "loss": 0.944, "step": 5199 }, { "epoch": 2.3561395559583147, "grad_norm": 0.22218955173542274, "learning_rate": 5.086416383595044e-05, "loss": 0.9552, "step": 5200 }, { "epoch": 2.356592659719076, "grad_norm": 0.20943245292516025, "learning_rate": 5.085198527519586e-05, "loss": 0.9309, "step": 5201 }, { "epoch": 2.357045763479837, "grad_norm": 0.18886551800549867, "learning_rate": 5.083980562845841e-05, "loss": 0.9294, "step": 5202 }, { "epoch": 2.357498867240598, "grad_norm": 0.18006007374278502, "learning_rate": 5.082762489695694e-05, "loss": 0.9537, "step": 5203 }, { "epoch": 2.3579519710013592, "grad_norm": 0.17353228134103643, "learning_rate": 5.0815443081910396e-05, "loss": 0.9393, "step": 5204 }, { "epoch": 2.3584050747621204, "grad_norm": 0.18848507888625274, "learning_rate": 5.080326018453785e-05, "loss": 0.9315, "step": 5205 }, { "epoch": 2.358858178522882, "grad_norm": 0.19676579019247292, "learning_rate": 5.0791076206058474e-05, "loss": 0.9325, "step": 5206 }, { "epoch": 2.359311282283643, "grad_norm": 0.20835077749018016, "learning_rate": 5.0778891147691536e-05, "loss": 0.9869, "step": 5207 }, { "epoch": 2.359764386044404, "grad_norm": 0.19625093279606795, "learning_rate": 5.0766705010656416e-05, "loss": 0.9701, "step": 5208 }, { "epoch": 2.3602174898051653, "grad_norm": 0.22188620006215531, "learning_rate": 5.075451779617263e-05, "loss": 0.9308, "step": 5209 }, { "epoch": 2.3606705935659265, "grad_norm": 0.23404600894672165, "learning_rate": 5.074232950545976e-05, "loss": 0.9362, "step": 5210 }, { "epoch": 2.361123697326688, "grad_norm": 0.21778131227931152, "learning_rate": 5.073014013973753e-05, "loss": 0.9361, "step": 5211 }, { "epoch": 2.361576801087449, "grad_norm": 0.21367634885843573, "learning_rate": 5.071794970022575e-05, "loss": 0.9422, "step": 5212 }, { "epoch": 2.3620299048482103, "grad_norm": 0.20783028011152588, "learning_rate": 5.070575818814435e-05, "loss": 0.9463, "step": 5213 }, { "epoch": 2.3624830086089714, "grad_norm": 0.205530929483306, "learning_rate": 5.069356560471337e-05, "loss": 0.9302, "step": 5214 }, { "epoch": 2.3629361123697326, "grad_norm": 0.18810631244640116, "learning_rate": 5.0681371951152934e-05, "loss": 0.96, "step": 5215 }, { "epoch": 2.3633892161304937, "grad_norm": 0.19566005158297123, "learning_rate": 5.066917722868329e-05, "loss": 0.9539, "step": 5216 }, { "epoch": 2.3638423198912553, "grad_norm": 0.19910083608749995, "learning_rate": 5.06569814385248e-05, "loss": 0.9671, "step": 5217 }, { "epoch": 2.3642954236520164, "grad_norm": 0.2253378922733282, "learning_rate": 5.0644784581897916e-05, "loss": 0.9458, "step": 5218 }, { "epoch": 2.3647485274127775, "grad_norm": 0.21078853237207876, "learning_rate": 5.0632586660023206e-05, "loss": 0.963, "step": 5219 }, { "epoch": 2.3652016311735387, "grad_norm": 0.18107301415634103, "learning_rate": 5.0620387674121354e-05, "loss": 0.9337, "step": 5220 }, { "epoch": 2.3656547349343, "grad_norm": 0.1859155463778867, "learning_rate": 5.060818762541313e-05, "loss": 0.9415, "step": 5221 }, { "epoch": 2.3661078386950614, "grad_norm": 0.19024408760075537, "learning_rate": 5.059598651511944e-05, "loss": 0.9561, "step": 5222 }, { "epoch": 2.3665609424558225, "grad_norm": 0.19601892241978594, "learning_rate": 5.058378434446124e-05, "loss": 0.9365, "step": 5223 }, { "epoch": 2.3670140462165836, "grad_norm": 0.2755274507585097, "learning_rate": 5.057158111465966e-05, "loss": 0.9454, "step": 5224 }, { "epoch": 2.3674671499773448, "grad_norm": 0.19100223254290363, "learning_rate": 5.0559376826935895e-05, "loss": 0.9433, "step": 5225 }, { "epoch": 2.367920253738106, "grad_norm": 0.3095886278504125, "learning_rate": 5.054717148251126e-05, "loss": 0.9572, "step": 5226 }, { "epoch": 2.368373357498867, "grad_norm": 0.3755504880565492, "learning_rate": 5.0534965082607146e-05, "loss": 0.9546, "step": 5227 }, { "epoch": 2.3688264612596286, "grad_norm": 0.4179684359043026, "learning_rate": 5.052275762844513e-05, "loss": 0.9362, "step": 5228 }, { "epoch": 2.3692795650203897, "grad_norm": 0.41460723963579854, "learning_rate": 5.051054912124679e-05, "loss": 0.9484, "step": 5229 }, { "epoch": 2.369732668781151, "grad_norm": 0.3679395571537325, "learning_rate": 5.049833956223388e-05, "loss": 0.9394, "step": 5230 }, { "epoch": 2.370185772541912, "grad_norm": 0.342308678590856, "learning_rate": 5.048612895262822e-05, "loss": 0.9616, "step": 5231 }, { "epoch": 2.3706388763026736, "grad_norm": 0.3387362321623024, "learning_rate": 5.0473917293651786e-05, "loss": 0.962, "step": 5232 }, { "epoch": 2.3710919800634347, "grad_norm": 0.33354928363405545, "learning_rate": 5.0461704586526605e-05, "loss": 0.9513, "step": 5233 }, { "epoch": 2.371545083824196, "grad_norm": 0.2928572891371414, "learning_rate": 5.044949083247482e-05, "loss": 0.9554, "step": 5234 }, { "epoch": 2.371998187584957, "grad_norm": 0.2183140065862006, "learning_rate": 5.0437276032718714e-05, "loss": 0.9364, "step": 5235 }, { "epoch": 2.372451291345718, "grad_norm": 0.2104211105132973, "learning_rate": 5.0425060188480636e-05, "loss": 0.9381, "step": 5236 }, { "epoch": 2.372904395106479, "grad_norm": 0.23049406911427117, "learning_rate": 5.041284330098306e-05, "loss": 0.9484, "step": 5237 }, { "epoch": 2.373357498867241, "grad_norm": 0.23540746825360692, "learning_rate": 5.040062537144854e-05, "loss": 0.9598, "step": 5238 }, { "epoch": 2.373810602628002, "grad_norm": 0.23846229745359804, "learning_rate": 5.038840640109978e-05, "loss": 0.9543, "step": 5239 }, { "epoch": 2.374263706388763, "grad_norm": 0.2809694745637124, "learning_rate": 5.0376186391159546e-05, "loss": 0.9682, "step": 5240 }, { "epoch": 2.374716810149524, "grad_norm": 0.3002941318748023, "learning_rate": 5.036396534285072e-05, "loss": 0.9502, "step": 5241 }, { "epoch": 2.3751699139102853, "grad_norm": 0.35246257860129077, "learning_rate": 5.035174325739628e-05, "loss": 0.9433, "step": 5242 }, { "epoch": 2.375623017671047, "grad_norm": 0.36028926897220176, "learning_rate": 5.033952013601934e-05, "loss": 0.9506, "step": 5243 }, { "epoch": 2.376076121431808, "grad_norm": 0.35669112757012383, "learning_rate": 5.0327295979943084e-05, "loss": 0.9312, "step": 5244 }, { "epoch": 2.376529225192569, "grad_norm": 0.3834090758133131, "learning_rate": 5.031507079039081e-05, "loss": 0.9121, "step": 5245 }, { "epoch": 2.3769823289533303, "grad_norm": 0.3673986553518731, "learning_rate": 5.0302844568585925e-05, "loss": 0.9405, "step": 5246 }, { "epoch": 2.3774354327140914, "grad_norm": 0.29689147298953716, "learning_rate": 5.029061731575193e-05, "loss": 0.9345, "step": 5247 }, { "epoch": 2.3778885364748525, "grad_norm": 0.22617040933910737, "learning_rate": 5.027838903311244e-05, "loss": 0.9407, "step": 5248 }, { "epoch": 2.378341640235614, "grad_norm": 0.20383888228774255, "learning_rate": 5.026615972189118e-05, "loss": 0.9154, "step": 5249 }, { "epoch": 2.3787947439963752, "grad_norm": 0.23498265613266275, "learning_rate": 5.025392938331193e-05, "loss": 0.9273, "step": 5250 }, { "epoch": 2.3792478477571364, "grad_norm": 0.27893561672576, "learning_rate": 5.024169801859863e-05, "loss": 0.9637, "step": 5251 }, { "epoch": 2.3797009515178975, "grad_norm": 0.28810376616783834, "learning_rate": 5.02294656289753e-05, "loss": 0.9596, "step": 5252 }, { "epoch": 2.3801540552786586, "grad_norm": 0.2890616958742276, "learning_rate": 5.021723221566606e-05, "loss": 0.9666, "step": 5253 }, { "epoch": 2.38060715903942, "grad_norm": 0.3068697060853258, "learning_rate": 5.020499777989514e-05, "loss": 0.9389, "step": 5254 }, { "epoch": 2.3810602628001813, "grad_norm": 0.34231403635111485, "learning_rate": 5.019276232288686e-05, "loss": 0.9439, "step": 5255 }, { "epoch": 2.3815133665609425, "grad_norm": 0.30472989371139836, "learning_rate": 5.0180525845865655e-05, "loss": 0.9621, "step": 5256 }, { "epoch": 2.3819664703217036, "grad_norm": 0.24535945410815663, "learning_rate": 5.0168288350056057e-05, "loss": 0.9496, "step": 5257 }, { "epoch": 2.3824195740824647, "grad_norm": 0.1904544774929958, "learning_rate": 5.01560498366827e-05, "loss": 0.9345, "step": 5258 }, { "epoch": 2.382872677843226, "grad_norm": 0.19212954597136186, "learning_rate": 5.0143810306970324e-05, "loss": 0.9315, "step": 5259 }, { "epoch": 2.3833257816039874, "grad_norm": 0.21436818221383494, "learning_rate": 5.013156976214375e-05, "loss": 0.9359, "step": 5260 }, { "epoch": 2.3837788853647486, "grad_norm": 0.24569655647254154, "learning_rate": 5.011932820342793e-05, "loss": 0.9326, "step": 5261 }, { "epoch": 2.3842319891255097, "grad_norm": 0.23579553811841258, "learning_rate": 5.010708563204791e-05, "loss": 0.9256, "step": 5262 }, { "epoch": 2.384685092886271, "grad_norm": 0.2965650563932622, "learning_rate": 5.0094842049228825e-05, "loss": 0.9291, "step": 5263 }, { "epoch": 2.3851381966470324, "grad_norm": 0.26374376971601593, "learning_rate": 5.008259745619591e-05, "loss": 0.9437, "step": 5264 }, { "epoch": 2.3855913004077935, "grad_norm": 0.2481377839502217, "learning_rate": 5.0070351854174526e-05, "loss": 0.9498, "step": 5265 }, { "epoch": 2.3860444041685547, "grad_norm": 0.26221342495224587, "learning_rate": 5.00581052443901e-05, "loss": 0.9539, "step": 5266 }, { "epoch": 2.386497507929316, "grad_norm": 0.2714373442882191, "learning_rate": 5.004585762806821e-05, "loss": 0.95, "step": 5267 }, { "epoch": 2.386950611690077, "grad_norm": 0.25416471999895024, "learning_rate": 5.003360900643447e-05, "loss": 0.9721, "step": 5268 }, { "epoch": 2.387403715450838, "grad_norm": 0.21166156456834714, "learning_rate": 5.0021359380714635e-05, "loss": 0.9609, "step": 5269 }, { "epoch": 2.3878568192115996, "grad_norm": 0.21157586773437675, "learning_rate": 5.000910875213455e-05, "loss": 0.9559, "step": 5270 }, { "epoch": 2.3883099229723608, "grad_norm": 0.20063901617814425, "learning_rate": 4.999685712192018e-05, "loss": 0.9472, "step": 5271 }, { "epoch": 2.388763026733122, "grad_norm": 0.19812985417847503, "learning_rate": 4.998460449129757e-05, "loss": 0.9524, "step": 5272 }, { "epoch": 2.389216130493883, "grad_norm": 0.20587412866959848, "learning_rate": 4.997235086149284e-05, "loss": 0.9244, "step": 5273 }, { "epoch": 2.389669234254644, "grad_norm": 0.25476702004771024, "learning_rate": 4.996009623373228e-05, "loss": 0.9806, "step": 5274 }, { "epoch": 2.3901223380154057, "grad_norm": 0.27725173685953036, "learning_rate": 4.9947840609242216e-05, "loss": 0.9425, "step": 5275 }, { "epoch": 2.390575441776167, "grad_norm": 0.26437352502611466, "learning_rate": 4.99355839892491e-05, "loss": 0.9341, "step": 5276 }, { "epoch": 2.391028545536928, "grad_norm": 0.2588804295716993, "learning_rate": 4.9923326374979475e-05, "loss": 0.9496, "step": 5277 }, { "epoch": 2.391481649297689, "grad_norm": 0.2710675334119439, "learning_rate": 4.991106776765999e-05, "loss": 0.9292, "step": 5278 }, { "epoch": 2.3919347530584503, "grad_norm": 0.27143497520333776, "learning_rate": 4.98988081685174e-05, "loss": 0.9369, "step": 5279 }, { "epoch": 2.3923878568192114, "grad_norm": 0.2792679136616877, "learning_rate": 4.9886547578778534e-05, "loss": 0.9338, "step": 5280 }, { "epoch": 2.392840960579973, "grad_norm": 0.23103917487323405, "learning_rate": 4.987428599967036e-05, "loss": 0.9313, "step": 5281 }, { "epoch": 2.393294064340734, "grad_norm": 0.24117594976352602, "learning_rate": 4.9862023432419916e-05, "loss": 0.9515, "step": 5282 }, { "epoch": 2.3937471681014952, "grad_norm": 0.23681066834783177, "learning_rate": 4.984975987825433e-05, "loss": 0.9551, "step": 5283 }, { "epoch": 2.3942002718622564, "grad_norm": 0.20159962846557164, "learning_rate": 4.983749533840086e-05, "loss": 0.9203, "step": 5284 }, { "epoch": 2.3946533756230175, "grad_norm": 0.2027037895662948, "learning_rate": 4.982522981408684e-05, "loss": 0.9337, "step": 5285 }, { "epoch": 2.395106479383779, "grad_norm": 0.19918306313814682, "learning_rate": 4.98129633065397e-05, "loss": 0.9515, "step": 5286 }, { "epoch": 2.39555958314454, "grad_norm": 0.21564343092603372, "learning_rate": 4.9800695816986986e-05, "loss": 0.9511, "step": 5287 }, { "epoch": 2.3960126869053013, "grad_norm": 0.20728585333967392, "learning_rate": 4.978842734665634e-05, "loss": 0.9584, "step": 5288 }, { "epoch": 2.3964657906660625, "grad_norm": 0.246283758977161, "learning_rate": 4.977615789677549e-05, "loss": 0.9457, "step": 5289 }, { "epoch": 2.3969188944268236, "grad_norm": 0.25343019342907636, "learning_rate": 4.976388746857226e-05, "loss": 0.9404, "step": 5290 }, { "epoch": 2.3973719981875847, "grad_norm": 0.20495445322898817, "learning_rate": 4.975161606327459e-05, "loss": 0.9363, "step": 5291 }, { "epoch": 2.3978251019483463, "grad_norm": 0.20094672489655188, "learning_rate": 4.97393436821105e-05, "loss": 0.9531, "step": 5292 }, { "epoch": 2.3982782057091074, "grad_norm": 0.2049691844055876, "learning_rate": 4.972707032630814e-05, "loss": 0.9561, "step": 5293 }, { "epoch": 2.3987313094698686, "grad_norm": 0.19552774948507243, "learning_rate": 4.971479599709569e-05, "loss": 0.9293, "step": 5294 }, { "epoch": 2.3991844132306297, "grad_norm": 0.2183152230086541, "learning_rate": 4.9702520695701494e-05, "loss": 0.9271, "step": 5295 }, { "epoch": 2.3996375169913913, "grad_norm": 0.22226084775121058, "learning_rate": 4.969024442335397e-05, "loss": 0.9746, "step": 5296 }, { "epoch": 2.4000906207521524, "grad_norm": 0.26070224646523305, "learning_rate": 4.9677967181281636e-05, "loss": 0.9537, "step": 5297 }, { "epoch": 2.4005437245129135, "grad_norm": 0.2382502800054207, "learning_rate": 4.9665688970713095e-05, "loss": 0.948, "step": 5298 }, { "epoch": 2.4009968282736747, "grad_norm": 0.1998378779064115, "learning_rate": 4.965340979287706e-05, "loss": 0.9436, "step": 5299 }, { "epoch": 2.401449932034436, "grad_norm": 0.18714865377792186, "learning_rate": 4.9641129649002323e-05, "loss": 0.9427, "step": 5300 }, { "epoch": 2.401903035795197, "grad_norm": 0.2329949392280846, "learning_rate": 4.962884854031782e-05, "loss": 0.9404, "step": 5301 }, { "epoch": 2.4023561395559585, "grad_norm": 0.2764637438377823, "learning_rate": 4.961656646805251e-05, "loss": 0.9521, "step": 5302 }, { "epoch": 2.4028092433167196, "grad_norm": 0.25187781350133553, "learning_rate": 4.960428343343551e-05, "loss": 0.9461, "step": 5303 }, { "epoch": 2.4032623470774808, "grad_norm": 0.20066319650534237, "learning_rate": 4.9591999437695994e-05, "loss": 0.9536, "step": 5304 }, { "epoch": 2.403715450838242, "grad_norm": 0.19697509333117086, "learning_rate": 4.957971448206327e-05, "loss": 0.943, "step": 5305 }, { "epoch": 2.404168554599003, "grad_norm": 0.2601705056257675, "learning_rate": 4.956742856776671e-05, "loss": 0.9376, "step": 5306 }, { "epoch": 2.4046216583597646, "grad_norm": 0.322135993969517, "learning_rate": 4.955514169603581e-05, "loss": 0.9411, "step": 5307 }, { "epoch": 2.4050747621205257, "grad_norm": 0.3086149585417656, "learning_rate": 4.954285386810011e-05, "loss": 0.9657, "step": 5308 }, { "epoch": 2.405527865881287, "grad_norm": 0.2592134596231675, "learning_rate": 4.953056508518932e-05, "loss": 0.9245, "step": 5309 }, { "epoch": 2.405980969642048, "grad_norm": 0.26439406068148164, "learning_rate": 4.951827534853318e-05, "loss": 0.9702, "step": 5310 }, { "epoch": 2.406434073402809, "grad_norm": 0.30013361080020723, "learning_rate": 4.950598465936157e-05, "loss": 0.9423, "step": 5311 }, { "epoch": 2.4068871771635703, "grad_norm": 0.2997673407439418, "learning_rate": 4.9493693018904425e-05, "loss": 0.9408, "step": 5312 }, { "epoch": 2.407340280924332, "grad_norm": 0.2540974702335248, "learning_rate": 4.9481400428391816e-05, "loss": 0.9644, "step": 5313 }, { "epoch": 2.407793384685093, "grad_norm": 0.22328738097753092, "learning_rate": 4.946910688905389e-05, "loss": 0.9452, "step": 5314 }, { "epoch": 2.408246488445854, "grad_norm": 0.22997343318108485, "learning_rate": 4.9456812402120886e-05, "loss": 0.9523, "step": 5315 }, { "epoch": 2.4086995922066152, "grad_norm": 0.23365208937097826, "learning_rate": 4.9444516968823133e-05, "loss": 0.9616, "step": 5316 }, { "epoch": 2.4091526959673764, "grad_norm": 0.22672323856287616, "learning_rate": 4.943222059039107e-05, "loss": 0.9557, "step": 5317 }, { "epoch": 2.409605799728138, "grad_norm": 0.24334809281160633, "learning_rate": 4.941992326805522e-05, "loss": 0.9706, "step": 5318 }, { "epoch": 2.410058903488899, "grad_norm": 0.21843645354359445, "learning_rate": 4.9407625003046215e-05, "loss": 0.9495, "step": 5319 }, { "epoch": 2.41051200724966, "grad_norm": 0.20430223525478455, "learning_rate": 4.9395325796594764e-05, "loss": 0.9519, "step": 5320 }, { "epoch": 2.4109651110104213, "grad_norm": 0.2081819504020259, "learning_rate": 4.938302564993167e-05, "loss": 0.9335, "step": 5321 }, { "epoch": 2.4114182147711825, "grad_norm": 0.2200366223700552, "learning_rate": 4.937072456428784e-05, "loss": 0.9601, "step": 5322 }, { "epoch": 2.4118713185319436, "grad_norm": 0.20643323404242447, "learning_rate": 4.935842254089427e-05, "loss": 0.943, "step": 5323 }, { "epoch": 2.412324422292705, "grad_norm": 0.21068811936997733, "learning_rate": 4.934611958098206e-05, "loss": 0.9441, "step": 5324 }, { "epoch": 2.4127775260534663, "grad_norm": 0.2556546119183275, "learning_rate": 4.933381568578239e-05, "loss": 0.9475, "step": 5325 }, { "epoch": 2.4132306298142274, "grad_norm": 0.22880594801900123, "learning_rate": 4.932151085652653e-05, "loss": 0.9331, "step": 5326 }, { "epoch": 2.4136837335749886, "grad_norm": 0.2492955953975364, "learning_rate": 4.930920509444588e-05, "loss": 0.9543, "step": 5327 }, { "epoch": 2.41413683733575, "grad_norm": 0.2597019479052959, "learning_rate": 4.9296898400771884e-05, "loss": 0.9651, "step": 5328 }, { "epoch": 2.4145899410965113, "grad_norm": 0.26410208228376253, "learning_rate": 4.92845907767361e-05, "loss": 0.9429, "step": 5329 }, { "epoch": 2.4150430448572724, "grad_norm": 0.31195909313237646, "learning_rate": 4.927228222357018e-05, "loss": 0.9458, "step": 5330 }, { "epoch": 2.4154961486180335, "grad_norm": 0.326707292057796, "learning_rate": 4.925997274250588e-05, "loss": 0.9449, "step": 5331 }, { "epoch": 2.4159492523787947, "grad_norm": 0.26975819669308, "learning_rate": 4.9247662334775016e-05, "loss": 0.9365, "step": 5332 }, { "epoch": 2.416402356139556, "grad_norm": 0.23013255923685996, "learning_rate": 4.923535100160954e-05, "loss": 0.9631, "step": 5333 }, { "epoch": 2.4168554599003174, "grad_norm": 0.23244159582841864, "learning_rate": 4.922303874424147e-05, "loss": 0.9482, "step": 5334 }, { "epoch": 2.4173085636610785, "grad_norm": 0.23586619854986754, "learning_rate": 4.921072556390292e-05, "loss": 0.9343, "step": 5335 }, { "epoch": 2.4177616674218396, "grad_norm": 0.259050533345318, "learning_rate": 4.91984114618261e-05, "loss": 0.9858, "step": 5336 }, { "epoch": 2.4182147711826008, "grad_norm": 0.21874882000822377, "learning_rate": 4.918609643924331e-05, "loss": 0.9397, "step": 5337 }, { "epoch": 2.418667874943362, "grad_norm": 0.2072406819959031, "learning_rate": 4.917378049738694e-05, "loss": 0.9311, "step": 5338 }, { "epoch": 2.4191209787041235, "grad_norm": 0.21287810239584512, "learning_rate": 4.916146363748946e-05, "loss": 0.9401, "step": 5339 }, { "epoch": 2.4195740824648846, "grad_norm": 0.1666854354780887, "learning_rate": 4.914914586078347e-05, "loss": 0.9386, "step": 5340 }, { "epoch": 2.4200271862256457, "grad_norm": 0.16024119914100293, "learning_rate": 4.9136827168501634e-05, "loss": 0.9567, "step": 5341 }, { "epoch": 2.420480289986407, "grad_norm": 0.20444120452607392, "learning_rate": 4.912450756187671e-05, "loss": 0.9436, "step": 5342 }, { "epoch": 2.420933393747168, "grad_norm": 0.21319002528216105, "learning_rate": 4.911218704214154e-05, "loss": 0.9536, "step": 5343 }, { "epoch": 2.421386497507929, "grad_norm": 0.20200109686282516, "learning_rate": 4.909986561052907e-05, "loss": 0.9427, "step": 5344 }, { "epoch": 2.4218396012686907, "grad_norm": 0.21845653970350645, "learning_rate": 4.9087543268272334e-05, "loss": 0.934, "step": 5345 }, { "epoch": 2.422292705029452, "grad_norm": 0.2729032582249965, "learning_rate": 4.9075220016604463e-05, "loss": 0.9469, "step": 5346 }, { "epoch": 2.422745808790213, "grad_norm": 0.3382969096143951, "learning_rate": 4.906289585675867e-05, "loss": 0.9515, "step": 5347 }, { "epoch": 2.423198912550974, "grad_norm": 0.35699502222423274, "learning_rate": 4.9050570789968255e-05, "loss": 0.9244, "step": 5348 }, { "epoch": 2.423652016311735, "grad_norm": 0.35290623171997365, "learning_rate": 4.903824481746662e-05, "loss": 0.9496, "step": 5349 }, { "epoch": 2.424105120072497, "grad_norm": 0.34247910376335106, "learning_rate": 4.902591794048725e-05, "loss": 0.9558, "step": 5350 }, { "epoch": 2.424558223833258, "grad_norm": 0.3251600606375128, "learning_rate": 4.9013590160263737e-05, "loss": 0.9434, "step": 5351 }, { "epoch": 2.425011327594019, "grad_norm": 0.2996300679562962, "learning_rate": 4.900126147802973e-05, "loss": 0.9274, "step": 5352 }, { "epoch": 2.42546443135478, "grad_norm": 0.26894443001195595, "learning_rate": 4.8988931895019e-05, "loss": 0.912, "step": 5353 }, { "epoch": 2.4259175351155413, "grad_norm": 0.2112546163569498, "learning_rate": 4.8976601412465415e-05, "loss": 0.9232, "step": 5354 }, { "epoch": 2.4263706388763024, "grad_norm": 0.16866158337943188, "learning_rate": 4.896427003160287e-05, "loss": 0.9515, "step": 5355 }, { "epoch": 2.426823742637064, "grad_norm": 0.21318322822243627, "learning_rate": 4.895193775366543e-05, "loss": 0.9168, "step": 5356 }, { "epoch": 2.427276846397825, "grad_norm": 0.3110554617835189, "learning_rate": 4.89396045798872e-05, "loss": 0.9738, "step": 5357 }, { "epoch": 2.4277299501585863, "grad_norm": 0.41057156944762785, "learning_rate": 4.8927270511502386e-05, "loss": 0.9637, "step": 5358 }, { "epoch": 2.4281830539193474, "grad_norm": 0.47592347647485894, "learning_rate": 4.8914935549745305e-05, "loss": 0.9544, "step": 5359 }, { "epoch": 2.428636157680109, "grad_norm": 0.47159219091578203, "learning_rate": 4.890259969585032e-05, "loss": 0.9303, "step": 5360 }, { "epoch": 2.42908926144087, "grad_norm": 0.44191284279254456, "learning_rate": 4.8890262951051925e-05, "loss": 0.9415, "step": 5361 }, { "epoch": 2.4295423652016312, "grad_norm": 0.36859805898856657, "learning_rate": 4.8877925316584686e-05, "loss": 0.9542, "step": 5362 }, { "epoch": 2.4299954689623924, "grad_norm": 0.2713969484759124, "learning_rate": 4.8865586793683245e-05, "loss": 0.939, "step": 5363 }, { "epoch": 2.4304485727231535, "grad_norm": 0.17582216846698134, "learning_rate": 4.8853247383582364e-05, "loss": 0.9551, "step": 5364 }, { "epoch": 2.4309016764839146, "grad_norm": 0.2281957288303571, "learning_rate": 4.8840907087516855e-05, "loss": 0.9446, "step": 5365 }, { "epoch": 2.431354780244676, "grad_norm": 0.357490320829992, "learning_rate": 4.882856590672165e-05, "loss": 0.9508, "step": 5366 }, { "epoch": 2.4318078840054373, "grad_norm": 0.3928406338633433, "learning_rate": 4.881622384243177e-05, "loss": 0.9594, "step": 5367 }, { "epoch": 2.4322609877661985, "grad_norm": 0.4037928265736473, "learning_rate": 4.880388089588229e-05, "loss": 0.9594, "step": 5368 }, { "epoch": 2.4327140915269596, "grad_norm": 0.4010174846129544, "learning_rate": 4.879153706830842e-05, "loss": 0.9484, "step": 5369 }, { "epoch": 2.4331671952877207, "grad_norm": 0.3452654029808446, "learning_rate": 4.877919236094542e-05, "loss": 0.939, "step": 5370 }, { "epoch": 2.4336202990484823, "grad_norm": 0.2419733011257948, "learning_rate": 4.8766846775028656e-05, "loss": 0.9264, "step": 5371 }, { "epoch": 2.4340734028092434, "grad_norm": 0.1866614993144044, "learning_rate": 4.8754500311793596e-05, "loss": 0.9326, "step": 5372 }, { "epoch": 2.4345265065700046, "grad_norm": 0.18341903150886568, "learning_rate": 4.8742152972475746e-05, "loss": 0.9427, "step": 5373 }, { "epoch": 2.4349796103307657, "grad_norm": 0.1820839098910212, "learning_rate": 4.872980475831075e-05, "loss": 0.9635, "step": 5374 }, { "epoch": 2.435432714091527, "grad_norm": 0.21081340996570047, "learning_rate": 4.8717455670534326e-05, "loss": 0.9528, "step": 5375 }, { "epoch": 2.435885817852288, "grad_norm": 0.24153989961956748, "learning_rate": 4.870510571038227e-05, "loss": 0.9717, "step": 5376 }, { "epoch": 2.4363389216130495, "grad_norm": 0.2613008059508991, "learning_rate": 4.869275487909047e-05, "loss": 0.943, "step": 5377 }, { "epoch": 2.4367920253738107, "grad_norm": 0.26700198576371287, "learning_rate": 4.868040317789491e-05, "loss": 0.9591, "step": 5378 }, { "epoch": 2.437245129134572, "grad_norm": 0.29966327258080133, "learning_rate": 4.866805060803163e-05, "loss": 0.9649, "step": 5379 }, { "epoch": 2.437698232895333, "grad_norm": 0.329006560454664, "learning_rate": 4.865569717073682e-05, "loss": 0.9678, "step": 5380 }, { "epoch": 2.438151336656094, "grad_norm": 0.34589294814609006, "learning_rate": 4.864334286724668e-05, "loss": 0.9299, "step": 5381 }, { "epoch": 2.4386044404168556, "grad_norm": 0.3102783678790292, "learning_rate": 4.863098769879754e-05, "loss": 0.9585, "step": 5382 }, { "epoch": 2.4390575441776168, "grad_norm": 0.243702821047354, "learning_rate": 4.861863166662581e-05, "loss": 0.9457, "step": 5383 }, { "epoch": 2.439510647938378, "grad_norm": 0.19520667463030747, "learning_rate": 4.8606274771968005e-05, "loss": 0.9535, "step": 5384 }, { "epoch": 2.439963751699139, "grad_norm": 0.19177936686456806, "learning_rate": 4.859391701606069e-05, "loss": 0.9529, "step": 5385 }, { "epoch": 2.4404168554599, "grad_norm": 0.19945148550501932, "learning_rate": 4.858155840014054e-05, "loss": 0.9738, "step": 5386 }, { "epoch": 2.4408699592206613, "grad_norm": 0.22680290677239703, "learning_rate": 4.85691989254443e-05, "loss": 0.9638, "step": 5387 }, { "epoch": 2.441323062981423, "grad_norm": 0.2558449784871807, "learning_rate": 4.8556838593208845e-05, "loss": 0.9458, "step": 5388 }, { "epoch": 2.441776166742184, "grad_norm": 0.25167311817291016, "learning_rate": 4.8544477404671054e-05, "loss": 0.9292, "step": 5389 }, { "epoch": 2.442229270502945, "grad_norm": 0.25543097441580626, "learning_rate": 4.853211536106796e-05, "loss": 0.942, "step": 5390 }, { "epoch": 2.4426823742637063, "grad_norm": 0.2357124127967925, "learning_rate": 4.851975246363666e-05, "loss": 0.9757, "step": 5391 }, { "epoch": 2.443135478024468, "grad_norm": 0.23473866309958058, "learning_rate": 4.850738871361434e-05, "loss": 0.9564, "step": 5392 }, { "epoch": 2.443588581785229, "grad_norm": 0.27606485535991226, "learning_rate": 4.849502411223827e-05, "loss": 0.9465, "step": 5393 }, { "epoch": 2.44404168554599, "grad_norm": 0.31911414578704983, "learning_rate": 4.84826586607458e-05, "loss": 0.9426, "step": 5394 }, { "epoch": 2.4444947893067512, "grad_norm": 0.23730759460362713, "learning_rate": 4.847029236037437e-05, "loss": 0.9498, "step": 5395 }, { "epoch": 2.4449478930675124, "grad_norm": 0.2693721903808171, "learning_rate": 4.84579252123615e-05, "loss": 0.9594, "step": 5396 }, { "epoch": 2.4454009968282735, "grad_norm": 0.2834037722063052, "learning_rate": 4.844555721794482e-05, "loss": 0.9659, "step": 5397 }, { "epoch": 2.445854100589035, "grad_norm": 0.22183396507123726, "learning_rate": 4.843318837836199e-05, "loss": 0.9407, "step": 5398 }, { "epoch": 2.446307204349796, "grad_norm": 0.25073701723204017, "learning_rate": 4.842081869485081e-05, "loss": 0.9455, "step": 5399 }, { "epoch": 2.4467603081105573, "grad_norm": 0.26806710796931815, "learning_rate": 4.840844816864914e-05, "loss": 0.9626, "step": 5400 }, { "epoch": 2.4472134118713185, "grad_norm": 0.2545477305920185, "learning_rate": 4.839607680099491e-05, "loss": 0.9426, "step": 5401 }, { "epoch": 2.4476665156320796, "grad_norm": 0.1983849870849048, "learning_rate": 4.838370459312616e-05, "loss": 0.949, "step": 5402 }, { "epoch": 2.448119619392841, "grad_norm": 0.19626891221696774, "learning_rate": 4.8371331546281025e-05, "loss": 0.9477, "step": 5403 }, { "epoch": 2.4485727231536023, "grad_norm": 0.2461634596002457, "learning_rate": 4.835895766169768e-05, "loss": 0.9535, "step": 5404 }, { "epoch": 2.4490258269143634, "grad_norm": 0.24209600258793226, "learning_rate": 4.8346582940614414e-05, "loss": 0.9493, "step": 5405 }, { "epoch": 2.4494789306751246, "grad_norm": 0.1880346112840367, "learning_rate": 4.8334207384269596e-05, "loss": 0.9393, "step": 5406 }, { "epoch": 2.4499320344358857, "grad_norm": 0.21357662576525133, "learning_rate": 4.832183099390168e-05, "loss": 0.9469, "step": 5407 }, { "epoch": 2.450385138196647, "grad_norm": 0.20818344142601508, "learning_rate": 4.830945377074919e-05, "loss": 0.9453, "step": 5408 }, { "epoch": 2.4508382419574084, "grad_norm": 0.19196056827688548, "learning_rate": 4.829707571605075e-05, "loss": 0.9198, "step": 5409 }, { "epoch": 2.4512913457181695, "grad_norm": 0.1910527298998238, "learning_rate": 4.828469683104505e-05, "loss": 0.9665, "step": 5410 }, { "epoch": 2.4517444494789307, "grad_norm": 0.1890457567645241, "learning_rate": 4.827231711697089e-05, "loss": 0.9377, "step": 5411 }, { "epoch": 2.452197553239692, "grad_norm": 0.19323235330641256, "learning_rate": 4.825993657506712e-05, "loss": 0.9229, "step": 5412 }, { "epoch": 2.452650657000453, "grad_norm": 0.17548399417160931, "learning_rate": 4.82475552065727e-05, "loss": 0.9653, "step": 5413 }, { "epoch": 2.4531037607612145, "grad_norm": 0.16995496242660654, "learning_rate": 4.823517301272666e-05, "loss": 0.9542, "step": 5414 }, { "epoch": 2.4535568645219756, "grad_norm": 0.16456433157889444, "learning_rate": 4.822278999476811e-05, "loss": 0.9505, "step": 5415 }, { "epoch": 2.4540099682827368, "grad_norm": 0.1931237947104999, "learning_rate": 4.821040615393624e-05, "loss": 0.9627, "step": 5416 }, { "epoch": 2.454463072043498, "grad_norm": 0.22187189697071483, "learning_rate": 4.8198021491470355e-05, "loss": 0.9498, "step": 5417 }, { "epoch": 2.454916175804259, "grad_norm": 0.2834931496571174, "learning_rate": 4.8185636008609785e-05, "loss": 0.9433, "step": 5418 }, { "epoch": 2.45536927956502, "grad_norm": 0.33959653327789746, "learning_rate": 4.8173249706594e-05, "loss": 0.9289, "step": 5419 }, { "epoch": 2.4558223833257817, "grad_norm": 0.36906943712237, "learning_rate": 4.81608625866625e-05, "loss": 0.9585, "step": 5420 }, { "epoch": 2.456275487086543, "grad_norm": 0.34565546307439066, "learning_rate": 4.814847465005491e-05, "loss": 0.9409, "step": 5421 }, { "epoch": 2.456728590847304, "grad_norm": 0.29441909040658953, "learning_rate": 4.813608589801092e-05, "loss": 0.9359, "step": 5422 }, { "epoch": 2.457181694608065, "grad_norm": 0.2519608263801979, "learning_rate": 4.81236963317703e-05, "loss": 0.9441, "step": 5423 }, { "epoch": 2.4576347983688267, "grad_norm": 0.2335787322597531, "learning_rate": 4.811130595257289e-05, "loss": 0.9428, "step": 5424 }, { "epoch": 2.458087902129588, "grad_norm": 0.1961887981526555, "learning_rate": 4.809891476165863e-05, "loss": 0.9305, "step": 5425 }, { "epoch": 2.458541005890349, "grad_norm": 0.20025739381415428, "learning_rate": 4.8086522760267535e-05, "loss": 0.9536, "step": 5426 }, { "epoch": 2.45899410965111, "grad_norm": 0.18597282588226804, "learning_rate": 4.807412994963969e-05, "loss": 0.9392, "step": 5427 }, { "epoch": 2.4594472134118712, "grad_norm": 0.18344973127530517, "learning_rate": 4.806173633101531e-05, "loss": 0.9597, "step": 5428 }, { "epoch": 2.4599003171726324, "grad_norm": 0.2126323898156542, "learning_rate": 4.804934190563461e-05, "loss": 0.929, "step": 5429 }, { "epoch": 2.460353420933394, "grad_norm": 0.19428645235693762, "learning_rate": 4.803694667473794e-05, "loss": 0.9533, "step": 5430 }, { "epoch": 2.460806524694155, "grad_norm": 0.1884634296058354, "learning_rate": 4.802455063956572e-05, "loss": 0.9396, "step": 5431 }, { "epoch": 2.461259628454916, "grad_norm": 0.2587065788506596, "learning_rate": 4.801215380135847e-05, "loss": 0.9466, "step": 5432 }, { "epoch": 2.4617127322156773, "grad_norm": 0.24608282286306069, "learning_rate": 4.799975616135674e-05, "loss": 0.9641, "step": 5433 }, { "epoch": 2.4621658359764385, "grad_norm": 0.25505996479544035, "learning_rate": 4.798735772080121e-05, "loss": 0.9477, "step": 5434 }, { "epoch": 2.4626189397372, "grad_norm": 0.29186440138612485, "learning_rate": 4.7974958480932594e-05, "loss": 0.961, "step": 5435 }, { "epoch": 2.463072043497961, "grad_norm": 0.2595783524601156, "learning_rate": 4.7962558442991745e-05, "loss": 0.9518, "step": 5436 }, { "epoch": 2.4635251472587223, "grad_norm": 0.6267036361909841, "learning_rate": 4.7950157608219546e-05, "loss": 0.9444, "step": 5437 }, { "epoch": 2.4639782510194834, "grad_norm": 0.17651896038624174, "learning_rate": 4.793775597785698e-05, "loss": 0.9291, "step": 5438 }, { "epoch": 2.4644313547802446, "grad_norm": 0.22666403132640342, "learning_rate": 4.792535355314511e-05, "loss": 0.9518, "step": 5439 }, { "epoch": 2.4648844585410057, "grad_norm": 0.3396735321125199, "learning_rate": 4.791295033532506e-05, "loss": 0.957, "step": 5440 }, { "epoch": 2.4653375623017673, "grad_norm": 0.20589605292211258, "learning_rate": 4.7900546325638075e-05, "loss": 0.9333, "step": 5441 }, { "epoch": 2.4657906660625284, "grad_norm": 0.179107464044486, "learning_rate": 4.788814152532542e-05, "loss": 0.9558, "step": 5442 }, { "epoch": 2.4662437698232895, "grad_norm": 0.21293292209952822, "learning_rate": 4.787573593562849e-05, "loss": 0.9462, "step": 5443 }, { "epoch": 2.4666968735840507, "grad_norm": 0.24705149096716106, "learning_rate": 4.786332955778875e-05, "loss": 0.9451, "step": 5444 }, { "epoch": 2.467149977344812, "grad_norm": 0.23956655091443052, "learning_rate": 4.785092239304771e-05, "loss": 0.9472, "step": 5445 }, { "epoch": 2.4676030811055734, "grad_norm": 0.25385396831601253, "learning_rate": 4.7838514442647e-05, "loss": 0.9394, "step": 5446 }, { "epoch": 2.4680561848663345, "grad_norm": 0.2828125372776677, "learning_rate": 4.782610570782831e-05, "loss": 0.9424, "step": 5447 }, { "epoch": 2.4685092886270956, "grad_norm": 0.28601880619692494, "learning_rate": 4.78136961898334e-05, "loss": 0.9553, "step": 5448 }, { "epoch": 2.4689623923878568, "grad_norm": 0.2580659798523909, "learning_rate": 4.780128588990414e-05, "loss": 0.9409, "step": 5449 }, { "epoch": 2.469415496148618, "grad_norm": 0.23895016466442728, "learning_rate": 4.7788874809282445e-05, "loss": 0.9568, "step": 5450 }, { "epoch": 2.469868599909379, "grad_norm": 0.2490703393882188, "learning_rate": 4.777646294921031e-05, "loss": 0.9312, "step": 5451 }, { "epoch": 2.4703217036701406, "grad_norm": 0.234508605134918, "learning_rate": 4.776405031092983e-05, "loss": 0.9532, "step": 5452 }, { "epoch": 2.4707748074309017, "grad_norm": 0.1987231098569375, "learning_rate": 4.775163689568317e-05, "loss": 0.9412, "step": 5453 }, { "epoch": 2.471227911191663, "grad_norm": 0.24932048677124063, "learning_rate": 4.773922270471255e-05, "loss": 0.9487, "step": 5454 }, { "epoch": 2.471681014952424, "grad_norm": 0.2800575862450085, "learning_rate": 4.772680773926031e-05, "loss": 0.9358, "step": 5455 }, { "epoch": 2.4721341187131856, "grad_norm": 0.23747981527656778, "learning_rate": 4.771439200056882e-05, "loss": 0.9325, "step": 5456 }, { "epoch": 2.4725872224739467, "grad_norm": 0.22286503832540905, "learning_rate": 4.770197548988057e-05, "loss": 0.9463, "step": 5457 }, { "epoch": 2.473040326234708, "grad_norm": 0.24694703915225907, "learning_rate": 4.76895582084381e-05, "loss": 0.9136, "step": 5458 }, { "epoch": 2.473493429995469, "grad_norm": 0.19915831591012825, "learning_rate": 4.767714015748405e-05, "loss": 0.9429, "step": 5459 }, { "epoch": 2.47394653375623, "grad_norm": 0.1718179742334371, "learning_rate": 4.76647213382611e-05, "loss": 0.9481, "step": 5460 }, { "epoch": 2.474399637516991, "grad_norm": 0.1986100379004953, "learning_rate": 4.765230175201203e-05, "loss": 0.9349, "step": 5461 }, { "epoch": 2.474852741277753, "grad_norm": 0.21858375999394758, "learning_rate": 4.7639881399979724e-05, "loss": 0.9337, "step": 5462 }, { "epoch": 2.475305845038514, "grad_norm": 0.19948664047345457, "learning_rate": 4.762746028340708e-05, "loss": 0.9516, "step": 5463 }, { "epoch": 2.475758948799275, "grad_norm": 0.18476342163904094, "learning_rate": 4.761503840353713e-05, "loss": 0.9464, "step": 5464 }, { "epoch": 2.476212052560036, "grad_norm": 0.2140281739059919, "learning_rate": 4.760261576161295e-05, "loss": 0.9144, "step": 5465 }, { "epoch": 2.4766651563207973, "grad_norm": 0.2195466292961448, "learning_rate": 4.759019235887771e-05, "loss": 0.9451, "step": 5466 }, { "epoch": 2.477118260081559, "grad_norm": 0.2108424021126099, "learning_rate": 4.7577768196574646e-05, "loss": 0.9438, "step": 5467 }, { "epoch": 2.47757136384232, "grad_norm": 0.22929427639905997, "learning_rate": 4.756534327594708e-05, "loss": 0.9509, "step": 5468 }, { "epoch": 2.478024467603081, "grad_norm": 0.2987431955223949, "learning_rate": 4.7552917598238374e-05, "loss": 0.9393, "step": 5469 }, { "epoch": 2.4784775713638423, "grad_norm": 0.3656449830412961, "learning_rate": 4.754049116469202e-05, "loss": 0.9556, "step": 5470 }, { "epoch": 2.4789306751246034, "grad_norm": 0.3888824763431453, "learning_rate": 4.752806397655155e-05, "loss": 0.9438, "step": 5471 }, { "epoch": 2.4793837788853645, "grad_norm": 0.3461249312821175, "learning_rate": 4.7515636035060586e-05, "loss": 0.9535, "step": 5472 }, { "epoch": 2.479836882646126, "grad_norm": 0.2948408346329638, "learning_rate": 4.7503207341462815e-05, "loss": 0.9291, "step": 5473 }, { "epoch": 2.4802899864068872, "grad_norm": 0.2610647258797239, "learning_rate": 4.7490777897002e-05, "loss": 0.9545, "step": 5474 }, { "epoch": 2.4807430901676484, "grad_norm": 0.23406586686477787, "learning_rate": 4.7478347702922005e-05, "loss": 0.9204, "step": 5475 }, { "epoch": 2.4811961939284095, "grad_norm": 0.22026423539536935, "learning_rate": 4.746591676046673e-05, "loss": 0.9423, "step": 5476 }, { "epoch": 2.4816492976891706, "grad_norm": 0.22676037251222234, "learning_rate": 4.745348507088017e-05, "loss": 0.9441, "step": 5477 }, { "epoch": 2.482102401449932, "grad_norm": 0.19880361393480186, "learning_rate": 4.744105263540639e-05, "loss": 0.9549, "step": 5478 }, { "epoch": 2.4825555052106933, "grad_norm": 0.18751759810588997, "learning_rate": 4.742861945528954e-05, "loss": 0.9593, "step": 5479 }, { "epoch": 2.4830086089714545, "grad_norm": 0.20323208536721432, "learning_rate": 4.7416185531773824e-05, "loss": 0.9419, "step": 5480 }, { "epoch": 2.4834617127322156, "grad_norm": 0.27610249583303204, "learning_rate": 4.7403750866103553e-05, "loss": 0.9322, "step": 5481 }, { "epoch": 2.4839148164929767, "grad_norm": 0.3231628170875361, "learning_rate": 4.739131545952306e-05, "loss": 0.9311, "step": 5482 }, { "epoch": 2.484367920253738, "grad_norm": 0.2683600174185857, "learning_rate": 4.737887931327683e-05, "loss": 0.9551, "step": 5483 }, { "epoch": 2.4848210240144994, "grad_norm": 0.20967116283356715, "learning_rate": 4.7366442428609347e-05, "loss": 0.9429, "step": 5484 }, { "epoch": 2.4852741277752606, "grad_norm": 0.2066722020975361, "learning_rate": 4.7354004806765196e-05, "loss": 0.9291, "step": 5485 }, { "epoch": 2.4857272315360217, "grad_norm": 0.275997873878797, "learning_rate": 4.7341566448989046e-05, "loss": 0.9491, "step": 5486 }, { "epoch": 2.486180335296783, "grad_norm": 0.29320471780642793, "learning_rate": 4.732912735652564e-05, "loss": 0.9561, "step": 5487 }, { "epoch": 2.4866334390575444, "grad_norm": 0.275260921853944, "learning_rate": 4.731668753061977e-05, "loss": 0.9405, "step": 5488 }, { "epoch": 2.4870865428183055, "grad_norm": 0.25024882020186967, "learning_rate": 4.7304246972516324e-05, "loss": 0.923, "step": 5489 }, { "epoch": 2.4875396465790667, "grad_norm": 0.2633831022752552, "learning_rate": 4.7291805683460254e-05, "loss": 0.9483, "step": 5490 }, { "epoch": 2.487992750339828, "grad_norm": 0.28107334357414326, "learning_rate": 4.727936366469659e-05, "loss": 0.9319, "step": 5491 }, { "epoch": 2.488445854100589, "grad_norm": 0.23522445514632934, "learning_rate": 4.726692091747044e-05, "loss": 0.952, "step": 5492 }, { "epoch": 2.48889895786135, "grad_norm": 0.2242684279923168, "learning_rate": 4.725447744302697e-05, "loss": 0.9492, "step": 5493 }, { "epoch": 2.4893520616221116, "grad_norm": 0.2447809078102598, "learning_rate": 4.724203324261145e-05, "loss": 0.9493, "step": 5494 }, { "epoch": 2.4898051653828728, "grad_norm": 0.25959825737378145, "learning_rate": 4.7229588317469155e-05, "loss": 0.9595, "step": 5495 }, { "epoch": 2.490258269143634, "grad_norm": 0.20749071093218346, "learning_rate": 4.7217142668845507e-05, "loss": 0.9404, "step": 5496 }, { "epoch": 2.490711372904395, "grad_norm": 0.19768930399822796, "learning_rate": 4.7204696297985975e-05, "loss": 0.9449, "step": 5497 }, { "epoch": 2.491164476665156, "grad_norm": 0.22433099426318037, "learning_rate": 4.7192249206136065e-05, "loss": 0.9683, "step": 5498 }, { "epoch": 2.4916175804259177, "grad_norm": 0.206844228586513, "learning_rate": 4.717980139454143e-05, "loss": 0.9282, "step": 5499 }, { "epoch": 2.492070684186679, "grad_norm": 0.20695753417583324, "learning_rate": 4.71673528644477e-05, "loss": 0.9583, "step": 5500 }, { "epoch": 2.49252378794744, "grad_norm": 0.20835183274503705, "learning_rate": 4.7154903617100676e-05, "loss": 0.9512, "step": 5501 }, { "epoch": 2.492976891708201, "grad_norm": 0.22325622251658817, "learning_rate": 4.7142453653746164e-05, "loss": 0.924, "step": 5502 }, { "epoch": 2.4934299954689623, "grad_norm": 0.2426662985503918, "learning_rate": 4.713000297563005e-05, "loss": 0.9693, "step": 5503 }, { "epoch": 2.4938830992297234, "grad_norm": 0.2700025911005813, "learning_rate": 4.711755158399831e-05, "loss": 0.9499, "step": 5504 }, { "epoch": 2.494336202990485, "grad_norm": 0.22818589337981315, "learning_rate": 4.710509948009698e-05, "loss": 0.9454, "step": 5505 }, { "epoch": 2.494789306751246, "grad_norm": 0.19071730909064497, "learning_rate": 4.709264666517217e-05, "loss": 0.9403, "step": 5506 }, { "epoch": 2.4952424105120072, "grad_norm": 0.24946057208372382, "learning_rate": 4.7080193140470065e-05, "loss": 0.9695, "step": 5507 }, { "epoch": 2.4956955142727684, "grad_norm": 0.29384147975461034, "learning_rate": 4.7067738907236926e-05, "loss": 0.9494, "step": 5508 }, { "epoch": 2.4961486180335295, "grad_norm": 0.2561314949314844, "learning_rate": 4.705528396671906e-05, "loss": 0.9353, "step": 5509 }, { "epoch": 2.496601721794291, "grad_norm": 0.21566270864523576, "learning_rate": 4.7042828320162875e-05, "loss": 0.9434, "step": 5510 }, { "epoch": 2.497054825555052, "grad_norm": 0.2163611976275105, "learning_rate": 4.703037196881482e-05, "loss": 0.9515, "step": 5511 }, { "epoch": 2.4975079293158133, "grad_norm": 0.1903646421221169, "learning_rate": 4.7017914913921446e-05, "loss": 0.9563, "step": 5512 }, { "epoch": 2.4979610330765745, "grad_norm": 0.2100931324536131, "learning_rate": 4.7005457156729344e-05, "loss": 0.9087, "step": 5513 }, { "epoch": 2.4984141368373356, "grad_norm": 0.20594721117584466, "learning_rate": 4.699299869848521e-05, "loss": 0.9519, "step": 5514 }, { "epoch": 2.4988672405980967, "grad_norm": 0.1976370441854954, "learning_rate": 4.698053954043579e-05, "loss": 0.9644, "step": 5515 }, { "epoch": 2.4993203443588583, "grad_norm": 0.20589269969973578, "learning_rate": 4.696807968382786e-05, "loss": 0.9419, "step": 5516 }, { "epoch": 2.4997734481196194, "grad_norm": 0.19656372240094022, "learning_rate": 4.695561912990835e-05, "loss": 0.9439, "step": 5517 }, { "epoch": 2.5002265518803806, "grad_norm": 0.19023547589536427, "learning_rate": 4.69431578799242e-05, "loss": 0.9411, "step": 5518 }, { "epoch": 2.5006796556411417, "grad_norm": 0.2274329546623226, "learning_rate": 4.6930695935122437e-05, "loss": 0.9476, "step": 5519 }, { "epoch": 2.5011327594019033, "grad_norm": 0.19054186125477182, "learning_rate": 4.691823329675017e-05, "loss": 0.9387, "step": 5520 }, { "epoch": 2.5015858631626644, "grad_norm": 0.1883139500002388, "learning_rate": 4.690576996605454e-05, "loss": 0.9336, "step": 5521 }, { "epoch": 2.5020389669234255, "grad_norm": 0.23907946111906125, "learning_rate": 4.6893305944282786e-05, "loss": 0.9685, "step": 5522 }, { "epoch": 2.5024920706841867, "grad_norm": 0.2747103136896211, "learning_rate": 4.688084123268222e-05, "loss": 0.9197, "step": 5523 }, { "epoch": 2.502945174444948, "grad_norm": 0.2619182230643188, "learning_rate": 4.68683758325002e-05, "loss": 0.9328, "step": 5524 }, { "epoch": 2.503398278205709, "grad_norm": 0.23345773758458804, "learning_rate": 4.6855909744984184e-05, "loss": 0.943, "step": 5525 }, { "epoch": 2.50385138196647, "grad_norm": 0.1834493559574495, "learning_rate": 4.684344297138168e-05, "loss": 0.9379, "step": 5526 }, { "epoch": 2.5043044857272316, "grad_norm": 0.16957416183442678, "learning_rate": 4.6830975512940253e-05, "loss": 0.9551, "step": 5527 }, { "epoch": 2.5047575894879928, "grad_norm": 0.22392699287381382, "learning_rate": 4.681850737090757e-05, "loss": 0.9473, "step": 5528 }, { "epoch": 2.505210693248754, "grad_norm": 1.974521305627745, "learning_rate": 4.680603854653132e-05, "loss": 0.9686, "step": 5529 }, { "epoch": 2.505663797009515, "grad_norm": 0.25252864228680205, "learning_rate": 4.679356904105931e-05, "loss": 0.9397, "step": 5530 }, { "epoch": 2.5061169007702766, "grad_norm": 0.4229455503374792, "learning_rate": 4.678109885573938e-05, "loss": 0.9624, "step": 5531 }, { "epoch": 2.5065700045310377, "grad_norm": 0.4365704358017845, "learning_rate": 4.676862799181946e-05, "loss": 0.9177, "step": 5532 }, { "epoch": 2.507023108291799, "grad_norm": 0.3778427480984631, "learning_rate": 4.6756156450547515e-05, "loss": 0.9204, "step": 5533 }, { "epoch": 2.50747621205256, "grad_norm": 0.383273837826787, "learning_rate": 4.674368423317162e-05, "loss": 0.9254, "step": 5534 }, { "epoch": 2.507929315813321, "grad_norm": 0.3667302526313347, "learning_rate": 4.6731211340939897e-05, "loss": 0.954, "step": 5535 }, { "epoch": 2.5083824195740823, "grad_norm": 0.4090694704693409, "learning_rate": 4.6718737775100536e-05, "loss": 0.9488, "step": 5536 }, { "epoch": 2.508835523334844, "grad_norm": 0.35838734632939406, "learning_rate": 4.6706263536901804e-05, "loss": 0.9583, "step": 5537 }, { "epoch": 2.509288627095605, "grad_norm": 0.32790062207586845, "learning_rate": 4.6693788627592004e-05, "loss": 0.9578, "step": 5538 }, { "epoch": 2.509741730856366, "grad_norm": 0.34855329993114903, "learning_rate": 4.668131304841954e-05, "loss": 0.9503, "step": 5539 }, { "epoch": 2.5101948346171272, "grad_norm": 0.2982435704195751, "learning_rate": 4.666883680063286e-05, "loss": 0.939, "step": 5540 }, { "epoch": 2.510647938377889, "grad_norm": 0.22914733234409773, "learning_rate": 4.6656359885480515e-05, "loss": 0.9635, "step": 5541 }, { "epoch": 2.51110104213865, "grad_norm": 0.24843575610565088, "learning_rate": 4.664388230421109e-05, "loss": 0.9434, "step": 5542 }, { "epoch": 2.511554145899411, "grad_norm": 0.26324557694653594, "learning_rate": 4.6631404058073236e-05, "loss": 0.9519, "step": 5543 }, { "epoch": 2.512007249660172, "grad_norm": 0.2502633722824147, "learning_rate": 4.6618925148315676e-05, "loss": 0.9456, "step": 5544 }, { "epoch": 2.5124603534209333, "grad_norm": 0.2681913620707465, "learning_rate": 4.660644557618722e-05, "loss": 0.9512, "step": 5545 }, { "epoch": 2.5129134571816945, "grad_norm": 0.2485247784088327, "learning_rate": 4.659396534293671e-05, "loss": 0.9416, "step": 5546 }, { "epoch": 2.5133665609424556, "grad_norm": 0.2288396989674444, "learning_rate": 4.65814844498131e-05, "loss": 0.9117, "step": 5547 }, { "epoch": 2.513819664703217, "grad_norm": 0.23347491031543344, "learning_rate": 4.656900289806534e-05, "loss": 0.9289, "step": 5548 }, { "epoch": 2.5142727684639783, "grad_norm": 0.22922314665829377, "learning_rate": 4.65565206889425e-05, "loss": 0.9405, "step": 5549 }, { "epoch": 2.5147258722247394, "grad_norm": 0.21992804002524535, "learning_rate": 4.6544037823693735e-05, "loss": 0.9334, "step": 5550 }, { "epoch": 2.5151789759855006, "grad_norm": 0.21857244257810848, "learning_rate": 4.6531554303568186e-05, "loss": 0.9567, "step": 5551 }, { "epoch": 2.515632079746262, "grad_norm": 0.20771179761228845, "learning_rate": 4.651907012981514e-05, "loss": 0.947, "step": 5552 }, { "epoch": 2.5160851835070233, "grad_norm": 0.21357317502148818, "learning_rate": 4.6506585303683916e-05, "loss": 0.9576, "step": 5553 }, { "epoch": 2.5165382872677844, "grad_norm": 0.1993227963655666, "learning_rate": 4.649409982642389e-05, "loss": 0.9386, "step": 5554 }, { "epoch": 2.5169913910285455, "grad_norm": 0.21913682434151702, "learning_rate": 4.648161369928451e-05, "loss": 0.9459, "step": 5555 }, { "epoch": 2.5174444947893067, "grad_norm": 0.2745346218433436, "learning_rate": 4.646912692351529e-05, "loss": 0.933, "step": 5556 }, { "epoch": 2.517897598550068, "grad_norm": 0.23808866965120593, "learning_rate": 4.645663950036581e-05, "loss": 0.9507, "step": 5557 }, { "epoch": 2.518350702310829, "grad_norm": 0.23022766014894924, "learning_rate": 4.644415143108573e-05, "loss": 0.9531, "step": 5558 }, { "epoch": 2.5188038060715905, "grad_norm": 0.2575484016215423, "learning_rate": 4.6431662716924743e-05, "loss": 0.9429, "step": 5559 }, { "epoch": 2.5192569098323516, "grad_norm": 0.20751701066015654, "learning_rate": 4.641917335913262e-05, "loss": 0.9455, "step": 5560 }, { "epoch": 2.5197100135931128, "grad_norm": 0.17875824542309676, "learning_rate": 4.640668335895922e-05, "loss": 0.9531, "step": 5561 }, { "epoch": 2.520163117353874, "grad_norm": 0.2254929299599564, "learning_rate": 4.6394192717654426e-05, "loss": 0.9605, "step": 5562 }, { "epoch": 2.5206162211146355, "grad_norm": 0.22077336174727838, "learning_rate": 4.6381701436468224e-05, "loss": 0.9807, "step": 5563 }, { "epoch": 2.5210693248753966, "grad_norm": 0.18785735638258247, "learning_rate": 4.636920951665063e-05, "loss": 0.9482, "step": 5564 }, { "epoch": 2.5215224286361577, "grad_norm": 0.18009188664081324, "learning_rate": 4.6356716959451736e-05, "loss": 0.9276, "step": 5565 }, { "epoch": 2.521975532396919, "grad_norm": 0.19224297435867577, "learning_rate": 4.634422376612171e-05, "loss": 0.9605, "step": 5566 }, { "epoch": 2.52242863615768, "grad_norm": 0.22339399312313216, "learning_rate": 4.633172993791077e-05, "loss": 0.9748, "step": 5567 }, { "epoch": 2.522881739918441, "grad_norm": 0.22296539871713325, "learning_rate": 4.63192354760692e-05, "loss": 0.9359, "step": 5568 }, { "epoch": 2.5233348436792027, "grad_norm": 0.19207258931689164, "learning_rate": 4.6306740381847356e-05, "loss": 0.9349, "step": 5569 }, { "epoch": 2.523787947439964, "grad_norm": 0.22374305736723202, "learning_rate": 4.629424465649564e-05, "loss": 0.9685, "step": 5570 }, { "epoch": 2.524241051200725, "grad_norm": 0.2195109111296499, "learning_rate": 4.628174830126455e-05, "loss": 0.9346, "step": 5571 }, { "epoch": 2.524694154961486, "grad_norm": 0.20973809885294226, "learning_rate": 4.6269251317404604e-05, "loss": 0.9297, "step": 5572 }, { "epoch": 2.5251472587222477, "grad_norm": 0.2108726857628751, "learning_rate": 4.6256753706166413e-05, "loss": 0.9426, "step": 5573 }, { "epoch": 2.525600362483009, "grad_norm": 0.186909721363838, "learning_rate": 4.624425546880063e-05, "loss": 0.9374, "step": 5574 }, { "epoch": 2.52605346624377, "grad_norm": 0.21643853194285806, "learning_rate": 4.623175660655799e-05, "loss": 0.9392, "step": 5575 }, { "epoch": 2.526506570004531, "grad_norm": 0.2048245083874745, "learning_rate": 4.6219257120689275e-05, "loss": 0.9469, "step": 5576 }, { "epoch": 2.526959673765292, "grad_norm": 0.1909976310667556, "learning_rate": 4.620675701244535e-05, "loss": 0.9347, "step": 5577 }, { "epoch": 2.5274127775260533, "grad_norm": 0.15715933297810344, "learning_rate": 4.6194256283077116e-05, "loss": 0.9651, "step": 5578 }, { "epoch": 2.5278658812868144, "grad_norm": 0.1972830630099313, "learning_rate": 4.618175493383556e-05, "loss": 0.9557, "step": 5579 }, { "epoch": 2.528318985047576, "grad_norm": 0.20934715854351352, "learning_rate": 4.616925296597172e-05, "loss": 0.9595, "step": 5580 }, { "epoch": 2.528772088808337, "grad_norm": 0.1670230974253069, "learning_rate": 4.6156750380736703e-05, "loss": 0.9448, "step": 5581 }, { "epoch": 2.5292251925690983, "grad_norm": 0.17258054313096322, "learning_rate": 4.6144247179381656e-05, "loss": 0.9227, "step": 5582 }, { "epoch": 2.5296782963298594, "grad_norm": 0.20236856097050915, "learning_rate": 4.613174336315781e-05, "loss": 0.9377, "step": 5583 }, { "epoch": 2.530131400090621, "grad_norm": 0.1811568349648225, "learning_rate": 4.6119238933316434e-05, "loss": 0.9398, "step": 5584 }, { "epoch": 2.530584503851382, "grad_norm": 0.19741979567050272, "learning_rate": 4.61067338911089e-05, "loss": 0.9628, "step": 5585 }, { "epoch": 2.5310376076121432, "grad_norm": 0.2044023918759804, "learning_rate": 4.609422823778661e-05, "loss": 0.9291, "step": 5586 }, { "epoch": 2.5314907113729044, "grad_norm": 0.19692144846351822, "learning_rate": 4.6081721974601016e-05, "loss": 0.9265, "step": 5587 }, { "epoch": 2.5319438151336655, "grad_norm": 0.23021197354726272, "learning_rate": 4.6069215102803674e-05, "loss": 0.9515, "step": 5588 }, { "epoch": 2.5323969188944266, "grad_norm": 0.25457286487782704, "learning_rate": 4.605670762364616e-05, "loss": 0.952, "step": 5589 }, { "epoch": 2.5328500226551878, "grad_norm": 0.26829373697615366, "learning_rate": 4.604419953838014e-05, "loss": 0.9478, "step": 5590 }, { "epoch": 2.5333031264159493, "grad_norm": 0.2981602476163503, "learning_rate": 4.6031690848257284e-05, "loss": 0.944, "step": 5591 }, { "epoch": 2.5337562301767105, "grad_norm": 0.2826449399781984, "learning_rate": 4.6019181554529414e-05, "loss": 0.9515, "step": 5592 }, { "epoch": 2.5342093339374716, "grad_norm": 0.2472134146565376, "learning_rate": 4.600667165844835e-05, "loss": 0.934, "step": 5593 }, { "epoch": 2.5346624376982327, "grad_norm": 0.19178953006253602, "learning_rate": 4.599416116126597e-05, "loss": 0.9375, "step": 5594 }, { "epoch": 2.5351155414589943, "grad_norm": 0.19358084696939046, "learning_rate": 4.598165006423424e-05, "loss": 0.9432, "step": 5595 }, { "epoch": 2.5355686452197554, "grad_norm": 0.22691379865102607, "learning_rate": 4.596913836860517e-05, "loss": 0.9301, "step": 5596 }, { "epoch": 2.5360217489805166, "grad_norm": 0.27096582747690634, "learning_rate": 4.5956626075630844e-05, "loss": 0.9663, "step": 5597 }, { "epoch": 2.5364748527412777, "grad_norm": 0.320687512043896, "learning_rate": 4.59441131865634e-05, "loss": 0.9544, "step": 5598 }, { "epoch": 2.536927956502039, "grad_norm": 0.30636199793159086, "learning_rate": 4.593159970265501e-05, "loss": 0.9472, "step": 5599 }, { "epoch": 2.5373810602628, "grad_norm": 0.23198581666988594, "learning_rate": 4.5919085625157936e-05, "loss": 0.961, "step": 5600 }, { "epoch": 2.5378341640235615, "grad_norm": 0.20917973923163405, "learning_rate": 4.5906570955324486e-05, "loss": 0.949, "step": 5601 }, { "epoch": 2.5382872677843227, "grad_norm": 0.24658541687024665, "learning_rate": 4.589405569440704e-05, "loss": 0.9408, "step": 5602 }, { "epoch": 2.538740371545084, "grad_norm": 0.2547289151814682, "learning_rate": 4.588153984365802e-05, "loss": 0.9439, "step": 5603 }, { "epoch": 2.539193475305845, "grad_norm": 0.23100282344250553, "learning_rate": 4.586902340432992e-05, "loss": 0.9462, "step": 5604 }, { "epoch": 2.5396465790666065, "grad_norm": 0.18698061630645174, "learning_rate": 4.5856506377675295e-05, "loss": 0.9453, "step": 5605 }, { "epoch": 2.5400996828273676, "grad_norm": 0.1898939195220462, "learning_rate": 4.584398876494674e-05, "loss": 0.9543, "step": 5606 }, { "epoch": 2.5405527865881288, "grad_norm": 0.18658891177205542, "learning_rate": 4.583147056739695e-05, "loss": 0.9476, "step": 5607 }, { "epoch": 2.54100589034889, "grad_norm": 0.21253037337676337, "learning_rate": 4.5818951786278597e-05, "loss": 0.9332, "step": 5608 }, { "epoch": 2.541458994109651, "grad_norm": 0.24122070568187642, "learning_rate": 4.580643242284451e-05, "loss": 0.931, "step": 5609 }, { "epoch": 2.541912097870412, "grad_norm": 0.18597767413154576, "learning_rate": 4.5793912478347506e-05, "loss": 0.9574, "step": 5610 }, { "epoch": 2.5423652016311733, "grad_norm": 0.19055387674323052, "learning_rate": 4.5781391954040494e-05, "loss": 0.9344, "step": 5611 }, { "epoch": 2.542818305391935, "grad_norm": 0.21097591351568376, "learning_rate": 4.5768870851176435e-05, "loss": 0.9499, "step": 5612 }, { "epoch": 2.543271409152696, "grad_norm": 0.2238537413936067, "learning_rate": 4.575634917100833e-05, "loss": 0.937, "step": 5613 }, { "epoch": 2.543724512913457, "grad_norm": 0.20506905564946815, "learning_rate": 4.5743826914789266e-05, "loss": 0.9344, "step": 5614 }, { "epoch": 2.5441776166742183, "grad_norm": 0.2020592616442736, "learning_rate": 4.573130408377237e-05, "loss": 0.9549, "step": 5615 }, { "epoch": 2.54463072043498, "grad_norm": 0.21174860977620544, "learning_rate": 4.571878067921083e-05, "loss": 0.9579, "step": 5616 }, { "epoch": 2.545083824195741, "grad_norm": 0.18007930809863162, "learning_rate": 4.5706256702357896e-05, "loss": 0.9399, "step": 5617 }, { "epoch": 2.545536927956502, "grad_norm": 0.18855769416996412, "learning_rate": 4.569373215446685e-05, "loss": 0.9303, "step": 5618 }, { "epoch": 2.5459900317172632, "grad_norm": 0.23107725974675475, "learning_rate": 4.568120703679108e-05, "loss": 0.9532, "step": 5619 }, { "epoch": 2.5464431354780244, "grad_norm": 0.2613230262148369, "learning_rate": 4.566868135058399e-05, "loss": 0.9675, "step": 5620 }, { "epoch": 2.5468962392387855, "grad_norm": 0.31846486878476143, "learning_rate": 4.565615509709905e-05, "loss": 0.9276, "step": 5621 }, { "epoch": 2.5473493429995466, "grad_norm": 0.3807636350005718, "learning_rate": 4.564362827758979e-05, "loss": 0.9375, "step": 5622 }, { "epoch": 2.547802446760308, "grad_norm": 0.3896916079730604, "learning_rate": 4.563110089330981e-05, "loss": 0.9322, "step": 5623 }, { "epoch": 2.5482555505210693, "grad_norm": 0.33308105476409766, "learning_rate": 4.561857294551275e-05, "loss": 0.9392, "step": 5624 }, { "epoch": 2.5487086542818305, "grad_norm": 0.24693464067261803, "learning_rate": 4.560604443545231e-05, "loss": 0.954, "step": 5625 }, { "epoch": 2.5491617580425916, "grad_norm": 0.1841421968997698, "learning_rate": 4.559351536438223e-05, "loss": 0.9555, "step": 5626 }, { "epoch": 2.549614861803353, "grad_norm": 0.19781009625452384, "learning_rate": 4.558098573355634e-05, "loss": 0.9546, "step": 5627 }, { "epoch": 2.5500679655641143, "grad_norm": 0.24619852202959283, "learning_rate": 4.5568455544228514e-05, "loss": 0.9442, "step": 5628 }, { "epoch": 2.5505210693248754, "grad_norm": 0.2610443747236109, "learning_rate": 4.555592479765265e-05, "loss": 0.9507, "step": 5629 }, { "epoch": 2.5509741730856366, "grad_norm": 0.2441802327954335, "learning_rate": 4.554339349508276e-05, "loss": 0.9249, "step": 5630 }, { "epoch": 2.5514272768463977, "grad_norm": 0.20581718059079784, "learning_rate": 4.553086163777285e-05, "loss": 0.9559, "step": 5631 }, { "epoch": 2.551880380607159, "grad_norm": 0.20601963646009355, "learning_rate": 4.5518329226977034e-05, "loss": 0.959, "step": 5632 }, { "epoch": 2.5523334843679204, "grad_norm": 0.1974239732713669, "learning_rate": 4.550579626394946e-05, "loss": 0.9377, "step": 5633 }, { "epoch": 2.5527865881286815, "grad_norm": 0.19739512329423842, "learning_rate": 4.549326274994432e-05, "loss": 0.9397, "step": 5634 }, { "epoch": 2.5532396918894427, "grad_norm": 0.19993632171786774, "learning_rate": 4.548072868621587e-05, "loss": 0.9608, "step": 5635 }, { "epoch": 2.553692795650204, "grad_norm": 0.20402667313015999, "learning_rate": 4.546819407401842e-05, "loss": 0.9578, "step": 5636 }, { "epoch": 2.5541458994109654, "grad_norm": 0.17845332441407025, "learning_rate": 4.545565891460635e-05, "loss": 0.9614, "step": 5637 }, { "epoch": 2.5545990031717265, "grad_norm": 0.22891109138283466, "learning_rate": 4.544312320923406e-05, "loss": 0.9199, "step": 5638 }, { "epoch": 2.5550521069324876, "grad_norm": 0.2010454744587308, "learning_rate": 4.543058695915605e-05, "loss": 0.95, "step": 5639 }, { "epoch": 2.5555052106932488, "grad_norm": 0.24033653902669053, "learning_rate": 4.541805016562684e-05, "loss": 0.9408, "step": 5640 }, { "epoch": 2.55595831445401, "grad_norm": 0.28400879252064803, "learning_rate": 4.5405512829901025e-05, "loss": 0.9476, "step": 5641 }, { "epoch": 2.556411418214771, "grad_norm": 0.21350092303275992, "learning_rate": 4.5392974953233246e-05, "loss": 0.9606, "step": 5642 }, { "epoch": 2.556864521975532, "grad_norm": 0.2005557554718147, "learning_rate": 4.5380436536878175e-05, "loss": 0.957, "step": 5643 }, { "epoch": 2.5573176257362937, "grad_norm": 0.19218228089427158, "learning_rate": 4.5367897582090564e-05, "loss": 0.9382, "step": 5644 }, { "epoch": 2.557770729497055, "grad_norm": 0.1931955295867188, "learning_rate": 4.5355358090125235e-05, "loss": 0.9402, "step": 5645 }, { "epoch": 2.558223833257816, "grad_norm": 0.20012516766372368, "learning_rate": 4.534281806223703e-05, "loss": 0.9435, "step": 5646 }, { "epoch": 2.558676937018577, "grad_norm": 0.17813973764292576, "learning_rate": 4.533027749968086e-05, "loss": 0.9473, "step": 5647 }, { "epoch": 2.5591300407793387, "grad_norm": 0.19688873442287916, "learning_rate": 4.531773640371169e-05, "loss": 0.9363, "step": 5648 }, { "epoch": 2.5595831445401, "grad_norm": 0.18848638284979918, "learning_rate": 4.530519477558454e-05, "loss": 0.9587, "step": 5649 }, { "epoch": 2.560036248300861, "grad_norm": 0.17224106250867438, "learning_rate": 4.5292652616554465e-05, "loss": 0.9604, "step": 5650 }, { "epoch": 2.560489352061622, "grad_norm": 0.20479933366425093, "learning_rate": 4.5280109927876596e-05, "loss": 0.9305, "step": 5651 }, { "epoch": 2.5609424558223832, "grad_norm": 0.1931384685411603, "learning_rate": 4.526756671080611e-05, "loss": 0.9192, "step": 5652 }, { "epoch": 2.5613955595831444, "grad_norm": 0.1704177795210219, "learning_rate": 4.5255022966598234e-05, "loss": 0.9248, "step": 5653 }, { "epoch": 2.5618486633439055, "grad_norm": 0.1930874335635507, "learning_rate": 4.5242478696508246e-05, "loss": 0.9386, "step": 5654 }, { "epoch": 2.562301767104667, "grad_norm": 0.21100601867998753, "learning_rate": 4.522993390179149e-05, "loss": 0.9449, "step": 5655 }, { "epoch": 2.562754870865428, "grad_norm": 0.1994261974664805, "learning_rate": 4.521738858370333e-05, "loss": 0.9176, "step": 5656 }, { "epoch": 2.5632079746261893, "grad_norm": 0.21345637135196904, "learning_rate": 4.5204842743499235e-05, "loss": 0.952, "step": 5657 }, { "epoch": 2.5636610783869505, "grad_norm": 0.1734887555029129, "learning_rate": 4.5192296382434664e-05, "loss": 0.9337, "step": 5658 }, { "epoch": 2.564114182147712, "grad_norm": 0.2003391728286632, "learning_rate": 4.517974950176519e-05, "loss": 0.939, "step": 5659 }, { "epoch": 2.564567285908473, "grad_norm": 0.17884804610373464, "learning_rate": 4.516720210274639e-05, "loss": 0.9336, "step": 5660 }, { "epoch": 2.5650203896692343, "grad_norm": 2.1744149438777525, "learning_rate": 4.5154654186633916e-05, "loss": 0.9741, "step": 5661 }, { "epoch": 2.5654734934299954, "grad_norm": 0.21172411978027905, "learning_rate": 4.5142105754683455e-05, "loss": 0.942, "step": 5662 }, { "epoch": 2.5659265971907566, "grad_norm": 0.4050973461860857, "learning_rate": 4.512955680815076e-05, "loss": 0.9525, "step": 5663 }, { "epoch": 2.5663797009515177, "grad_norm": 0.4650633663820827, "learning_rate": 4.511700734829166e-05, "loss": 0.9484, "step": 5664 }, { "epoch": 2.5668328047122793, "grad_norm": 0.4155604888088717, "learning_rate": 4.510445737636198e-05, "loss": 0.9428, "step": 5665 }, { "epoch": 2.5672859084730404, "grad_norm": 0.33612266101828286, "learning_rate": 4.509190689361762e-05, "loss": 0.9497, "step": 5666 }, { "epoch": 2.5677390122338015, "grad_norm": 0.29997151600164895, "learning_rate": 4.507935590131456e-05, "loss": 0.9502, "step": 5667 }, { "epoch": 2.5681921159945627, "grad_norm": 0.26044658544378835, "learning_rate": 4.50668044007088e-05, "loss": 0.941, "step": 5668 }, { "epoch": 2.5686452197553242, "grad_norm": 0.33253214470761255, "learning_rate": 4.505425239305638e-05, "loss": 0.9591, "step": 5669 }, { "epoch": 2.5690983235160854, "grad_norm": 0.24845380766829805, "learning_rate": 4.5041699879613416e-05, "loss": 0.9413, "step": 5670 }, { "epoch": 2.5695514272768465, "grad_norm": 0.2016919303566665, "learning_rate": 4.502914686163608e-05, "loss": 0.9299, "step": 5671 }, { "epoch": 2.5700045310376076, "grad_norm": 0.21658224531014064, "learning_rate": 4.5016593340380555e-05, "loss": 0.9421, "step": 5672 }, { "epoch": 2.5704576347983688, "grad_norm": 0.22213810728332445, "learning_rate": 4.5004039317103126e-05, "loss": 0.957, "step": 5673 }, { "epoch": 2.57091073855913, "grad_norm": 0.1926441551911884, "learning_rate": 4.49914847930601e-05, "loss": 0.9645, "step": 5674 }, { "epoch": 2.571363842319891, "grad_norm": 0.20294792238916615, "learning_rate": 4.497892976950781e-05, "loss": 0.9242, "step": 5675 }, { "epoch": 2.5718169460806526, "grad_norm": 0.1983352020208915, "learning_rate": 4.4966374247702696e-05, "loss": 0.9476, "step": 5676 }, { "epoch": 2.5722700498414137, "grad_norm": 0.17551600699261474, "learning_rate": 4.495381822890122e-05, "loss": 0.9429, "step": 5677 }, { "epoch": 2.572723153602175, "grad_norm": 0.18554628534788617, "learning_rate": 4.4941261714359855e-05, "loss": 0.9501, "step": 5678 }, { "epoch": 2.573176257362936, "grad_norm": 0.20804573974043525, "learning_rate": 4.49287047053352e-05, "loss": 0.9359, "step": 5679 }, { "epoch": 2.5736293611236976, "grad_norm": 0.2875077300191955, "learning_rate": 4.491614720308383e-05, "loss": 0.9406, "step": 5680 }, { "epoch": 2.5740824648844587, "grad_norm": 0.36426628389835375, "learning_rate": 4.4903589208862426e-05, "loss": 0.9306, "step": 5681 }, { "epoch": 2.57453556864522, "grad_norm": 0.3806676916267181, "learning_rate": 4.4891030723927696e-05, "loss": 0.93, "step": 5682 }, { "epoch": 2.574988672405981, "grad_norm": 0.3521927964237562, "learning_rate": 4.487847174953638e-05, "loss": 0.9437, "step": 5683 }, { "epoch": 2.575441776166742, "grad_norm": 0.30917077243499336, "learning_rate": 4.48659122869453e-05, "loss": 0.9544, "step": 5684 }, { "epoch": 2.575894879927503, "grad_norm": 0.7636675798791257, "learning_rate": 4.485335233741131e-05, "loss": 0.9638, "step": 5685 }, { "epoch": 2.5763479836882643, "grad_norm": 0.18610423998929668, "learning_rate": 4.484079190219128e-05, "loss": 0.9461, "step": 5686 }, { "epoch": 2.576801087449026, "grad_norm": 0.2719207726426694, "learning_rate": 4.482823098254221e-05, "loss": 0.9301, "step": 5687 }, { "epoch": 2.577254191209787, "grad_norm": 0.26831524766033, "learning_rate": 4.4815669579721066e-05, "loss": 0.9564, "step": 5688 }, { "epoch": 2.577707294970548, "grad_norm": 0.32650925536381625, "learning_rate": 4.4803107694984906e-05, "loss": 0.9488, "step": 5689 }, { "epoch": 2.5781603987313093, "grad_norm": 0.3167866456483274, "learning_rate": 4.479054532959082e-05, "loss": 0.9391, "step": 5690 }, { "epoch": 2.578613502492071, "grad_norm": 0.2870827364527292, "learning_rate": 4.477798248479596e-05, "loss": 0.9572, "step": 5691 }, { "epoch": 2.579066606252832, "grad_norm": 0.18910649223165815, "learning_rate": 4.476541916185753e-05, "loss": 0.9467, "step": 5692 }, { "epoch": 2.579519710013593, "grad_norm": 0.1717162513313077, "learning_rate": 4.475285536203274e-05, "loss": 0.9259, "step": 5693 }, { "epoch": 2.5799728137743543, "grad_norm": 0.2510575024059809, "learning_rate": 4.47402910865789e-05, "loss": 0.9012, "step": 5694 }, { "epoch": 2.5804259175351154, "grad_norm": 0.30811144710365296, "learning_rate": 4.472772633675336e-05, "loss": 0.9353, "step": 5695 }, { "epoch": 2.5808790212958765, "grad_norm": 0.35373909324011, "learning_rate": 4.471516111381347e-05, "loss": 0.9573, "step": 5696 }, { "epoch": 2.581332125056638, "grad_norm": 0.3109557667505129, "learning_rate": 4.470259541901666e-05, "loss": 0.9459, "step": 5697 }, { "epoch": 2.5817852288173992, "grad_norm": 0.26384314016036914, "learning_rate": 4.469002925362045e-05, "loss": 0.9252, "step": 5698 }, { "epoch": 2.5822383325781604, "grad_norm": 0.2067733547744488, "learning_rate": 4.467746261888232e-05, "loss": 0.9394, "step": 5699 }, { "epoch": 2.5826914363389215, "grad_norm": 0.17094801797705678, "learning_rate": 4.4664895516059864e-05, "loss": 0.9334, "step": 5700 }, { "epoch": 2.583144540099683, "grad_norm": 0.19987304517711324, "learning_rate": 4.4652327946410695e-05, "loss": 0.9733, "step": 5701 }, { "epoch": 2.583597643860444, "grad_norm": 0.2417311004646793, "learning_rate": 4.463975991119248e-05, "loss": 0.9542, "step": 5702 }, { "epoch": 2.5840507476212053, "grad_norm": 0.2766389969078786, "learning_rate": 4.462719141166293e-05, "loss": 0.948, "step": 5703 }, { "epoch": 2.5845038513819665, "grad_norm": 0.24288832507939367, "learning_rate": 4.4614622449079804e-05, "loss": 0.959, "step": 5704 }, { "epoch": 2.5849569551427276, "grad_norm": 0.3334283586334024, "learning_rate": 4.4602053024700906e-05, "loss": 0.9464, "step": 5705 }, { "epoch": 2.5854100589034887, "grad_norm": 0.23835332924495553, "learning_rate": 4.458948313978408e-05, "loss": 0.9453, "step": 5706 }, { "epoch": 2.58586316266425, "grad_norm": 0.20268819290787404, "learning_rate": 4.457691279558724e-05, "loss": 0.9261, "step": 5707 }, { "epoch": 2.5863162664250114, "grad_norm": 0.15616626909733747, "learning_rate": 4.4564341993368306e-05, "loss": 0.9453, "step": 5708 }, { "epoch": 2.5867693701857726, "grad_norm": 0.1897762944861974, "learning_rate": 4.4551770734385284e-05, "loss": 0.9533, "step": 5709 }, { "epoch": 2.5872224739465337, "grad_norm": 0.23535136240463816, "learning_rate": 4.4539199019896214e-05, "loss": 0.9218, "step": 5710 }, { "epoch": 2.587675577707295, "grad_norm": 0.25514151286124465, "learning_rate": 4.452662685115916e-05, "loss": 0.9419, "step": 5711 }, { "epoch": 2.5881286814680564, "grad_norm": 0.2535806181084725, "learning_rate": 4.451405422943226e-05, "loss": 0.9418, "step": 5712 }, { "epoch": 2.5885817852288175, "grad_norm": 0.2138124324349007, "learning_rate": 4.450148115597367e-05, "loss": 0.9648, "step": 5713 }, { "epoch": 2.5890348889895787, "grad_norm": 0.21491612912351082, "learning_rate": 4.448890763204162e-05, "loss": 0.9549, "step": 5714 }, { "epoch": 2.58948799275034, "grad_norm": 0.2367715093579166, "learning_rate": 4.447633365889436e-05, "loss": 0.943, "step": 5715 }, { "epoch": 2.589941096511101, "grad_norm": 0.21524201883095379, "learning_rate": 4.44637592377902e-05, "loss": 0.9442, "step": 5716 }, { "epoch": 2.590394200271862, "grad_norm": 0.22207952496547434, "learning_rate": 4.44511843699875e-05, "loss": 0.9422, "step": 5717 }, { "epoch": 2.590847304032623, "grad_norm": 0.21574293359518265, "learning_rate": 4.4438609056744644e-05, "loss": 0.9702, "step": 5718 }, { "epoch": 2.5913004077933848, "grad_norm": 0.2560202238153639, "learning_rate": 4.442603329932007e-05, "loss": 0.9369, "step": 5719 }, { "epoch": 2.591753511554146, "grad_norm": 0.27071046040059604, "learning_rate": 4.441345709897228e-05, "loss": 0.9584, "step": 5720 }, { "epoch": 2.592206615314907, "grad_norm": 0.2486128114500181, "learning_rate": 4.44008804569598e-05, "loss": 0.9752, "step": 5721 }, { "epoch": 2.592659719075668, "grad_norm": 0.19901892841864544, "learning_rate": 4.438830337454119e-05, "loss": 0.945, "step": 5722 }, { "epoch": 2.5931128228364297, "grad_norm": 0.2392896818824029, "learning_rate": 4.4375725852975064e-05, "loss": 0.9335, "step": 5723 }, { "epoch": 2.593565926597191, "grad_norm": 0.3092687943466376, "learning_rate": 4.436314789352009e-05, "loss": 0.9427, "step": 5724 }, { "epoch": 2.594019030357952, "grad_norm": 0.3682762168118018, "learning_rate": 4.435056949743499e-05, "loss": 0.9521, "step": 5725 }, { "epoch": 2.594472134118713, "grad_norm": 0.2712748332996306, "learning_rate": 4.433799066597848e-05, "loss": 0.9558, "step": 5726 }, { "epoch": 2.5949252378794743, "grad_norm": 0.19124326250571405, "learning_rate": 4.432541140040938e-05, "loss": 0.9413, "step": 5727 }, { "epoch": 2.5953783416402354, "grad_norm": 0.2104403509475996, "learning_rate": 4.4312831701986516e-05, "loss": 0.9171, "step": 5728 }, { "epoch": 2.595831445400997, "grad_norm": 0.2888004422399631, "learning_rate": 4.4300251571968775e-05, "loss": 0.9382, "step": 5729 }, { "epoch": 2.596284549161758, "grad_norm": 0.35283941154838544, "learning_rate": 4.428767101161506e-05, "loss": 0.9421, "step": 5730 }, { "epoch": 2.5967376529225192, "grad_norm": 0.38093282851410687, "learning_rate": 4.4275090022184354e-05, "loss": 0.9359, "step": 5731 }, { "epoch": 2.5971907566832804, "grad_norm": 0.3480013358510412, "learning_rate": 4.426250860493565e-05, "loss": 0.9461, "step": 5732 }, { "epoch": 2.597643860444042, "grad_norm": 0.28484297737805536, "learning_rate": 4.4249926761128013e-05, "loss": 0.9292, "step": 5733 }, { "epoch": 2.598096964204803, "grad_norm": 0.25258413392004575, "learning_rate": 4.4237344492020525e-05, "loss": 0.9296, "step": 5734 }, { "epoch": 2.598550067965564, "grad_norm": 0.19444883283710956, "learning_rate": 4.422476179887234e-05, "loss": 0.9387, "step": 5735 }, { "epoch": 2.5990031717263253, "grad_norm": 0.21672709052269612, "learning_rate": 4.421217868294261e-05, "loss": 0.9294, "step": 5736 }, { "epoch": 2.5994562754870865, "grad_norm": 0.2697048793158359, "learning_rate": 4.419959514549059e-05, "loss": 0.9615, "step": 5737 }, { "epoch": 2.5999093792478476, "grad_norm": 0.2906448831746704, "learning_rate": 4.418701118777552e-05, "loss": 0.9671, "step": 5738 }, { "epoch": 2.6003624830086087, "grad_norm": 0.3111375711315011, "learning_rate": 4.4174426811056706e-05, "loss": 0.9377, "step": 5739 }, { "epoch": 2.6008155867693703, "grad_norm": 0.2840361785358121, "learning_rate": 4.41618420165935e-05, "loss": 0.9753, "step": 5740 }, { "epoch": 2.6012686905301314, "grad_norm": 0.23544196536562528, "learning_rate": 4.414925680564529e-05, "loss": 0.943, "step": 5741 }, { "epoch": 2.6017217942908926, "grad_norm": 0.18542754566063754, "learning_rate": 4.413667117947151e-05, "loss": 0.9263, "step": 5742 }, { "epoch": 2.6021748980516537, "grad_norm": 0.3195051566311767, "learning_rate": 4.412408513933163e-05, "loss": 0.9489, "step": 5743 }, { "epoch": 2.6026280018124153, "grad_norm": 0.24109893006030206, "learning_rate": 4.4111498686485164e-05, "loss": 0.9346, "step": 5744 }, { "epoch": 2.6030811055731764, "grad_norm": 0.28810218819455785, "learning_rate": 4.4098911822191655e-05, "loss": 0.9523, "step": 5745 }, { "epoch": 2.6035342093339375, "grad_norm": 0.2760652259770382, "learning_rate": 4.408632454771072e-05, "loss": 0.9269, "step": 5746 }, { "epoch": 2.6039873130946987, "grad_norm": 0.25524599404590836, "learning_rate": 4.407373686430199e-05, "loss": 0.9395, "step": 5747 }, { "epoch": 2.60444041685546, "grad_norm": 0.21637200853122182, "learning_rate": 4.4061148773225134e-05, "loss": 0.9478, "step": 5748 }, { "epoch": 2.604893520616221, "grad_norm": 0.17217914159920672, "learning_rate": 4.404856027573987e-05, "loss": 0.9435, "step": 5749 }, { "epoch": 2.605346624376982, "grad_norm": 0.15186104703730052, "learning_rate": 4.403597137310597e-05, "loss": 0.9535, "step": 5750 }, { "epoch": 2.6057997281377436, "grad_norm": 0.1561759406499342, "learning_rate": 4.402338206658322e-05, "loss": 0.9335, "step": 5751 }, { "epoch": 2.6062528318985048, "grad_norm": 0.1927262887249084, "learning_rate": 4.4010792357431485e-05, "loss": 0.9396, "step": 5752 }, { "epoch": 2.606705935659266, "grad_norm": 0.22094750131327773, "learning_rate": 4.399820224691062e-05, "loss": 0.9459, "step": 5753 }, { "epoch": 2.607159039420027, "grad_norm": 0.22277784366702963, "learning_rate": 4.3985611736280555e-05, "loss": 0.926, "step": 5754 }, { "epoch": 2.6076121431807886, "grad_norm": 0.2249627995019141, "learning_rate": 4.397302082680125e-05, "loss": 0.9317, "step": 5755 }, { "epoch": 2.6080652469415497, "grad_norm": 0.19506855600340078, "learning_rate": 4.3960429519732714e-05, "loss": 0.9467, "step": 5756 }, { "epoch": 2.608518350702311, "grad_norm": 0.18015467796619195, "learning_rate": 4.394783781633498e-05, "loss": 0.9502, "step": 5757 }, { "epoch": 2.608971454463072, "grad_norm": 0.18388304623277937, "learning_rate": 4.393524571786813e-05, "loss": 0.9445, "step": 5758 }, { "epoch": 2.609424558223833, "grad_norm": 0.23966257915370662, "learning_rate": 4.392265322559228e-05, "loss": 0.9607, "step": 5759 }, { "epoch": 2.6098776619845943, "grad_norm": 0.27579273317183656, "learning_rate": 4.39100603407676e-05, "loss": 0.9341, "step": 5760 }, { "epoch": 2.610330765745356, "grad_norm": 0.28285824972993506, "learning_rate": 4.389746706465427e-05, "loss": 0.9348, "step": 5761 }, { "epoch": 2.610783869506117, "grad_norm": 0.26747843616359235, "learning_rate": 4.3884873398512546e-05, "loss": 0.9603, "step": 5762 }, { "epoch": 2.611236973266878, "grad_norm": 0.20092331829878338, "learning_rate": 4.3872279343602695e-05, "loss": 0.9529, "step": 5763 }, { "epoch": 2.6116900770276392, "grad_norm": 0.16914884987414924, "learning_rate": 4.385968490118505e-05, "loss": 0.944, "step": 5764 }, { "epoch": 2.612143180788401, "grad_norm": 0.28543809446453505, "learning_rate": 4.384709007251994e-05, "loss": 0.9643, "step": 5765 }, { "epoch": 2.612596284549162, "grad_norm": 0.35895021297840624, "learning_rate": 4.383449485886777e-05, "loss": 0.9204, "step": 5766 }, { "epoch": 2.613049388309923, "grad_norm": 0.3106547853247225, "learning_rate": 4.3821899261488976e-05, "loss": 0.9339, "step": 5767 }, { "epoch": 2.613502492070684, "grad_norm": 0.28040089069360397, "learning_rate": 4.380930328164402e-05, "loss": 0.9525, "step": 5768 }, { "epoch": 2.6139555958314453, "grad_norm": 0.28261307837884436, "learning_rate": 4.3796706920593405e-05, "loss": 0.9407, "step": 5769 }, { "epoch": 2.6144086995922065, "grad_norm": 0.21040799404269325, "learning_rate": 4.3784110179597704e-05, "loss": 0.9364, "step": 5770 }, { "epoch": 2.6148618033529676, "grad_norm": 0.17693891649750473, "learning_rate": 4.3771513059917486e-05, "loss": 0.9485, "step": 5771 }, { "epoch": 2.615314907113729, "grad_norm": 0.20109737205730235, "learning_rate": 4.375891556281335e-05, "loss": 0.956, "step": 5772 }, { "epoch": 2.6157680108744903, "grad_norm": 0.19467989109062594, "learning_rate": 4.3746317689546e-05, "loss": 0.9403, "step": 5773 }, { "epoch": 2.6162211146352514, "grad_norm": 0.2342631401941167, "learning_rate": 4.3733719441376105e-05, "loss": 0.9298, "step": 5774 }, { "epoch": 2.6166742183960126, "grad_norm": 0.22143939838235036, "learning_rate": 4.3721120819564406e-05, "loss": 0.9504, "step": 5775 }, { "epoch": 2.617127322156774, "grad_norm": 0.23192201464907503, "learning_rate": 4.370852182537167e-05, "loss": 0.9484, "step": 5776 }, { "epoch": 2.6175804259175353, "grad_norm": 0.22011657531026813, "learning_rate": 4.369592246005872e-05, "loss": 0.961, "step": 5777 }, { "epoch": 2.6180335296782964, "grad_norm": 0.18854419685105941, "learning_rate": 4.368332272488639e-05, "loss": 0.9336, "step": 5778 }, { "epoch": 2.6184866334390575, "grad_norm": 0.19770218062254072, "learning_rate": 4.367072262111558e-05, "loss": 0.9408, "step": 5779 }, { "epoch": 2.6189397371998187, "grad_norm": 0.1868317032095863, "learning_rate": 4.365812215000719e-05, "loss": 0.9557, "step": 5780 }, { "epoch": 2.61939284096058, "grad_norm": 0.18855313060406106, "learning_rate": 4.36455213128222e-05, "loss": 0.9687, "step": 5781 }, { "epoch": 2.619845944721341, "grad_norm": 0.23052844536348602, "learning_rate": 4.36329201108216e-05, "loss": 0.9218, "step": 5782 }, { "epoch": 2.6202990484821025, "grad_norm": 0.2220876838543353, "learning_rate": 4.362031854526641e-05, "loss": 0.9493, "step": 5783 }, { "epoch": 2.6207521522428636, "grad_norm": 0.19108117602424152, "learning_rate": 4.36077166174177e-05, "loss": 0.9386, "step": 5784 }, { "epoch": 2.6212052560036248, "grad_norm": 0.15892668289706038, "learning_rate": 4.359511432853658e-05, "loss": 0.9446, "step": 5785 }, { "epoch": 2.621658359764386, "grad_norm": 0.182667058421785, "learning_rate": 4.358251167988418e-05, "loss": 0.9461, "step": 5786 }, { "epoch": 2.6221114635251475, "grad_norm": 0.21771157832288748, "learning_rate": 4.356990867272168e-05, "loss": 0.928, "step": 5787 }, { "epoch": 2.6225645672859086, "grad_norm": 0.20939481504959898, "learning_rate": 4.3557305308310293e-05, "loss": 0.9496, "step": 5788 }, { "epoch": 2.6230176710466697, "grad_norm": 0.1894731517167312, "learning_rate": 4.354470158791127e-05, "loss": 0.9693, "step": 5789 }, { "epoch": 2.623470774807431, "grad_norm": 0.1730062830479484, "learning_rate": 4.353209751278591e-05, "loss": 0.9605, "step": 5790 }, { "epoch": 2.623923878568192, "grad_norm": 0.20239753369867752, "learning_rate": 4.351949308419549e-05, "loss": 0.9479, "step": 5791 }, { "epoch": 2.624376982328953, "grad_norm": 0.2107570360033853, "learning_rate": 4.350688830340139e-05, "loss": 0.9251, "step": 5792 }, { "epoch": 2.6248300860897147, "grad_norm": 0.24512093573838262, "learning_rate": 4.349428317166499e-05, "loss": 0.9531, "step": 5793 }, { "epoch": 2.625283189850476, "grad_norm": 0.2431692611879219, "learning_rate": 4.348167769024773e-05, "loss": 0.9373, "step": 5794 }, { "epoch": 2.625736293611237, "grad_norm": 0.2617639248465907, "learning_rate": 4.3469071860411054e-05, "loss": 0.9557, "step": 5795 }, { "epoch": 2.626189397371998, "grad_norm": 0.2205877637165084, "learning_rate": 4.345646568341646e-05, "loss": 0.9351, "step": 5796 }, { "epoch": 2.6266425011327597, "grad_norm": 0.17681973462625178, "learning_rate": 4.3443859160525477e-05, "loss": 0.9197, "step": 5797 }, { "epoch": 2.627095604893521, "grad_norm": 0.1698040137635792, "learning_rate": 4.343125229299968e-05, "loss": 0.9364, "step": 5798 }, { "epoch": 2.627548708654282, "grad_norm": 0.18012899248827854, "learning_rate": 4.341864508210065e-05, "loss": 0.9288, "step": 5799 }, { "epoch": 2.628001812415043, "grad_norm": 0.19275703494308452, "learning_rate": 4.340603752909003e-05, "loss": 0.9322, "step": 5800 }, { "epoch": 2.628454916175804, "grad_norm": 0.17072184171695787, "learning_rate": 4.3393429635229475e-05, "loss": 0.9558, "step": 5801 }, { "epoch": 2.6289080199365653, "grad_norm": 0.16215363066164534, "learning_rate": 4.338082140178069e-05, "loss": 0.9335, "step": 5802 }, { "epoch": 2.6293611236973264, "grad_norm": 0.18649545986689883, "learning_rate": 4.336821283000541e-05, "loss": 0.9331, "step": 5803 }, { "epoch": 2.629814227458088, "grad_norm": 0.18566647791203905, "learning_rate": 4.33556039211654e-05, "loss": 0.9564, "step": 5804 }, { "epoch": 2.630267331218849, "grad_norm": 0.20094805369045388, "learning_rate": 4.334299467652248e-05, "loss": 0.9728, "step": 5805 }, { "epoch": 2.6307204349796103, "grad_norm": 0.23239993724970923, "learning_rate": 4.333038509733848e-05, "loss": 0.9446, "step": 5806 }, { "epoch": 2.6311735387403714, "grad_norm": 0.21050593877820087, "learning_rate": 4.331777518487524e-05, "loss": 0.9448, "step": 5807 }, { "epoch": 2.631626642501133, "grad_norm": 0.1836803088914215, "learning_rate": 4.3305164940394716e-05, "loss": 0.9709, "step": 5808 }, { "epoch": 2.632079746261894, "grad_norm": 0.21262631423194198, "learning_rate": 4.329255436515879e-05, "loss": 0.968, "step": 5809 }, { "epoch": 2.6325328500226552, "grad_norm": 0.22197771406879108, "learning_rate": 4.327994346042946e-05, "loss": 0.9274, "step": 5810 }, { "epoch": 2.6329859537834164, "grad_norm": 0.20930988127100558, "learning_rate": 4.326733222746872e-05, "loss": 0.9272, "step": 5811 }, { "epoch": 2.6334390575441775, "grad_norm": 0.19059059510663756, "learning_rate": 4.325472066753861e-05, "loss": 0.9366, "step": 5812 }, { "epoch": 2.6338921613049386, "grad_norm": 0.16253147350547703, "learning_rate": 4.324210878190119e-05, "loss": 0.9382, "step": 5813 }, { "epoch": 2.6343452650656998, "grad_norm": 0.19276332824852768, "learning_rate": 4.322949657181857e-05, "loss": 0.9559, "step": 5814 }, { "epoch": 2.6347983688264613, "grad_norm": 0.20059513806273577, "learning_rate": 4.321688403855287e-05, "loss": 0.9344, "step": 5815 }, { "epoch": 2.6352514725872225, "grad_norm": 0.1977934295067851, "learning_rate": 4.3204271183366285e-05, "loss": 0.9587, "step": 5816 }, { "epoch": 2.6357045763479836, "grad_norm": 0.2008553249457216, "learning_rate": 4.319165800752097e-05, "loss": 0.9617, "step": 5817 }, { "epoch": 2.6361576801087447, "grad_norm": 0.21272092823676933, "learning_rate": 4.317904451227917e-05, "loss": 0.9439, "step": 5818 }, { "epoch": 2.6366107838695063, "grad_norm": 0.1945817958997329, "learning_rate": 4.3166430698903164e-05, "loss": 0.9315, "step": 5819 }, { "epoch": 2.6370638876302674, "grad_norm": 0.2177725504241592, "learning_rate": 4.315381656865522e-05, "loss": 0.9492, "step": 5820 }, { "epoch": 2.6375169913910286, "grad_norm": 0.17717910475973284, "learning_rate": 4.314120212279767e-05, "loss": 0.9389, "step": 5821 }, { "epoch": 2.6379700951517897, "grad_norm": 0.16230564626256014, "learning_rate": 4.312858736259288e-05, "loss": 0.9634, "step": 5822 }, { "epoch": 2.638423198912551, "grad_norm": 0.1714943091222838, "learning_rate": 4.311597228930323e-05, "loss": 0.9528, "step": 5823 }, { "epoch": 2.638876302673312, "grad_norm": 0.2031846239073825, "learning_rate": 4.3103356904191135e-05, "loss": 0.9403, "step": 5824 }, { "epoch": 2.6393294064340735, "grad_norm": 0.2245499030875706, "learning_rate": 4.309074120851907e-05, "loss": 0.9396, "step": 5825 }, { "epoch": 2.6397825101948347, "grad_norm": 0.24044418201188422, "learning_rate": 4.307812520354948e-05, "loss": 0.9351, "step": 5826 }, { "epoch": 2.640235613955596, "grad_norm": 0.2643542137202415, "learning_rate": 4.3065508890544906e-05, "loss": 0.9264, "step": 5827 }, { "epoch": 2.640688717716357, "grad_norm": 0.2335268701757262, "learning_rate": 4.305289227076787e-05, "loss": 0.9403, "step": 5828 }, { "epoch": 2.6411418214771185, "grad_norm": 0.1653301291470191, "learning_rate": 4.304027534548096e-05, "loss": 0.9514, "step": 5829 }, { "epoch": 2.6415949252378796, "grad_norm": 0.15693917862902027, "learning_rate": 4.302765811594678e-05, "loss": 0.9357, "step": 5830 }, { "epoch": 2.6420480289986408, "grad_norm": 0.17848428308727432, "learning_rate": 4.3015040583427956e-05, "loss": 0.9574, "step": 5831 }, { "epoch": 2.642501132759402, "grad_norm": 0.1977727531040749, "learning_rate": 4.3002422749187144e-05, "loss": 0.953, "step": 5832 }, { "epoch": 2.642954236520163, "grad_norm": 0.17872630405764306, "learning_rate": 4.2989804614487076e-05, "loss": 0.941, "step": 5833 }, { "epoch": 2.643407340280924, "grad_norm": 0.1934874339603688, "learning_rate": 4.297718618059044e-05, "loss": 0.9401, "step": 5834 }, { "epoch": 2.6438604440416853, "grad_norm": 0.226704946682049, "learning_rate": 4.296456744876002e-05, "loss": 0.9204, "step": 5835 }, { "epoch": 2.644313547802447, "grad_norm": 0.24978351838814314, "learning_rate": 4.295194842025858e-05, "loss": 0.94, "step": 5836 }, { "epoch": 2.644766651563208, "grad_norm": 0.19451704157798946, "learning_rate": 4.293932909634894e-05, "loss": 0.9541, "step": 5837 }, { "epoch": 2.645219755323969, "grad_norm": 0.16731241019809936, "learning_rate": 4.2926709478293954e-05, "loss": 0.929, "step": 5838 }, { "epoch": 2.6456728590847303, "grad_norm": 0.17124672627495136, "learning_rate": 4.291408956735648e-05, "loss": 0.936, "step": 5839 }, { "epoch": 2.646125962845492, "grad_norm": 0.20484206063597651, "learning_rate": 4.290146936479943e-05, "loss": 0.9393, "step": 5840 }, { "epoch": 2.646579066606253, "grad_norm": 0.24871521117269454, "learning_rate": 4.288884887188574e-05, "loss": 0.9569, "step": 5841 }, { "epoch": 2.647032170367014, "grad_norm": 0.20204350330952717, "learning_rate": 4.2876228089878366e-05, "loss": 0.948, "step": 5842 }, { "epoch": 2.6474852741277752, "grad_norm": 0.1660532053661265, "learning_rate": 4.286360702004031e-05, "loss": 0.9478, "step": 5843 }, { "epoch": 2.6479383778885364, "grad_norm": 0.18613249849394428, "learning_rate": 4.285098566363457e-05, "loss": 0.9372, "step": 5844 }, { "epoch": 2.6483914816492975, "grad_norm": 0.17667130601569048, "learning_rate": 4.2838364021924216e-05, "loss": 0.949, "step": 5845 }, { "epoch": 2.6488445854100586, "grad_norm": 0.18507434175102092, "learning_rate": 4.282574209617231e-05, "loss": 0.9587, "step": 5846 }, { "epoch": 2.64929768917082, "grad_norm": 0.19662070686836897, "learning_rate": 4.281311988764196e-05, "loss": 0.9519, "step": 5847 }, { "epoch": 2.6497507929315813, "grad_norm": 0.1915146968889399, "learning_rate": 4.280049739759629e-05, "loss": 0.9472, "step": 5848 }, { "epoch": 2.6502038966923425, "grad_norm": 0.1566962842887867, "learning_rate": 4.278787462729848e-05, "loss": 0.9503, "step": 5849 }, { "epoch": 2.6506570004531036, "grad_norm": 0.21600216662161395, "learning_rate": 4.277525157801171e-05, "loss": 0.9447, "step": 5850 }, { "epoch": 2.651110104213865, "grad_norm": 0.3083126429324391, "learning_rate": 4.2762628250999206e-05, "loss": 0.9596, "step": 5851 }, { "epoch": 2.6515632079746263, "grad_norm": 0.2815156058294197, "learning_rate": 4.2750004647524196e-05, "loss": 0.9464, "step": 5852 }, { "epoch": 2.6520163117353874, "grad_norm": 0.2373007762584992, "learning_rate": 4.273738076884996e-05, "loss": 0.9268, "step": 5853 }, { "epoch": 2.6524694154961486, "grad_norm": 0.19665885452305096, "learning_rate": 4.2724756616239794e-05, "loss": 0.9554, "step": 5854 }, { "epoch": 2.6529225192569097, "grad_norm": 0.2162122721692876, "learning_rate": 4.271213219095704e-05, "loss": 0.9549, "step": 5855 }, { "epoch": 2.653375623017671, "grad_norm": 0.19032432806546576, "learning_rate": 4.269950749426505e-05, "loss": 0.9394, "step": 5856 }, { "epoch": 2.6538287267784324, "grad_norm": 0.18225362304310594, "learning_rate": 4.268688252742719e-05, "loss": 0.9501, "step": 5857 }, { "epoch": 2.6542818305391935, "grad_norm": 0.17879126703105466, "learning_rate": 4.267425729170688e-05, "loss": 0.9213, "step": 5858 }, { "epoch": 2.6547349342999547, "grad_norm": 0.19359784664968632, "learning_rate": 4.2661631788367564e-05, "loss": 0.9292, "step": 5859 }, { "epoch": 2.655188038060716, "grad_norm": 0.2026647930248494, "learning_rate": 4.264900601867269e-05, "loss": 0.977, "step": 5860 }, { "epoch": 2.6556411418214774, "grad_norm": 0.226210302663474, "learning_rate": 4.2636379983885756e-05, "loss": 0.9248, "step": 5861 }, { "epoch": 2.6560942455822385, "grad_norm": 0.23241522663770559, "learning_rate": 4.262375368527028e-05, "loss": 0.9442, "step": 5862 }, { "epoch": 2.6565473493429996, "grad_norm": 0.17865959060606423, "learning_rate": 4.26111271240898e-05, "loss": 0.9633, "step": 5863 }, { "epoch": 2.6570004531037608, "grad_norm": 0.19289998540873332, "learning_rate": 4.259850030160788e-05, "loss": 0.938, "step": 5864 }, { "epoch": 2.657453556864522, "grad_norm": 0.24991771472099303, "learning_rate": 4.2585873219088115e-05, "loss": 0.9562, "step": 5865 }, { "epoch": 2.657906660625283, "grad_norm": 0.24634350411198097, "learning_rate": 4.257324587779414e-05, "loss": 0.9458, "step": 5866 }, { "epoch": 2.658359764386044, "grad_norm": 0.22799276620562128, "learning_rate": 4.2560618278989596e-05, "loss": 0.9332, "step": 5867 }, { "epoch": 2.6588128681468057, "grad_norm": 0.2006090034383811, "learning_rate": 4.2547990423938146e-05, "loss": 0.9398, "step": 5868 }, { "epoch": 2.659265971907567, "grad_norm": 0.18430388287433974, "learning_rate": 4.2535362313903494e-05, "loss": 0.9631, "step": 5869 }, { "epoch": 2.659719075668328, "grad_norm": 0.20786797395218276, "learning_rate": 4.252273395014937e-05, "loss": 0.9338, "step": 5870 }, { "epoch": 2.660172179429089, "grad_norm": 0.18896516830498977, "learning_rate": 4.251010533393951e-05, "loss": 0.9491, "step": 5871 }, { "epoch": 2.6606252831898507, "grad_norm": 0.1987537285749006, "learning_rate": 4.24974764665377e-05, "loss": 0.9458, "step": 5872 }, { "epoch": 2.661078386950612, "grad_norm": 0.21495203623330303, "learning_rate": 4.2484847349207736e-05, "loss": 0.9603, "step": 5873 }, { "epoch": 2.661531490711373, "grad_norm": 0.21146369699360665, "learning_rate": 4.247221798321344e-05, "loss": 0.9453, "step": 5874 }, { "epoch": 2.661984594472134, "grad_norm": 0.1987685689665894, "learning_rate": 4.245958836981866e-05, "loss": 0.9445, "step": 5875 }, { "epoch": 2.6624376982328952, "grad_norm": 0.2118904743089966, "learning_rate": 4.244695851028728e-05, "loss": 0.9351, "step": 5876 }, { "epoch": 2.6628908019936564, "grad_norm": 0.2422645725667947, "learning_rate": 4.24343284058832e-05, "loss": 0.9696, "step": 5877 }, { "epoch": 2.6633439057544175, "grad_norm": 0.18086865877719252, "learning_rate": 4.242169805787033e-05, "loss": 0.9264, "step": 5878 }, { "epoch": 2.663797009515179, "grad_norm": 0.16368323313962183, "learning_rate": 4.240906746751261e-05, "loss": 0.947, "step": 5879 }, { "epoch": 2.66425011327594, "grad_norm": 0.19166932747020612, "learning_rate": 4.239643663607403e-05, "loss": 0.9425, "step": 5880 }, { "epoch": 2.6647032170367013, "grad_norm": 0.19052727429781735, "learning_rate": 4.2383805564818584e-05, "loss": 0.959, "step": 5881 }, { "epoch": 2.6651563207974625, "grad_norm": 0.22188733412341327, "learning_rate": 4.237117425501029e-05, "loss": 0.9436, "step": 5882 }, { "epoch": 2.665609424558224, "grad_norm": 0.7091142356327578, "learning_rate": 4.235854270791319e-05, "loss": 0.9567, "step": 5883 }, { "epoch": 2.666062528318985, "grad_norm": 0.17282855433477984, "learning_rate": 4.2345910924791356e-05, "loss": 0.9311, "step": 5884 }, { "epoch": 2.6665156320797463, "grad_norm": 0.18771655724673275, "learning_rate": 4.2333278906908876e-05, "loss": 0.9369, "step": 5885 }, { "epoch": 2.6669687358405074, "grad_norm": 0.2125704003808977, "learning_rate": 4.2320646655529865e-05, "loss": 0.9568, "step": 5886 }, { "epoch": 2.6674218396012686, "grad_norm": 0.19306346347358452, "learning_rate": 4.230801417191847e-05, "loss": 0.9391, "step": 5887 }, { "epoch": 2.6678749433620297, "grad_norm": 0.2172589292692943, "learning_rate": 4.229538145733883e-05, "loss": 0.9325, "step": 5888 }, { "epoch": 2.6683280471227913, "grad_norm": 0.2444856540043642, "learning_rate": 4.2282748513055146e-05, "loss": 0.9505, "step": 5889 }, { "epoch": 2.6687811508835524, "grad_norm": 0.29046861180711425, "learning_rate": 4.227011534033162e-05, "loss": 0.9545, "step": 5890 }, { "epoch": 2.6692342546443135, "grad_norm": 0.29857414366816426, "learning_rate": 4.2257481940432487e-05, "loss": 0.9298, "step": 5891 }, { "epoch": 2.6696873584050747, "grad_norm": 0.2968033830814379, "learning_rate": 4.2244848314621996e-05, "loss": 0.969, "step": 5892 }, { "epoch": 2.6701404621658362, "grad_norm": 0.3015332304652636, "learning_rate": 4.223221446416442e-05, "loss": 0.9387, "step": 5893 }, { "epoch": 2.6705935659265974, "grad_norm": 0.28169161301909523, "learning_rate": 4.221958039032407e-05, "loss": 0.9202, "step": 5894 }, { "epoch": 2.6710466696873585, "grad_norm": 0.19904996460225155, "learning_rate": 4.2206946094365254e-05, "loss": 0.9602, "step": 5895 }, { "epoch": 2.6714997734481196, "grad_norm": 0.17581576963655438, "learning_rate": 4.219431157755232e-05, "loss": 0.9609, "step": 5896 }, { "epoch": 2.6719528772088808, "grad_norm": 0.21417480814262788, "learning_rate": 4.2181676841149643e-05, "loss": 0.9367, "step": 5897 }, { "epoch": 2.672405980969642, "grad_norm": 0.2287413270347817, "learning_rate": 4.2169041886421586e-05, "loss": 0.9437, "step": 5898 }, { "epoch": 2.672859084730403, "grad_norm": 0.3593001272540887, "learning_rate": 4.2156406714632575e-05, "loss": 0.9468, "step": 5899 }, { "epoch": 2.6733121884911646, "grad_norm": 0.21888506543007052, "learning_rate": 4.2143771327047026e-05, "loss": 0.9669, "step": 5900 }, { "epoch": 2.6737652922519257, "grad_norm": 0.20998493643943897, "learning_rate": 4.2131135724929404e-05, "loss": 0.9478, "step": 5901 }, { "epoch": 2.674218396012687, "grad_norm": 0.24112107502277783, "learning_rate": 4.2118499909544186e-05, "loss": 0.9721, "step": 5902 }, { "epoch": 2.674671499773448, "grad_norm": 0.266468381777115, "learning_rate": 4.210586388215586e-05, "loss": 0.9441, "step": 5903 }, { "epoch": 2.6751246035342096, "grad_norm": 0.2642246802719887, "learning_rate": 4.2093227644028944e-05, "loss": 0.9268, "step": 5904 }, { "epoch": 2.6755777072949707, "grad_norm": 0.24045023894213471, "learning_rate": 4.208059119642797e-05, "loss": 0.9535, "step": 5905 }, { "epoch": 2.676030811055732, "grad_norm": 0.2231201608960572, "learning_rate": 4.206795454061749e-05, "loss": 0.9451, "step": 5906 }, { "epoch": 2.676483914816493, "grad_norm": 0.2059667767154422, "learning_rate": 4.205531767786211e-05, "loss": 0.9648, "step": 5907 }, { "epoch": 2.676937018577254, "grad_norm": 0.15591045321566835, "learning_rate": 4.20426806094264e-05, "loss": 0.9285, "step": 5908 }, { "epoch": 2.677390122338015, "grad_norm": 0.19583566484765844, "learning_rate": 4.2030043336574986e-05, "loss": 0.9493, "step": 5909 }, { "epoch": 2.677843226098777, "grad_norm": 0.2492654349533564, "learning_rate": 4.201740586057252e-05, "loss": 0.9419, "step": 5910 }, { "epoch": 2.678296329859538, "grad_norm": 0.27407853538610916, "learning_rate": 4.200476818268366e-05, "loss": 0.9134, "step": 5911 }, { "epoch": 2.678749433620299, "grad_norm": 0.3134122003675789, "learning_rate": 4.199213030417309e-05, "loss": 0.9205, "step": 5912 }, { "epoch": 2.67920253738106, "grad_norm": 0.3016257796029566, "learning_rate": 4.1979492226305504e-05, "loss": 0.9462, "step": 5913 }, { "epoch": 2.6796556411418218, "grad_norm": 0.2566217965620852, "learning_rate": 4.1966853950345635e-05, "loss": 0.9494, "step": 5914 }, { "epoch": 2.680108744902583, "grad_norm": 0.19106995737235336, "learning_rate": 4.195421547755819e-05, "loss": 0.953, "step": 5915 }, { "epoch": 2.680561848663344, "grad_norm": 0.1610802712683873, "learning_rate": 4.194157680920797e-05, "loss": 0.9432, "step": 5916 }, { "epoch": 2.681014952424105, "grad_norm": 0.1700532921176455, "learning_rate": 4.192893794655973e-05, "loss": 0.9261, "step": 5917 }, { "epoch": 2.6814680561848663, "grad_norm": 0.19683722200960244, "learning_rate": 4.1916298890878285e-05, "loss": 0.9226, "step": 5918 }, { "epoch": 2.6819211599456274, "grad_norm": 0.23162607320980216, "learning_rate": 4.190365964342844e-05, "loss": 0.9603, "step": 5919 }, { "epoch": 2.6823742637063885, "grad_norm": 0.24197302839542742, "learning_rate": 4.189102020547505e-05, "loss": 0.9387, "step": 5920 }, { "epoch": 2.68282736746715, "grad_norm": 0.20737878013209654, "learning_rate": 4.187838057828297e-05, "loss": 0.9533, "step": 5921 }, { "epoch": 2.6832804712279112, "grad_norm": 0.18295534245871137, "learning_rate": 4.186574076311706e-05, "loss": 0.9281, "step": 5922 }, { "epoch": 2.6837335749886724, "grad_norm": 0.17232149352669954, "learning_rate": 4.185310076124223e-05, "loss": 0.9357, "step": 5923 }, { "epoch": 2.6841866787494335, "grad_norm": 0.16996385018899718, "learning_rate": 4.1840460573923387e-05, "loss": 0.9658, "step": 5924 }, { "epoch": 2.684639782510195, "grad_norm": 0.17134632803913608, "learning_rate": 4.182782020242546e-05, "loss": 0.9261, "step": 5925 }, { "epoch": 2.685092886270956, "grad_norm": 0.16268664725129345, "learning_rate": 4.181517964801341e-05, "loss": 0.9544, "step": 5926 }, { "epoch": 2.6855459900317173, "grad_norm": 0.1612653358913028, "learning_rate": 4.18025389119522e-05, "loss": 0.9329, "step": 5927 }, { "epoch": 2.6859990937924785, "grad_norm": 0.17721956640586556, "learning_rate": 4.1789897995506814e-05, "loss": 0.9516, "step": 5928 }, { "epoch": 2.6864521975532396, "grad_norm": 0.1687953055898851, "learning_rate": 4.1777256899942274e-05, "loss": 0.922, "step": 5929 }, { "epoch": 2.6869053013140007, "grad_norm": 0.16843981230994712, "learning_rate": 4.1764615626523584e-05, "loss": 0.9425, "step": 5930 }, { "epoch": 2.687358405074762, "grad_norm": 0.21063766237134407, "learning_rate": 4.175197417651579e-05, "loss": 0.9337, "step": 5931 }, { "epoch": 2.6878115088355234, "grad_norm": 0.17131295353295656, "learning_rate": 4.173933255118395e-05, "loss": 0.9371, "step": 5932 }, { "epoch": 2.6882646125962846, "grad_norm": 0.18907778538302966, "learning_rate": 4.1726690751793146e-05, "loss": 0.9441, "step": 5933 }, { "epoch": 2.6887177163570457, "grad_norm": 0.20521563586961555, "learning_rate": 4.1714048779608466e-05, "loss": 0.9208, "step": 5934 }, { "epoch": 2.689170820117807, "grad_norm": 0.2340060117590188, "learning_rate": 4.1701406635895027e-05, "loss": 0.9505, "step": 5935 }, { "epoch": 2.6896239238785684, "grad_norm": 0.24034103978408125, "learning_rate": 4.168876432191795e-05, "loss": 0.9498, "step": 5936 }, { "epoch": 2.6900770276393295, "grad_norm": 0.22663977760091436, "learning_rate": 4.167612183894238e-05, "loss": 0.9441, "step": 5937 }, { "epoch": 2.6905301314000907, "grad_norm": 0.24144260405273088, "learning_rate": 4.166347918823349e-05, "loss": 0.9468, "step": 5938 }, { "epoch": 2.690983235160852, "grad_norm": 0.2341251069622031, "learning_rate": 4.165083637105645e-05, "loss": 0.9528, "step": 5939 }, { "epoch": 2.691436338921613, "grad_norm": 0.17417696087759962, "learning_rate": 4.1638193388676455e-05, "loss": 0.9368, "step": 5940 }, { "epoch": 2.691889442682374, "grad_norm": 0.21063393028346358, "learning_rate": 4.162555024235872e-05, "loss": 0.9562, "step": 5941 }, { "epoch": 2.6923425464431356, "grad_norm": 0.26195223532532663, "learning_rate": 4.161290693336848e-05, "loss": 0.9652, "step": 5942 }, { "epoch": 2.6927956502038968, "grad_norm": 0.2845493095310242, "learning_rate": 4.1600263462970964e-05, "loss": 0.9217, "step": 5943 }, { "epoch": 2.693248753964658, "grad_norm": 0.24466052894549245, "learning_rate": 4.1587619832431434e-05, "loss": 0.9552, "step": 5944 }, { "epoch": 2.693701857725419, "grad_norm": 0.1942411475058703, "learning_rate": 4.157497604301518e-05, "loss": 0.9335, "step": 5945 }, { "epoch": 2.6941549614861806, "grad_norm": 0.16332064160384732, "learning_rate": 4.156233209598749e-05, "loss": 0.94, "step": 5946 }, { "epoch": 2.6946080652469417, "grad_norm": 0.16821910506364804, "learning_rate": 4.154968799261367e-05, "loss": 0.9347, "step": 5947 }, { "epoch": 2.695061169007703, "grad_norm": 0.21095065458071202, "learning_rate": 4.153704373415905e-05, "loss": 0.9338, "step": 5948 }, { "epoch": 2.695514272768464, "grad_norm": 0.24796101975239443, "learning_rate": 4.1524399321888955e-05, "loss": 0.9362, "step": 5949 }, { "epoch": 2.695967376529225, "grad_norm": 0.21515123911656214, "learning_rate": 4.151175475706876e-05, "loss": 0.9371, "step": 5950 }, { "epoch": 2.6964204802899863, "grad_norm": 0.1994179793501732, "learning_rate": 4.149911004096382e-05, "loss": 0.928, "step": 5951 }, { "epoch": 2.6968735840507474, "grad_norm": 0.202128978076105, "learning_rate": 4.1486465174839514e-05, "loss": 0.9501, "step": 5952 }, { "epoch": 2.697326687811509, "grad_norm": 0.18379600354960698, "learning_rate": 4.147382015996127e-05, "loss": 0.9497, "step": 5953 }, { "epoch": 2.69777979157227, "grad_norm": 0.1961717219343874, "learning_rate": 4.146117499759449e-05, "loss": 0.9338, "step": 5954 }, { "epoch": 2.6982328953330312, "grad_norm": 0.15323293071372956, "learning_rate": 4.144852968900459e-05, "loss": 0.9238, "step": 5955 }, { "epoch": 2.6986859990937924, "grad_norm": 0.1813059231712377, "learning_rate": 4.143588423545704e-05, "loss": 0.9431, "step": 5956 }, { "epoch": 2.699139102854554, "grad_norm": 0.22330094820747137, "learning_rate": 4.142323863821727e-05, "loss": 0.9376, "step": 5957 }, { "epoch": 2.699592206615315, "grad_norm": 0.2621613678817773, "learning_rate": 4.141059289855078e-05, "loss": 0.9395, "step": 5958 }, { "epoch": 2.700045310376076, "grad_norm": 0.2753314075847928, "learning_rate": 4.1397947017723054e-05, "loss": 0.9541, "step": 5959 }, { "epoch": 2.7004984141368373, "grad_norm": 0.20522035711706282, "learning_rate": 4.138530099699958e-05, "loss": 0.931, "step": 5960 }, { "epoch": 2.7009515178975985, "grad_norm": 0.17452764436687784, "learning_rate": 4.137265483764588e-05, "loss": 0.9271, "step": 5961 }, { "epoch": 2.7014046216583596, "grad_norm": 0.22981964899428559, "learning_rate": 4.136000854092749e-05, "loss": 0.9228, "step": 5962 }, { "epoch": 2.7018577254191207, "grad_norm": 0.2353897803664448, "learning_rate": 4.134736210810995e-05, "loss": 0.9382, "step": 5963 }, { "epoch": 2.7023108291798823, "grad_norm": 0.258546872833548, "learning_rate": 4.1334715540458815e-05, "loss": 0.9499, "step": 5964 }, { "epoch": 2.7027639329406434, "grad_norm": 0.24949677155789826, "learning_rate": 4.132206883923966e-05, "loss": 0.9442, "step": 5965 }, { "epoch": 2.7032170367014046, "grad_norm": 0.19510260180083208, "learning_rate": 4.130942200571808e-05, "loss": 0.9435, "step": 5966 }, { "epoch": 2.7036701404621657, "grad_norm": 0.16779460796981152, "learning_rate": 4.1296775041159634e-05, "loss": 0.9575, "step": 5967 }, { "epoch": 2.7041232442229273, "grad_norm": 0.16315714953083452, "learning_rate": 4.128412794682997e-05, "loss": 0.9377, "step": 5968 }, { "epoch": 2.7045763479836884, "grad_norm": 0.38130762755351255, "learning_rate": 4.12714807239947e-05, "loss": 0.9338, "step": 5969 }, { "epoch": 2.7050294517444495, "grad_norm": 0.13782180717516335, "learning_rate": 4.1258833373919465e-05, "loss": 0.9311, "step": 5970 }, { "epoch": 2.7054825555052107, "grad_norm": 0.15283914878692703, "learning_rate": 4.1246185897869904e-05, "loss": 0.9429, "step": 5971 }, { "epoch": 2.705935659265972, "grad_norm": 0.18298490449583385, "learning_rate": 4.123353829711168e-05, "loss": 0.9618, "step": 5972 }, { "epoch": 2.706388763026733, "grad_norm": 0.2115668179446055, "learning_rate": 4.1220890572910476e-05, "loss": 0.9323, "step": 5973 }, { "epoch": 2.7068418667874945, "grad_norm": 0.18615760833530023, "learning_rate": 4.120824272653197e-05, "loss": 0.9297, "step": 5974 }, { "epoch": 2.7072949705482556, "grad_norm": 0.18481421946412968, "learning_rate": 4.119559475924187e-05, "loss": 0.9557, "step": 5975 }, { "epoch": 2.7077480743090168, "grad_norm": 0.1805445398656486, "learning_rate": 4.1182946672305874e-05, "loss": 0.9687, "step": 5976 }, { "epoch": 2.708201178069778, "grad_norm": 0.16662419744783688, "learning_rate": 4.117029846698971e-05, "loss": 0.9469, "step": 5977 }, { "epoch": 2.7086542818305395, "grad_norm": 0.22056630482196465, "learning_rate": 4.115765014455912e-05, "loss": 0.9504, "step": 5978 }, { "epoch": 2.7091073855913006, "grad_norm": 0.23005731353823042, "learning_rate": 4.1145001706279845e-05, "loss": 0.9452, "step": 5979 }, { "epoch": 2.7095604893520617, "grad_norm": 0.19322693019372758, "learning_rate": 4.113235315341764e-05, "loss": 0.9329, "step": 5980 }, { "epoch": 2.710013593112823, "grad_norm": 0.17174892689083562, "learning_rate": 4.111970448723828e-05, "loss": 0.9397, "step": 5981 }, { "epoch": 2.710466696873584, "grad_norm": 0.18462020004268637, "learning_rate": 4.110705570900754e-05, "loss": 0.9555, "step": 5982 }, { "epoch": 2.710919800634345, "grad_norm": 0.1914839176363616, "learning_rate": 4.1094406819991224e-05, "loss": 0.9246, "step": 5983 }, { "epoch": 2.7113729043951063, "grad_norm": 0.16159317647727237, "learning_rate": 4.1081757821455127e-05, "loss": 0.9517, "step": 5984 }, { "epoch": 2.711826008155868, "grad_norm": 0.18602838324546914, "learning_rate": 4.106910871466505e-05, "loss": 0.9257, "step": 5985 }, { "epoch": 2.712279111916629, "grad_norm": 0.18199754453793995, "learning_rate": 4.105645950088684e-05, "loss": 0.9402, "step": 5986 }, { "epoch": 2.71273221567739, "grad_norm": 0.17201860175676423, "learning_rate": 4.1043810181386314e-05, "loss": 0.9342, "step": 5987 }, { "epoch": 2.713185319438151, "grad_norm": 0.23888897471204143, "learning_rate": 4.103116075742933e-05, "loss": 0.9416, "step": 5988 }, { "epoch": 2.713638423198913, "grad_norm": 0.2538049684653299, "learning_rate": 4.101851123028175e-05, "loss": 0.9245, "step": 5989 }, { "epoch": 2.714091526959674, "grad_norm": 0.21996317535920465, "learning_rate": 4.1005861601209424e-05, "loss": 0.9393, "step": 5990 }, { "epoch": 2.714544630720435, "grad_norm": 0.21699389811919673, "learning_rate": 4.0993211871478246e-05, "loss": 0.9467, "step": 5991 }, { "epoch": 2.714997734481196, "grad_norm": 0.2266566813417079, "learning_rate": 4.098056204235409e-05, "loss": 0.9255, "step": 5992 }, { "epoch": 2.7154508382419573, "grad_norm": 0.28435980412402584, "learning_rate": 4.096791211510286e-05, "loss": 0.9746, "step": 5993 }, { "epoch": 2.7159039420027185, "grad_norm": 0.31376447984047273, "learning_rate": 4.095526209099046e-05, "loss": 0.9446, "step": 5994 }, { "epoch": 2.7163570457634796, "grad_norm": 0.32242408699399194, "learning_rate": 4.094261197128281e-05, "loss": 0.9638, "step": 5995 }, { "epoch": 2.716810149524241, "grad_norm": 0.2597179562959269, "learning_rate": 4.092996175724583e-05, "loss": 0.9507, "step": 5996 }, { "epoch": 2.7172632532850023, "grad_norm": 0.19805797809498193, "learning_rate": 4.091731145014547e-05, "loss": 0.9445, "step": 5997 }, { "epoch": 2.7177163570457634, "grad_norm": 0.25417284104962, "learning_rate": 4.0904661051247656e-05, "loss": 0.9385, "step": 5998 }, { "epoch": 2.7181694608065246, "grad_norm": 0.2545793336465508, "learning_rate": 4.0892010561818355e-05, "loss": 0.915, "step": 5999 }, { "epoch": 2.718622564567286, "grad_norm": 0.33731759129990213, "learning_rate": 4.087935998312353e-05, "loss": 0.9256, "step": 6000 }, { "epoch": 2.7190756683280473, "grad_norm": 0.34377257811164497, "learning_rate": 4.086670931642916e-05, "loss": 0.9404, "step": 6001 }, { "epoch": 2.7195287720888084, "grad_norm": 0.3051978324702256, "learning_rate": 4.0854058563001195e-05, "loss": 0.9524, "step": 6002 }, { "epoch": 2.7199818758495695, "grad_norm": 0.2669819076197626, "learning_rate": 4.0841407724105655e-05, "loss": 0.9337, "step": 6003 }, { "epoch": 2.7204349796103307, "grad_norm": 0.20600407513199767, "learning_rate": 4.082875680100852e-05, "loss": 0.9482, "step": 6004 }, { "epoch": 2.720888083371092, "grad_norm": 0.16745328437165063, "learning_rate": 4.081610579497581e-05, "loss": 0.9408, "step": 6005 }, { "epoch": 2.7213411871318534, "grad_norm": 0.1953373089253392, "learning_rate": 4.0803454707273545e-05, "loss": 0.9403, "step": 6006 }, { "epoch": 2.7217942908926145, "grad_norm": 0.1975440212821099, "learning_rate": 4.079080353916773e-05, "loss": 0.9267, "step": 6007 }, { "epoch": 2.7222473946533756, "grad_norm": 0.19877029964299064, "learning_rate": 4.077815229192441e-05, "loss": 0.9385, "step": 6008 }, { "epoch": 2.7227004984141367, "grad_norm": 0.18089410258566357, "learning_rate": 4.0765500966809614e-05, "loss": 0.9586, "step": 6009 }, { "epoch": 2.7231536021748983, "grad_norm": 0.16295568123150156, "learning_rate": 4.07528495650894e-05, "loss": 0.9595, "step": 6010 }, { "epoch": 2.7236067059356595, "grad_norm": 0.17845501157916677, "learning_rate": 4.0740198088029804e-05, "loss": 0.93, "step": 6011 }, { "epoch": 2.7240598096964206, "grad_norm": 0.20449734969035935, "learning_rate": 4.072754653689691e-05, "loss": 0.9562, "step": 6012 }, { "epoch": 2.7245129134571817, "grad_norm": 0.20524187971417285, "learning_rate": 4.071489491295678e-05, "loss": 0.961, "step": 6013 }, { "epoch": 2.724966017217943, "grad_norm": 0.23460852272849128, "learning_rate": 4.0702243217475474e-05, "loss": 0.9481, "step": 6014 }, { "epoch": 2.725419120978704, "grad_norm": 0.21489644482053277, "learning_rate": 4.0689591451719106e-05, "loss": 0.9491, "step": 6015 }, { "epoch": 2.725872224739465, "grad_norm": 0.1982658675373089, "learning_rate": 4.0676939616953744e-05, "loss": 0.9249, "step": 6016 }, { "epoch": 2.7263253285002267, "grad_norm": 0.1660148363173646, "learning_rate": 4.0664287714445505e-05, "loss": 0.9458, "step": 6017 }, { "epoch": 2.726778432260988, "grad_norm": 0.1955203162328406, "learning_rate": 4.0651635745460474e-05, "loss": 0.9478, "step": 6018 }, { "epoch": 2.727231536021749, "grad_norm": 0.19907477515069733, "learning_rate": 4.0638983711264776e-05, "loss": 0.9368, "step": 6019 }, { "epoch": 2.72768463978251, "grad_norm": 0.18313035577988765, "learning_rate": 4.062633161312451e-05, "loss": 0.9285, "step": 6020 }, { "epoch": 2.7281377435432717, "grad_norm": 0.18512319867160382, "learning_rate": 4.0613679452305826e-05, "loss": 0.9411, "step": 6021 }, { "epoch": 2.728590847304033, "grad_norm": 0.22923634737517834, "learning_rate": 4.0601027230074843e-05, "loss": 0.953, "step": 6022 }, { "epoch": 2.729043951064794, "grad_norm": 0.227882765579301, "learning_rate": 4.058837494769769e-05, "loss": 0.9543, "step": 6023 }, { "epoch": 2.729497054825555, "grad_norm": 0.19195800943184968, "learning_rate": 4.057572260644052e-05, "loss": 0.929, "step": 6024 }, { "epoch": 2.729950158586316, "grad_norm": 0.1649449418207911, "learning_rate": 4.056307020756948e-05, "loss": 0.9326, "step": 6025 }, { "epoch": 2.7304032623470773, "grad_norm": 0.18183160543545746, "learning_rate": 4.055041775235072e-05, "loss": 0.9394, "step": 6026 }, { "epoch": 2.7308563661078384, "grad_norm": 0.16679330516151333, "learning_rate": 4.053776524205041e-05, "loss": 0.9394, "step": 6027 }, { "epoch": 2.7313094698686, "grad_norm": 0.13363141024510794, "learning_rate": 4.052511267793471e-05, "loss": 0.9291, "step": 6028 }, { "epoch": 2.731762573629361, "grad_norm": 0.1647888734552007, "learning_rate": 4.0512460061269775e-05, "loss": 0.9433, "step": 6029 }, { "epoch": 2.7322156773901223, "grad_norm": 0.20043871532696764, "learning_rate": 4.04998073933218e-05, "loss": 0.9479, "step": 6030 }, { "epoch": 2.7326687811508834, "grad_norm": 0.1569332968210533, "learning_rate": 4.048715467535696e-05, "loss": 0.9366, "step": 6031 }, { "epoch": 2.733121884911645, "grad_norm": 0.1625690482095216, "learning_rate": 4.047450190864145e-05, "loss": 0.9379, "step": 6032 }, { "epoch": 2.733574988672406, "grad_norm": 0.19106728509371196, "learning_rate": 4.046184909444144e-05, "loss": 0.9279, "step": 6033 }, { "epoch": 2.7340280924331672, "grad_norm": 0.18976733045353955, "learning_rate": 4.044919623402316e-05, "loss": 0.9559, "step": 6034 }, { "epoch": 2.7344811961939284, "grad_norm": 0.17018334144367242, "learning_rate": 4.043654332865278e-05, "loss": 0.9289, "step": 6035 }, { "epoch": 2.7349342999546895, "grad_norm": 0.17723598019193568, "learning_rate": 4.042389037959652e-05, "loss": 0.9497, "step": 6036 }, { "epoch": 2.7353874037154506, "grad_norm": 0.21651630213268772, "learning_rate": 4.041123738812058e-05, "loss": 0.929, "step": 6037 }, { "epoch": 2.735840507476212, "grad_norm": 0.17722900682928083, "learning_rate": 4.0398584355491176e-05, "loss": 0.9489, "step": 6038 }, { "epoch": 2.7362936112369733, "grad_norm": 0.17208153270251475, "learning_rate": 4.0385931282974525e-05, "loss": 0.9642, "step": 6039 }, { "epoch": 2.7367467149977345, "grad_norm": 0.18105081013814506, "learning_rate": 4.037327817183686e-05, "loss": 0.9295, "step": 6040 }, { "epoch": 2.7371998187584956, "grad_norm": 0.18264091148774117, "learning_rate": 4.0360625023344394e-05, "loss": 0.9468, "step": 6041 }, { "epoch": 2.737652922519257, "grad_norm": 0.15337048010275342, "learning_rate": 4.0347971838763366e-05, "loss": 0.9577, "step": 6042 }, { "epoch": 2.7381060262800183, "grad_norm": 0.15338255815475232, "learning_rate": 4.033531861936e-05, "loss": 0.931, "step": 6043 }, { "epoch": 2.7385591300407794, "grad_norm": 0.15990914458274402, "learning_rate": 4.0322665366400544e-05, "loss": 0.9423, "step": 6044 }, { "epoch": 2.7390122338015406, "grad_norm": 0.161973027172897, "learning_rate": 4.031001208115123e-05, "loss": 0.9658, "step": 6045 }, { "epoch": 2.7394653375623017, "grad_norm": 0.17673023982672212, "learning_rate": 4.029735876487829e-05, "loss": 0.9701, "step": 6046 }, { "epoch": 2.739918441323063, "grad_norm": 0.18155484165965777, "learning_rate": 4.028470541884799e-05, "loss": 0.9119, "step": 6047 }, { "epoch": 2.740371545083824, "grad_norm": 0.20318942378081992, "learning_rate": 4.027205204432656e-05, "loss": 0.9608, "step": 6048 }, { "epoch": 2.7408246488445855, "grad_norm": 0.21768130532033572, "learning_rate": 4.025939864258027e-05, "loss": 0.9294, "step": 6049 }, { "epoch": 2.7412777526053467, "grad_norm": 0.291840478483286, "learning_rate": 4.024674521487537e-05, "loss": 0.9389, "step": 6050 }, { "epoch": 2.741730856366108, "grad_norm": 0.20244304241895128, "learning_rate": 4.023409176247811e-05, "loss": 0.9522, "step": 6051 }, { "epoch": 2.742183960126869, "grad_norm": 0.19960778548515384, "learning_rate": 4.0221438286654765e-05, "loss": 0.9403, "step": 6052 }, { "epoch": 2.7426370638876305, "grad_norm": 0.19217218132833974, "learning_rate": 4.020878478867159e-05, "loss": 0.9358, "step": 6053 }, { "epoch": 2.7430901676483916, "grad_norm": 0.25222235971228374, "learning_rate": 4.019613126979483e-05, "loss": 0.9707, "step": 6054 }, { "epoch": 2.7435432714091528, "grad_norm": 0.23363263199612916, "learning_rate": 4.018347773129077e-05, "loss": 0.9114, "step": 6055 }, { "epoch": 2.743996375169914, "grad_norm": 0.19608938686233046, "learning_rate": 4.017082417442569e-05, "loss": 0.9553, "step": 6056 }, { "epoch": 2.744449478930675, "grad_norm": 0.21674668161051078, "learning_rate": 4.015817060046583e-05, "loss": 0.9394, "step": 6057 }, { "epoch": 2.744902582691436, "grad_norm": 0.20261971295631462, "learning_rate": 4.0145517010677484e-05, "loss": 0.9424, "step": 6058 }, { "epoch": 2.7453556864521973, "grad_norm": 0.18400775105400974, "learning_rate": 4.013286340632692e-05, "loss": 0.9401, "step": 6059 }, { "epoch": 2.745808790212959, "grad_norm": 0.19017341168863966, "learning_rate": 4.012020978868041e-05, "loss": 0.9456, "step": 6060 }, { "epoch": 2.74626189397372, "grad_norm": 0.20780202496082759, "learning_rate": 4.0107556159004234e-05, "loss": 0.9388, "step": 6061 }, { "epoch": 2.746714997734481, "grad_norm": 0.2135231931298424, "learning_rate": 4.009490251856467e-05, "loss": 0.9518, "step": 6062 }, { "epoch": 2.7471681014952423, "grad_norm": 0.2352760848177765, "learning_rate": 4.008224886862799e-05, "loss": 0.9452, "step": 6063 }, { "epoch": 2.747621205256004, "grad_norm": 0.19598352378759015, "learning_rate": 4.006959521046049e-05, "loss": 0.9238, "step": 6064 }, { "epoch": 2.748074309016765, "grad_norm": 0.23205069411296436, "learning_rate": 4.005694154532842e-05, "loss": 0.9434, "step": 6065 }, { "epoch": 2.748527412777526, "grad_norm": 0.23982354445682516, "learning_rate": 4.004428787449809e-05, "loss": 0.9529, "step": 6066 }, { "epoch": 2.7489805165382872, "grad_norm": 0.18408923140376202, "learning_rate": 4.003163419923577e-05, "loss": 0.9247, "step": 6067 }, { "epoch": 2.7494336202990484, "grad_norm": 0.2161353182277104, "learning_rate": 4.0018980520807745e-05, "loss": 0.9433, "step": 6068 }, { "epoch": 2.7498867240598095, "grad_norm": 0.27245222977227906, "learning_rate": 4.0006326840480295e-05, "loss": 0.9473, "step": 6069 }, { "epoch": 2.750339827820571, "grad_norm": 0.3265396215364662, "learning_rate": 3.999367315951971e-05, "loss": 0.9512, "step": 6070 }, { "epoch": 2.750792931581332, "grad_norm": 0.34752737053014204, "learning_rate": 3.998101947919227e-05, "loss": 0.9531, "step": 6071 }, { "epoch": 2.7512460353420933, "grad_norm": 0.366773493424702, "learning_rate": 3.996836580076424e-05, "loss": 0.9638, "step": 6072 }, { "epoch": 2.7516991391028545, "grad_norm": 0.32424545772711594, "learning_rate": 3.995571212550192e-05, "loss": 0.9483, "step": 6073 }, { "epoch": 2.752152242863616, "grad_norm": 0.24452891902977142, "learning_rate": 3.994305845467159e-05, "loss": 0.9326, "step": 6074 }, { "epoch": 2.752605346624377, "grad_norm": 0.18886141646404814, "learning_rate": 3.993040478953953e-05, "loss": 0.9322, "step": 6075 }, { "epoch": 2.7530584503851383, "grad_norm": 0.22345677063861552, "learning_rate": 3.991775113137202e-05, "loss": 0.9706, "step": 6076 }, { "epoch": 2.7535115541458994, "grad_norm": 0.3036779589781224, "learning_rate": 3.990509748143534e-05, "loss": 0.9616, "step": 6077 }, { "epoch": 2.7539646579066606, "grad_norm": 0.3221381625918407, "learning_rate": 3.989244384099578e-05, "loss": 0.9531, "step": 6078 }, { "epoch": 2.7544177616674217, "grad_norm": 0.28554452532627406, "learning_rate": 3.9879790211319606e-05, "loss": 0.9437, "step": 6079 }, { "epoch": 2.754870865428183, "grad_norm": 0.2394896896885771, "learning_rate": 3.9867136593673096e-05, "loss": 0.9263, "step": 6080 }, { "epoch": 2.7553239691889444, "grad_norm": 0.17281986896576595, "learning_rate": 3.985448298932253e-05, "loss": 0.9134, "step": 6081 }, { "epoch": 2.7557770729497055, "grad_norm": 0.22477364089064536, "learning_rate": 3.984182939953418e-05, "loss": 0.962, "step": 6082 }, { "epoch": 2.7562301767104667, "grad_norm": 0.33734591527304064, "learning_rate": 3.982917582557432e-05, "loss": 0.928, "step": 6083 }, { "epoch": 2.756683280471228, "grad_norm": 0.3244199260640858, "learning_rate": 3.981652226870923e-05, "loss": 0.9364, "step": 6084 }, { "epoch": 2.7571363842319894, "grad_norm": 0.24508147034339708, "learning_rate": 3.9803868730205175e-05, "loss": 0.9427, "step": 6085 }, { "epoch": 2.7575894879927505, "grad_norm": 0.22724853298393952, "learning_rate": 3.9791215211328426e-05, "loss": 0.9563, "step": 6086 }, { "epoch": 2.7580425917535116, "grad_norm": 0.22089075288626117, "learning_rate": 3.977856171334524e-05, "loss": 0.9284, "step": 6087 }, { "epoch": 2.7584956955142728, "grad_norm": 0.2540047347040992, "learning_rate": 3.976590823752188e-05, "loss": 0.9271, "step": 6088 }, { "epoch": 2.758948799275034, "grad_norm": 0.30560385875156076, "learning_rate": 3.975325478512463e-05, "loss": 0.952, "step": 6089 }, { "epoch": 2.759401903035795, "grad_norm": 0.32518279351403473, "learning_rate": 3.974060135741972e-05, "loss": 0.9337, "step": 6090 }, { "epoch": 2.759855006796556, "grad_norm": 0.26149430578813265, "learning_rate": 3.972794795567343e-05, "loss": 0.9475, "step": 6091 }, { "epoch": 2.7603081105573177, "grad_norm": 0.15282926871577066, "learning_rate": 3.9715294581152015e-05, "loss": 0.9188, "step": 6092 }, { "epoch": 2.760761214318079, "grad_norm": 0.23322881906168982, "learning_rate": 3.970264123512171e-05, "loss": 0.9397, "step": 6093 }, { "epoch": 2.76121431807884, "grad_norm": 0.3524443774819629, "learning_rate": 3.968998791884878e-05, "loss": 0.9341, "step": 6094 }, { "epoch": 2.761667421839601, "grad_norm": 0.39924002052605406, "learning_rate": 3.967733463359946e-05, "loss": 0.9536, "step": 6095 }, { "epoch": 2.7621205256003627, "grad_norm": 0.33123903017086, "learning_rate": 3.966468138064e-05, "loss": 0.9201, "step": 6096 }, { "epoch": 2.762573629361124, "grad_norm": 0.23749002711076572, "learning_rate": 3.965202816123664e-05, "loss": 0.9387, "step": 6097 }, { "epoch": 2.763026733121885, "grad_norm": 0.17499825298654736, "learning_rate": 3.9639374976655606e-05, "loss": 0.9493, "step": 6098 }, { "epoch": 2.763479836882646, "grad_norm": 0.1935287445756804, "learning_rate": 3.9626721828163156e-05, "loss": 0.9257, "step": 6099 }, { "epoch": 2.763932940643407, "grad_norm": 0.22859730300470324, "learning_rate": 3.961406871702549e-05, "loss": 0.9406, "step": 6100 }, { "epoch": 2.7643860444041684, "grad_norm": 0.23144452180453773, "learning_rate": 3.9601415644508844e-05, "loss": 0.9205, "step": 6101 }, { "epoch": 2.76483914816493, "grad_norm": 0.21627175499374335, "learning_rate": 3.958876261187944e-05, "loss": 0.9326, "step": 6102 }, { "epoch": 2.765292251925691, "grad_norm": 0.20186995053488013, "learning_rate": 3.95761096204035e-05, "loss": 0.9411, "step": 6103 }, { "epoch": 2.765745355686452, "grad_norm": 0.18002994740181671, "learning_rate": 3.9563456671347236e-05, "loss": 0.9402, "step": 6104 }, { "epoch": 2.7661984594472133, "grad_norm": 0.162925550427835, "learning_rate": 3.955080376597686e-05, "loss": 0.9408, "step": 6105 }, { "epoch": 2.766651563207975, "grad_norm": 0.2317611079368925, "learning_rate": 3.953815090555858e-05, "loss": 0.9367, "step": 6106 }, { "epoch": 2.767104666968736, "grad_norm": 0.19667628416707741, "learning_rate": 3.952549809135857e-05, "loss": 0.9503, "step": 6107 }, { "epoch": 2.767557770729497, "grad_norm": 0.1956761653891428, "learning_rate": 3.951284532464306e-05, "loss": 0.9586, "step": 6108 }, { "epoch": 2.7680108744902583, "grad_norm": 0.20526501870029887, "learning_rate": 3.950019260667822e-05, "loss": 0.9438, "step": 6109 }, { "epoch": 2.7684639782510194, "grad_norm": 0.22878003323790283, "learning_rate": 3.9487539938730245e-05, "loss": 0.9672, "step": 6110 }, { "epoch": 2.7689170820117806, "grad_norm": 0.2305049428838569, "learning_rate": 3.947488732206531e-05, "loss": 0.932, "step": 6111 }, { "epoch": 2.7693701857725417, "grad_norm": 0.19638925250197098, "learning_rate": 3.946223475794961e-05, "loss": 0.9307, "step": 6112 }, { "epoch": 2.7698232895333033, "grad_norm": 0.18957304796059685, "learning_rate": 3.9449582247649286e-05, "loss": 0.9363, "step": 6113 }, { "epoch": 2.7702763932940644, "grad_norm": 0.17130469579350285, "learning_rate": 3.9436929792430535e-05, "loss": 0.9304, "step": 6114 }, { "epoch": 2.7707294970548255, "grad_norm": 0.15072635259442776, "learning_rate": 3.942427739355949e-05, "loss": 0.9514, "step": 6115 }, { "epoch": 2.7711826008155866, "grad_norm": 0.17392484739266462, "learning_rate": 3.941162505230232e-05, "loss": 0.9408, "step": 6116 }, { "epoch": 2.7716357045763482, "grad_norm": 0.2506690269824079, "learning_rate": 3.939897276992517e-05, "loss": 0.9328, "step": 6117 }, { "epoch": 2.7720888083371094, "grad_norm": 0.2783209752200637, "learning_rate": 3.938632054769418e-05, "loss": 0.9569, "step": 6118 }, { "epoch": 2.7725419120978705, "grad_norm": 0.31126834959640676, "learning_rate": 3.937366838687549e-05, "loss": 0.9359, "step": 6119 }, { "epoch": 2.7729950158586316, "grad_norm": 0.23920762330062084, "learning_rate": 3.936101628873524e-05, "loss": 0.9527, "step": 6120 }, { "epoch": 2.7734481196193927, "grad_norm": 0.16711997168164258, "learning_rate": 3.934836425453953e-05, "loss": 0.9344, "step": 6121 }, { "epoch": 2.773901223380154, "grad_norm": 0.2149917680738688, "learning_rate": 3.93357122855545e-05, "loss": 0.9472, "step": 6122 }, { "epoch": 2.774354327140915, "grad_norm": 0.31963003509410304, "learning_rate": 3.932306038304626e-05, "loss": 0.9323, "step": 6123 }, { "epoch": 2.7748074309016766, "grad_norm": 0.36662017544400166, "learning_rate": 3.93104085482809e-05, "loss": 0.9365, "step": 6124 }, { "epoch": 2.7752605346624377, "grad_norm": 0.30086097023325087, "learning_rate": 3.929775678252453e-05, "loss": 0.9367, "step": 6125 }, { "epoch": 2.775713638423199, "grad_norm": 0.21374229177524168, "learning_rate": 3.9285105087043235e-05, "loss": 0.9408, "step": 6126 }, { "epoch": 2.77616674218396, "grad_norm": 0.18741239158514633, "learning_rate": 3.92724534631031e-05, "loss": 0.9448, "step": 6127 }, { "epoch": 2.7766198459447216, "grad_norm": 0.20097686426507597, "learning_rate": 3.92598019119702e-05, "loss": 0.9529, "step": 6128 }, { "epoch": 2.7770729497054827, "grad_norm": 0.26067973088012913, "learning_rate": 3.9247150434910614e-05, "loss": 0.9507, "step": 6129 }, { "epoch": 2.777526053466244, "grad_norm": 0.2767154652550661, "learning_rate": 3.92344990331904e-05, "loss": 0.9427, "step": 6130 }, { "epoch": 2.777979157227005, "grad_norm": 0.255073127780222, "learning_rate": 3.9221847708075606e-05, "loss": 0.9452, "step": 6131 }, { "epoch": 2.778432260987766, "grad_norm": 0.20195492359354975, "learning_rate": 3.920919646083228e-05, "loss": 0.9542, "step": 6132 }, { "epoch": 2.778885364748527, "grad_norm": 0.1680164741155152, "learning_rate": 3.919654529272647e-05, "loss": 0.9401, "step": 6133 }, { "epoch": 2.779338468509289, "grad_norm": 0.1564701341434974, "learning_rate": 3.9183894205024196e-05, "loss": 0.9172, "step": 6134 }, { "epoch": 2.77979157227005, "grad_norm": 0.17699975771984514, "learning_rate": 3.9171243198991485e-05, "loss": 0.9482, "step": 6135 }, { "epoch": 2.780244676030811, "grad_norm": 0.1983306578961993, "learning_rate": 3.915859227589436e-05, "loss": 0.9526, "step": 6136 }, { "epoch": 2.780697779791572, "grad_norm": 0.20160908110098846, "learning_rate": 3.914594143699881e-05, "loss": 0.9285, "step": 6137 }, { "epoch": 2.7811508835523338, "grad_norm": 0.17339869824085535, "learning_rate": 3.9133290683570855e-05, "loss": 0.9473, "step": 6138 }, { "epoch": 2.781603987313095, "grad_norm": 0.1497245449829219, "learning_rate": 3.912064001687648e-05, "loss": 0.926, "step": 6139 }, { "epoch": 2.782057091073856, "grad_norm": 0.1702565382544993, "learning_rate": 3.910798943818165e-05, "loss": 0.9269, "step": 6140 }, { "epoch": 2.782510194834617, "grad_norm": 0.17549245777825107, "learning_rate": 3.909533894875235e-05, "loss": 0.9382, "step": 6141 }, { "epoch": 2.7829632985953783, "grad_norm": 0.2012329431996123, "learning_rate": 3.908268854985453e-05, "loss": 0.9274, "step": 6142 }, { "epoch": 2.7834164023561394, "grad_norm": 0.17604629252597864, "learning_rate": 3.907003824275417e-05, "loss": 0.9348, "step": 6143 }, { "epoch": 2.7838695061169005, "grad_norm": 0.16807903379119885, "learning_rate": 3.905738802871719e-05, "loss": 0.9465, "step": 6144 }, { "epoch": 2.784322609877662, "grad_norm": 0.2398256344957672, "learning_rate": 3.9044737909009545e-05, "loss": 0.9543, "step": 6145 }, { "epoch": 2.7847757136384232, "grad_norm": 0.24779549195127154, "learning_rate": 3.903208788489714e-05, "loss": 0.9549, "step": 6146 }, { "epoch": 2.7852288173991844, "grad_norm": 0.20869902137476917, "learning_rate": 3.901943795764592e-05, "loss": 0.9481, "step": 6147 }, { "epoch": 2.7856819211599455, "grad_norm": 0.19325401954478733, "learning_rate": 3.900678812852176e-05, "loss": 0.9419, "step": 6148 }, { "epoch": 2.786135024920707, "grad_norm": 0.2300179946839362, "learning_rate": 3.899413839879058e-05, "loss": 0.9368, "step": 6149 }, { "epoch": 2.786588128681468, "grad_norm": 0.2391054449367105, "learning_rate": 3.898148876971826e-05, "loss": 0.9352, "step": 6150 }, { "epoch": 2.7870412324422293, "grad_norm": 0.21527493066668965, "learning_rate": 3.896883924257067e-05, "loss": 0.9398, "step": 6151 }, { "epoch": 2.7874943362029905, "grad_norm": 0.19396716241106982, "learning_rate": 3.8956189818613685e-05, "loss": 0.9541, "step": 6152 }, { "epoch": 2.7879474399637516, "grad_norm": 0.1981289199079839, "learning_rate": 3.894354049911317e-05, "loss": 0.936, "step": 6153 }, { "epoch": 2.7884005437245127, "grad_norm": 0.17570373621205385, "learning_rate": 3.893089128533495e-05, "loss": 0.9447, "step": 6154 }, { "epoch": 2.788853647485274, "grad_norm": 0.18544653304300107, "learning_rate": 3.8918242178544894e-05, "loss": 0.9554, "step": 6155 }, { "epoch": 2.7893067512460354, "grad_norm": 0.24935406827854398, "learning_rate": 3.8905593180008796e-05, "loss": 0.9267, "step": 6156 }, { "epoch": 2.7897598550067966, "grad_norm": 0.282545403299615, "learning_rate": 3.889294429099247e-05, "loss": 0.9421, "step": 6157 }, { "epoch": 2.7902129587675577, "grad_norm": 0.21872494283450855, "learning_rate": 3.8880295512761736e-05, "loss": 0.9454, "step": 6158 }, { "epoch": 2.790666062528319, "grad_norm": 0.1851226236358803, "learning_rate": 3.886764684658238e-05, "loss": 0.9466, "step": 6159 }, { "epoch": 2.7911191662890804, "grad_norm": 0.20624226998694054, "learning_rate": 3.8854998293720175e-05, "loss": 0.9163, "step": 6160 }, { "epoch": 2.7915722700498415, "grad_norm": 0.19047800798126022, "learning_rate": 3.88423498554409e-05, "loss": 0.9477, "step": 6161 }, { "epoch": 2.7920253738106027, "grad_norm": 0.19298472227262808, "learning_rate": 3.882970153301031e-05, "loss": 0.9505, "step": 6162 }, { "epoch": 2.792478477571364, "grad_norm": 0.18656634014691448, "learning_rate": 3.8817053327694146e-05, "loss": 0.9617, "step": 6163 }, { "epoch": 2.792931581332125, "grad_norm": 0.18310231709046457, "learning_rate": 3.880440524075815e-05, "loss": 0.936, "step": 6164 }, { "epoch": 2.793384685092886, "grad_norm": 0.1419909108673105, "learning_rate": 3.879175727346804e-05, "loss": 0.942, "step": 6165 }, { "epoch": 2.7938377888536476, "grad_norm": 0.1783450002698297, "learning_rate": 3.877910942708954e-05, "loss": 0.9433, "step": 6166 }, { "epoch": 2.7942908926144088, "grad_norm": 0.18743067840206515, "learning_rate": 3.876646170288834e-05, "loss": 0.9513, "step": 6167 }, { "epoch": 2.79474399637517, "grad_norm": 0.18061433885026462, "learning_rate": 3.8753814102130117e-05, "loss": 0.9616, "step": 6168 }, { "epoch": 2.795197100135931, "grad_norm": 0.15463077241303125, "learning_rate": 3.8741166626080555e-05, "loss": 0.9492, "step": 6169 }, { "epoch": 2.7956502038966926, "grad_norm": 0.19250657778959432, "learning_rate": 3.872851927600531e-05, "loss": 0.9494, "step": 6170 }, { "epoch": 2.7961033076574537, "grad_norm": 0.20630394477412686, "learning_rate": 3.8715872053170043e-05, "loss": 0.9654, "step": 6171 }, { "epoch": 2.796556411418215, "grad_norm": 0.198916914105253, "learning_rate": 3.870322495884037e-05, "loss": 0.9361, "step": 6172 }, { "epoch": 2.797009515178976, "grad_norm": 0.2614115648725533, "learning_rate": 3.8690577994281936e-05, "loss": 0.9525, "step": 6173 }, { "epoch": 2.797462618939737, "grad_norm": 0.20693165458384807, "learning_rate": 3.8677931160760344e-05, "loss": 0.9359, "step": 6174 }, { "epoch": 2.7979157227004983, "grad_norm": 0.17740763433243073, "learning_rate": 3.86652844595412e-05, "loss": 0.9544, "step": 6175 }, { "epoch": 2.7983688264612594, "grad_norm": 0.2017246398634792, "learning_rate": 3.865263789189006e-05, "loss": 0.9347, "step": 6176 }, { "epoch": 2.798821930222021, "grad_norm": 0.20712753887932886, "learning_rate": 3.863999145907252e-05, "loss": 0.9508, "step": 6177 }, { "epoch": 2.799275033982782, "grad_norm": 0.19703090887368882, "learning_rate": 3.862734516235413e-05, "loss": 0.9573, "step": 6178 }, { "epoch": 2.7997281377435432, "grad_norm": 0.1762872029757361, "learning_rate": 3.861469900300043e-05, "loss": 0.9459, "step": 6179 }, { "epoch": 2.8001812415043044, "grad_norm": 0.14674089266115092, "learning_rate": 3.860205298227696e-05, "loss": 0.9442, "step": 6180 }, { "epoch": 2.800634345265066, "grad_norm": 0.1587655863344216, "learning_rate": 3.8589407101449224e-05, "loss": 0.9362, "step": 6181 }, { "epoch": 2.801087449025827, "grad_norm": 0.17427830551680365, "learning_rate": 3.8576761361782733e-05, "loss": 0.9491, "step": 6182 }, { "epoch": 2.801540552786588, "grad_norm": 0.19453450522596977, "learning_rate": 3.856411576454297e-05, "loss": 0.9357, "step": 6183 }, { "epoch": 2.8019936565473493, "grad_norm": 0.15265529873308728, "learning_rate": 3.855147031099543e-05, "loss": 0.9396, "step": 6184 }, { "epoch": 2.8024467603081105, "grad_norm": 0.17319590369660506, "learning_rate": 3.8538825002405526e-05, "loss": 0.9515, "step": 6185 }, { "epoch": 2.8028998640688716, "grad_norm": 0.19529505134984065, "learning_rate": 3.8526179840038745e-05, "loss": 0.9226, "step": 6186 }, { "epoch": 2.8033529678296327, "grad_norm": 0.1820396472780763, "learning_rate": 3.851353482516049e-05, "loss": 0.938, "step": 6187 }, { "epoch": 2.8038060715903943, "grad_norm": 0.15333676759025155, "learning_rate": 3.85008899590362e-05, "loss": 0.9195, "step": 6188 }, { "epoch": 2.8042591753511554, "grad_norm": 0.188351190513239, "learning_rate": 3.848824524293126e-05, "loss": 0.9492, "step": 6189 }, { "epoch": 2.8047122791119166, "grad_norm": 0.2066526743762348, "learning_rate": 3.847560067811106e-05, "loss": 0.9603, "step": 6190 }, { "epoch": 2.8051653828726777, "grad_norm": 0.20451362530160686, "learning_rate": 3.8462956265840966e-05, "loss": 0.9504, "step": 6191 }, { "epoch": 2.8056184866334393, "grad_norm": 0.20390587069485028, "learning_rate": 3.8450312007386344e-05, "loss": 0.9562, "step": 6192 }, { "epoch": 2.8060715903942004, "grad_norm": 0.1801876707175782, "learning_rate": 3.8437667904012524e-05, "loss": 0.9359, "step": 6193 }, { "epoch": 2.8065246941549615, "grad_norm": 0.20652441879740577, "learning_rate": 3.842502395698483e-05, "loss": 0.9424, "step": 6194 }, { "epoch": 2.8069777979157227, "grad_norm": 0.25070804010957704, "learning_rate": 3.841238016756858e-05, "loss": 0.948, "step": 6195 }, { "epoch": 2.807430901676484, "grad_norm": 0.2691447760031136, "learning_rate": 3.839973653702905e-05, "loss": 0.9346, "step": 6196 }, { "epoch": 2.807884005437245, "grad_norm": 0.1883205089253929, "learning_rate": 3.838709306663153e-05, "loss": 0.9618, "step": 6197 }, { "epoch": 2.8083371091980065, "grad_norm": 0.1828015506598667, "learning_rate": 3.837444975764128e-05, "loss": 0.9356, "step": 6198 }, { "epoch": 2.8087902129587676, "grad_norm": 0.18839330166927407, "learning_rate": 3.8361806611323545e-05, "loss": 0.9233, "step": 6199 }, { "epoch": 2.8092433167195288, "grad_norm": 0.20139111494339776, "learning_rate": 3.834916362894356e-05, "loss": 0.9503, "step": 6200 }, { "epoch": 2.80969642048029, "grad_norm": 0.21215688730769952, "learning_rate": 3.8336520811766516e-05, "loss": 0.9632, "step": 6201 }, { "epoch": 2.8101495242410515, "grad_norm": 0.19865377953629546, "learning_rate": 3.8323878161057623e-05, "loss": 0.9349, "step": 6202 }, { "epoch": 2.8106026280018126, "grad_norm": 0.21679908462459702, "learning_rate": 3.831123567808206e-05, "loss": 0.9228, "step": 6203 }, { "epoch": 2.8110557317625737, "grad_norm": 0.19896503361941348, "learning_rate": 3.829859336410498e-05, "loss": 0.9561, "step": 6204 }, { "epoch": 2.811508835523335, "grad_norm": 0.17508914851902083, "learning_rate": 3.828595122039154e-05, "loss": 0.9397, "step": 6205 }, { "epoch": 2.811961939284096, "grad_norm": 0.2349638494501595, "learning_rate": 3.8273309248206854e-05, "loss": 0.9469, "step": 6206 }, { "epoch": 2.812415043044857, "grad_norm": 0.29710481735263555, "learning_rate": 3.8260667448816054e-05, "loss": 0.9578, "step": 6207 }, { "epoch": 2.8128681468056183, "grad_norm": 0.260656187092394, "learning_rate": 3.8248025823484224e-05, "loss": 0.9393, "step": 6208 }, { "epoch": 2.81332125056638, "grad_norm": 0.2574133719990056, "learning_rate": 3.823538437347643e-05, "loss": 0.9367, "step": 6209 }, { "epoch": 2.813774354327141, "grad_norm": 0.2215219220014139, "learning_rate": 3.822274310005773e-05, "loss": 0.9339, "step": 6210 }, { "epoch": 2.814227458087902, "grad_norm": 0.1790547930069395, "learning_rate": 3.8210102004493185e-05, "loss": 0.9583, "step": 6211 }, { "epoch": 2.814680561848663, "grad_norm": 0.2150618861527549, "learning_rate": 3.81974610880478e-05, "loss": 0.9644, "step": 6212 }, { "epoch": 2.815133665609425, "grad_norm": 0.3060861099024422, "learning_rate": 3.818482035198661e-05, "loss": 0.9306, "step": 6213 }, { "epoch": 2.815586769370186, "grad_norm": 0.3293594862519032, "learning_rate": 3.817217979757456e-05, "loss": 0.9377, "step": 6214 }, { "epoch": 2.816039873130947, "grad_norm": 0.32651110274180684, "learning_rate": 3.815953942607664e-05, "loss": 0.9469, "step": 6215 }, { "epoch": 2.816492976891708, "grad_norm": 0.25969616618564856, "learning_rate": 3.8146899238757794e-05, "loss": 0.9443, "step": 6216 }, { "epoch": 2.8169460806524693, "grad_norm": 0.2008599214822308, "learning_rate": 3.8134259236882955e-05, "loss": 0.9457, "step": 6217 }, { "epoch": 2.8173991844132305, "grad_norm": 0.15939906615108024, "learning_rate": 3.812161942171705e-05, "loss": 0.9518, "step": 6218 }, { "epoch": 2.8178522881739916, "grad_norm": 0.15301383632737237, "learning_rate": 3.810897979452496e-05, "loss": 0.9247, "step": 6219 }, { "epoch": 2.818305391934753, "grad_norm": 0.18541086096065432, "learning_rate": 3.8096340356571576e-05, "loss": 0.9457, "step": 6220 }, { "epoch": 2.8187584956955143, "grad_norm": 0.19651468156935056, "learning_rate": 3.8083701109121735e-05, "loss": 0.9249, "step": 6221 }, { "epoch": 2.8192115994562754, "grad_norm": 0.18331229011002909, "learning_rate": 3.807106205344029e-05, "loss": 0.9275, "step": 6222 }, { "epoch": 2.8196647032170365, "grad_norm": 0.14956432143849518, "learning_rate": 3.8058423190792055e-05, "loss": 0.9296, "step": 6223 }, { "epoch": 2.820117806977798, "grad_norm": 0.15485505738364952, "learning_rate": 3.804578452244183e-05, "loss": 0.9466, "step": 6224 }, { "epoch": 2.8205709107385593, "grad_norm": 0.15911113912473618, "learning_rate": 3.803314604965439e-05, "loss": 0.9283, "step": 6225 }, { "epoch": 2.8210240144993204, "grad_norm": 0.17373106604463232, "learning_rate": 3.802050777369451e-05, "loss": 0.9281, "step": 6226 }, { "epoch": 2.8214771182600815, "grad_norm": 0.17342554241303865, "learning_rate": 3.8007869695826915e-05, "loss": 0.9334, "step": 6227 }, { "epoch": 2.8219302220208426, "grad_norm": 0.1773988665615082, "learning_rate": 3.799523181731635e-05, "loss": 0.9476, "step": 6228 }, { "epoch": 2.822383325781604, "grad_norm": 0.19760653292236913, "learning_rate": 3.798259413942749e-05, "loss": 0.9476, "step": 6229 }, { "epoch": 2.8228364295423654, "grad_norm": 0.19013628969657995, "learning_rate": 3.796995666342503e-05, "loss": 0.9317, "step": 6230 }, { "epoch": 2.8232895333031265, "grad_norm": 0.1607698703137353, "learning_rate": 3.7957319390573616e-05, "loss": 0.9341, "step": 6231 }, { "epoch": 2.8237426370638876, "grad_norm": 0.18643202722886473, "learning_rate": 3.7944682322137906e-05, "loss": 0.9442, "step": 6232 }, { "epoch": 2.8241957408246487, "grad_norm": 0.2310188985994892, "learning_rate": 3.793204545938252e-05, "loss": 0.9287, "step": 6233 }, { "epoch": 2.8246488445854103, "grad_norm": 0.24346939209184404, "learning_rate": 3.791940880357204e-05, "loss": 0.9355, "step": 6234 }, { "epoch": 2.8251019483461715, "grad_norm": 0.24218562945879227, "learning_rate": 3.790677235597107e-05, "loss": 0.9424, "step": 6235 }, { "epoch": 2.8255550521069326, "grad_norm": 0.2923435948595009, "learning_rate": 3.7894136117844155e-05, "loss": 0.9535, "step": 6236 }, { "epoch": 2.8260081558676937, "grad_norm": 0.32762345543108673, "learning_rate": 3.788150009045582e-05, "loss": 0.9697, "step": 6237 }, { "epoch": 2.826461259628455, "grad_norm": 0.15267345566518062, "learning_rate": 3.78688642750706e-05, "loss": 0.9617, "step": 6238 }, { "epoch": 2.826914363389216, "grad_norm": 0.2641076293377013, "learning_rate": 3.785622867295299e-05, "loss": 0.9369, "step": 6239 }, { "epoch": 2.827367467149977, "grad_norm": 0.3036585758094229, "learning_rate": 3.784359328536744e-05, "loss": 0.9246, "step": 6240 }, { "epoch": 2.8278205709107387, "grad_norm": 0.35446863334855805, "learning_rate": 3.783095811357843e-05, "loss": 0.9455, "step": 6241 }, { "epoch": 2.8282736746715, "grad_norm": 0.32199646371293056, "learning_rate": 3.781832315885037e-05, "loss": 0.9434, "step": 6242 }, { "epoch": 2.828726778432261, "grad_norm": 0.20785880313169827, "learning_rate": 3.7805688422447684e-05, "loss": 0.9497, "step": 6243 }, { "epoch": 2.829179882193022, "grad_norm": 0.19079570665283763, "learning_rate": 3.779305390563475e-05, "loss": 0.9405, "step": 6244 }, { "epoch": 2.8296329859537837, "grad_norm": 0.27112901003594936, "learning_rate": 3.7780419609675945e-05, "loss": 0.9446, "step": 6245 }, { "epoch": 2.830086089714545, "grad_norm": 0.3133534921928195, "learning_rate": 3.776778553583559e-05, "loss": 0.9491, "step": 6246 }, { "epoch": 2.830539193475306, "grad_norm": 0.3301054781920361, "learning_rate": 3.775515168537802e-05, "loss": 0.9432, "step": 6247 }, { "epoch": 2.830992297236067, "grad_norm": 0.21073505292633624, "learning_rate": 3.774251805956753e-05, "loss": 0.9339, "step": 6248 }, { "epoch": 2.831445400996828, "grad_norm": 0.17736619350959226, "learning_rate": 3.7729884659668394e-05, "loss": 0.9218, "step": 6249 }, { "epoch": 2.8318985047575893, "grad_norm": 0.2361205398252794, "learning_rate": 3.771725148694487e-05, "loss": 0.9337, "step": 6250 }, { "epoch": 2.8323516085183504, "grad_norm": 0.2716996418045679, "learning_rate": 3.770461854266118e-05, "loss": 0.9314, "step": 6251 }, { "epoch": 2.832804712279112, "grad_norm": 0.337725103511505, "learning_rate": 3.7691985828081544e-05, "loss": 0.9334, "step": 6252 }, { "epoch": 2.833257816039873, "grad_norm": 0.30052730646261344, "learning_rate": 3.767935334447014e-05, "loss": 0.9264, "step": 6253 }, { "epoch": 2.8337109198006343, "grad_norm": 0.2243999058463976, "learning_rate": 3.766672109309113e-05, "loss": 0.9454, "step": 6254 }, { "epoch": 2.8341640235613954, "grad_norm": 0.17390768147219662, "learning_rate": 3.765408907520865e-05, "loss": 0.9484, "step": 6255 }, { "epoch": 2.834617127322157, "grad_norm": 0.21976545803709754, "learning_rate": 3.764145729208681e-05, "loss": 0.9572, "step": 6256 }, { "epoch": 2.835070231082918, "grad_norm": 0.2782520427843211, "learning_rate": 3.762882574498971e-05, "loss": 0.9212, "step": 6257 }, { "epoch": 2.8355233348436792, "grad_norm": 0.269704899193921, "learning_rate": 3.7616194435181416e-05, "loss": 0.9618, "step": 6258 }, { "epoch": 2.8359764386044404, "grad_norm": 0.23607592103081027, "learning_rate": 3.760356336392597e-05, "loss": 0.9229, "step": 6259 }, { "epoch": 2.8364295423652015, "grad_norm": 0.18755153498097327, "learning_rate": 3.759093253248739e-05, "loss": 0.954, "step": 6260 }, { "epoch": 2.8368826461259626, "grad_norm": 0.1430366995874861, "learning_rate": 3.757830194212969e-05, "loss": 0.9618, "step": 6261 }, { "epoch": 2.837335749886724, "grad_norm": 0.23027000555050384, "learning_rate": 3.7565671594116816e-05, "loss": 0.9461, "step": 6262 }, { "epoch": 2.8377888536474853, "grad_norm": 0.2935689657443989, "learning_rate": 3.7553041489712725e-05, "loss": 0.9254, "step": 6263 }, { "epoch": 2.8382419574082465, "grad_norm": 0.30482761205458736, "learning_rate": 3.754041163018134e-05, "loss": 0.9371, "step": 6264 }, { "epoch": 2.8386950611690076, "grad_norm": 0.2698351240150186, "learning_rate": 3.752778201678657e-05, "loss": 0.9125, "step": 6265 }, { "epoch": 2.839148164929769, "grad_norm": 0.2190236056195924, "learning_rate": 3.751515265079227e-05, "loss": 0.9419, "step": 6266 }, { "epoch": 2.8396012686905303, "grad_norm": 0.2163828721097432, "learning_rate": 3.75025235334623e-05, "loss": 0.9487, "step": 6267 }, { "epoch": 2.8400543724512914, "grad_norm": 0.16843395746476958, "learning_rate": 3.7489894666060494e-05, "loss": 0.9276, "step": 6268 }, { "epoch": 2.8405074762120526, "grad_norm": 0.1921443699172697, "learning_rate": 3.747726604985064e-05, "loss": 0.9342, "step": 6269 }, { "epoch": 2.8409605799728137, "grad_norm": 0.22536203207759745, "learning_rate": 3.746463768609652e-05, "loss": 0.9231, "step": 6270 }, { "epoch": 2.841413683733575, "grad_norm": 0.16245532029285625, "learning_rate": 3.7452009576061874e-05, "loss": 0.9228, "step": 6271 }, { "epoch": 2.841866787494336, "grad_norm": 0.1792647782556479, "learning_rate": 3.7439381721010424e-05, "loss": 0.9463, "step": 6272 }, { "epoch": 2.8423198912550975, "grad_norm": 0.19173645552540197, "learning_rate": 3.742675412220588e-05, "loss": 0.926, "step": 6273 }, { "epoch": 2.8427729950158587, "grad_norm": 0.1453411974126866, "learning_rate": 3.7414126780911905e-05, "loss": 0.9249, "step": 6274 }, { "epoch": 2.84322609877662, "grad_norm": 0.15163530811511003, "learning_rate": 3.7401499698392145e-05, "loss": 0.9393, "step": 6275 }, { "epoch": 2.843679202537381, "grad_norm": 0.16668898655719674, "learning_rate": 3.7388872875910225e-05, "loss": 0.9612, "step": 6276 }, { "epoch": 2.8441323062981425, "grad_norm": 0.1927697538464099, "learning_rate": 3.737624631472974e-05, "loss": 0.936, "step": 6277 }, { "epoch": 2.8445854100589036, "grad_norm": 0.18634052234140194, "learning_rate": 3.736362001611426e-05, "loss": 0.9521, "step": 6278 }, { "epoch": 2.8450385138196648, "grad_norm": 0.1636806038576799, "learning_rate": 3.735099398132733e-05, "loss": 0.9469, "step": 6279 }, { "epoch": 2.845491617580426, "grad_norm": 0.1890227312487151, "learning_rate": 3.733836821163245e-05, "loss": 0.9347, "step": 6280 }, { "epoch": 2.845944721341187, "grad_norm": 0.17267224516396062, "learning_rate": 3.7325742708293134e-05, "loss": 0.9609, "step": 6281 }, { "epoch": 2.846397825101948, "grad_norm": 0.18969146293003547, "learning_rate": 3.731311747257283e-05, "loss": 0.9568, "step": 6282 }, { "epoch": 2.8468509288627093, "grad_norm": 0.20322627860466685, "learning_rate": 3.730049250573497e-05, "loss": 0.9331, "step": 6283 }, { "epoch": 2.847304032623471, "grad_norm": 0.16720976381971647, "learning_rate": 3.728786780904297e-05, "loss": 0.9571, "step": 6284 }, { "epoch": 2.847757136384232, "grad_norm": 0.18585351549246915, "learning_rate": 3.727524338376021e-05, "loss": 0.9284, "step": 6285 }, { "epoch": 2.848210240144993, "grad_norm": 0.2090494488018941, "learning_rate": 3.7262619231150057e-05, "loss": 0.945, "step": 6286 }, { "epoch": 2.8486633439057543, "grad_norm": 0.1415570682385262, "learning_rate": 3.724999535247582e-05, "loss": 0.9528, "step": 6287 }, { "epoch": 2.849116447666516, "grad_norm": 0.2172744222503447, "learning_rate": 3.723737174900081e-05, "loss": 0.9624, "step": 6288 }, { "epoch": 2.849569551427277, "grad_norm": 0.22136210129519512, "learning_rate": 3.72247484219883e-05, "loss": 0.936, "step": 6289 }, { "epoch": 2.850022655188038, "grad_norm": 0.2322639521595417, "learning_rate": 3.721212537270153e-05, "loss": 0.947, "step": 6290 }, { "epoch": 2.8504757589487992, "grad_norm": 0.1664527752211536, "learning_rate": 3.719950260240372e-05, "loss": 0.9472, "step": 6291 }, { "epoch": 2.8509288627095604, "grad_norm": 0.20499263033298723, "learning_rate": 3.718688011235806e-05, "loss": 0.9451, "step": 6292 }, { "epoch": 2.8513819664703215, "grad_norm": 0.18212418775021497, "learning_rate": 3.717425790382771e-05, "loss": 0.9329, "step": 6293 }, { "epoch": 2.851835070231083, "grad_norm": 0.19053268324319306, "learning_rate": 3.71616359780758e-05, "loss": 0.9387, "step": 6294 }, { "epoch": 2.852288173991844, "grad_norm": 0.20602266609353315, "learning_rate": 3.7149014336365434e-05, "loss": 0.9632, "step": 6295 }, { "epoch": 2.8527412777526053, "grad_norm": 0.19463597477311803, "learning_rate": 3.7136392979959695e-05, "loss": 0.9194, "step": 6296 }, { "epoch": 2.8531943815133665, "grad_norm": 0.17866593876060233, "learning_rate": 3.712377191012164e-05, "loss": 0.9476, "step": 6297 }, { "epoch": 2.853647485274128, "grad_norm": 0.18140947889467335, "learning_rate": 3.7111151128114266e-05, "loss": 0.9421, "step": 6298 }, { "epoch": 2.854100589034889, "grad_norm": 0.16369248516972396, "learning_rate": 3.709853063520058e-05, "loss": 0.9234, "step": 6299 }, { "epoch": 2.8545536927956503, "grad_norm": 0.17885043939913367, "learning_rate": 3.708591043264353e-05, "loss": 0.9425, "step": 6300 }, { "epoch": 2.8550067965564114, "grad_norm": 0.19053093995745304, "learning_rate": 3.707329052170606e-05, "loss": 0.9482, "step": 6301 }, { "epoch": 2.8554599003171726, "grad_norm": 0.19406431935655535, "learning_rate": 3.706067090365107e-05, "loss": 0.9492, "step": 6302 }, { "epoch": 2.8559130040779337, "grad_norm": 0.1878936568752391, "learning_rate": 3.704805157974143e-05, "loss": 0.9356, "step": 6303 }, { "epoch": 2.856366107838695, "grad_norm": 0.17626448708087272, "learning_rate": 3.703543255123999e-05, "loss": 0.9412, "step": 6304 }, { "epoch": 2.8568192115994564, "grad_norm": 0.1812515053459801, "learning_rate": 3.702281381940957e-05, "loss": 0.9304, "step": 6305 }, { "epoch": 2.8572723153602175, "grad_norm": 0.17977843827528392, "learning_rate": 3.701019538551294e-05, "loss": 0.9331, "step": 6306 }, { "epoch": 2.8577254191209787, "grad_norm": 0.19471044534758936, "learning_rate": 3.699757725081286e-05, "loss": 0.9373, "step": 6307 }, { "epoch": 2.85817852288174, "grad_norm": 0.1826072932753648, "learning_rate": 3.6984959416572064e-05, "loss": 0.9452, "step": 6308 }, { "epoch": 2.8586316266425014, "grad_norm": 0.15599213900015305, "learning_rate": 3.697234188405324e-05, "loss": 0.9358, "step": 6309 }, { "epoch": 2.8590847304032625, "grad_norm": 0.15308393902209633, "learning_rate": 3.6959724654519045e-05, "loss": 0.9561, "step": 6310 }, { "epoch": 2.8595378341640236, "grad_norm": 0.18523900222973194, "learning_rate": 3.6947107729232135e-05, "loss": 0.928, "step": 6311 }, { "epoch": 2.8599909379247848, "grad_norm": 0.158837930074131, "learning_rate": 3.69344911094551e-05, "loss": 0.9564, "step": 6312 }, { "epoch": 2.860444041685546, "grad_norm": 0.21435595990988152, "learning_rate": 3.6921874796450523e-05, "loss": 0.9336, "step": 6313 }, { "epoch": 2.860897145446307, "grad_norm": 0.2520186466268998, "learning_rate": 3.690925879148094e-05, "loss": 0.963, "step": 6314 }, { "epoch": 2.861350249207068, "grad_norm": 0.18591563758983978, "learning_rate": 3.6896643095808865e-05, "loss": 0.9534, "step": 6315 }, { "epoch": 2.8618033529678297, "grad_norm": 0.21644384778915626, "learning_rate": 3.6884027710696776e-05, "loss": 0.9535, "step": 6316 }, { "epoch": 2.862256456728591, "grad_norm": 0.2556637193945497, "learning_rate": 3.687141263740712e-05, "loss": 0.9434, "step": 6317 }, { "epoch": 2.862709560489352, "grad_norm": 0.18865720374923112, "learning_rate": 3.6858797877202335e-05, "loss": 0.9536, "step": 6318 }, { "epoch": 2.863162664250113, "grad_norm": 0.18221154116385427, "learning_rate": 3.6846183431344786e-05, "loss": 0.937, "step": 6319 }, { "epoch": 2.8636157680108747, "grad_norm": 0.20007861184309253, "learning_rate": 3.683356930109684e-05, "loss": 0.9283, "step": 6320 }, { "epoch": 2.864068871771636, "grad_norm": 0.18531816231500658, "learning_rate": 3.682095548772083e-05, "loss": 0.9663, "step": 6321 }, { "epoch": 2.864521975532397, "grad_norm": 0.1741009588719043, "learning_rate": 3.6808341992479045e-05, "loss": 0.9583, "step": 6322 }, { "epoch": 2.864975079293158, "grad_norm": 0.19691964365809805, "learning_rate": 3.679572881663373e-05, "loss": 0.9329, "step": 6323 }, { "epoch": 2.865428183053919, "grad_norm": 0.3562002489605617, "learning_rate": 3.678311596144713e-05, "loss": 0.9378, "step": 6324 }, { "epoch": 2.8658812868146804, "grad_norm": 0.16523347797008356, "learning_rate": 3.677050342818143e-05, "loss": 0.9201, "step": 6325 }, { "epoch": 2.866334390575442, "grad_norm": 0.17211142209306204, "learning_rate": 3.6757891218098805e-05, "loss": 0.9247, "step": 6326 }, { "epoch": 2.866787494336203, "grad_norm": 0.20342489880205347, "learning_rate": 3.674527933246141e-05, "loss": 0.9427, "step": 6327 }, { "epoch": 2.867240598096964, "grad_norm": 0.20882337977175555, "learning_rate": 3.6732667772531295e-05, "loss": 0.9426, "step": 6328 }, { "epoch": 2.8676937018577253, "grad_norm": 0.23220002121259795, "learning_rate": 3.672005653957056e-05, "loss": 0.9152, "step": 6329 }, { "epoch": 2.868146805618487, "grad_norm": 0.25247704377526675, "learning_rate": 3.670744563484123e-05, "loss": 0.9454, "step": 6330 }, { "epoch": 2.868599909379248, "grad_norm": 0.22383304502927207, "learning_rate": 3.6694835059605304e-05, "loss": 0.9381, "step": 6331 }, { "epoch": 2.869053013140009, "grad_norm": 0.18196496253416075, "learning_rate": 3.668222481512477e-05, "loss": 0.9397, "step": 6332 }, { "epoch": 2.8695061169007703, "grad_norm": 0.20076118350227368, "learning_rate": 3.666961490266155e-05, "loss": 0.9553, "step": 6333 }, { "epoch": 2.8699592206615314, "grad_norm": 0.20839551602652487, "learning_rate": 3.665700532347754e-05, "loss": 0.9536, "step": 6334 }, { "epoch": 2.8704123244222925, "grad_norm": 0.28852399017691877, "learning_rate": 3.664439607883462e-05, "loss": 0.9441, "step": 6335 }, { "epoch": 2.8708654281830537, "grad_norm": 0.3013358371176152, "learning_rate": 3.6631787169994615e-05, "loss": 0.944, "step": 6336 }, { "epoch": 2.8713185319438153, "grad_norm": 0.27513332820766034, "learning_rate": 3.661917859821933e-05, "loss": 0.9449, "step": 6337 }, { "epoch": 2.8717716357045764, "grad_norm": 0.23359081389116115, "learning_rate": 3.660657036477055e-05, "loss": 0.9441, "step": 6338 }, { "epoch": 2.8722247394653375, "grad_norm": 0.19777046834373446, "learning_rate": 3.659396247090999e-05, "loss": 0.9644, "step": 6339 }, { "epoch": 2.8726778432260986, "grad_norm": 0.17337564643030842, "learning_rate": 3.6581354917899365e-05, "loss": 0.9364, "step": 6340 }, { "epoch": 2.8731309469868602, "grad_norm": 0.2109067841537543, "learning_rate": 3.656874770700033e-05, "loss": 0.9344, "step": 6341 }, { "epoch": 2.8735840507476214, "grad_norm": 0.2715622644789916, "learning_rate": 3.655614083947453e-05, "loss": 0.9457, "step": 6342 }, { "epoch": 2.8740371545083825, "grad_norm": 0.26823375581737763, "learning_rate": 3.654353431658355e-05, "loss": 0.9366, "step": 6343 }, { "epoch": 2.8744902582691436, "grad_norm": 0.20145107654235192, "learning_rate": 3.653092813958896e-05, "loss": 0.9432, "step": 6344 }, { "epoch": 2.8749433620299047, "grad_norm": 0.13953866636915246, "learning_rate": 3.6518322309752284e-05, "loss": 0.9458, "step": 6345 }, { "epoch": 2.875396465790666, "grad_norm": 0.21776960330016543, "learning_rate": 3.6505716828335016e-05, "loss": 0.9561, "step": 6346 }, { "epoch": 2.875849569551427, "grad_norm": 0.2551687022511646, "learning_rate": 3.649311169659862e-05, "loss": 0.9295, "step": 6347 }, { "epoch": 2.8763026733121886, "grad_norm": 0.22641787773943248, "learning_rate": 3.648050691580452e-05, "loss": 0.9586, "step": 6348 }, { "epoch": 2.8767557770729497, "grad_norm": 0.20715283432121898, "learning_rate": 3.6467902487214106e-05, "loss": 0.9645, "step": 6349 }, { "epoch": 2.877208880833711, "grad_norm": 0.1985639311333701, "learning_rate": 3.6455298412088736e-05, "loss": 0.9672, "step": 6350 }, { "epoch": 2.877661984594472, "grad_norm": 0.1771252566718833, "learning_rate": 3.644269469168971e-05, "loss": 0.9524, "step": 6351 }, { "epoch": 2.8781150883552336, "grad_norm": 0.2227709815130766, "learning_rate": 3.643009132727833e-05, "loss": 0.9506, "step": 6352 }, { "epoch": 2.8785681921159947, "grad_norm": 0.20654829745970965, "learning_rate": 3.641748832011584e-05, "loss": 0.9228, "step": 6353 }, { "epoch": 2.879021295876756, "grad_norm": 0.21912000857794478, "learning_rate": 3.640488567146344e-05, "loss": 0.9361, "step": 6354 }, { "epoch": 2.879474399637517, "grad_norm": 0.17607164176106835, "learning_rate": 3.639228338258232e-05, "loss": 0.9749, "step": 6355 }, { "epoch": 2.879927503398278, "grad_norm": 0.16110509509649998, "learning_rate": 3.63796814547336e-05, "loss": 0.9545, "step": 6356 }, { "epoch": 2.880380607159039, "grad_norm": 0.18626858480681288, "learning_rate": 3.636707988917841e-05, "loss": 0.9519, "step": 6357 }, { "epoch": 2.880833710919801, "grad_norm": 0.21323170354860066, "learning_rate": 3.635447868717781e-05, "loss": 0.9481, "step": 6358 }, { "epoch": 2.881286814680562, "grad_norm": 0.20934392087388748, "learning_rate": 3.6341877849992815e-05, "loss": 0.9401, "step": 6359 }, { "epoch": 2.881739918441323, "grad_norm": 0.18408798705220836, "learning_rate": 3.632927737888443e-05, "loss": 0.935, "step": 6360 }, { "epoch": 2.882193022202084, "grad_norm": 0.18083173568137464, "learning_rate": 3.631667727511362e-05, "loss": 0.9524, "step": 6361 }, { "epoch": 2.8826461259628458, "grad_norm": 0.21720140231385918, "learning_rate": 3.6304077539941295e-05, "loss": 0.9241, "step": 6362 }, { "epoch": 2.883099229723607, "grad_norm": 0.3081567870225384, "learning_rate": 3.629147817462834e-05, "loss": 0.9373, "step": 6363 }, { "epoch": 2.883552333484368, "grad_norm": 0.21556261273386013, "learning_rate": 3.627887918043561e-05, "loss": 0.9349, "step": 6364 }, { "epoch": 2.884005437245129, "grad_norm": 0.18522389717110432, "learning_rate": 3.626628055862391e-05, "loss": 0.9361, "step": 6365 }, { "epoch": 2.8844585410058903, "grad_norm": 0.1654721665059535, "learning_rate": 3.625368231045402e-05, "loss": 0.9265, "step": 6366 }, { "epoch": 2.8849116447666514, "grad_norm": 0.18576229421229315, "learning_rate": 3.6241084437186654e-05, "loss": 0.9576, "step": 6367 }, { "epoch": 2.8853647485274125, "grad_norm": 0.16176258067934057, "learning_rate": 3.622848694008253e-05, "loss": 0.944, "step": 6368 }, { "epoch": 2.885817852288174, "grad_norm": 0.16910184480643103, "learning_rate": 3.62158898204023e-05, "loss": 0.9219, "step": 6369 }, { "epoch": 2.8862709560489352, "grad_norm": 0.19356425544522562, "learning_rate": 3.620329307940659e-05, "loss": 0.929, "step": 6370 }, { "epoch": 2.8867240598096964, "grad_norm": 0.16452675676207182, "learning_rate": 3.6190696718355986e-05, "loss": 0.9544, "step": 6371 }, { "epoch": 2.8871771635704575, "grad_norm": 0.22959856878489593, "learning_rate": 3.617810073851103e-05, "loss": 0.9397, "step": 6372 }, { "epoch": 2.887630267331219, "grad_norm": 0.20789226328740135, "learning_rate": 3.616550514113223e-05, "loss": 0.9548, "step": 6373 }, { "epoch": 2.88808337109198, "grad_norm": 0.20462431925848926, "learning_rate": 3.615290992748007e-05, "loss": 0.9308, "step": 6374 }, { "epoch": 2.8885364748527413, "grad_norm": 0.2038855133062246, "learning_rate": 3.614031509881496e-05, "loss": 0.9173, "step": 6375 }, { "epoch": 2.8889895786135025, "grad_norm": 0.15539542988760013, "learning_rate": 3.6127720656397304e-05, "loss": 0.9558, "step": 6376 }, { "epoch": 2.8894426823742636, "grad_norm": 0.1869033104993341, "learning_rate": 3.6115126601487454e-05, "loss": 0.9456, "step": 6377 }, { "epoch": 2.8898957861350247, "grad_norm": 0.16140290508578672, "learning_rate": 3.610253293534573e-05, "loss": 0.935, "step": 6378 }, { "epoch": 2.890348889895786, "grad_norm": 0.16957746234522456, "learning_rate": 3.608993965923241e-05, "loss": 0.9291, "step": 6379 }, { "epoch": 2.8908019936565474, "grad_norm": 0.1550164746988269, "learning_rate": 3.607734677440772e-05, "loss": 0.9559, "step": 6380 }, { "epoch": 2.8912550974173086, "grad_norm": 0.195914652628551, "learning_rate": 3.606475428213188e-05, "loss": 0.931, "step": 6381 }, { "epoch": 2.8917082011780697, "grad_norm": 0.21411193705677334, "learning_rate": 3.605216218366502e-05, "loss": 0.9602, "step": 6382 }, { "epoch": 2.892161304938831, "grad_norm": 0.24149460649554916, "learning_rate": 3.603957048026729e-05, "loss": 0.9401, "step": 6383 }, { "epoch": 2.8926144086995924, "grad_norm": 0.2104141413736035, "learning_rate": 3.6026979173198764e-05, "loss": 0.9406, "step": 6384 }, { "epoch": 2.8930675124603535, "grad_norm": 0.18120300049799107, "learning_rate": 3.601438826371946e-05, "loss": 0.9295, "step": 6385 }, { "epoch": 2.8935206162211147, "grad_norm": 0.18294011210251593, "learning_rate": 3.60017977530894e-05, "loss": 0.9491, "step": 6386 }, { "epoch": 2.893973719981876, "grad_norm": 0.18111459585983658, "learning_rate": 3.5989207642568535e-05, "loss": 0.9422, "step": 6387 }, { "epoch": 2.894426823742637, "grad_norm": 0.18211935525015505, "learning_rate": 3.597661793341679e-05, "loss": 0.9294, "step": 6388 }, { "epoch": 2.894879927503398, "grad_norm": 0.17975378127554753, "learning_rate": 3.596402862689405e-05, "loss": 0.9335, "step": 6389 }, { "epoch": 2.8953330312641596, "grad_norm": 0.17040265743736036, "learning_rate": 3.5951439724260146e-05, "loss": 0.9427, "step": 6390 }, { "epoch": 2.8957861350249208, "grad_norm": 0.17001682233289594, "learning_rate": 3.593885122677489e-05, "loss": 0.9529, "step": 6391 }, { "epoch": 2.896239238785682, "grad_norm": 0.1731711163934781, "learning_rate": 3.592626313569803e-05, "loss": 0.9355, "step": 6392 }, { "epoch": 2.896692342546443, "grad_norm": 0.1840873553965688, "learning_rate": 3.5913675452289295e-05, "loss": 0.9219, "step": 6393 }, { "epoch": 2.8971454463072046, "grad_norm": 0.1618745147746183, "learning_rate": 3.5901088177808365e-05, "loss": 0.9305, "step": 6394 }, { "epoch": 2.8975985500679657, "grad_norm": 0.17377592585641588, "learning_rate": 3.588850131351486e-05, "loss": 0.9141, "step": 6395 }, { "epoch": 2.898051653828727, "grad_norm": 0.16234757926438312, "learning_rate": 3.587591486066839e-05, "loss": 0.938, "step": 6396 }, { "epoch": 2.898504757589488, "grad_norm": 0.17535703176004946, "learning_rate": 3.586332882052851e-05, "loss": 0.9334, "step": 6397 }, { "epoch": 2.898957861350249, "grad_norm": 0.17406093928104102, "learning_rate": 3.585074319435472e-05, "loss": 0.9341, "step": 6398 }, { "epoch": 2.8994109651110103, "grad_norm": 0.22252011865036234, "learning_rate": 3.583815798340651e-05, "loss": 0.9562, "step": 6399 }, { "epoch": 2.8998640688717714, "grad_norm": 0.20908942532948435, "learning_rate": 3.58255731889433e-05, "loss": 0.9509, "step": 6400 }, { "epoch": 2.900317172632533, "grad_norm": 0.2606428339422496, "learning_rate": 3.581298881222449e-05, "loss": 0.9546, "step": 6401 }, { "epoch": 2.900770276393294, "grad_norm": 0.16095624317833213, "learning_rate": 3.580040485450942e-05, "loss": 0.9281, "step": 6402 }, { "epoch": 2.9012233801540552, "grad_norm": 0.18359174991451832, "learning_rate": 3.5787821317057396e-05, "loss": 0.9505, "step": 6403 }, { "epoch": 2.9016764839148164, "grad_norm": 0.1879986601227827, "learning_rate": 3.5775238201127675e-05, "loss": 0.964, "step": 6404 }, { "epoch": 2.902129587675578, "grad_norm": 0.1654310401596091, "learning_rate": 3.576265550797948e-05, "loss": 0.9271, "step": 6405 }, { "epoch": 2.902582691436339, "grad_norm": 0.20939153009902317, "learning_rate": 3.5750073238872e-05, "loss": 0.9555, "step": 6406 }, { "epoch": 2.9030357951971, "grad_norm": 0.25516811792333116, "learning_rate": 3.573749139506436e-05, "loss": 0.9272, "step": 6407 }, { "epoch": 2.9034888989578613, "grad_norm": 0.23962909545678285, "learning_rate": 3.572490997781566e-05, "loss": 0.9414, "step": 6408 }, { "epoch": 2.9039420027186225, "grad_norm": 0.19905294141690344, "learning_rate": 3.571232898838495e-05, "loss": 0.957, "step": 6409 }, { "epoch": 2.9043951064793836, "grad_norm": 0.14974579917928368, "learning_rate": 3.569974842803124e-05, "loss": 0.9219, "step": 6410 }, { "epoch": 2.9048482102401447, "grad_norm": 0.17448133952619152, "learning_rate": 3.56871682980135e-05, "loss": 0.9357, "step": 6411 }, { "epoch": 2.9053013140009063, "grad_norm": 0.2519165917306121, "learning_rate": 3.567458859959063e-05, "loss": 0.9368, "step": 6412 }, { "epoch": 2.9057544177616674, "grad_norm": 0.2529341553844849, "learning_rate": 3.566200933402152e-05, "loss": 0.9443, "step": 6413 }, { "epoch": 2.9062075215224286, "grad_norm": 0.3038941715505987, "learning_rate": 3.5649430502565024e-05, "loss": 0.9548, "step": 6414 }, { "epoch": 2.9066606252831897, "grad_norm": 0.1933204108597005, "learning_rate": 3.5636852106479915e-05, "loss": 0.9561, "step": 6415 }, { "epoch": 2.9071137290439513, "grad_norm": 0.18071806542632468, "learning_rate": 3.562427414702494e-05, "loss": 0.9401, "step": 6416 }, { "epoch": 2.9075668328047124, "grad_norm": 0.16845456988154764, "learning_rate": 3.5611696625458824e-05, "loss": 0.9387, "step": 6417 }, { "epoch": 2.9080199365654735, "grad_norm": 0.22430654392277918, "learning_rate": 3.55991195430402e-05, "loss": 0.9653, "step": 6418 }, { "epoch": 2.9084730403262347, "grad_norm": 0.1419049091023493, "learning_rate": 3.5586542901027724e-05, "loss": 0.9438, "step": 6419 }, { "epoch": 2.908926144086996, "grad_norm": 0.16132543986173373, "learning_rate": 3.5573966700679934e-05, "loss": 0.9357, "step": 6420 }, { "epoch": 2.909379247847757, "grad_norm": 0.13904146029259937, "learning_rate": 3.556139094325536e-05, "loss": 0.9591, "step": 6421 }, { "epoch": 2.9098323516085185, "grad_norm": 0.16247154740902173, "learning_rate": 3.554881563001251e-05, "loss": 0.9388, "step": 6422 }, { "epoch": 2.9102854553692796, "grad_norm": 0.1545900404132985, "learning_rate": 3.5536240762209804e-05, "loss": 0.9528, "step": 6423 }, { "epoch": 2.9107385591300408, "grad_norm": 0.16052624478742766, "learning_rate": 3.552366634110565e-05, "loss": 0.9307, "step": 6424 }, { "epoch": 2.911191662890802, "grad_norm": 0.1495113604400517, "learning_rate": 3.551109236795839e-05, "loss": 0.9477, "step": 6425 }, { "epoch": 2.9116447666515635, "grad_norm": 0.20108624684917764, "learning_rate": 3.5498518844026334e-05, "loss": 0.9478, "step": 6426 }, { "epoch": 2.9120978704123246, "grad_norm": 0.18691183523039404, "learning_rate": 3.5485945770567754e-05, "loss": 0.947, "step": 6427 }, { "epoch": 2.9125509741730857, "grad_norm": 0.158553265112801, "learning_rate": 3.5473373148840844e-05, "loss": 0.9258, "step": 6428 }, { "epoch": 2.913004077933847, "grad_norm": 0.19831181693310904, "learning_rate": 3.546080098010379e-05, "loss": 0.941, "step": 6429 }, { "epoch": 2.913457181694608, "grad_norm": 0.2079491962121435, "learning_rate": 3.5448229265614716e-05, "loss": 0.9662, "step": 6430 }, { "epoch": 2.913910285455369, "grad_norm": 0.17163035576280927, "learning_rate": 3.5435658006631694e-05, "loss": 0.952, "step": 6431 }, { "epoch": 2.9143633892161303, "grad_norm": 0.18417499408105928, "learning_rate": 3.542308720441277e-05, "loss": 0.948, "step": 6432 }, { "epoch": 2.914816492976892, "grad_norm": 0.18709499006040267, "learning_rate": 3.5410516860215924e-05, "loss": 0.9124, "step": 6433 }, { "epoch": 2.915269596737653, "grad_norm": 0.16391542244210194, "learning_rate": 3.53979469752991e-05, "loss": 0.9537, "step": 6434 }, { "epoch": 2.915722700498414, "grad_norm": 0.1627486153586155, "learning_rate": 3.538537755092021e-05, "loss": 0.9491, "step": 6435 }, { "epoch": 2.916175804259175, "grad_norm": 0.17973883447066655, "learning_rate": 3.537280858833708e-05, "loss": 0.9383, "step": 6436 }, { "epoch": 2.916628908019937, "grad_norm": 0.1564324924214812, "learning_rate": 3.5360240088807525e-05, "loss": 0.9601, "step": 6437 }, { "epoch": 2.917082011780698, "grad_norm": 0.1798264590316204, "learning_rate": 3.534767205358931e-05, "loss": 0.9385, "step": 6438 }, { "epoch": 2.917535115541459, "grad_norm": 0.23217251162989672, "learning_rate": 3.533510448394014e-05, "loss": 0.9465, "step": 6439 }, { "epoch": 2.91798821930222, "grad_norm": 0.18095711160271763, "learning_rate": 3.5322537381117684e-05, "loss": 0.9252, "step": 6440 }, { "epoch": 2.9184413230629813, "grad_norm": 0.15328322380533693, "learning_rate": 3.530997074637958e-05, "loss": 0.9297, "step": 6441 }, { "epoch": 2.9188944268237424, "grad_norm": 0.20249980511864388, "learning_rate": 3.529740458098335e-05, "loss": 0.9549, "step": 6442 }, { "epoch": 2.9193475305845036, "grad_norm": 0.1516065081378582, "learning_rate": 3.528483888618655e-05, "loss": 0.9105, "step": 6443 }, { "epoch": 2.919800634345265, "grad_norm": 0.15913219785552501, "learning_rate": 3.527227366324666e-05, "loss": 0.9333, "step": 6444 }, { "epoch": 2.9202537381060263, "grad_norm": 0.1821939331166661, "learning_rate": 3.5259708913421105e-05, "loss": 0.9351, "step": 6445 }, { "epoch": 2.9207068418667874, "grad_norm": 0.1626544436480549, "learning_rate": 3.524714463796727e-05, "loss": 0.9066, "step": 6446 }, { "epoch": 2.9211599456275485, "grad_norm": 0.16650760592375002, "learning_rate": 3.52345808381425e-05, "loss": 0.9246, "step": 6447 }, { "epoch": 2.92161304938831, "grad_norm": 0.15503624315514666, "learning_rate": 3.522201751520406e-05, "loss": 0.9452, "step": 6448 }, { "epoch": 2.9220661531490713, "grad_norm": 0.13795280248585615, "learning_rate": 3.52094546704092e-05, "loss": 0.9479, "step": 6449 }, { "epoch": 2.9225192569098324, "grad_norm": 0.14973705838214937, "learning_rate": 3.5196892305015114e-05, "loss": 0.9266, "step": 6450 }, { "epoch": 2.9229723606705935, "grad_norm": 0.1707174093061963, "learning_rate": 3.5184330420278954e-05, "loss": 0.9149, "step": 6451 }, { "epoch": 2.9234254644313546, "grad_norm": 0.1677366872148127, "learning_rate": 3.5171769017457806e-05, "loss": 0.9453, "step": 6452 }, { "epoch": 2.923878568192116, "grad_norm": 0.16132284039970565, "learning_rate": 3.5159208097808726e-05, "loss": 0.9387, "step": 6453 }, { "epoch": 2.9243316719528774, "grad_norm": 0.1619392950875924, "learning_rate": 3.5146647662588706e-05, "loss": 0.962, "step": 6454 }, { "epoch": 2.9247847757136385, "grad_norm": 0.1825312035914705, "learning_rate": 3.5134087713054714e-05, "loss": 0.9552, "step": 6455 }, { "epoch": 2.9252378794743996, "grad_norm": 0.2668897137063027, "learning_rate": 3.5121528250463627e-05, "loss": 0.9313, "step": 6456 }, { "epoch": 2.9256909832351607, "grad_norm": 0.26714595090619175, "learning_rate": 3.510896927607232e-05, "loss": 0.9115, "step": 6457 }, { "epoch": 2.9261440869959223, "grad_norm": 0.1866315275018167, "learning_rate": 3.509641079113758e-05, "loss": 0.9569, "step": 6458 }, { "epoch": 2.9265971907566835, "grad_norm": 0.16761557861694573, "learning_rate": 3.5083852796916176e-05, "loss": 0.9505, "step": 6459 }, { "epoch": 2.9270502945174446, "grad_norm": 0.1776330243667112, "learning_rate": 3.5071295294664816e-05, "loss": 0.9354, "step": 6460 }, { "epoch": 2.9275033982782057, "grad_norm": 0.18695622595773057, "learning_rate": 3.505873828564015e-05, "loss": 0.9351, "step": 6461 }, { "epoch": 2.927956502038967, "grad_norm": 0.17916481787724875, "learning_rate": 3.50461817710988e-05, "loss": 0.9437, "step": 6462 }, { "epoch": 2.928409605799728, "grad_norm": 0.24672265430957563, "learning_rate": 3.503362575229732e-05, "loss": 0.944, "step": 6463 }, { "epoch": 2.928862709560489, "grad_norm": 0.23647648514201028, "learning_rate": 3.50210702304922e-05, "loss": 0.9435, "step": 6464 }, { "epoch": 2.9293158133212507, "grad_norm": 0.15276300481128485, "learning_rate": 3.5008515206939923e-05, "loss": 0.9264, "step": 6465 }, { "epoch": 2.929768917082012, "grad_norm": 0.2216971956925572, "learning_rate": 3.499596068289688e-05, "loss": 0.9321, "step": 6466 }, { "epoch": 2.930222020842773, "grad_norm": 0.21743378660617935, "learning_rate": 3.498340665961945e-05, "loss": 0.9361, "step": 6467 }, { "epoch": 2.930675124603534, "grad_norm": 0.1984507686064301, "learning_rate": 3.4970853138363935e-05, "loss": 0.941, "step": 6468 }, { "epoch": 2.9311282283642957, "grad_norm": 0.22778013674484218, "learning_rate": 3.49583001203866e-05, "loss": 0.9646, "step": 6469 }, { "epoch": 2.931581332125057, "grad_norm": 0.20873512921539197, "learning_rate": 3.494574760694363e-05, "loss": 0.946, "step": 6470 }, { "epoch": 2.932034435885818, "grad_norm": 0.16263039107005464, "learning_rate": 3.493319559929121e-05, "loss": 0.9454, "step": 6471 }, { "epoch": 2.932487539646579, "grad_norm": 0.16195228873773473, "learning_rate": 3.492064409868545e-05, "loss": 0.9323, "step": 6472 }, { "epoch": 2.93294064340734, "grad_norm": 0.1872417275648728, "learning_rate": 3.4908093106382384e-05, "loss": 0.9574, "step": 6473 }, { "epoch": 2.9333937471681013, "grad_norm": 0.18884778026576504, "learning_rate": 3.4895542623638036e-05, "loss": 0.9534, "step": 6474 }, { "epoch": 2.9338468509288624, "grad_norm": 0.167350401937392, "learning_rate": 3.488299265170835e-05, "loss": 0.9423, "step": 6475 }, { "epoch": 2.934299954689624, "grad_norm": 0.18247472426876443, "learning_rate": 3.4870443191849244e-05, "loss": 0.9448, "step": 6476 }, { "epoch": 2.934753058450385, "grad_norm": 0.27366767625710753, "learning_rate": 3.485789424531656e-05, "loss": 0.9312, "step": 6477 }, { "epoch": 2.9352061622111463, "grad_norm": 0.21375772104850188, "learning_rate": 3.4845345813366105e-05, "loss": 0.961, "step": 6478 }, { "epoch": 2.9356592659719074, "grad_norm": 0.18881610608277702, "learning_rate": 3.483279789725362e-05, "loss": 0.9386, "step": 6479 }, { "epoch": 2.936112369732669, "grad_norm": 0.2145808711410035, "learning_rate": 3.482025049823483e-05, "loss": 0.9315, "step": 6480 }, { "epoch": 2.93656547349343, "grad_norm": 0.1646501628593373, "learning_rate": 3.480770361756534e-05, "loss": 0.9503, "step": 6481 }, { "epoch": 2.9370185772541912, "grad_norm": 0.1839463394761029, "learning_rate": 3.479515725650078e-05, "loss": 0.9369, "step": 6482 }, { "epoch": 2.9374716810149524, "grad_norm": 0.16228638344908536, "learning_rate": 3.4782611416296675e-05, "loss": 0.9591, "step": 6483 }, { "epoch": 2.9379247847757135, "grad_norm": 0.16320712863329095, "learning_rate": 3.477006609820852e-05, "loss": 0.9541, "step": 6484 }, { "epoch": 2.9383778885364746, "grad_norm": 0.15433288209308943, "learning_rate": 3.4757521303491754e-05, "loss": 0.9587, "step": 6485 }, { "epoch": 2.938830992297236, "grad_norm": 0.14221742906595417, "learning_rate": 3.474497703340177e-05, "loss": 0.9377, "step": 6486 }, { "epoch": 2.9392840960579973, "grad_norm": 0.18039057688099655, "learning_rate": 3.473243328919389e-05, "loss": 0.9531, "step": 6487 }, { "epoch": 2.9397371998187585, "grad_norm": 0.1962590796906167, "learning_rate": 3.471989007212341e-05, "loss": 0.935, "step": 6488 }, { "epoch": 2.9401903035795196, "grad_norm": 0.18174419757552873, "learning_rate": 3.470734738344554e-05, "loss": 0.9525, "step": 6489 }, { "epoch": 2.940643407340281, "grad_norm": 0.15277312310246985, "learning_rate": 3.469480522441547e-05, "loss": 0.9563, "step": 6490 }, { "epoch": 2.9410965111010423, "grad_norm": 0.1566636202540107, "learning_rate": 3.4682263596288315e-05, "loss": 0.9228, "step": 6491 }, { "epoch": 2.9415496148618034, "grad_norm": 0.18487826607932542, "learning_rate": 3.466972250031914e-05, "loss": 0.9349, "step": 6492 }, { "epoch": 2.9420027186225646, "grad_norm": 0.18145056601879397, "learning_rate": 3.465718193776297e-05, "loss": 0.9397, "step": 6493 }, { "epoch": 2.9424558223833257, "grad_norm": 0.163786948785562, "learning_rate": 3.4644641909874765e-05, "loss": 0.9261, "step": 6494 }, { "epoch": 2.942908926144087, "grad_norm": 0.15463046206013192, "learning_rate": 3.463210241790943e-05, "loss": 0.9509, "step": 6495 }, { "epoch": 2.943362029904848, "grad_norm": 0.22341018552616074, "learning_rate": 3.461956346312184e-05, "loss": 0.9476, "step": 6496 }, { "epoch": 2.9438151336656095, "grad_norm": 0.16389515793819592, "learning_rate": 3.4607025046766774e-05, "loss": 0.9478, "step": 6497 }, { "epoch": 2.9442682374263707, "grad_norm": 0.1890752277564199, "learning_rate": 3.459448717009899e-05, "loss": 0.9387, "step": 6498 }, { "epoch": 2.944721341187132, "grad_norm": 0.22252159316862738, "learning_rate": 3.458194983437317e-05, "loss": 0.9309, "step": 6499 }, { "epoch": 2.945174444947893, "grad_norm": 0.19226041489387208, "learning_rate": 3.4569413040843966e-05, "loss": 0.9486, "step": 6500 }, { "epoch": 2.9456275487086545, "grad_norm": 0.17311026755651643, "learning_rate": 3.455687679076596e-05, "loss": 0.9574, "step": 6501 }, { "epoch": 2.9460806524694156, "grad_norm": 0.17598237290172342, "learning_rate": 3.4544341085393673e-05, "loss": 0.9374, "step": 6502 }, { "epoch": 2.9465337562301768, "grad_norm": 0.20093028210800118, "learning_rate": 3.45318059259816e-05, "loss": 0.9717, "step": 6503 }, { "epoch": 2.946986859990938, "grad_norm": 0.2050925701063166, "learning_rate": 3.4519271313784156e-05, "loss": 0.9432, "step": 6504 }, { "epoch": 2.947439963751699, "grad_norm": 0.19468687462693518, "learning_rate": 3.45067372500557e-05, "loss": 0.9368, "step": 6505 }, { "epoch": 2.94789306751246, "grad_norm": 0.4103734109765124, "learning_rate": 3.449420373605056e-05, "loss": 0.9413, "step": 6506 }, { "epoch": 2.9483461712732213, "grad_norm": 0.14181505902146138, "learning_rate": 3.448167077302297e-05, "loss": 0.9383, "step": 6507 }, { "epoch": 2.948799275033983, "grad_norm": 0.1815309629288713, "learning_rate": 3.4469138362227164e-05, "loss": 0.9501, "step": 6508 }, { "epoch": 2.949252378794744, "grad_norm": 0.16600372707733774, "learning_rate": 3.445660650491727e-05, "loss": 0.9428, "step": 6509 }, { "epoch": 2.949705482555505, "grad_norm": 0.21340934876822637, "learning_rate": 3.4444075202347365e-05, "loss": 0.9498, "step": 6510 }, { "epoch": 2.9501585863162663, "grad_norm": 0.24424286358289213, "learning_rate": 3.4431544455771506e-05, "loss": 0.9292, "step": 6511 }, { "epoch": 2.950611690077028, "grad_norm": 0.25016495309937725, "learning_rate": 3.4419014266443675e-05, "loss": 0.9337, "step": 6512 }, { "epoch": 2.951064793837789, "grad_norm": 0.21900063163713537, "learning_rate": 3.4406484635617785e-05, "loss": 0.9462, "step": 6513 }, { "epoch": 2.95151789759855, "grad_norm": 0.17138129170235203, "learning_rate": 3.439395556454771e-05, "loss": 0.9292, "step": 6514 }, { "epoch": 2.9519710013593112, "grad_norm": 0.15019350478075863, "learning_rate": 3.438142705448726e-05, "loss": 0.9536, "step": 6515 }, { "epoch": 2.9524241051200724, "grad_norm": 0.20149554818321128, "learning_rate": 3.43688991066902e-05, "loss": 0.9423, "step": 6516 }, { "epoch": 2.9528772088808335, "grad_norm": 0.24324548675019125, "learning_rate": 3.435637172241022e-05, "loss": 0.9419, "step": 6517 }, { "epoch": 2.953330312641595, "grad_norm": 0.20928394125460983, "learning_rate": 3.4343844902900965e-05, "loss": 0.9344, "step": 6518 }, { "epoch": 2.953783416402356, "grad_norm": 0.18165132223362185, "learning_rate": 3.4331318649416024e-05, "loss": 0.9407, "step": 6519 }, { "epoch": 2.9542365201631173, "grad_norm": 0.17205683072779884, "learning_rate": 3.431879296320893e-05, "loss": 0.9262, "step": 6520 }, { "epoch": 2.9546896239238785, "grad_norm": 0.1624155534987249, "learning_rate": 3.4306267845533154e-05, "loss": 0.9436, "step": 6521 }, { "epoch": 2.95514272768464, "grad_norm": 0.17229020360900982, "learning_rate": 3.429374329764212e-05, "loss": 0.9618, "step": 6522 }, { "epoch": 2.955595831445401, "grad_norm": 0.18569543230450317, "learning_rate": 3.428121932078918e-05, "loss": 0.9427, "step": 6523 }, { "epoch": 2.9560489352061623, "grad_norm": 0.16811886491344005, "learning_rate": 3.4268695916227644e-05, "loss": 0.9109, "step": 6524 }, { "epoch": 2.9565020389669234, "grad_norm": 0.1385135442623148, "learning_rate": 3.425617308521075e-05, "loss": 0.956, "step": 6525 }, { "epoch": 2.9569551427276846, "grad_norm": 0.14105842728557402, "learning_rate": 3.424365082899168e-05, "loss": 0.9458, "step": 6526 }, { "epoch": 2.9574082464884457, "grad_norm": 0.19015887677280807, "learning_rate": 3.4231129148823585e-05, "loss": 0.933, "step": 6527 }, { "epoch": 2.957861350249207, "grad_norm": 0.20181156270932554, "learning_rate": 3.421860804595951e-05, "loss": 0.9299, "step": 6528 }, { "epoch": 2.9583144540099684, "grad_norm": 0.1990069241692, "learning_rate": 3.420608752165251e-05, "loss": 0.9342, "step": 6529 }, { "epoch": 2.9587675577707295, "grad_norm": 0.16268164384840067, "learning_rate": 3.4193567577155505e-05, "loss": 0.9591, "step": 6530 }, { "epoch": 2.9592206615314907, "grad_norm": 0.1501874740015094, "learning_rate": 3.418104821372141e-05, "loss": 0.9375, "step": 6531 }, { "epoch": 2.959673765292252, "grad_norm": 0.18450395813760728, "learning_rate": 3.416852943260307e-05, "loss": 0.9455, "step": 6532 }, { "epoch": 2.9601268690530134, "grad_norm": 0.2175626473887387, "learning_rate": 3.415601123505327e-05, "loss": 0.932, "step": 6533 }, { "epoch": 2.9605799728137745, "grad_norm": 0.18850300158493052, "learning_rate": 3.414349362232472e-05, "loss": 0.9355, "step": 6534 }, { "epoch": 2.9610330765745356, "grad_norm": 0.15358625053870245, "learning_rate": 3.413097659567009e-05, "loss": 0.9546, "step": 6535 }, { "epoch": 2.9614861803352968, "grad_norm": 0.18069340326912164, "learning_rate": 3.411846015634199e-05, "loss": 0.9463, "step": 6536 }, { "epoch": 2.961939284096058, "grad_norm": 0.23211579148575778, "learning_rate": 3.4105944305592976e-05, "loss": 0.9444, "step": 6537 }, { "epoch": 2.962392387856819, "grad_norm": 0.22039577548900877, "learning_rate": 3.409342904467553e-05, "loss": 0.9333, "step": 6538 }, { "epoch": 2.96284549161758, "grad_norm": 0.18369568968914113, "learning_rate": 3.408091437484208e-05, "loss": 0.9255, "step": 6539 }, { "epoch": 2.9632985953783417, "grad_norm": 0.12297691397223758, "learning_rate": 3.4068400297345e-05, "loss": 0.9455, "step": 6540 }, { "epoch": 2.963751699139103, "grad_norm": 0.2077397217333601, "learning_rate": 3.4055886813436614e-05, "loss": 0.9631, "step": 6541 }, { "epoch": 2.964204802899864, "grad_norm": 0.275909877295981, "learning_rate": 3.4043373924369155e-05, "loss": 0.9346, "step": 6542 }, { "epoch": 2.964657906660625, "grad_norm": 0.204266351595878, "learning_rate": 3.4030861631394825e-05, "loss": 0.9393, "step": 6543 }, { "epoch": 2.9651110104213867, "grad_norm": 0.1755243203383119, "learning_rate": 3.401834993576577e-05, "loss": 0.9509, "step": 6544 }, { "epoch": 2.965564114182148, "grad_norm": 0.18121102401656963, "learning_rate": 3.400583883873403e-05, "loss": 0.9448, "step": 6545 }, { "epoch": 2.966017217942909, "grad_norm": 0.24563354848923474, "learning_rate": 3.399332834155166e-05, "loss": 0.943, "step": 6546 }, { "epoch": 2.96647032170367, "grad_norm": 0.28344139624972514, "learning_rate": 3.3980818445470586e-05, "loss": 0.9273, "step": 6547 }, { "epoch": 2.966923425464431, "grad_norm": 0.2717878413232148, "learning_rate": 3.3968309151742716e-05, "loss": 0.9371, "step": 6548 }, { "epoch": 2.9673765292251923, "grad_norm": 0.1727518792856758, "learning_rate": 3.395580046161988e-05, "loss": 0.9466, "step": 6549 }, { "epoch": 2.967829632985954, "grad_norm": 0.19216738655976226, "learning_rate": 3.3943292376353844e-05, "loss": 0.9203, "step": 6550 }, { "epoch": 2.968282736746715, "grad_norm": 0.2105133840432589, "learning_rate": 3.393078489719633e-05, "loss": 0.9316, "step": 6551 }, { "epoch": 2.968735840507476, "grad_norm": 0.2166363403920426, "learning_rate": 3.391827802539898e-05, "loss": 0.9194, "step": 6552 }, { "epoch": 2.9691889442682373, "grad_norm": 0.19046754608798086, "learning_rate": 3.3905771762213396e-05, "loss": 0.966, "step": 6553 }, { "epoch": 2.969642048028999, "grad_norm": 0.17066066536259344, "learning_rate": 3.3893266108891096e-05, "loss": 0.9446, "step": 6554 }, { "epoch": 2.97009515178976, "grad_norm": 0.17903963623274255, "learning_rate": 3.3880761066683586e-05, "loss": 0.9486, "step": 6555 }, { "epoch": 2.970548255550521, "grad_norm": 0.20267766838680004, "learning_rate": 3.386825663684222e-05, "loss": 0.9306, "step": 6556 }, { "epoch": 2.9710013593112823, "grad_norm": 0.1965847273175392, "learning_rate": 3.3855752820618364e-05, "loss": 0.9427, "step": 6557 }, { "epoch": 2.9714544630720434, "grad_norm": 0.15427595758745608, "learning_rate": 3.384324961926331e-05, "loss": 0.9482, "step": 6558 }, { "epoch": 2.9719075668328045, "grad_norm": 0.15824790947335007, "learning_rate": 3.3830747034028285e-05, "loss": 0.9104, "step": 6559 }, { "epoch": 2.9723606705935657, "grad_norm": 0.19571008864132963, "learning_rate": 3.3818245066164445e-05, "loss": 0.953, "step": 6560 }, { "epoch": 2.9728137743543273, "grad_norm": 0.23546494915044988, "learning_rate": 3.38057437169229e-05, "loss": 0.9456, "step": 6561 }, { "epoch": 2.9732668781150884, "grad_norm": 0.20651013365818283, "learning_rate": 3.379324298755467e-05, "loss": 0.9481, "step": 6562 }, { "epoch": 2.9737199818758495, "grad_norm": 0.16247877763677004, "learning_rate": 3.3780742879310745e-05, "loss": 0.9364, "step": 6563 }, { "epoch": 2.9741730856366106, "grad_norm": 0.15692733962599525, "learning_rate": 3.376824339344204e-05, "loss": 0.9382, "step": 6564 }, { "epoch": 2.9746261893973722, "grad_norm": 0.23282669390175212, "learning_rate": 3.375574453119939e-05, "loss": 0.9439, "step": 6565 }, { "epoch": 2.9750792931581334, "grad_norm": 0.2009800427191312, "learning_rate": 3.374324629383361e-05, "loss": 0.9484, "step": 6566 }, { "epoch": 2.9755323969188945, "grad_norm": 0.17333838042897978, "learning_rate": 3.373074868259541e-05, "loss": 0.9432, "step": 6567 }, { "epoch": 2.9759855006796556, "grad_norm": 0.1951302094680639, "learning_rate": 3.371825169873546e-05, "loss": 0.9668, "step": 6568 }, { "epoch": 2.9764386044404167, "grad_norm": 0.23096835104171237, "learning_rate": 3.370575534350437e-05, "loss": 0.9448, "step": 6569 }, { "epoch": 2.976891708201178, "grad_norm": 0.2527220273001515, "learning_rate": 3.369325961815266e-05, "loss": 0.962, "step": 6570 }, { "epoch": 2.977344811961939, "grad_norm": 0.2086530926053117, "learning_rate": 3.368076452393081e-05, "loss": 0.9312, "step": 6571 }, { "epoch": 2.9777979157227006, "grad_norm": 0.14024427096247502, "learning_rate": 3.3668270062089245e-05, "loss": 0.9429, "step": 6572 }, { "epoch": 2.9782510194834617, "grad_norm": 0.188203289737109, "learning_rate": 3.3655776233878304e-05, "loss": 0.9308, "step": 6573 }, { "epoch": 2.978704123244223, "grad_norm": 0.21280135245853646, "learning_rate": 3.364328304054828e-05, "loss": 0.9301, "step": 6574 }, { "epoch": 2.979157227004984, "grad_norm": 0.2006771451388068, "learning_rate": 3.3630790483349386e-05, "loss": 0.9611, "step": 6575 }, { "epoch": 2.9796103307657456, "grad_norm": 0.13802698464486418, "learning_rate": 3.361829856353178e-05, "loss": 0.9587, "step": 6576 }, { "epoch": 2.9800634345265067, "grad_norm": 0.1595790672268439, "learning_rate": 3.360580728234559e-05, "loss": 0.9183, "step": 6577 }, { "epoch": 2.980516538287268, "grad_norm": 0.13622607686515817, "learning_rate": 3.359331664104079e-05, "loss": 0.9483, "step": 6578 }, { "epoch": 2.980969642048029, "grad_norm": 0.15233807858713516, "learning_rate": 3.358082664086739e-05, "loss": 0.9248, "step": 6579 }, { "epoch": 2.98142274580879, "grad_norm": 0.15743030080129697, "learning_rate": 3.356833728307527e-05, "loss": 0.9171, "step": 6580 }, { "epoch": 2.981875849569551, "grad_norm": 0.16291334873491864, "learning_rate": 3.355584856891428e-05, "loss": 0.9282, "step": 6581 }, { "epoch": 2.982328953330313, "grad_norm": 0.1391840611248419, "learning_rate": 3.3543360499634194e-05, "loss": 0.9449, "step": 6582 }, { "epoch": 2.982782057091074, "grad_norm": 0.13057686022324452, "learning_rate": 3.3530873076484724e-05, "loss": 0.9244, "step": 6583 }, { "epoch": 2.983235160851835, "grad_norm": 0.1478627556517053, "learning_rate": 3.35183863007155e-05, "loss": 0.9523, "step": 6584 }, { "epoch": 2.983688264612596, "grad_norm": 0.1410766410627531, "learning_rate": 3.3505900173576127e-05, "loss": 0.9557, "step": 6585 }, { "epoch": 2.9841413683733577, "grad_norm": 0.12456333793520498, "learning_rate": 3.349341469631609e-05, "loss": 0.9451, "step": 6586 }, { "epoch": 2.984594472134119, "grad_norm": 0.13293828784745323, "learning_rate": 3.3480929870184864e-05, "loss": 0.9402, "step": 6587 }, { "epoch": 2.98504757589488, "grad_norm": 0.14389682301412404, "learning_rate": 3.346844569643182e-05, "loss": 0.9303, "step": 6588 }, { "epoch": 2.985500679655641, "grad_norm": 0.12243230542514535, "learning_rate": 3.345596217630628e-05, "loss": 0.9409, "step": 6589 }, { "epoch": 2.9859537834164023, "grad_norm": 0.14317289318070572, "learning_rate": 3.3443479311057503e-05, "loss": 0.9434, "step": 6590 }, { "epoch": 2.9864068871771634, "grad_norm": 0.14254752965589096, "learning_rate": 3.343099710193467e-05, "loss": 0.9503, "step": 6591 }, { "epoch": 2.9868599909379245, "grad_norm": 0.15334203498292392, "learning_rate": 3.341851555018692e-05, "loss": 0.926, "step": 6592 }, { "epoch": 2.987313094698686, "grad_norm": 0.15126771592615743, "learning_rate": 3.3406034657063296e-05, "loss": 0.9702, "step": 6593 }, { "epoch": 2.9877661984594472, "grad_norm": 0.29075937412474406, "learning_rate": 3.3393554423812785e-05, "loss": 0.9403, "step": 6594 }, { "epoch": 2.9882193022202084, "grad_norm": 0.6249294713404769, "learning_rate": 3.338107485168433e-05, "loss": 0.9394, "step": 6595 }, { "epoch": 2.9886724059809695, "grad_norm": 0.13639057986387157, "learning_rate": 3.336859594192677e-05, "loss": 0.9496, "step": 6596 }, { "epoch": 2.989125509741731, "grad_norm": 0.17117285051140863, "learning_rate": 3.3356117695788917e-05, "loss": 0.9339, "step": 6597 }, { "epoch": 2.989578613502492, "grad_norm": 0.1873237802628228, "learning_rate": 3.3343640114519484e-05, "loss": 0.9417, "step": 6598 }, { "epoch": 2.9900317172632533, "grad_norm": 0.1674650381249836, "learning_rate": 3.3331163199367137e-05, "loss": 0.9468, "step": 6599 }, { "epoch": 2.9904848210240145, "grad_norm": 0.14547300050368261, "learning_rate": 3.331868695158047e-05, "loss": 0.9359, "step": 6600 }, { "epoch": 2.9909379247847756, "grad_norm": 0.20090244066713592, "learning_rate": 3.3306211372408e-05, "loss": 0.9561, "step": 6601 }, { "epoch": 2.9913910285455367, "grad_norm": 0.1876277132653228, "learning_rate": 3.329373646309821e-05, "loss": 0.9317, "step": 6602 }, { "epoch": 2.991844132306298, "grad_norm": 0.17319804179201956, "learning_rate": 3.3281262224899464e-05, "loss": 0.9596, "step": 6603 }, { "epoch": 2.9922972360670594, "grad_norm": 0.18325827622058574, "learning_rate": 3.32687886590601e-05, "loss": 0.9291, "step": 6604 }, { "epoch": 2.9927503398278206, "grad_norm": 0.18521364334411264, "learning_rate": 3.3256315766828376e-05, "loss": 0.9555, "step": 6605 }, { "epoch": 2.9932034435885817, "grad_norm": 0.17612709054428966, "learning_rate": 3.324384354945249e-05, "loss": 0.9304, "step": 6606 }, { "epoch": 2.993656547349343, "grad_norm": 0.16760129401570198, "learning_rate": 3.323137200818055e-05, "loss": 0.9649, "step": 6607 }, { "epoch": 2.9941096511101044, "grad_norm": 0.16784357460475763, "learning_rate": 3.321890114426062e-05, "loss": 0.9276, "step": 6608 }, { "epoch": 2.9945627548708655, "grad_norm": 0.15621931127547464, "learning_rate": 3.320643095894069e-05, "loss": 0.9558, "step": 6609 }, { "epoch": 2.9950158586316267, "grad_norm": 0.17439212254053124, "learning_rate": 3.3193961453468684e-05, "loss": 0.9192, "step": 6610 }, { "epoch": 2.995468962392388, "grad_norm": 0.19533993850971287, "learning_rate": 3.318149262909244e-05, "loss": 0.9309, "step": 6611 }, { "epoch": 2.995922066153149, "grad_norm": 0.17683205834060547, "learning_rate": 3.316902448705976e-05, "loss": 0.9516, "step": 6612 }, { "epoch": 2.99637516991391, "grad_norm": 0.16731679571364683, "learning_rate": 3.3156557028618346e-05, "loss": 0.9201, "step": 6613 }, { "epoch": 2.9968282736746716, "grad_norm": 0.1508429937809996, "learning_rate": 3.3144090255015836e-05, "loss": 0.941, "step": 6614 }, { "epoch": 2.9972813774354328, "grad_norm": 0.20271520615099423, "learning_rate": 3.3131624167499815e-05, "loss": 0.9421, "step": 6615 }, { "epoch": 2.997734481196194, "grad_norm": 0.18414171676104407, "learning_rate": 3.311915876731781e-05, "loss": 0.9474, "step": 6616 }, { "epoch": 2.998187584956955, "grad_norm": 0.1837742535493843, "learning_rate": 3.310669405571724e-05, "loss": 0.9287, "step": 6617 }, { "epoch": 2.9986406887177166, "grad_norm": 0.1838009174071029, "learning_rate": 3.309423003394549e-05, "loss": 0.9218, "step": 6618 }, { "epoch": 2.9990937924784777, "grad_norm": 0.1840922915988931, "learning_rate": 3.308176670324985e-05, "loss": 0.9652, "step": 6619 }, { "epoch": 2.999546896239239, "grad_norm": 0.1772181193464211, "learning_rate": 3.306930406487757e-05, "loss": 0.9396, "step": 6620 }, { "epoch": 3.0, "grad_norm": 0.1952025411764821, "learning_rate": 3.305684212007581e-05, "loss": 0.9407, "step": 6621 }, { "epoch": 3.000453103760761, "grad_norm": 0.19055569611900505, "learning_rate": 3.304438087009167e-05, "loss": 0.9087, "step": 6622 }, { "epoch": 3.0009062075215223, "grad_norm": 0.16790475129872903, "learning_rate": 3.303192031617216e-05, "loss": 0.9327, "step": 6623 }, { "epoch": 3.001359311282284, "grad_norm": 0.15404471554414023, "learning_rate": 3.301946045956424e-05, "loss": 0.9437, "step": 6624 }, { "epoch": 3.001812415043045, "grad_norm": 0.18299480453909373, "learning_rate": 3.30070013015148e-05, "loss": 0.9407, "step": 6625 }, { "epoch": 3.002265518803806, "grad_norm": 0.17271846025117346, "learning_rate": 3.299454284327066e-05, "loss": 0.9122, "step": 6626 }, { "epoch": 3.0027186225645672, "grad_norm": 0.15293930054080287, "learning_rate": 3.298208508607857e-05, "loss": 0.9506, "step": 6627 }, { "epoch": 3.0031717263253284, "grad_norm": 0.16318868362991643, "learning_rate": 3.2969628031185185e-05, "loss": 0.9326, "step": 6628 }, { "epoch": 3.00362483008609, "grad_norm": 0.15477473121651994, "learning_rate": 3.295717167983714e-05, "loss": 0.9239, "step": 6629 }, { "epoch": 3.004077933846851, "grad_norm": 0.18201142567865103, "learning_rate": 3.2944716033280955e-05, "loss": 0.9357, "step": 6630 }, { "epoch": 3.004531037607612, "grad_norm": 0.18363461465258166, "learning_rate": 3.293226109276309e-05, "loss": 0.946, "step": 6631 }, { "epoch": 3.0049841413683733, "grad_norm": 0.1910290462330428, "learning_rate": 3.291980685952994e-05, "loss": 0.9457, "step": 6632 }, { "epoch": 3.0054372451291345, "grad_norm": 0.21243091387777255, "learning_rate": 3.290735333482784e-05, "loss": 0.8896, "step": 6633 }, { "epoch": 3.0058903488898956, "grad_norm": 0.1627975949700988, "learning_rate": 3.289490051990303e-05, "loss": 0.9456, "step": 6634 }, { "epoch": 3.006343452650657, "grad_norm": 0.16189277743809524, "learning_rate": 3.2882448416001706e-05, "loss": 0.9373, "step": 6635 }, { "epoch": 3.0067965564114183, "grad_norm": 0.16063782187051223, "learning_rate": 3.2869997024369965e-05, "loss": 0.9163, "step": 6636 }, { "epoch": 3.0072496601721794, "grad_norm": 0.17089616745952382, "learning_rate": 3.285754634625384e-05, "loss": 0.9263, "step": 6637 }, { "epoch": 3.0077027639329406, "grad_norm": 0.16326369743135624, "learning_rate": 3.284509638289933e-05, "loss": 0.9207, "step": 6638 }, { "epoch": 3.0081558676937017, "grad_norm": 0.18228666221497028, "learning_rate": 3.2832647135552305e-05, "loss": 0.9516, "step": 6639 }, { "epoch": 3.0086089714544633, "grad_norm": 0.1613999889290782, "learning_rate": 3.282019860545859e-05, "loss": 0.9362, "step": 6640 }, { "epoch": 3.0090620752152244, "grad_norm": 0.15450722379452403, "learning_rate": 3.280775079386394e-05, "loss": 0.9384, "step": 6641 }, { "epoch": 3.0095151789759855, "grad_norm": 0.1759330686942657, "learning_rate": 3.2795303702014045e-05, "loss": 0.9232, "step": 6642 }, { "epoch": 3.0099682827367467, "grad_norm": 0.17960964726853854, "learning_rate": 3.27828573311545e-05, "loss": 0.9269, "step": 6643 }, { "epoch": 3.010421386497508, "grad_norm": 0.17726544528445268, "learning_rate": 3.277041168253086e-05, "loss": 0.9382, "step": 6644 }, { "epoch": 3.0108744902582694, "grad_norm": 0.15419315909371079, "learning_rate": 3.275796675738857e-05, "loss": 0.9367, "step": 6645 }, { "epoch": 3.0113275940190305, "grad_norm": 0.14397303970802078, "learning_rate": 3.274552255697304e-05, "loss": 0.9438, "step": 6646 }, { "epoch": 3.0117806977797916, "grad_norm": 0.16456696404632581, "learning_rate": 3.2733079082529574e-05, "loss": 0.9329, "step": 6647 }, { "epoch": 3.0122338015405528, "grad_norm": 0.19907422593859012, "learning_rate": 3.272063633530342e-05, "loss": 0.9222, "step": 6648 }, { "epoch": 3.012686905301314, "grad_norm": 0.2058746144022533, "learning_rate": 3.270819431653976e-05, "loss": 0.9431, "step": 6649 }, { "epoch": 3.013140009062075, "grad_norm": 0.23515695784877025, "learning_rate": 3.2695753027483696e-05, "loss": 0.936, "step": 6650 }, { "epoch": 3.0135931128228366, "grad_norm": 0.21659553373618926, "learning_rate": 3.268331246938025e-05, "loss": 0.9235, "step": 6651 }, { "epoch": 3.0140462165835977, "grad_norm": 0.1841193743494822, "learning_rate": 3.267087264347438e-05, "loss": 0.92, "step": 6652 }, { "epoch": 3.014499320344359, "grad_norm": 0.18458802959642756, "learning_rate": 3.265843355101096e-05, "loss": 0.9066, "step": 6653 }, { "epoch": 3.01495242410512, "grad_norm": 0.18204516235669652, "learning_rate": 3.264599519323482e-05, "loss": 0.9243, "step": 6654 }, { "epoch": 3.015405527865881, "grad_norm": 0.20330026826350722, "learning_rate": 3.263355757139067e-05, "loss": 0.9308, "step": 6655 }, { "epoch": 3.0158586316266427, "grad_norm": 0.1552820518370922, "learning_rate": 3.262112068672317e-05, "loss": 0.9319, "step": 6656 }, { "epoch": 3.016311735387404, "grad_norm": 0.1548381457183761, "learning_rate": 3.260868454047693e-05, "loss": 0.932, "step": 6657 }, { "epoch": 3.016764839148165, "grad_norm": 0.2028505530960422, "learning_rate": 3.259624913389645e-05, "loss": 0.9216, "step": 6658 }, { "epoch": 3.017217942908926, "grad_norm": 0.16268479719622558, "learning_rate": 3.2583814468226176e-05, "loss": 0.932, "step": 6659 }, { "epoch": 3.017671046669687, "grad_norm": 0.15858177768069912, "learning_rate": 3.2571380544710466e-05, "loss": 0.9197, "step": 6660 }, { "epoch": 3.018124150430449, "grad_norm": 0.1681122015908539, "learning_rate": 3.2558947364593615e-05, "loss": 0.9462, "step": 6661 }, { "epoch": 3.01857725419121, "grad_norm": 0.13432078140352566, "learning_rate": 3.2546514929119834e-05, "loss": 0.928, "step": 6662 }, { "epoch": 3.019030357951971, "grad_norm": 0.1622607790749794, "learning_rate": 3.253408323953328e-05, "loss": 0.9312, "step": 6663 }, { "epoch": 3.019483461712732, "grad_norm": 0.15887444316597144, "learning_rate": 3.2521652297078e-05, "loss": 0.9191, "step": 6664 }, { "epoch": 3.0199365654734933, "grad_norm": 0.17596047476400598, "learning_rate": 3.2509222102998e-05, "loss": 0.9236, "step": 6665 }, { "epoch": 3.0203896692342544, "grad_norm": 0.19010473092272212, "learning_rate": 3.249679265853719e-05, "loss": 0.9293, "step": 6666 }, { "epoch": 3.020842772995016, "grad_norm": 0.14603664139386896, "learning_rate": 3.248436396493942e-05, "loss": 0.9224, "step": 6667 }, { "epoch": 3.021295876755777, "grad_norm": 0.16040924387797487, "learning_rate": 3.247193602344845e-05, "loss": 0.909, "step": 6668 }, { "epoch": 3.0217489805165383, "grad_norm": 0.16364766272259573, "learning_rate": 3.2459508835308e-05, "loss": 0.9401, "step": 6669 }, { "epoch": 3.0222020842772994, "grad_norm": 0.15803119265254362, "learning_rate": 3.2447082401761646e-05, "loss": 0.9295, "step": 6670 }, { "epoch": 3.0226551880380605, "grad_norm": 0.16453731641633637, "learning_rate": 3.243465672405294e-05, "loss": 0.931, "step": 6671 }, { "epoch": 3.023108291798822, "grad_norm": 0.17866143249051272, "learning_rate": 3.242223180342537e-05, "loss": 0.9361, "step": 6672 }, { "epoch": 3.0235613955595833, "grad_norm": 0.17166884603206056, "learning_rate": 3.24098076411223e-05, "loss": 0.9461, "step": 6673 }, { "epoch": 3.0240144993203444, "grad_norm": 0.14563012299042974, "learning_rate": 3.239738423838707e-05, "loss": 0.9259, "step": 6674 }, { "epoch": 3.0244676030811055, "grad_norm": 0.21720621652039812, "learning_rate": 3.238496159646289e-05, "loss": 0.9326, "step": 6675 }, { "epoch": 3.0249207068418666, "grad_norm": 0.16952674351189997, "learning_rate": 3.237253971659294e-05, "loss": 0.9288, "step": 6676 }, { "epoch": 3.025373810602628, "grad_norm": 0.16817420118941576, "learning_rate": 3.23601186000203e-05, "loss": 0.9393, "step": 6677 }, { "epoch": 3.0258269143633894, "grad_norm": 0.16049564100399769, "learning_rate": 3.234769824798799e-05, "loss": 0.9308, "step": 6678 }, { "epoch": 3.0262800181241505, "grad_norm": 0.19889128669996556, "learning_rate": 3.233527866173892e-05, "loss": 0.9461, "step": 6679 }, { "epoch": 3.0267331218849116, "grad_norm": 0.16068775148604064, "learning_rate": 3.232285984251597e-05, "loss": 0.937, "step": 6680 }, { "epoch": 3.0271862256456727, "grad_norm": 0.165449039262753, "learning_rate": 3.2310441791561905e-05, "loss": 0.9184, "step": 6681 }, { "epoch": 3.027639329406434, "grad_norm": 0.22012573718855513, "learning_rate": 3.229802451011945e-05, "loss": 0.9606, "step": 6682 }, { "epoch": 3.0280924331671955, "grad_norm": 0.2066497912901376, "learning_rate": 3.22856079994312e-05, "loss": 0.9568, "step": 6683 }, { "epoch": 3.0285455369279566, "grad_norm": 0.1821297246910906, "learning_rate": 3.227319226073971e-05, "loss": 0.9614, "step": 6684 }, { "epoch": 3.0289986406887177, "grad_norm": 0.1725508679584051, "learning_rate": 3.226077729528746e-05, "loss": 0.9415, "step": 6685 }, { "epoch": 3.029451744449479, "grad_norm": 0.14772494903618788, "learning_rate": 3.2248363104316844e-05, "loss": 0.926, "step": 6686 }, { "epoch": 3.02990484821024, "grad_norm": 0.19029964749508216, "learning_rate": 3.223594968907018e-05, "loss": 0.9189, "step": 6687 }, { "epoch": 3.0303579519710016, "grad_norm": 0.2144455569350751, "learning_rate": 3.22235370507897e-05, "loss": 0.936, "step": 6688 }, { "epoch": 3.0308110557317627, "grad_norm": 0.17130921238450816, "learning_rate": 3.221112519071756e-05, "loss": 0.9252, "step": 6689 }, { "epoch": 3.031264159492524, "grad_norm": 0.1347314540671984, "learning_rate": 3.2198714110095864e-05, "loss": 0.9469, "step": 6690 }, { "epoch": 3.031717263253285, "grad_norm": 0.17756997738322805, "learning_rate": 3.2186303810166606e-05, "loss": 0.9226, "step": 6691 }, { "epoch": 3.032170367014046, "grad_norm": 0.26408467995128826, "learning_rate": 3.217389429217171e-05, "loss": 0.9631, "step": 6692 }, { "epoch": 3.0326234707748077, "grad_norm": 0.2865163355383909, "learning_rate": 3.216148555735301e-05, "loss": 0.9278, "step": 6693 }, { "epoch": 3.033076574535569, "grad_norm": 0.22371587448217015, "learning_rate": 3.21490776069523e-05, "loss": 0.9309, "step": 6694 }, { "epoch": 3.03352967829633, "grad_norm": 0.15409929462174685, "learning_rate": 3.213667044221127e-05, "loss": 0.928, "step": 6695 }, { "epoch": 3.033982782057091, "grad_norm": 0.2173169188165155, "learning_rate": 3.212426406437152e-05, "loss": 0.9122, "step": 6696 }, { "epoch": 3.034435885817852, "grad_norm": 0.31157557048170004, "learning_rate": 3.211185847467459e-05, "loss": 0.917, "step": 6697 }, { "epoch": 3.0348889895786133, "grad_norm": 0.25710451224617054, "learning_rate": 3.209945367436194e-05, "loss": 0.9384, "step": 6698 }, { "epoch": 3.035342093339375, "grad_norm": 0.1980767482880625, "learning_rate": 3.208704966467495e-05, "loss": 0.9484, "step": 6699 }, { "epoch": 3.035795197100136, "grad_norm": 0.27252684276261807, "learning_rate": 3.2074646446854907e-05, "loss": 0.9157, "step": 6700 }, { "epoch": 3.036248300860897, "grad_norm": 0.25114086340310365, "learning_rate": 3.2062244022143033e-05, "loss": 0.9242, "step": 6701 }, { "epoch": 3.0367014046216583, "grad_norm": 0.22470275191481012, "learning_rate": 3.204984239178046e-05, "loss": 0.9315, "step": 6702 }, { "epoch": 3.0371545083824194, "grad_norm": 0.16396879314301127, "learning_rate": 3.203744155700826e-05, "loss": 0.9313, "step": 6703 }, { "epoch": 3.037607612143181, "grad_norm": 0.1886559122621621, "learning_rate": 3.202504151906741e-05, "loss": 0.9362, "step": 6704 }, { "epoch": 3.038060715903942, "grad_norm": 0.2320223533066141, "learning_rate": 3.2012642279198804e-05, "loss": 0.9098, "step": 6705 }, { "epoch": 3.0385138196647032, "grad_norm": 0.22878534003799803, "learning_rate": 3.2000243838643265e-05, "loss": 0.9326, "step": 6706 }, { "epoch": 3.0389669234254644, "grad_norm": 0.1816288072837526, "learning_rate": 3.198784619864155e-05, "loss": 0.9246, "step": 6707 }, { "epoch": 3.0394200271862255, "grad_norm": 0.13895718655912637, "learning_rate": 3.1975449360434284e-05, "loss": 0.9403, "step": 6708 }, { "epoch": 3.0398731309469866, "grad_norm": 0.1945527637031588, "learning_rate": 3.196305332526207e-05, "loss": 0.9304, "step": 6709 }, { "epoch": 3.040326234707748, "grad_norm": 0.18370637866262277, "learning_rate": 3.1950658094365404e-05, "loss": 0.9363, "step": 6710 }, { "epoch": 3.0407793384685093, "grad_norm": 0.16072526392764988, "learning_rate": 3.193826366898471e-05, "loss": 0.9239, "step": 6711 }, { "epoch": 3.0412324422292705, "grad_norm": 0.15207394652089534, "learning_rate": 3.1925870050360307e-05, "loss": 0.9333, "step": 6712 }, { "epoch": 3.0416855459900316, "grad_norm": 0.1554777798407512, "learning_rate": 3.191347723973247e-05, "loss": 0.9434, "step": 6713 }, { "epoch": 3.0421386497507927, "grad_norm": 0.15446366307538278, "learning_rate": 3.1901085238341374e-05, "loss": 0.9328, "step": 6714 }, { "epoch": 3.0425917535115543, "grad_norm": 0.16883140394841767, "learning_rate": 3.1888694047427125e-05, "loss": 0.9612, "step": 6715 }, { "epoch": 3.0430448572723154, "grad_norm": 0.15857837202563224, "learning_rate": 3.187630366822971e-05, "loss": 0.9233, "step": 6716 }, { "epoch": 3.0434979610330766, "grad_norm": 0.1905544644221646, "learning_rate": 3.1863914101989086e-05, "loss": 0.9376, "step": 6717 }, { "epoch": 3.0439510647938377, "grad_norm": 0.20198350798078835, "learning_rate": 3.185152534994509e-05, "loss": 0.9005, "step": 6718 }, { "epoch": 3.044404168554599, "grad_norm": 0.16758588421244403, "learning_rate": 3.18391374133375e-05, "loss": 0.9408, "step": 6719 }, { "epoch": 3.0448572723153604, "grad_norm": 0.16189303673778185, "learning_rate": 3.182675029340601e-05, "loss": 0.9232, "step": 6720 }, { "epoch": 3.0453103760761215, "grad_norm": 0.1850153141270898, "learning_rate": 3.1814363991390215e-05, "loss": 0.9465, "step": 6721 }, { "epoch": 3.0457634798368827, "grad_norm": 0.13434841898625238, "learning_rate": 3.180197850852965e-05, "loss": 0.9149, "step": 6722 }, { "epoch": 3.046216583597644, "grad_norm": 0.1551003225009602, "learning_rate": 3.1789593846063764e-05, "loss": 0.9389, "step": 6723 }, { "epoch": 3.046669687358405, "grad_norm": 0.17428239641276985, "learning_rate": 3.17772100052319e-05, "loss": 0.9327, "step": 6724 }, { "epoch": 3.0471227911191665, "grad_norm": 0.1451533222473835, "learning_rate": 3.176482698727335e-05, "loss": 0.9546, "step": 6725 }, { "epoch": 3.0475758948799276, "grad_norm": 0.18154404576155814, "learning_rate": 3.175244479342731e-05, "loss": 0.9329, "step": 6726 }, { "epoch": 3.0480289986406888, "grad_norm": 0.1700680223302758, "learning_rate": 3.17400634249329e-05, "loss": 0.942, "step": 6727 }, { "epoch": 3.04848210240145, "grad_norm": 0.1353758092316655, "learning_rate": 3.172768288302913e-05, "loss": 0.9228, "step": 6728 }, { "epoch": 3.048935206162211, "grad_norm": 0.17114575385137903, "learning_rate": 3.171530316895497e-05, "loss": 0.925, "step": 6729 }, { "epoch": 3.049388309922972, "grad_norm": 0.15618792373703808, "learning_rate": 3.170292428394927e-05, "loss": 0.9221, "step": 6730 }, { "epoch": 3.0498414136837337, "grad_norm": 0.13486181236401787, "learning_rate": 3.1690546229250826e-05, "loss": 0.9322, "step": 6731 }, { "epoch": 3.050294517444495, "grad_norm": 0.18913278284991183, "learning_rate": 3.167816900609834e-05, "loss": 0.9391, "step": 6732 }, { "epoch": 3.050747621205256, "grad_norm": 0.1536111279644896, "learning_rate": 3.166579261573042e-05, "loss": 0.9521, "step": 6733 }, { "epoch": 3.051200724966017, "grad_norm": 0.146365295220383, "learning_rate": 3.16534170593856e-05, "loss": 0.9305, "step": 6734 }, { "epoch": 3.0516538287267783, "grad_norm": 0.16772395588473035, "learning_rate": 3.164104233830234e-05, "loss": 0.9375, "step": 6735 }, { "epoch": 3.05210693248754, "grad_norm": 0.1562481250078232, "learning_rate": 3.1628668453719e-05, "loss": 0.9572, "step": 6736 }, { "epoch": 3.052560036248301, "grad_norm": 0.15248011760614613, "learning_rate": 3.161629540687385e-05, "loss": 0.9478, "step": 6737 }, { "epoch": 3.053013140009062, "grad_norm": 0.15349984309496403, "learning_rate": 3.160392319900511e-05, "loss": 0.9509, "step": 6738 }, { "epoch": 3.0534662437698232, "grad_norm": 0.13969126383757657, "learning_rate": 3.159155183135089e-05, "loss": 0.9312, "step": 6739 }, { "epoch": 3.0539193475305844, "grad_norm": 0.15655554051428439, "learning_rate": 3.157918130514921e-05, "loss": 0.9262, "step": 6740 }, { "epoch": 3.054372451291346, "grad_norm": 0.19479875517887765, "learning_rate": 3.156681162163802e-05, "loss": 0.9128, "step": 6741 }, { "epoch": 3.054825555052107, "grad_norm": 0.1671244389863229, "learning_rate": 3.155444278205519e-05, "loss": 0.9349, "step": 6742 }, { "epoch": 3.055278658812868, "grad_norm": 0.17036364510577703, "learning_rate": 3.154207478763851e-05, "loss": 0.9217, "step": 6743 }, { "epoch": 3.0557317625736293, "grad_norm": 0.19630485607701387, "learning_rate": 3.1529707639625644e-05, "loss": 0.9373, "step": 6744 }, { "epoch": 3.0561848663343905, "grad_norm": 0.18390347458699988, "learning_rate": 3.1517341339254215e-05, "loss": 0.9438, "step": 6745 }, { "epoch": 3.0566379700951516, "grad_norm": 0.14729348745650486, "learning_rate": 3.150497588776174e-05, "loss": 0.9474, "step": 6746 }, { "epoch": 3.057091073855913, "grad_norm": 0.167609238962297, "learning_rate": 3.1492611286385676e-05, "loss": 0.9634, "step": 6747 }, { "epoch": 3.0575441776166743, "grad_norm": 0.1815487064534789, "learning_rate": 3.148024753636335e-05, "loss": 0.9315, "step": 6748 }, { "epoch": 3.0579972813774354, "grad_norm": 0.17301651430080509, "learning_rate": 3.146788463893206e-05, "loss": 0.9269, "step": 6749 }, { "epoch": 3.0584503851381966, "grad_norm": 0.17302112265946856, "learning_rate": 3.145552259532896e-05, "loss": 0.9473, "step": 6750 }, { "epoch": 3.0589034888989577, "grad_norm": 0.16184476193762742, "learning_rate": 3.1443161406791175e-05, "loss": 0.9237, "step": 6751 }, { "epoch": 3.0593565926597193, "grad_norm": 0.15017894222846254, "learning_rate": 3.1430801074555705e-05, "loss": 0.939, "step": 6752 }, { "epoch": 3.0598096964204804, "grad_norm": 0.13538021703395767, "learning_rate": 3.141844159985947e-05, "loss": 0.947, "step": 6753 }, { "epoch": 3.0602628001812415, "grad_norm": 0.1682507409095479, "learning_rate": 3.140608298393932e-05, "loss": 0.9421, "step": 6754 }, { "epoch": 3.0607159039420027, "grad_norm": 0.18439920648373723, "learning_rate": 3.1393725228032e-05, "loss": 0.9169, "step": 6755 }, { "epoch": 3.061169007702764, "grad_norm": 0.1295631167915577, "learning_rate": 3.1381368333374194e-05, "loss": 0.9465, "step": 6756 }, { "epoch": 3.0616221114635254, "grad_norm": 0.1825605120164677, "learning_rate": 3.136901230120247e-05, "loss": 0.9351, "step": 6757 }, { "epoch": 3.0620752152242865, "grad_norm": 0.20927900835456115, "learning_rate": 3.135665713275334e-05, "loss": 0.9342, "step": 6758 }, { "epoch": 3.0625283189850476, "grad_norm": 0.14552939488944863, "learning_rate": 3.134430282926319e-05, "loss": 0.9215, "step": 6759 }, { "epoch": 3.0629814227458088, "grad_norm": 0.1809394636100303, "learning_rate": 3.133194939196838e-05, "loss": 0.9167, "step": 6760 }, { "epoch": 3.06343452650657, "grad_norm": 0.22955471294707808, "learning_rate": 3.1319596822105106e-05, "loss": 0.9291, "step": 6761 }, { "epoch": 3.063887630267331, "grad_norm": 0.2156941744566662, "learning_rate": 3.130724512090954e-05, "loss": 0.8995, "step": 6762 }, { "epoch": 3.0643407340280926, "grad_norm": 0.19599338223607038, "learning_rate": 3.129489428961774e-05, "loss": 0.9105, "step": 6763 }, { "epoch": 3.0647938377888537, "grad_norm": 0.12882018821769706, "learning_rate": 3.128254432946568e-05, "loss": 0.9216, "step": 6764 }, { "epoch": 3.065246941549615, "grad_norm": 0.22654891643271186, "learning_rate": 3.1270195241689254e-05, "loss": 0.9409, "step": 6765 }, { "epoch": 3.065700045310376, "grad_norm": 0.22298689281209724, "learning_rate": 3.125784702752427e-05, "loss": 0.9085, "step": 6766 }, { "epoch": 3.066153149071137, "grad_norm": 0.1548538036456685, "learning_rate": 3.124549968820642e-05, "loss": 0.9551, "step": 6767 }, { "epoch": 3.0666062528318987, "grad_norm": 0.1656683395812958, "learning_rate": 3.123315322497135e-05, "loss": 0.9442, "step": 6768 }, { "epoch": 3.06705935659266, "grad_norm": 0.2108620453712115, "learning_rate": 3.122080763905459e-05, "loss": 0.9331, "step": 6769 }, { "epoch": 3.067512460353421, "grad_norm": 0.23758864305880517, "learning_rate": 3.120846293169159e-05, "loss": 0.9398, "step": 6770 }, { "epoch": 3.067965564114182, "grad_norm": 0.1842402286676886, "learning_rate": 3.119611910411771e-05, "loss": 0.9141, "step": 6771 }, { "epoch": 3.068418667874943, "grad_norm": 0.15288444982543445, "learning_rate": 3.118377615756824e-05, "loss": 0.9403, "step": 6772 }, { "epoch": 3.0688717716357043, "grad_norm": 0.2346707535988224, "learning_rate": 3.1171434093278356e-05, "loss": 0.9373, "step": 6773 }, { "epoch": 3.069324875396466, "grad_norm": 0.2121777972914222, "learning_rate": 3.115909291248315e-05, "loss": 0.9192, "step": 6774 }, { "epoch": 3.069777979157227, "grad_norm": 0.15341545902443732, "learning_rate": 3.114675261641764e-05, "loss": 0.9505, "step": 6775 }, { "epoch": 3.070231082917988, "grad_norm": 0.17851286967287414, "learning_rate": 3.113441320631676e-05, "loss": 0.9472, "step": 6776 }, { "epoch": 3.0706841866787493, "grad_norm": 0.18426688104717026, "learning_rate": 3.112207468341532e-05, "loss": 0.9361, "step": 6777 }, { "epoch": 3.0711372904395104, "grad_norm": 0.17178816981562603, "learning_rate": 3.1109737048948075e-05, "loss": 0.9428, "step": 6778 }, { "epoch": 3.071590394200272, "grad_norm": 0.17030438018738908, "learning_rate": 3.109740030414968e-05, "loss": 0.9558, "step": 6779 }, { "epoch": 3.072043497961033, "grad_norm": 0.16794169264642844, "learning_rate": 3.10850644502547e-05, "loss": 0.9312, "step": 6780 }, { "epoch": 3.0724966017217943, "grad_norm": 0.13689226580406666, "learning_rate": 3.107272948849761e-05, "loss": 0.9279, "step": 6781 }, { "epoch": 3.0729497054825554, "grad_norm": 0.17071551802830748, "learning_rate": 3.10603954201128e-05, "loss": 0.9126, "step": 6782 }, { "epoch": 3.0734028092433165, "grad_norm": 0.1396009649375111, "learning_rate": 3.104806224633459e-05, "loss": 0.9273, "step": 6783 }, { "epoch": 3.073855913004078, "grad_norm": 0.13196201446560077, "learning_rate": 3.103572996839715e-05, "loss": 0.9385, "step": 6784 }, { "epoch": 3.0743090167648393, "grad_norm": 0.12712153710064747, "learning_rate": 3.1023398587534605e-05, "loss": 0.9437, "step": 6785 }, { "epoch": 3.0747621205256004, "grad_norm": 0.1567258836419047, "learning_rate": 3.1011068104981004e-05, "loss": 0.9252, "step": 6786 }, { "epoch": 3.0752152242863615, "grad_norm": 0.1462528600704233, "learning_rate": 3.099873852197028e-05, "loss": 0.9566, "step": 6787 }, { "epoch": 3.0756683280471226, "grad_norm": 0.14287809249445546, "learning_rate": 3.0986409839736284e-05, "loss": 0.9323, "step": 6788 }, { "epoch": 3.076121431807884, "grad_norm": 0.16770768822798865, "learning_rate": 3.097408205951276e-05, "loss": 0.9417, "step": 6789 }, { "epoch": 3.0765745355686454, "grad_norm": 0.16544615227213563, "learning_rate": 3.09617551825334e-05, "loss": 0.9379, "step": 6790 }, { "epoch": 3.0770276393294065, "grad_norm": 0.14048071492013858, "learning_rate": 3.0949429210031765e-05, "loss": 0.9154, "step": 6791 }, { "epoch": 3.0774807430901676, "grad_norm": 0.16893724345787864, "learning_rate": 3.093710414324135e-05, "loss": 0.934, "step": 6792 }, { "epoch": 3.0779338468509287, "grad_norm": 0.1521164895035623, "learning_rate": 3.092477998339555e-05, "loss": 0.9175, "step": 6793 }, { "epoch": 3.07838695061169, "grad_norm": 0.14760035456006762, "learning_rate": 3.091245673172768e-05, "loss": 0.9404, "step": 6794 }, { "epoch": 3.0788400543724515, "grad_norm": 0.14595661041161026, "learning_rate": 3.090013438947094e-05, "loss": 0.9505, "step": 6795 }, { "epoch": 3.0792931581332126, "grad_norm": 0.15947343504439995, "learning_rate": 3.088781295785848e-05, "loss": 0.9379, "step": 6796 }, { "epoch": 3.0797462618939737, "grad_norm": 0.15133954027204216, "learning_rate": 3.0875492438123304e-05, "loss": 0.9385, "step": 6797 }, { "epoch": 3.080199365654735, "grad_norm": 0.1376077480467302, "learning_rate": 3.086317283149837e-05, "loss": 0.9463, "step": 6798 }, { "epoch": 3.080652469415496, "grad_norm": 0.1556521535087603, "learning_rate": 3.085085413921653e-05, "loss": 0.948, "step": 6799 }, { "epoch": 3.0811055731762576, "grad_norm": 0.15156003763578568, "learning_rate": 3.083853636251054e-05, "loss": 0.9401, "step": 6800 }, { "epoch": 3.0815586769370187, "grad_norm": 0.1210892060783824, "learning_rate": 3.082621950261308e-05, "loss": 0.9345, "step": 6801 }, { "epoch": 3.08201178069778, "grad_norm": 0.16256269856732505, "learning_rate": 3.08139035607567e-05, "loss": 0.9215, "step": 6802 }, { "epoch": 3.082464884458541, "grad_norm": 0.2992518504414058, "learning_rate": 3.08015885381739e-05, "loss": 0.9671, "step": 6803 }, { "epoch": 3.082917988219302, "grad_norm": 0.143710039949148, "learning_rate": 3.078927443609709e-05, "loss": 0.9443, "step": 6804 }, { "epoch": 3.0833710919800636, "grad_norm": 0.1512265478491929, "learning_rate": 3.077696125575854e-05, "loss": 0.927, "step": 6805 }, { "epoch": 3.083824195740825, "grad_norm": 0.15959259638134074, "learning_rate": 3.076464899839047e-05, "loss": 0.9344, "step": 6806 }, { "epoch": 3.084277299501586, "grad_norm": 0.15445373822504416, "learning_rate": 3.075233766522499e-05, "loss": 0.9435, "step": 6807 }, { "epoch": 3.084730403262347, "grad_norm": 0.13765984229258202, "learning_rate": 3.074002725749414e-05, "loss": 0.9292, "step": 6808 }, { "epoch": 3.085183507023108, "grad_norm": 0.1678775048305443, "learning_rate": 3.072771777642983e-05, "loss": 0.9243, "step": 6809 }, { "epoch": 3.0856366107838693, "grad_norm": 0.156342507282808, "learning_rate": 3.071540922326391e-05, "loss": 0.928, "step": 6810 }, { "epoch": 3.086089714544631, "grad_norm": 0.15366693973018566, "learning_rate": 3.070310159922813e-05, "loss": 0.9377, "step": 6811 }, { "epoch": 3.086542818305392, "grad_norm": 0.1763071207218951, "learning_rate": 3.069079490555413e-05, "loss": 0.9113, "step": 6812 }, { "epoch": 3.086995922066153, "grad_norm": 0.15733116131859617, "learning_rate": 3.0678489143473474e-05, "loss": 0.9471, "step": 6813 }, { "epoch": 3.0874490258269143, "grad_norm": 0.15204286482520213, "learning_rate": 3.066618431421762e-05, "loss": 0.9189, "step": 6814 }, { "epoch": 3.0879021295876754, "grad_norm": 0.18506377328369036, "learning_rate": 3.065388041901795e-05, "loss": 0.9324, "step": 6815 }, { "epoch": 3.088355233348437, "grad_norm": 0.16551875536719576, "learning_rate": 3.064157745910574e-05, "loss": 0.9419, "step": 6816 }, { "epoch": 3.088808337109198, "grad_norm": 0.15550951273587477, "learning_rate": 3.0629275435712176e-05, "loss": 0.9437, "step": 6817 }, { "epoch": 3.0892614408699592, "grad_norm": 0.15452235290097402, "learning_rate": 3.061697435006835e-05, "loss": 0.9049, "step": 6818 }, { "epoch": 3.0897145446307204, "grad_norm": 0.13731077938279818, "learning_rate": 3.0604674203405256e-05, "loss": 0.9432, "step": 6819 }, { "epoch": 3.0901676483914815, "grad_norm": 0.15986755599731123, "learning_rate": 3.059237499695379e-05, "loss": 0.9137, "step": 6820 }, { "epoch": 3.090620752152243, "grad_norm": 0.17271255267508406, "learning_rate": 3.0580076731944795e-05, "loss": 0.9162, "step": 6821 }, { "epoch": 3.091073855913004, "grad_norm": 0.14408902730739173, "learning_rate": 3.056777940960895e-05, "loss": 0.9252, "step": 6822 }, { "epoch": 3.0915269596737653, "grad_norm": 0.13343019399481096, "learning_rate": 3.055548303117689e-05, "loss": 0.9179, "step": 6823 }, { "epoch": 3.0919800634345265, "grad_norm": 0.15777347332118377, "learning_rate": 3.054318759787913e-05, "loss": 0.9444, "step": 6824 }, { "epoch": 3.0924331671952876, "grad_norm": 0.14440270929134263, "learning_rate": 3.0530893110946114e-05, "loss": 0.9229, "step": 6825 }, { "epoch": 3.0928862709560487, "grad_norm": 0.13324526806502954, "learning_rate": 3.051859957160819e-05, "loss": 0.9302, "step": 6826 }, { "epoch": 3.0933393747168103, "grad_norm": 0.14843789922725512, "learning_rate": 3.0506306981095575e-05, "loss": 0.9562, "step": 6827 }, { "epoch": 3.0937924784775714, "grad_norm": 0.12816768536652456, "learning_rate": 3.0494015340638437e-05, "loss": 0.9197, "step": 6828 }, { "epoch": 3.0942455822383326, "grad_norm": 0.164432994061955, "learning_rate": 3.0481724651466828e-05, "loss": 0.9207, "step": 6829 }, { "epoch": 3.0946986859990937, "grad_norm": 0.1519734909254641, "learning_rate": 3.0469434914810687e-05, "loss": 0.9118, "step": 6830 }, { "epoch": 3.095151789759855, "grad_norm": 0.12777775861938068, "learning_rate": 3.0457146131899892e-05, "loss": 0.9321, "step": 6831 }, { "epoch": 3.0956048935206164, "grad_norm": 0.15424137886679748, "learning_rate": 3.0444858303964203e-05, "loss": 0.943, "step": 6832 }, { "epoch": 3.0960579972813775, "grad_norm": 0.12895214528571575, "learning_rate": 3.043257143223329e-05, "loss": 0.9328, "step": 6833 }, { "epoch": 3.0965111010421387, "grad_norm": 0.13522600007660499, "learning_rate": 3.0420285517936734e-05, "loss": 0.9313, "step": 6834 }, { "epoch": 3.0969642048029, "grad_norm": 0.14332355980271744, "learning_rate": 3.040800056230401e-05, "loss": 0.9355, "step": 6835 }, { "epoch": 3.097417308563661, "grad_norm": 0.14062604031569784, "learning_rate": 3.03957165665645e-05, "loss": 0.9236, "step": 6836 }, { "epoch": 3.0978704123244225, "grad_norm": 0.15299670869397433, "learning_rate": 3.0383433531947504e-05, "loss": 0.9352, "step": 6837 }, { "epoch": 3.0983235160851836, "grad_norm": 0.14038049752907827, "learning_rate": 3.0371151459682196e-05, "loss": 0.9543, "step": 6838 }, { "epoch": 3.0987766198459448, "grad_norm": 0.1864505302567306, "learning_rate": 3.0358870350997676e-05, "loss": 0.942, "step": 6839 }, { "epoch": 3.099229723606706, "grad_norm": 0.14360450707513522, "learning_rate": 3.034659020712295e-05, "loss": 0.9152, "step": 6840 }, { "epoch": 3.099682827367467, "grad_norm": 0.1551462690423586, "learning_rate": 3.033431102928693e-05, "loss": 0.9562, "step": 6841 }, { "epoch": 3.100135931128228, "grad_norm": 0.16855665925026403, "learning_rate": 3.032203281871838e-05, "loss": 0.927, "step": 6842 }, { "epoch": 3.1005890348889897, "grad_norm": 0.14265146309374752, "learning_rate": 3.0309755576646044e-05, "loss": 0.9263, "step": 6843 }, { "epoch": 3.101042138649751, "grad_norm": 0.17063221483882282, "learning_rate": 3.029747930429852e-05, "loss": 0.9398, "step": 6844 }, { "epoch": 3.101495242410512, "grad_norm": 0.14806212704456467, "learning_rate": 3.028520400290433e-05, "loss": 0.9237, "step": 6845 }, { "epoch": 3.101948346171273, "grad_norm": 0.1718072341742586, "learning_rate": 3.0272929673691885e-05, "loss": 0.9383, "step": 6846 }, { "epoch": 3.1024014499320343, "grad_norm": 0.1639578150367734, "learning_rate": 3.0260656317889506e-05, "loss": 0.92, "step": 6847 }, { "epoch": 3.102854553692796, "grad_norm": 0.1814660435510652, "learning_rate": 3.0248383936725422e-05, "loss": 0.9261, "step": 6848 }, { "epoch": 3.103307657453557, "grad_norm": 0.13203268790757736, "learning_rate": 3.0236112531427763e-05, "loss": 0.934, "step": 6849 }, { "epoch": 3.103760761214318, "grad_norm": 0.15729732529674545, "learning_rate": 3.0223842103224532e-05, "loss": 0.9339, "step": 6850 }, { "epoch": 3.1042138649750792, "grad_norm": 0.15457234427445155, "learning_rate": 3.0211572653343676e-05, "loss": 0.9181, "step": 6851 }, { "epoch": 3.1046669687358404, "grad_norm": 0.129073432652238, "learning_rate": 3.0199304183013027e-05, "loss": 0.921, "step": 6852 }, { "epoch": 3.105120072496602, "grad_norm": 0.13371049937799567, "learning_rate": 3.0187036693460317e-05, "loss": 0.9529, "step": 6853 }, { "epoch": 3.105573176257363, "grad_norm": 0.14970993749052672, "learning_rate": 3.017477018591318e-05, "loss": 0.9274, "step": 6854 }, { "epoch": 3.106026280018124, "grad_norm": 0.1420947390767925, "learning_rate": 3.016250466159916e-05, "loss": 0.9352, "step": 6855 }, { "epoch": 3.1064793837788853, "grad_norm": 0.15013709608082096, "learning_rate": 3.0150240121745676e-05, "loss": 0.935, "step": 6856 }, { "epoch": 3.1069324875396465, "grad_norm": 0.1663835471503635, "learning_rate": 3.0137976567580104e-05, "loss": 0.9615, "step": 6857 }, { "epoch": 3.1073855913004076, "grad_norm": 0.13769723938830825, "learning_rate": 3.0125714000329645e-05, "loss": 0.9243, "step": 6858 }, { "epoch": 3.107838695061169, "grad_norm": 0.12508341191678266, "learning_rate": 3.011345242122147e-05, "loss": 0.9147, "step": 6859 }, { "epoch": 3.1082917988219303, "grad_norm": 0.1551769956987318, "learning_rate": 3.0101191831482613e-05, "loss": 0.9362, "step": 6860 }, { "epoch": 3.1087449025826914, "grad_norm": 0.14752349651736824, "learning_rate": 3.008893223234002e-05, "loss": 0.9425, "step": 6861 }, { "epoch": 3.1091980063434526, "grad_norm": 0.12539749681408216, "learning_rate": 3.0076673625020535e-05, "loss": 0.9347, "step": 6862 }, { "epoch": 3.1096511101042137, "grad_norm": 0.13172276815840475, "learning_rate": 3.0064416010750908e-05, "loss": 0.9575, "step": 6863 }, { "epoch": 3.1101042138649753, "grad_norm": 0.14992923940122704, "learning_rate": 3.0052159390757794e-05, "loss": 0.9286, "step": 6864 }, { "epoch": 3.1105573176257364, "grad_norm": 0.14694297846353763, "learning_rate": 3.0039903766267734e-05, "loss": 0.9323, "step": 6865 }, { "epoch": 3.1110104213864975, "grad_norm": 0.1568392402939755, "learning_rate": 3.0027649138507167e-05, "loss": 0.9411, "step": 6866 }, { "epoch": 3.1114635251472587, "grad_norm": 0.14883448176241718, "learning_rate": 3.001539550870245e-05, "loss": 0.9407, "step": 6867 }, { "epoch": 3.11191662890802, "grad_norm": 0.16208076607257593, "learning_rate": 3.000314287807983e-05, "loss": 0.958, "step": 6868 }, { "epoch": 3.1123697326687814, "grad_norm": 0.22026144584444277, "learning_rate": 2.9990891247865455e-05, "loss": 0.944, "step": 6869 }, { "epoch": 3.1128228364295425, "grad_norm": 0.22797537600488382, "learning_rate": 2.997864061928538e-05, "loss": 0.9293, "step": 6870 }, { "epoch": 3.1132759401903036, "grad_norm": 0.20569387446397894, "learning_rate": 2.9966390993565545e-05, "loss": 0.9304, "step": 6871 }, { "epoch": 3.1137290439510648, "grad_norm": 0.13994820662750035, "learning_rate": 2.99541423719318e-05, "loss": 0.9339, "step": 6872 }, { "epoch": 3.114182147711826, "grad_norm": 0.16288230487917918, "learning_rate": 2.9941894755609905e-05, "loss": 0.9543, "step": 6873 }, { "epoch": 3.114635251472587, "grad_norm": 0.1787637863618634, "learning_rate": 2.9929648145825484e-05, "loss": 0.9161, "step": 6874 }, { "epoch": 3.1150883552333486, "grad_norm": 0.17596056336223057, "learning_rate": 2.9917402543804098e-05, "loss": 0.9366, "step": 6875 }, { "epoch": 3.1155414589941097, "grad_norm": 0.16268276771973358, "learning_rate": 2.990515795077119e-05, "loss": 0.9434, "step": 6876 }, { "epoch": 3.115994562754871, "grad_norm": 0.13614099918650638, "learning_rate": 2.9892914367952096e-05, "loss": 0.925, "step": 6877 }, { "epoch": 3.116447666515632, "grad_norm": 0.16334979452345483, "learning_rate": 2.9880671796572075e-05, "loss": 0.9182, "step": 6878 }, { "epoch": 3.116900770276393, "grad_norm": 0.20704399203279308, "learning_rate": 2.9868430237856256e-05, "loss": 0.9367, "step": 6879 }, { "epoch": 3.1173538740371547, "grad_norm": 0.16844318505739742, "learning_rate": 2.9856189693029686e-05, "loss": 0.9181, "step": 6880 }, { "epoch": 3.117806977797916, "grad_norm": 0.1548490350741845, "learning_rate": 2.984395016331731e-05, "loss": 0.9405, "step": 6881 }, { "epoch": 3.118260081558677, "grad_norm": 0.14011027241009422, "learning_rate": 2.9831711649943953e-05, "loss": 0.934, "step": 6882 }, { "epoch": 3.118713185319438, "grad_norm": 0.16920603452781713, "learning_rate": 2.9819474154134352e-05, "loss": 0.9232, "step": 6883 }, { "epoch": 3.119166289080199, "grad_norm": 0.14535649926398153, "learning_rate": 2.9807237677113147e-05, "loss": 0.9345, "step": 6884 }, { "epoch": 3.119619392840961, "grad_norm": 0.1409195178131932, "learning_rate": 2.9795002220104866e-05, "loss": 0.9208, "step": 6885 }, { "epoch": 3.120072496601722, "grad_norm": 0.15639480293552727, "learning_rate": 2.978276778433394e-05, "loss": 0.9229, "step": 6886 }, { "epoch": 3.120525600362483, "grad_norm": 0.16126282639372067, "learning_rate": 2.97705343710247e-05, "loss": 0.9422, "step": 6887 }, { "epoch": 3.120978704123244, "grad_norm": 0.16344507229744268, "learning_rate": 2.9758301981401367e-05, "loss": 0.9263, "step": 6888 }, { "epoch": 3.1214318078840053, "grad_norm": 0.14254342814816431, "learning_rate": 2.974607061668807e-05, "loss": 0.9375, "step": 6889 }, { "epoch": 3.1218849116447664, "grad_norm": 0.17437583296846013, "learning_rate": 2.973384027810883e-05, "loss": 0.9383, "step": 6890 }, { "epoch": 3.122338015405528, "grad_norm": 0.1767039800227301, "learning_rate": 2.9721610966887558e-05, "loss": 0.9222, "step": 6891 }, { "epoch": 3.122791119166289, "grad_norm": 0.13784717549084524, "learning_rate": 2.9709382684248066e-05, "loss": 0.9232, "step": 6892 }, { "epoch": 3.1232442229270503, "grad_norm": 0.1499263930639814, "learning_rate": 2.9697155431414074e-05, "loss": 0.96, "step": 6893 }, { "epoch": 3.1236973266878114, "grad_norm": 0.18177184028661425, "learning_rate": 2.9684929209609193e-05, "loss": 0.9389, "step": 6894 }, { "epoch": 3.1241504304485725, "grad_norm": 0.1928515892777653, "learning_rate": 2.9672704020056923e-05, "loss": 0.9072, "step": 6895 }, { "epoch": 3.124603534209334, "grad_norm": 0.17777512340241672, "learning_rate": 2.966047986398066e-05, "loss": 0.9129, "step": 6896 }, { "epoch": 3.1250566379700953, "grad_norm": 0.13580337877365525, "learning_rate": 2.964825674260372e-05, "loss": 0.9193, "step": 6897 }, { "epoch": 3.1255097417308564, "grad_norm": 0.16048258601236273, "learning_rate": 2.9636034657149308e-05, "loss": 0.9347, "step": 6898 }, { "epoch": 3.1259628454916175, "grad_norm": 0.16706694581914963, "learning_rate": 2.962381360884048e-05, "loss": 0.9426, "step": 6899 }, { "epoch": 3.1264159492523786, "grad_norm": 0.15345831141040947, "learning_rate": 2.9611593598900236e-05, "loss": 0.9438, "step": 6900 }, { "epoch": 3.1268690530131398, "grad_norm": 0.1382736473359578, "learning_rate": 2.9599374628551477e-05, "loss": 0.938, "step": 6901 }, { "epoch": 3.1273221567739014, "grad_norm": 0.1635013755667165, "learning_rate": 2.958715669901697e-05, "loss": 0.9452, "step": 6902 }, { "epoch": 3.1277752605346625, "grad_norm": 0.18388827181401687, "learning_rate": 2.9574939811519384e-05, "loss": 0.9576, "step": 6903 }, { "epoch": 3.1282283642954236, "grad_norm": 0.14831303209567784, "learning_rate": 2.956272396728131e-05, "loss": 0.9252, "step": 6904 }, { "epoch": 3.1286814680561847, "grad_norm": 0.14537790404613168, "learning_rate": 2.9550509167525193e-05, "loss": 0.9386, "step": 6905 }, { "epoch": 3.129134571816946, "grad_norm": 0.16822469787244743, "learning_rate": 2.9538295413473422e-05, "loss": 0.9379, "step": 6906 }, { "epoch": 3.1295876755777075, "grad_norm": 0.1556069233568947, "learning_rate": 2.9526082706348228e-05, "loss": 0.9155, "step": 6907 }, { "epoch": 3.1300407793384686, "grad_norm": 0.14682417584050092, "learning_rate": 2.9513871047371787e-05, "loss": 0.9294, "step": 6908 }, { "epoch": 3.1304938830992297, "grad_norm": 0.1629797567070521, "learning_rate": 2.950166043776614e-05, "loss": 0.9484, "step": 6909 }, { "epoch": 3.130946986859991, "grad_norm": 0.13416998256380896, "learning_rate": 2.9489450878753233e-05, "loss": 0.9397, "step": 6910 }, { "epoch": 3.131400090620752, "grad_norm": 0.16921977991164053, "learning_rate": 2.9477242371554893e-05, "loss": 0.9453, "step": 6911 }, { "epoch": 3.1318531943815135, "grad_norm": 0.1635493936480428, "learning_rate": 2.946503491739286e-05, "loss": 0.918, "step": 6912 }, { "epoch": 3.1323062981422747, "grad_norm": 0.15063759503815619, "learning_rate": 2.945282851748876e-05, "loss": 0.9334, "step": 6913 }, { "epoch": 3.132759401903036, "grad_norm": 0.13641442771626588, "learning_rate": 2.944062317306412e-05, "loss": 0.9521, "step": 6914 }, { "epoch": 3.133212505663797, "grad_norm": 0.13417792483839974, "learning_rate": 2.942841888534035e-05, "loss": 0.9487, "step": 6915 }, { "epoch": 3.133665609424558, "grad_norm": 0.1377528073246441, "learning_rate": 2.941621565553877e-05, "loss": 0.9174, "step": 6916 }, { "epoch": 3.1341187131853196, "grad_norm": 0.14967641791868905, "learning_rate": 2.940401348488058e-05, "loss": 0.9291, "step": 6917 }, { "epoch": 3.134571816946081, "grad_norm": 0.16249906638327216, "learning_rate": 2.9391812374586882e-05, "loss": 0.9258, "step": 6918 }, { "epoch": 3.135024920706842, "grad_norm": 0.15040716331788115, "learning_rate": 2.9379612325878662e-05, "loss": 0.9447, "step": 6919 }, { "epoch": 3.135478024467603, "grad_norm": 0.15785546617413138, "learning_rate": 2.9367413339976804e-05, "loss": 0.9402, "step": 6920 }, { "epoch": 3.135931128228364, "grad_norm": 0.18172467896571734, "learning_rate": 2.9355215418102098e-05, "loss": 0.9291, "step": 6921 }, { "epoch": 3.1363842319891253, "grad_norm": 0.16942862133184405, "learning_rate": 2.9343018561475216e-05, "loss": 0.938, "step": 6922 }, { "epoch": 3.136837335749887, "grad_norm": 0.13908473687159778, "learning_rate": 2.9330822771316724e-05, "loss": 0.947, "step": 6923 }, { "epoch": 3.137290439510648, "grad_norm": 0.16555754216606097, "learning_rate": 2.9318628048847076e-05, "loss": 0.942, "step": 6924 }, { "epoch": 3.137743543271409, "grad_norm": 0.1728604518762245, "learning_rate": 2.930643439528664e-05, "loss": 0.9467, "step": 6925 }, { "epoch": 3.1381966470321703, "grad_norm": 0.15765954903525886, "learning_rate": 2.9294241811855656e-05, "loss": 0.9191, "step": 6926 }, { "epoch": 3.1386497507929314, "grad_norm": 0.1257224986758833, "learning_rate": 2.9282050299774252e-05, "loss": 0.9253, "step": 6927 }, { "epoch": 3.139102854553693, "grad_norm": 0.1736109536520621, "learning_rate": 2.9269859860262477e-05, "loss": 0.948, "step": 6928 }, { "epoch": 3.139555958314454, "grad_norm": 0.17065483342273965, "learning_rate": 2.9257670494540248e-05, "loss": 0.9263, "step": 6929 }, { "epoch": 3.1400090620752152, "grad_norm": 0.17600905440643644, "learning_rate": 2.9245482203827387e-05, "loss": 0.96, "step": 6930 }, { "epoch": 3.1404621658359764, "grad_norm": 0.1504500635419245, "learning_rate": 2.9233294989343593e-05, "loss": 0.9236, "step": 6931 }, { "epoch": 3.1409152695967375, "grad_norm": 0.13322850430198604, "learning_rate": 2.922110885230848e-05, "loss": 0.9519, "step": 6932 }, { "epoch": 3.1413683733574986, "grad_norm": 0.12724388109685705, "learning_rate": 2.920892379394154e-05, "loss": 0.9337, "step": 6933 }, { "epoch": 3.14182147711826, "grad_norm": 0.16609141097548882, "learning_rate": 2.9196739815462164e-05, "loss": 0.9173, "step": 6934 }, { "epoch": 3.1422745808790213, "grad_norm": 0.17269716141188549, "learning_rate": 2.9184556918089614e-05, "loss": 0.9212, "step": 6935 }, { "epoch": 3.1427276846397825, "grad_norm": 0.14580406064991872, "learning_rate": 2.9172375103043076e-05, "loss": 0.928, "step": 6936 }, { "epoch": 3.1431807884005436, "grad_norm": 0.12394476159066317, "learning_rate": 2.9160194371541606e-05, "loss": 0.9293, "step": 6937 }, { "epoch": 3.1436338921613047, "grad_norm": 0.16095845662884617, "learning_rate": 2.9148014724804156e-05, "loss": 0.9737, "step": 6938 }, { "epoch": 3.1440869959220663, "grad_norm": 0.1307576056540713, "learning_rate": 2.9135836164049572e-05, "loss": 0.916, "step": 6939 }, { "epoch": 3.1445400996828274, "grad_norm": 0.12545191235694655, "learning_rate": 2.9123658690496595e-05, "loss": 0.9441, "step": 6940 }, { "epoch": 3.1449932034435886, "grad_norm": 0.12653126132338935, "learning_rate": 2.9111482305363846e-05, "loss": 0.9016, "step": 6941 }, { "epoch": 3.1454463072043497, "grad_norm": 0.1463384240529873, "learning_rate": 2.909930700986986e-05, "loss": 0.9222, "step": 6942 }, { "epoch": 3.145899410965111, "grad_norm": 0.16661859979420057, "learning_rate": 2.908713280523302e-05, "loss": 0.9211, "step": 6943 }, { "epoch": 3.1463525147258724, "grad_norm": 0.17256002530462716, "learning_rate": 2.907495969267164e-05, "loss": 0.9183, "step": 6944 }, { "epoch": 3.1468056184866335, "grad_norm": 0.15234626560499023, "learning_rate": 2.906278767340391e-05, "loss": 0.9385, "step": 6945 }, { "epoch": 3.1472587222473947, "grad_norm": 0.13020263724291922, "learning_rate": 2.905061674864792e-05, "loss": 0.9217, "step": 6946 }, { "epoch": 3.147711826008156, "grad_norm": 0.1410940960388908, "learning_rate": 2.9038446919621627e-05, "loss": 0.9493, "step": 6947 }, { "epoch": 3.148164929768917, "grad_norm": 0.15679896729979734, "learning_rate": 2.9026278187542903e-05, "loss": 0.946, "step": 6948 }, { "epoch": 3.1486180335296785, "grad_norm": 0.14476178121851066, "learning_rate": 2.9014110553629494e-05, "loss": 0.9225, "step": 6949 }, { "epoch": 3.1490711372904396, "grad_norm": 0.17133802224046868, "learning_rate": 2.9001944019099056e-05, "loss": 0.9304, "step": 6950 }, { "epoch": 3.1495242410512008, "grad_norm": 0.13634130916083373, "learning_rate": 2.8989778585169115e-05, "loss": 0.9491, "step": 6951 }, { "epoch": 3.149977344811962, "grad_norm": 0.17906635798438705, "learning_rate": 2.8977614253057087e-05, "loss": 0.9369, "step": 6952 }, { "epoch": 3.150430448572723, "grad_norm": 0.14583142459175455, "learning_rate": 2.8965451023980287e-05, "loss": 0.9432, "step": 6953 }, { "epoch": 3.150883552333484, "grad_norm": 0.1684601673656105, "learning_rate": 2.8953288899155923e-05, "loss": 0.9249, "step": 6954 }, { "epoch": 3.1513366560942457, "grad_norm": 0.15778087337425667, "learning_rate": 2.8941127879801095e-05, "loss": 0.924, "step": 6955 }, { "epoch": 3.151789759855007, "grad_norm": 0.15944902243095962, "learning_rate": 2.8928967967132763e-05, "loss": 0.9341, "step": 6956 }, { "epoch": 3.152242863615768, "grad_norm": 0.13991724640737996, "learning_rate": 2.8916809162367802e-05, "loss": 0.9387, "step": 6957 }, { "epoch": 3.152695967376529, "grad_norm": 0.15397280258729587, "learning_rate": 2.890465146672297e-05, "loss": 0.9279, "step": 6958 }, { "epoch": 3.1531490711372903, "grad_norm": 0.17518489347169686, "learning_rate": 2.8892494881414924e-05, "loss": 0.9428, "step": 6959 }, { "epoch": 3.153602174898052, "grad_norm": 0.14431817250188803, "learning_rate": 2.8880339407660202e-05, "loss": 0.9202, "step": 6960 }, { "epoch": 3.154055278658813, "grad_norm": 0.12541909294793407, "learning_rate": 2.886818504667522e-05, "loss": 0.94, "step": 6961 }, { "epoch": 3.154508382419574, "grad_norm": 0.14382587300755248, "learning_rate": 2.885603179967631e-05, "loss": 0.9568, "step": 6962 }, { "epoch": 3.1549614861803352, "grad_norm": 0.15370216312300886, "learning_rate": 2.8843879667879647e-05, "loss": 0.9427, "step": 6963 }, { "epoch": 3.1554145899410964, "grad_norm": 0.15125513349771344, "learning_rate": 2.883172865250135e-05, "loss": 0.9277, "step": 6964 }, { "epoch": 3.1558676937018575, "grad_norm": 0.13903877871142453, "learning_rate": 2.8819578754757382e-05, "loss": 0.9425, "step": 6965 }, { "epoch": 3.156320797462619, "grad_norm": 0.17643629712130593, "learning_rate": 2.880742997586362e-05, "loss": 0.9376, "step": 6966 }, { "epoch": 3.15677390122338, "grad_norm": 0.14678075688802286, "learning_rate": 2.8795282317035805e-05, "loss": 0.9369, "step": 6967 }, { "epoch": 3.1572270049841413, "grad_norm": 0.15109021151332774, "learning_rate": 2.8783135779489603e-05, "loss": 0.9219, "step": 6968 }, { "epoch": 3.1576801087449025, "grad_norm": 0.14123606620159077, "learning_rate": 2.8770990364440536e-05, "loss": 0.9291, "step": 6969 }, { "epoch": 3.1581332125056636, "grad_norm": 0.1531849646891924, "learning_rate": 2.8758846073104025e-05, "loss": 0.9171, "step": 6970 }, { "epoch": 3.158586316266425, "grad_norm": 0.2037532382294448, "learning_rate": 2.874670290669537e-05, "loss": 0.9215, "step": 6971 }, { "epoch": 3.1590394200271863, "grad_norm": 0.20443792482218726, "learning_rate": 2.8734560866429766e-05, "loss": 0.9391, "step": 6972 }, { "epoch": 3.1594925237879474, "grad_norm": 0.1471317873805296, "learning_rate": 2.87224199535223e-05, "loss": 0.9138, "step": 6973 }, { "epoch": 3.1599456275487086, "grad_norm": 0.16828371118546728, "learning_rate": 2.8710280169187937e-05, "loss": 0.9244, "step": 6974 }, { "epoch": 3.1603987313094697, "grad_norm": 0.16725538806319812, "learning_rate": 2.8698141514641533e-05, "loss": 0.9502, "step": 6975 }, { "epoch": 3.1608518350702313, "grad_norm": 0.1929825468615082, "learning_rate": 2.8686003991097837e-05, "loss": 0.9316, "step": 6976 }, { "epoch": 3.1613049388309924, "grad_norm": 0.19420910775571693, "learning_rate": 2.867386759977147e-05, "loss": 0.9269, "step": 6977 }, { "epoch": 3.1617580425917535, "grad_norm": 0.1838261008351295, "learning_rate": 2.8661732341876956e-05, "loss": 0.9454, "step": 6978 }, { "epoch": 3.1622111463525147, "grad_norm": 0.1541592020592411, "learning_rate": 2.8649598218628703e-05, "loss": 0.9475, "step": 6979 }, { "epoch": 3.162664250113276, "grad_norm": 0.18775902305109543, "learning_rate": 2.863746523124098e-05, "loss": 0.9218, "step": 6980 }, { "epoch": 3.1631173538740374, "grad_norm": 0.19269073181964536, "learning_rate": 2.8625333380927974e-05, "loss": 0.9401, "step": 6981 }, { "epoch": 3.1635704576347985, "grad_norm": 0.14711532224450682, "learning_rate": 2.861320266890375e-05, "loss": 0.9251, "step": 6982 }, { "epoch": 3.1640235613955596, "grad_norm": 0.19961676749278204, "learning_rate": 2.860107309638225e-05, "loss": 0.9271, "step": 6983 }, { "epoch": 3.1644766651563208, "grad_norm": 0.25513022831764093, "learning_rate": 2.858894466457731e-05, "loss": 0.9566, "step": 6984 }, { "epoch": 3.164929768917082, "grad_norm": 0.2093272632179096, "learning_rate": 2.8576817374702653e-05, "loss": 0.931, "step": 6985 }, { "epoch": 3.165382872677843, "grad_norm": 0.1430246645788372, "learning_rate": 2.8564691227971884e-05, "loss": 0.9471, "step": 6986 }, { "epoch": 3.1658359764386046, "grad_norm": 0.15733202259523205, "learning_rate": 2.85525662255985e-05, "loss": 0.9129, "step": 6987 }, { "epoch": 3.1662890801993657, "grad_norm": 0.1678133843434852, "learning_rate": 2.854044236879586e-05, "loss": 0.9278, "step": 6988 }, { "epoch": 3.166742183960127, "grad_norm": 0.17769969175264066, "learning_rate": 2.852831965877724e-05, "loss": 0.9589, "step": 6989 }, { "epoch": 3.167195287720888, "grad_norm": 0.174080729909305, "learning_rate": 2.851619809675578e-05, "loss": 0.9521, "step": 6990 }, { "epoch": 3.167648391481649, "grad_norm": 0.13489413950515905, "learning_rate": 2.850407768394451e-05, "loss": 0.9276, "step": 6991 }, { "epoch": 3.1681014952424107, "grad_norm": 0.15849767233287762, "learning_rate": 2.849195842155636e-05, "loss": 0.9475, "step": 6992 }, { "epoch": 3.168554599003172, "grad_norm": 0.17621423739208364, "learning_rate": 2.8479840310804124e-05, "loss": 0.928, "step": 6993 }, { "epoch": 3.169007702763933, "grad_norm": 0.1387651929768513, "learning_rate": 2.846772335290049e-05, "loss": 0.9225, "step": 6994 }, { "epoch": 3.169460806524694, "grad_norm": 0.13410487147558364, "learning_rate": 2.8455607549058037e-05, "loss": 0.9171, "step": 6995 }, { "epoch": 3.169913910285455, "grad_norm": 0.20305495930256384, "learning_rate": 2.8443492900489204e-05, "loss": 0.9425, "step": 6996 }, { "epoch": 3.170367014046217, "grad_norm": 0.20583403817902265, "learning_rate": 2.8431379408406337e-05, "loss": 0.9306, "step": 6997 }, { "epoch": 3.170820117806978, "grad_norm": 0.14578533758396336, "learning_rate": 2.841926707402167e-05, "loss": 0.9366, "step": 6998 }, { "epoch": 3.171273221567739, "grad_norm": 0.1985587176773611, "learning_rate": 2.840715589854731e-05, "loss": 0.9097, "step": 6999 }, { "epoch": 3.1717263253285, "grad_norm": 0.19430860569573918, "learning_rate": 2.8395045883195238e-05, "loss": 0.9313, "step": 7000 }, { "epoch": 3.1721794290892613, "grad_norm": 0.18409083175668556, "learning_rate": 2.838293702917734e-05, "loss": 0.918, "step": 7001 }, { "epoch": 3.172632532850023, "grad_norm": 0.16268512387234918, "learning_rate": 2.837082933770538e-05, "loss": 0.931, "step": 7002 }, { "epoch": 3.173085636610784, "grad_norm": 0.15597032881678013, "learning_rate": 2.8358722809991007e-05, "loss": 0.9348, "step": 7003 }, { "epoch": 3.173538740371545, "grad_norm": 0.20576291237475156, "learning_rate": 2.834661744724573e-05, "loss": 0.9435, "step": 7004 }, { "epoch": 3.1739918441323063, "grad_norm": 0.19342694439067304, "learning_rate": 2.8334513250680973e-05, "loss": 0.9371, "step": 7005 }, { "epoch": 3.1744449478930674, "grad_norm": 0.13628805499752994, "learning_rate": 2.8322410221508022e-05, "loss": 0.9378, "step": 7006 }, { "epoch": 3.1748980516538285, "grad_norm": 0.18483245208025115, "learning_rate": 2.831030836093806e-05, "loss": 0.9331, "step": 7007 }, { "epoch": 3.17535115541459, "grad_norm": 0.1789321840742733, "learning_rate": 2.8298207670182153e-05, "loss": 0.949, "step": 7008 }, { "epoch": 3.1758042591753513, "grad_norm": 0.1656672681637774, "learning_rate": 2.8286108150451234e-05, "loss": 0.9361, "step": 7009 }, { "epoch": 3.1762573629361124, "grad_norm": 0.19145593719730106, "learning_rate": 2.8274009802956136e-05, "loss": 0.9412, "step": 7010 }, { "epoch": 3.1767104666968735, "grad_norm": 0.18108342658231566, "learning_rate": 2.8261912628907578e-05, "loss": 0.9305, "step": 7011 }, { "epoch": 3.1771635704576346, "grad_norm": 0.16003010781595264, "learning_rate": 2.824981662951614e-05, "loss": 0.948, "step": 7012 }, { "epoch": 3.177616674218396, "grad_norm": 0.18455461505676996, "learning_rate": 2.823772180599229e-05, "loss": 0.9441, "step": 7013 }, { "epoch": 3.1780697779791574, "grad_norm": 0.13677793056000614, "learning_rate": 2.8225628159546394e-05, "loss": 0.9343, "step": 7014 }, { "epoch": 3.1785228817399185, "grad_norm": 0.19339924411090037, "learning_rate": 2.821353569138869e-05, "loss": 0.9364, "step": 7015 }, { "epoch": 3.1789759855006796, "grad_norm": 0.2019506975899863, "learning_rate": 2.8201444402729293e-05, "loss": 0.9419, "step": 7016 }, { "epoch": 3.1794290892614407, "grad_norm": 0.15195042265761052, "learning_rate": 2.8189354294778212e-05, "loss": 0.9312, "step": 7017 }, { "epoch": 3.179882193022202, "grad_norm": 0.18471331094012763, "learning_rate": 2.8177265368745327e-05, "loss": 0.9234, "step": 7018 }, { "epoch": 3.1803352967829635, "grad_norm": 0.13775053158119283, "learning_rate": 2.8165177625840406e-05, "loss": 0.9156, "step": 7019 }, { "epoch": 3.1807884005437246, "grad_norm": 0.20091151678148464, "learning_rate": 2.81530910672731e-05, "loss": 0.9219, "step": 7020 }, { "epoch": 3.1812415043044857, "grad_norm": 0.14462202712971325, "learning_rate": 2.814100569425293e-05, "loss": 0.9197, "step": 7021 }, { "epoch": 3.181694608065247, "grad_norm": 0.1617398634061412, "learning_rate": 2.8128921507989323e-05, "loss": 0.9193, "step": 7022 }, { "epoch": 3.182147711826008, "grad_norm": 0.18917207836800567, "learning_rate": 2.811683850969156e-05, "loss": 0.9342, "step": 7023 }, { "epoch": 3.1826008155867695, "grad_norm": 0.20004217799475943, "learning_rate": 2.8104756700568816e-05, "loss": 0.9267, "step": 7024 }, { "epoch": 3.1830539193475307, "grad_norm": 0.18508858245327536, "learning_rate": 2.8092676081830136e-05, "loss": 0.9107, "step": 7025 }, { "epoch": 3.183507023108292, "grad_norm": 0.14764010542671852, "learning_rate": 2.808059665468447e-05, "loss": 0.9319, "step": 7026 }, { "epoch": 3.183960126869053, "grad_norm": 0.2085865991231278, "learning_rate": 2.8068518420340622e-05, "loss": 0.9236, "step": 7027 }, { "epoch": 3.184413230629814, "grad_norm": 0.20414869855963733, "learning_rate": 2.80564413800073e-05, "loss": 0.9023, "step": 7028 }, { "epoch": 3.1848663343905756, "grad_norm": 0.1459128436836094, "learning_rate": 2.804436553489307e-05, "loss": 0.9287, "step": 7029 }, { "epoch": 3.185319438151337, "grad_norm": 0.15242897227495164, "learning_rate": 2.8032290886206393e-05, "loss": 0.9388, "step": 7030 }, { "epoch": 3.185772541912098, "grad_norm": 0.14937693355994042, "learning_rate": 2.802021743515562e-05, "loss": 0.9365, "step": 7031 }, { "epoch": 3.186225645672859, "grad_norm": 0.1507699263441524, "learning_rate": 2.8008145182948942e-05, "loss": 0.9401, "step": 7032 }, { "epoch": 3.18667874943362, "grad_norm": 0.15292668999766407, "learning_rate": 2.7996074130794477e-05, "loss": 0.9401, "step": 7033 }, { "epoch": 3.1871318531943817, "grad_norm": 0.12646597777801757, "learning_rate": 2.7984004279900197e-05, "loss": 0.9341, "step": 7034 }, { "epoch": 3.187584956955143, "grad_norm": 0.1292056137056216, "learning_rate": 2.7971935631473953e-05, "loss": 0.9223, "step": 7035 }, { "epoch": 3.188038060715904, "grad_norm": 0.13650510584438275, "learning_rate": 2.795986818672349e-05, "loss": 0.9258, "step": 7036 }, { "epoch": 3.188491164476665, "grad_norm": 0.13489487038153977, "learning_rate": 2.7947801946856424e-05, "loss": 0.9345, "step": 7037 }, { "epoch": 3.1889442682374263, "grad_norm": 0.1303767183742804, "learning_rate": 2.7935736913080246e-05, "loss": 0.9292, "step": 7038 }, { "epoch": 3.1893973719981874, "grad_norm": 0.14540921963947903, "learning_rate": 2.792367308660234e-05, "loss": 0.9463, "step": 7039 }, { "epoch": 3.189850475758949, "grad_norm": 0.14307936059905854, "learning_rate": 2.7911610468629966e-05, "loss": 0.9331, "step": 7040 }, { "epoch": 3.19030357951971, "grad_norm": 0.1429447131102895, "learning_rate": 2.7899549060370237e-05, "loss": 0.9418, "step": 7041 }, { "epoch": 3.1907566832804712, "grad_norm": 0.14711158418998857, "learning_rate": 2.7887488863030175e-05, "loss": 0.9258, "step": 7042 }, { "epoch": 3.1912097870412324, "grad_norm": 0.12771918707976648, "learning_rate": 2.7875429877816673e-05, "loss": 0.9244, "step": 7043 }, { "epoch": 3.1916628908019935, "grad_norm": 0.13689946746383835, "learning_rate": 2.7863372105936498e-05, "loss": 0.9334, "step": 7044 }, { "epoch": 3.192115994562755, "grad_norm": 0.15677138915544214, "learning_rate": 2.7851315548596304e-05, "loss": 0.9073, "step": 7045 }, { "epoch": 3.192569098323516, "grad_norm": 0.33029572290145487, "learning_rate": 2.7839260207002615e-05, "loss": 0.9449, "step": 7046 }, { "epoch": 3.1930222020842773, "grad_norm": 0.14858729233316356, "learning_rate": 2.7827206082361838e-05, "loss": 0.9433, "step": 7047 }, { "epoch": 3.1934753058450385, "grad_norm": 0.16208605734692988, "learning_rate": 2.781515317588026e-05, "loss": 0.9324, "step": 7048 }, { "epoch": 3.1939284096057996, "grad_norm": 0.14043289358147443, "learning_rate": 2.7803101488764036e-05, "loss": 0.9109, "step": 7049 }, { "epoch": 3.1943815133665607, "grad_norm": 0.15065031850880978, "learning_rate": 2.7791051022219204e-05, "loss": 0.9214, "step": 7050 }, { "epoch": 3.1948346171273223, "grad_norm": 0.163566411172961, "learning_rate": 2.777900177745169e-05, "loss": 0.9264, "step": 7051 }, { "epoch": 3.1952877208880834, "grad_norm": 0.15090336406578161, "learning_rate": 2.7766953755667273e-05, "loss": 0.932, "step": 7052 }, { "epoch": 3.1957408246488446, "grad_norm": 0.12659497340741752, "learning_rate": 2.775490695807165e-05, "loss": 0.9382, "step": 7053 }, { "epoch": 3.1961939284096057, "grad_norm": 0.14574361745919595, "learning_rate": 2.7742861385870353e-05, "loss": 0.9583, "step": 7054 }, { "epoch": 3.196647032170367, "grad_norm": 0.14117756108108395, "learning_rate": 2.7730817040268823e-05, "loss": 0.9461, "step": 7055 }, { "epoch": 3.1971001359311284, "grad_norm": 0.13187683668586309, "learning_rate": 2.771877392247236e-05, "loss": 0.9534, "step": 7056 }, { "epoch": 3.1975532396918895, "grad_norm": 0.1430940856222342, "learning_rate": 2.7706732033686142e-05, "loss": 0.9449, "step": 7057 }, { "epoch": 3.1980063434526507, "grad_norm": 0.1509082871384907, "learning_rate": 2.769469137511523e-05, "loss": 0.9139, "step": 7058 }, { "epoch": 3.198459447213412, "grad_norm": 0.15966784169786746, "learning_rate": 2.7682651947964557e-05, "loss": 0.9516, "step": 7059 }, { "epoch": 3.198912550974173, "grad_norm": 0.14157671358236587, "learning_rate": 2.767061375343894e-05, "loss": 0.9043, "step": 7060 }, { "epoch": 3.1993656547349345, "grad_norm": 0.14674078904676938, "learning_rate": 2.765857679274307e-05, "loss": 0.9319, "step": 7061 }, { "epoch": 3.1998187584956956, "grad_norm": 0.1688002278899432, "learning_rate": 2.7646541067081512e-05, "loss": 0.9332, "step": 7062 }, { "epoch": 3.2002718622564568, "grad_norm": 0.1563469753976682, "learning_rate": 2.7634506577658707e-05, "loss": 0.9396, "step": 7063 }, { "epoch": 3.200724966017218, "grad_norm": 0.15090424303786149, "learning_rate": 2.7622473325678982e-05, "loss": 0.923, "step": 7064 }, { "epoch": 3.201178069777979, "grad_norm": 0.16258433881812412, "learning_rate": 2.7610441312346518e-05, "loss": 0.9414, "step": 7065 }, { "epoch": 3.2016311735387406, "grad_norm": 0.15056353698174474, "learning_rate": 2.7598410538865394e-05, "loss": 0.922, "step": 7066 }, { "epoch": 3.2020842772995017, "grad_norm": 0.14599029496017799, "learning_rate": 2.758638100643955e-05, "loss": 0.9276, "step": 7067 }, { "epoch": 3.202537381060263, "grad_norm": 0.32035230384720526, "learning_rate": 2.757435271627282e-05, "loss": 0.9301, "step": 7068 }, { "epoch": 3.202990484821024, "grad_norm": 0.17427435587266807, "learning_rate": 2.7562325669568913e-05, "loss": 0.9349, "step": 7069 }, { "epoch": 3.203443588581785, "grad_norm": 0.25529424964943237, "learning_rate": 2.755029986753136e-05, "loss": 0.9462, "step": 7070 }, { "epoch": 3.2038966923425463, "grad_norm": 0.13032035814456372, "learning_rate": 2.753827531136365e-05, "loss": 0.918, "step": 7071 }, { "epoch": 3.204349796103308, "grad_norm": 0.14887281465602542, "learning_rate": 2.7526252002269082e-05, "loss": 0.9615, "step": 7072 }, { "epoch": 3.204802899864069, "grad_norm": 0.16082390204996366, "learning_rate": 2.751422994145088e-05, "loss": 0.9474, "step": 7073 }, { "epoch": 3.20525600362483, "grad_norm": 0.14205226803743337, "learning_rate": 2.7502209130112098e-05, "loss": 0.9555, "step": 7074 }, { "epoch": 3.2057091073855912, "grad_norm": 0.1327123419200534, "learning_rate": 2.74901895694557e-05, "loss": 0.9513, "step": 7075 }, { "epoch": 3.2061622111463524, "grad_norm": 0.17950325822549837, "learning_rate": 2.7478171260684515e-05, "loss": 0.9406, "step": 7076 }, { "epoch": 3.206615314907114, "grad_norm": 0.2012478417420713, "learning_rate": 2.7466154205001223e-05, "loss": 0.9411, "step": 7077 }, { "epoch": 3.207068418667875, "grad_norm": 0.17200351453026716, "learning_rate": 2.7454138403608407e-05, "loss": 0.9264, "step": 7078 }, { "epoch": 3.207521522428636, "grad_norm": 0.15963231452786542, "learning_rate": 2.744212385770852e-05, "loss": 0.9277, "step": 7079 }, { "epoch": 3.2079746261893973, "grad_norm": 0.1775841869198814, "learning_rate": 2.7430110568503887e-05, "loss": 0.9198, "step": 7080 }, { "epoch": 3.2084277299501585, "grad_norm": 0.1791415296329042, "learning_rate": 2.7418098537196693e-05, "loss": 0.9451, "step": 7081 }, { "epoch": 3.2088808337109196, "grad_norm": 0.14934536390043138, "learning_rate": 2.7406087764989017e-05, "loss": 0.943, "step": 7082 }, { "epoch": 3.209333937471681, "grad_norm": 0.13929631420389918, "learning_rate": 2.7394078253082805e-05, "loss": 0.9403, "step": 7083 }, { "epoch": 3.2097870412324423, "grad_norm": 0.15197042561585986, "learning_rate": 2.7382070002679885e-05, "loss": 0.9368, "step": 7084 }, { "epoch": 3.2102401449932034, "grad_norm": 0.15958290096958988, "learning_rate": 2.737006301498194e-05, "loss": 0.9347, "step": 7085 }, { "epoch": 3.2106932487539646, "grad_norm": 0.11664934871672387, "learning_rate": 2.7358057291190524e-05, "loss": 0.9375, "step": 7086 }, { "epoch": 3.2111463525147257, "grad_norm": 0.155414706219392, "learning_rate": 2.7346052832507096e-05, "loss": 0.9152, "step": 7087 }, { "epoch": 3.2115994562754873, "grad_norm": 0.19325870443840618, "learning_rate": 2.7334049640132963e-05, "loss": 0.9283, "step": 7088 }, { "epoch": 3.2120525600362484, "grad_norm": 0.1619208956605121, "learning_rate": 2.7322047715269307e-05, "loss": 0.9355, "step": 7089 }, { "epoch": 3.2125056637970095, "grad_norm": 0.1222559944600719, "learning_rate": 2.73100470591172e-05, "loss": 0.9341, "step": 7090 }, { "epoch": 3.2129587675577707, "grad_norm": 0.15621806546806077, "learning_rate": 2.7298047672877566e-05, "loss": 0.9313, "step": 7091 }, { "epoch": 3.213411871318532, "grad_norm": 0.18031333149190573, "learning_rate": 2.7286049557751223e-05, "loss": 0.9112, "step": 7092 }, { "epoch": 3.2138649750792934, "grad_norm": 0.1664266684454254, "learning_rate": 2.727405271493883e-05, "loss": 0.9258, "step": 7093 }, { "epoch": 3.2143180788400545, "grad_norm": 0.13555930044729908, "learning_rate": 2.7262057145640946e-05, "loss": 0.9451, "step": 7094 }, { "epoch": 3.2147711826008156, "grad_norm": 0.1277182397071495, "learning_rate": 2.7250062851057998e-05, "loss": 0.9493, "step": 7095 }, { "epoch": 3.2152242863615768, "grad_norm": 0.17231484124670746, "learning_rate": 2.723806983239028e-05, "loss": 0.9181, "step": 7096 }, { "epoch": 3.215677390122338, "grad_norm": 0.1579871062379384, "learning_rate": 2.7226078090837965e-05, "loss": 0.9211, "step": 7097 }, { "epoch": 3.2161304938830995, "grad_norm": 0.3031910508255769, "learning_rate": 2.721408762760109e-05, "loss": 0.9752, "step": 7098 }, { "epoch": 3.2165835976438606, "grad_norm": 0.18174698335462267, "learning_rate": 2.720209844387957e-05, "loss": 0.959, "step": 7099 }, { "epoch": 3.2170367014046217, "grad_norm": 0.2154842385641122, "learning_rate": 2.7190110540873193e-05, "loss": 0.9503, "step": 7100 }, { "epoch": 3.217489805165383, "grad_norm": 0.14952481348377272, "learning_rate": 2.7178123919781608e-05, "loss": 0.9016, "step": 7101 }, { "epoch": 3.217942908926144, "grad_norm": 0.15072859187593965, "learning_rate": 2.716613858180434e-05, "loss": 0.9579, "step": 7102 }, { "epoch": 3.218396012686905, "grad_norm": 0.17392428495492124, "learning_rate": 2.7154154528140803e-05, "loss": 0.925, "step": 7103 }, { "epoch": 3.2188491164476667, "grad_norm": 0.15889550309381326, "learning_rate": 2.714217175999026e-05, "loss": 0.9353, "step": 7104 }, { "epoch": 3.219302220208428, "grad_norm": 0.15662931711746533, "learning_rate": 2.7130190278551853e-05, "loss": 0.9198, "step": 7105 }, { "epoch": 3.219755323969189, "grad_norm": 0.15672937956225444, "learning_rate": 2.71182100850246e-05, "loss": 0.9433, "step": 7106 }, { "epoch": 3.22020842772995, "grad_norm": 0.18028691933614124, "learning_rate": 2.7106231180607387e-05, "loss": 0.9474, "step": 7107 }, { "epoch": 3.220661531490711, "grad_norm": 0.16119057492351987, "learning_rate": 2.7094253566498966e-05, "loss": 0.9436, "step": 7108 }, { "epoch": 3.221114635251473, "grad_norm": 0.14187926146383073, "learning_rate": 2.708227724389797e-05, "loss": 0.9257, "step": 7109 }, { "epoch": 3.221567739012234, "grad_norm": 0.17589557729334632, "learning_rate": 2.707030221400289e-05, "loss": 0.9536, "step": 7110 }, { "epoch": 3.222020842772995, "grad_norm": 0.13036955862035268, "learning_rate": 2.7058328478012098e-05, "loss": 0.9397, "step": 7111 }, { "epoch": 3.222473946533756, "grad_norm": 0.14331154203631763, "learning_rate": 2.704635603712383e-05, "loss": 0.91, "step": 7112 }, { "epoch": 3.2229270502945173, "grad_norm": 0.13407088465041758, "learning_rate": 2.7034384892536204e-05, "loss": 0.9187, "step": 7113 }, { "epoch": 3.2233801540552784, "grad_norm": 0.11474716084700436, "learning_rate": 2.702241504544719e-05, "loss": 0.933, "step": 7114 }, { "epoch": 3.22383325781604, "grad_norm": 0.13417289869698784, "learning_rate": 2.7010446497054635e-05, "loss": 0.9351, "step": 7115 }, { "epoch": 3.224286361576801, "grad_norm": 0.15775871994575955, "learning_rate": 2.699847924855627e-05, "loss": 0.9206, "step": 7116 }, { "epoch": 3.2247394653375623, "grad_norm": 0.19566458901744965, "learning_rate": 2.698651330114969e-05, "loss": 0.9186, "step": 7117 }, { "epoch": 3.2251925690983234, "grad_norm": 0.1768628423306594, "learning_rate": 2.6974548656032332e-05, "loss": 0.9167, "step": 7118 }, { "epoch": 3.2256456728590845, "grad_norm": 0.14143536397174283, "learning_rate": 2.6962585314401544e-05, "loss": 0.9415, "step": 7119 }, { "epoch": 3.226098776619846, "grad_norm": 0.1554200325333146, "learning_rate": 2.6950623277454514e-05, "loss": 0.9374, "step": 7120 }, { "epoch": 3.2265518803806073, "grad_norm": 0.1834674554948315, "learning_rate": 2.693866254638832e-05, "loss": 0.9482, "step": 7121 }, { "epoch": 3.2270049841413684, "grad_norm": 0.16632807016659015, "learning_rate": 2.6926703122399892e-05, "loss": 0.9578, "step": 7122 }, { "epoch": 3.2274580879021295, "grad_norm": 0.1363504099856245, "learning_rate": 2.691474500668604e-05, "loss": 0.9469, "step": 7123 }, { "epoch": 3.2279111916628906, "grad_norm": 0.15568814004240725, "learning_rate": 2.690278820044344e-05, "loss": 0.9437, "step": 7124 }, { "epoch": 3.228364295423652, "grad_norm": 0.17276454878639744, "learning_rate": 2.6890832704868643e-05, "loss": 0.9344, "step": 7125 }, { "epoch": 3.2288173991844134, "grad_norm": 0.1483753389435438, "learning_rate": 2.6878878521158055e-05, "loss": 0.9187, "step": 7126 }, { "epoch": 3.2292705029451745, "grad_norm": 0.17270206028491608, "learning_rate": 2.6866925650507948e-05, "loss": 0.9146, "step": 7127 }, { "epoch": 3.2297236067059356, "grad_norm": 0.15704824320097704, "learning_rate": 2.6854974094114486e-05, "loss": 0.935, "step": 7128 }, { "epoch": 3.2301767104666967, "grad_norm": 0.14756146293573652, "learning_rate": 2.68430238531737e-05, "loss": 0.9635, "step": 7129 }, { "epoch": 3.2306298142274583, "grad_norm": 0.14918889276002956, "learning_rate": 2.6831074928881453e-05, "loss": 0.9259, "step": 7130 }, { "epoch": 3.2310829179882194, "grad_norm": 0.15892729853294454, "learning_rate": 2.6819127322433513e-05, "loss": 0.9251, "step": 7131 }, { "epoch": 3.2315360217489806, "grad_norm": 0.16327163843743256, "learning_rate": 2.68071810350255e-05, "loss": 0.9351, "step": 7132 }, { "epoch": 3.2319891255097417, "grad_norm": 0.2974591505113146, "learning_rate": 2.679523606785291e-05, "loss": 0.9347, "step": 7133 }, { "epoch": 3.232442229270503, "grad_norm": 0.16567211334901236, "learning_rate": 2.6783292422111107e-05, "loss": 0.9376, "step": 7134 }, { "epoch": 3.232895333031264, "grad_norm": 0.14837606762264405, "learning_rate": 2.677135009899531e-05, "loss": 0.9522, "step": 7135 }, { "epoch": 3.2333484367920255, "grad_norm": 0.14067517214030836, "learning_rate": 2.6759409099700624e-05, "loss": 0.922, "step": 7136 }, { "epoch": 3.2338015405527867, "grad_norm": 0.14877122969817072, "learning_rate": 2.6747469425422016e-05, "loss": 0.9204, "step": 7137 }, { "epoch": 3.234254644313548, "grad_norm": 0.12978012083324283, "learning_rate": 2.6735531077354294e-05, "loss": 0.9149, "step": 7138 }, { "epoch": 3.234707748074309, "grad_norm": 0.1755248648808948, "learning_rate": 2.6723594056692172e-05, "loss": 0.9289, "step": 7139 }, { "epoch": 3.23516085183507, "grad_norm": 0.1662147397445442, "learning_rate": 2.671165836463021e-05, "loss": 0.9248, "step": 7140 }, { "epoch": 3.2356139555958316, "grad_norm": 0.12957034519162286, "learning_rate": 2.6699724002362843e-05, "loss": 0.9246, "step": 7141 }, { "epoch": 3.236067059356593, "grad_norm": 0.16486607317197274, "learning_rate": 2.668779097108437e-05, "loss": 0.9187, "step": 7142 }, { "epoch": 3.236520163117354, "grad_norm": 0.1512087759287949, "learning_rate": 2.6675859271988955e-05, "loss": 0.9261, "step": 7143 }, { "epoch": 3.236973266878115, "grad_norm": 0.1520273747536424, "learning_rate": 2.666392890627063e-05, "loss": 0.941, "step": 7144 }, { "epoch": 3.237426370638876, "grad_norm": 0.14028547740817213, "learning_rate": 2.6651999875123303e-05, "loss": 0.9575, "step": 7145 }, { "epoch": 3.2378794743996373, "grad_norm": 0.1670010445476141, "learning_rate": 2.6640072179740716e-05, "loss": 0.9243, "step": 7146 }, { "epoch": 3.238332578160399, "grad_norm": 0.18834334727927, "learning_rate": 2.6628145821316525e-05, "loss": 0.9096, "step": 7147 }, { "epoch": 3.23878568192116, "grad_norm": 0.16551561405565446, "learning_rate": 2.6616220801044214e-05, "loss": 0.9232, "step": 7148 }, { "epoch": 3.239238785681921, "grad_norm": 0.1848260717750047, "learning_rate": 2.660429712011715e-05, "loss": 0.94, "step": 7149 }, { "epoch": 3.2396918894426823, "grad_norm": 0.1459632966095171, "learning_rate": 2.659237477972856e-05, "loss": 0.9262, "step": 7150 }, { "epoch": 3.2401449932034434, "grad_norm": 0.1949013392274339, "learning_rate": 2.6580453781071553e-05, "loss": 0.9397, "step": 7151 }, { "epoch": 3.240598096964205, "grad_norm": 0.17814646184582061, "learning_rate": 2.6568534125339075e-05, "loss": 0.925, "step": 7152 }, { "epoch": 3.241051200724966, "grad_norm": 0.1492039866088105, "learning_rate": 2.655661581372397e-05, "loss": 0.9372, "step": 7153 }, { "epoch": 3.2415043044857272, "grad_norm": 0.1890843087468702, "learning_rate": 2.654469884741891e-05, "loss": 0.9316, "step": 7154 }, { "epoch": 3.2419574082464884, "grad_norm": 0.177711489663385, "learning_rate": 2.6532783227616464e-05, "loss": 0.9388, "step": 7155 }, { "epoch": 3.2424105120072495, "grad_norm": 0.17328168973739325, "learning_rate": 2.6520868955509053e-05, "loss": 0.9447, "step": 7156 }, { "epoch": 3.242863615768011, "grad_norm": 0.15592652879834726, "learning_rate": 2.6508956032288967e-05, "loss": 0.9341, "step": 7157 }, { "epoch": 3.243316719528772, "grad_norm": 0.2235047036055957, "learning_rate": 2.649704445914836e-05, "loss": 0.983, "step": 7158 }, { "epoch": 3.2437698232895333, "grad_norm": 0.16838382767891136, "learning_rate": 2.6485134237279243e-05, "loss": 0.9268, "step": 7159 }, { "epoch": 3.2442229270502945, "grad_norm": 0.1571257629052705, "learning_rate": 2.6473225367873508e-05, "loss": 0.9228, "step": 7160 }, { "epoch": 3.2446760308110556, "grad_norm": 0.1737155739525815, "learning_rate": 2.646131785212291e-05, "loss": 0.9396, "step": 7161 }, { "epoch": 3.245129134571817, "grad_norm": 0.13619790012786132, "learning_rate": 2.644941169121904e-05, "loss": 0.9252, "step": 7162 }, { "epoch": 3.2455822383325783, "grad_norm": 0.14349130189055467, "learning_rate": 2.643750688635339e-05, "loss": 0.9463, "step": 7163 }, { "epoch": 3.2460353420933394, "grad_norm": 0.17365591261965774, "learning_rate": 2.642560343871729e-05, "loss": 0.9238, "step": 7164 }, { "epoch": 3.2464884458541006, "grad_norm": 0.15553857742608831, "learning_rate": 2.6413701349501956e-05, "loss": 0.9258, "step": 7165 }, { "epoch": 3.2469415496148617, "grad_norm": 0.14092548718249467, "learning_rate": 2.6401800619898445e-05, "loss": 0.9368, "step": 7166 }, { "epoch": 3.247394653375623, "grad_norm": 0.14127991628372205, "learning_rate": 2.6389901251097705e-05, "loss": 0.9404, "step": 7167 }, { "epoch": 3.2478477571363844, "grad_norm": 0.1795710338866038, "learning_rate": 2.637800324429052e-05, "loss": 0.9619, "step": 7168 }, { "epoch": 3.2483008608971455, "grad_norm": 0.15541812249775144, "learning_rate": 2.636610660066756e-05, "loss": 0.9176, "step": 7169 }, { "epoch": 3.2487539646579067, "grad_norm": 0.1440389279761607, "learning_rate": 2.635421132141936e-05, "loss": 0.9327, "step": 7170 }, { "epoch": 3.249207068418668, "grad_norm": 0.16798267069988976, "learning_rate": 2.6342317407736274e-05, "loss": 0.9328, "step": 7171 }, { "epoch": 3.249660172179429, "grad_norm": 0.15997848941131868, "learning_rate": 2.633042486080858e-05, "loss": 0.9269, "step": 7172 }, { "epoch": 3.2501132759401905, "grad_norm": 0.14766231605425276, "learning_rate": 2.631853368182638e-05, "loss": 0.9399, "step": 7173 }, { "epoch": 3.2505663797009516, "grad_norm": 0.1465554977549281, "learning_rate": 2.6306643871979657e-05, "loss": 0.9231, "step": 7174 }, { "epoch": 3.2510194834617128, "grad_norm": 0.17502285286944372, "learning_rate": 2.6294755432458246e-05, "loss": 0.9372, "step": 7175 }, { "epoch": 3.251472587222474, "grad_norm": 0.15987080001965576, "learning_rate": 2.6282868364451858e-05, "loss": 0.9279, "step": 7176 }, { "epoch": 3.251925690983235, "grad_norm": 0.13789321778336533, "learning_rate": 2.6270982669150052e-05, "loss": 0.9267, "step": 7177 }, { "epoch": 3.252378794743996, "grad_norm": 0.14155366202391156, "learning_rate": 2.6259098347742274e-05, "loss": 0.9296, "step": 7178 }, { "epoch": 3.2528318985047577, "grad_norm": 0.16775812970193602, "learning_rate": 2.6247215401417787e-05, "loss": 0.9327, "step": 7179 }, { "epoch": 3.253285002265519, "grad_norm": 0.14930210764536644, "learning_rate": 2.623533383136576e-05, "loss": 0.945, "step": 7180 }, { "epoch": 3.25373810602628, "grad_norm": 0.1395435459834368, "learning_rate": 2.6223453638775205e-05, "loss": 0.9414, "step": 7181 }, { "epoch": 3.254191209787041, "grad_norm": 0.16469544714788917, "learning_rate": 2.6211574824835006e-05, "loss": 0.9226, "step": 7182 }, { "epoch": 3.2546443135478023, "grad_norm": 0.1610018192950057, "learning_rate": 2.6199697390733914e-05, "loss": 0.9355, "step": 7183 }, { "epoch": 3.255097417308564, "grad_norm": 0.17438569417056013, "learning_rate": 2.61878213376605e-05, "loss": 0.9424, "step": 7184 }, { "epoch": 3.255550521069325, "grad_norm": 0.1345471732578318, "learning_rate": 2.617594666680324e-05, "loss": 0.9431, "step": 7185 }, { "epoch": 3.256003624830086, "grad_norm": 0.15848587952855606, "learning_rate": 2.616407337935047e-05, "loss": 0.914, "step": 7186 }, { "epoch": 3.2564567285908472, "grad_norm": 0.15271050679624393, "learning_rate": 2.615220147649036e-05, "loss": 0.9159, "step": 7187 }, { "epoch": 3.2569098323516084, "grad_norm": 0.16740092177107246, "learning_rate": 2.6140330959410977e-05, "loss": 0.9514, "step": 7188 }, { "epoch": 3.2573629361123695, "grad_norm": 0.12953264843095366, "learning_rate": 2.6128461829300223e-05, "loss": 0.9404, "step": 7189 }, { "epoch": 3.257816039873131, "grad_norm": 0.15557617112098768, "learning_rate": 2.6116594087345865e-05, "loss": 0.9507, "step": 7190 }, { "epoch": 3.258269143633892, "grad_norm": 0.16927370564799837, "learning_rate": 2.610472773473553e-05, "loss": 0.9389, "step": 7191 }, { "epoch": 3.2587222473946533, "grad_norm": 0.14165863240165075, "learning_rate": 2.6092862772656724e-05, "loss": 0.9366, "step": 7192 }, { "epoch": 3.2591753511554145, "grad_norm": 0.13784981808469338, "learning_rate": 2.6080999202296795e-05, "loss": 0.9492, "step": 7193 }, { "epoch": 3.259628454916176, "grad_norm": 0.1471527986850341, "learning_rate": 2.6069137024842953e-05, "loss": 0.9397, "step": 7194 }, { "epoch": 3.260081558676937, "grad_norm": 0.13960874821578229, "learning_rate": 2.6057276241482283e-05, "loss": 0.9346, "step": 7195 }, { "epoch": 3.2605346624376983, "grad_norm": 0.14236893763858824, "learning_rate": 2.6045416853401717e-05, "loss": 0.9183, "step": 7196 }, { "epoch": 3.2609877661984594, "grad_norm": 0.13367901081980665, "learning_rate": 2.6033558861788044e-05, "loss": 0.9269, "step": 7197 }, { "epoch": 3.2614408699592206, "grad_norm": 0.12496565550163924, "learning_rate": 2.6021702267827936e-05, "loss": 0.9482, "step": 7198 }, { "epoch": 3.2618939737199817, "grad_norm": 0.1381661781366572, "learning_rate": 2.6009847072707893e-05, "loss": 0.952, "step": 7199 }, { "epoch": 3.2623470774807433, "grad_norm": 0.13605871706310885, "learning_rate": 2.5997993277614295e-05, "loss": 0.93, "step": 7200 }, { "epoch": 3.2628001812415044, "grad_norm": 0.14282251766436566, "learning_rate": 2.598614088373338e-05, "loss": 0.9231, "step": 7201 }, { "epoch": 3.2632532850022655, "grad_norm": 0.11493362270242584, "learning_rate": 2.597428989225124e-05, "loss": 0.9169, "step": 7202 }, { "epoch": 3.2637063887630267, "grad_norm": 0.12269273678698944, "learning_rate": 2.5962440304353846e-05, "loss": 0.9361, "step": 7203 }, { "epoch": 3.264159492523788, "grad_norm": 0.11993289004571533, "learning_rate": 2.5950592121226993e-05, "loss": 0.9201, "step": 7204 }, { "epoch": 3.2646125962845494, "grad_norm": 0.12718071850518572, "learning_rate": 2.593874534405637e-05, "loss": 0.9333, "step": 7205 }, { "epoch": 3.2650657000453105, "grad_norm": 0.135345435223812, "learning_rate": 2.5926899974027514e-05, "loss": 0.9527, "step": 7206 }, { "epoch": 3.2655188038060716, "grad_norm": 0.15548858588617717, "learning_rate": 2.59150560123258e-05, "loss": 0.9261, "step": 7207 }, { "epoch": 3.2659719075668328, "grad_norm": 0.15632345738541711, "learning_rate": 2.5903213460136495e-05, "loss": 0.9513, "step": 7208 }, { "epoch": 3.266425011327594, "grad_norm": 0.14176855219935303, "learning_rate": 2.5891372318644695e-05, "loss": 0.9569, "step": 7209 }, { "epoch": 3.266878115088355, "grad_norm": 0.14689113859182149, "learning_rate": 2.587953258903539e-05, "loss": 0.9067, "step": 7210 }, { "epoch": 3.2673312188491166, "grad_norm": 0.1537942647094162, "learning_rate": 2.5867694272493396e-05, "loss": 0.9298, "step": 7211 }, { "epoch": 3.2677843226098777, "grad_norm": 0.13949444856119655, "learning_rate": 2.58558573702034e-05, "loss": 0.9386, "step": 7212 }, { "epoch": 3.268237426370639, "grad_norm": 0.14594282474224765, "learning_rate": 2.5844021883349955e-05, "loss": 0.944, "step": 7213 }, { "epoch": 3.2686905301314, "grad_norm": 0.1282878776376447, "learning_rate": 2.5832187813117475e-05, "loss": 0.9315, "step": 7214 }, { "epoch": 3.269143633892161, "grad_norm": 0.15317350083887293, "learning_rate": 2.5820355160690195e-05, "loss": 0.9384, "step": 7215 }, { "epoch": 3.2695967376529227, "grad_norm": 0.1271254203522758, "learning_rate": 2.580852392725225e-05, "loss": 0.9416, "step": 7216 }, { "epoch": 3.270049841413684, "grad_norm": 0.15216592760702796, "learning_rate": 2.5796694113987616e-05, "loss": 0.9423, "step": 7217 }, { "epoch": 3.270502945174445, "grad_norm": 0.11694872273312318, "learning_rate": 2.5784865722080136e-05, "loss": 0.9332, "step": 7218 }, { "epoch": 3.270956048935206, "grad_norm": 0.14071951696900004, "learning_rate": 2.5773038752713493e-05, "loss": 0.9338, "step": 7219 }, { "epoch": 3.271409152695967, "grad_norm": 0.11856239708909984, "learning_rate": 2.5761213207071247e-05, "loss": 0.927, "step": 7220 }, { "epoch": 3.2718622564567283, "grad_norm": 0.12597271861937331, "learning_rate": 2.5749389086336806e-05, "loss": 0.9211, "step": 7221 }, { "epoch": 3.27231536021749, "grad_norm": 0.13793218014719963, "learning_rate": 2.573756639169344e-05, "loss": 0.9228, "step": 7222 }, { "epoch": 3.272768463978251, "grad_norm": 0.13508793279473935, "learning_rate": 2.5725745124324267e-05, "loss": 0.9296, "step": 7223 }, { "epoch": 3.273221567739012, "grad_norm": 0.14926360924059498, "learning_rate": 2.5713925285412264e-05, "loss": 0.9331, "step": 7224 }, { "epoch": 3.2736746714997733, "grad_norm": 0.15360535494483868, "learning_rate": 2.5702106876140274e-05, "loss": 0.9265, "step": 7225 }, { "epoch": 3.274127775260535, "grad_norm": 0.1251732471933323, "learning_rate": 2.5690289897691e-05, "loss": 0.9299, "step": 7226 }, { "epoch": 3.274580879021296, "grad_norm": 0.1466417820958632, "learning_rate": 2.5678474351246986e-05, "loss": 0.9319, "step": 7227 }, { "epoch": 3.275033982782057, "grad_norm": 0.11677798546461751, "learning_rate": 2.566666023799064e-05, "loss": 0.9504, "step": 7228 }, { "epoch": 3.2754870865428183, "grad_norm": 0.12463577863922518, "learning_rate": 2.565484755910423e-05, "loss": 0.9383, "step": 7229 }, { "epoch": 3.2759401903035794, "grad_norm": 0.1315912563079768, "learning_rate": 2.564303631576988e-05, "loss": 0.9286, "step": 7230 }, { "epoch": 3.2763932940643405, "grad_norm": 0.1453795862408951, "learning_rate": 2.563122650916956e-05, "loss": 0.922, "step": 7231 }, { "epoch": 3.276846397825102, "grad_norm": 0.14248720660115935, "learning_rate": 2.5619418140485107e-05, "loss": 0.9402, "step": 7232 }, { "epoch": 3.2772995015858633, "grad_norm": 0.2279245431822027, "learning_rate": 2.5607611210898212e-05, "loss": 0.9541, "step": 7233 }, { "epoch": 3.2777526053466244, "grad_norm": 0.23856901778269166, "learning_rate": 2.559580572159043e-05, "loss": 0.9206, "step": 7234 }, { "epoch": 3.2782057091073855, "grad_norm": 0.14118043001415378, "learning_rate": 2.5584001673743153e-05, "loss": 0.9364, "step": 7235 }, { "epoch": 3.2786588128681466, "grad_norm": 0.182648672353728, "learning_rate": 2.5572199068537637e-05, "loss": 0.9282, "step": 7236 }, { "epoch": 3.279111916628908, "grad_norm": 0.23336819476895207, "learning_rate": 2.5560397907155002e-05, "loss": 0.9164, "step": 7237 }, { "epoch": 3.2795650203896693, "grad_norm": 0.1846086753734237, "learning_rate": 2.5548598190776213e-05, "loss": 0.9383, "step": 7238 }, { "epoch": 3.2800181241504305, "grad_norm": 0.14351281501746818, "learning_rate": 2.5536799920582112e-05, "loss": 0.939, "step": 7239 }, { "epoch": 3.2804712279111916, "grad_norm": 0.19365229178407894, "learning_rate": 2.5525003097753357e-05, "loss": 0.9124, "step": 7240 }, { "epoch": 3.2809243316719527, "grad_norm": 0.18361667575942794, "learning_rate": 2.5513207723470482e-05, "loss": 0.9196, "step": 7241 }, { "epoch": 3.281377435432714, "grad_norm": 0.1305160956904019, "learning_rate": 2.5501413798913895e-05, "loss": 0.9091, "step": 7242 }, { "epoch": 3.2818305391934754, "grad_norm": 0.1393079071962835, "learning_rate": 2.5489621325263825e-05, "loss": 0.9342, "step": 7243 }, { "epoch": 3.2822836429542366, "grad_norm": 0.17193559970256714, "learning_rate": 2.5477830303700374e-05, "loss": 0.9449, "step": 7244 }, { "epoch": 3.2827367467149977, "grad_norm": 0.1443171379953748, "learning_rate": 2.5466040735403494e-05, "loss": 0.9289, "step": 7245 }, { "epoch": 3.283189850475759, "grad_norm": 0.12469242272079367, "learning_rate": 2.5454252621553e-05, "loss": 0.9355, "step": 7246 }, { "epoch": 3.28364295423652, "grad_norm": 0.1759595117691237, "learning_rate": 2.5442465963328565e-05, "loss": 0.9377, "step": 7247 }, { "epoch": 3.2840960579972815, "grad_norm": 0.1729293016270462, "learning_rate": 2.5430680761909682e-05, "loss": 0.9305, "step": 7248 }, { "epoch": 3.2845491617580427, "grad_norm": 0.12652157321883592, "learning_rate": 2.5418897018475747e-05, "loss": 0.9247, "step": 7249 }, { "epoch": 3.285002265518804, "grad_norm": 0.1747359482226048, "learning_rate": 2.540711473420598e-05, "loss": 0.9315, "step": 7250 }, { "epoch": 3.285455369279565, "grad_norm": 0.2143010639800548, "learning_rate": 2.5395333910279455e-05, "loss": 0.9461, "step": 7251 }, { "epoch": 3.285908473040326, "grad_norm": 0.16248589749075923, "learning_rate": 2.5383554547875103e-05, "loss": 0.929, "step": 7252 }, { "epoch": 3.286361576801087, "grad_norm": 0.12716502918029707, "learning_rate": 2.537177664817172e-05, "loss": 0.9255, "step": 7253 }, { "epoch": 3.286814680561849, "grad_norm": 0.1792213280685856, "learning_rate": 2.5360000212347942e-05, "loss": 0.9194, "step": 7254 }, { "epoch": 3.28726778432261, "grad_norm": 0.17820956905767885, "learning_rate": 2.5348225241582264e-05, "loss": 0.9432, "step": 7255 }, { "epoch": 3.287720888083371, "grad_norm": 0.13978231891636758, "learning_rate": 2.5336451737053037e-05, "loss": 0.9175, "step": 7256 }, { "epoch": 3.288173991844132, "grad_norm": 0.18093467729689391, "learning_rate": 2.532467969993846e-05, "loss": 0.9161, "step": 7257 }, { "epoch": 3.2886270956048937, "grad_norm": 0.23095039946035179, "learning_rate": 2.5312909131416587e-05, "loss": 0.9362, "step": 7258 }, { "epoch": 3.289080199365655, "grad_norm": 0.1807027435681864, "learning_rate": 2.530114003266534e-05, "loss": 0.951, "step": 7259 }, { "epoch": 3.289533303126416, "grad_norm": 0.1402775225132639, "learning_rate": 2.5289372404862462e-05, "loss": 0.9401, "step": 7260 }, { "epoch": 3.289986406887177, "grad_norm": 0.18420884078827113, "learning_rate": 2.5277606249185563e-05, "loss": 0.9471, "step": 7261 }, { "epoch": 3.2904395106479383, "grad_norm": 0.1547395099736173, "learning_rate": 2.526584156681212e-05, "loss": 0.9482, "step": 7262 }, { "epoch": 3.2908926144086994, "grad_norm": 0.1271197306924919, "learning_rate": 2.5254078358919453e-05, "loss": 0.9174, "step": 7263 }, { "epoch": 3.291345718169461, "grad_norm": 0.12814121638330989, "learning_rate": 2.5242316626684723e-05, "loss": 0.9366, "step": 7264 }, { "epoch": 3.291798821930222, "grad_norm": 0.13989864483815723, "learning_rate": 2.523055637128496e-05, "loss": 0.918, "step": 7265 }, { "epoch": 3.2922519256909832, "grad_norm": 0.16224589327072478, "learning_rate": 2.521879759389704e-05, "loss": 0.909, "step": 7266 }, { "epoch": 3.2927050294517444, "grad_norm": 0.14768327325008848, "learning_rate": 2.5207040295697698e-05, "loss": 0.9462, "step": 7267 }, { "epoch": 3.2931581332125055, "grad_norm": 0.13687131274755363, "learning_rate": 2.51952844778635e-05, "loss": 0.9276, "step": 7268 }, { "epoch": 3.293611236973267, "grad_norm": 0.1317307044136696, "learning_rate": 2.5183530141570877e-05, "loss": 0.9338, "step": 7269 }, { "epoch": 3.294064340734028, "grad_norm": 0.14824055671933872, "learning_rate": 2.5171777287996114e-05, "loss": 0.944, "step": 7270 }, { "epoch": 3.2945174444947893, "grad_norm": 0.12573602141888868, "learning_rate": 2.5160025918315355e-05, "loss": 0.9465, "step": 7271 }, { "epoch": 3.2949705482555505, "grad_norm": 0.14846335956606416, "learning_rate": 2.514827603370457e-05, "loss": 0.9496, "step": 7272 }, { "epoch": 3.2954236520163116, "grad_norm": 0.1555277102593883, "learning_rate": 2.513652763533962e-05, "loss": 0.9372, "step": 7273 }, { "epoch": 3.2958767557770727, "grad_norm": 0.10433026434489133, "learning_rate": 2.512478072439618e-05, "loss": 0.951, "step": 7274 }, { "epoch": 3.2963298595378343, "grad_norm": 0.1648743543527108, "learning_rate": 2.5113035302049795e-05, "loss": 0.9371, "step": 7275 }, { "epoch": 3.2967829632985954, "grad_norm": 0.11681122541842669, "learning_rate": 2.5101291369475848e-05, "loss": 0.9298, "step": 7276 }, { "epoch": 3.2972360670593566, "grad_norm": 0.1390453522206736, "learning_rate": 2.5089548927849583e-05, "loss": 0.9181, "step": 7277 }, { "epoch": 3.2976891708201177, "grad_norm": 0.11496794541746848, "learning_rate": 2.5077807978346095e-05, "loss": 0.958, "step": 7278 }, { "epoch": 3.298142274580879, "grad_norm": 0.14129215298793493, "learning_rate": 2.5066068522140324e-05, "loss": 0.9368, "step": 7279 }, { "epoch": 3.2985953783416404, "grad_norm": 0.1264270035439029, "learning_rate": 2.5054330560407075e-05, "loss": 0.9366, "step": 7280 }, { "epoch": 3.2990484821024015, "grad_norm": 0.13295753496415902, "learning_rate": 2.5042594094320985e-05, "loss": 0.9363, "step": 7281 }, { "epoch": 3.2995015858631627, "grad_norm": 0.1597413150480437, "learning_rate": 2.5030859125056543e-05, "loss": 0.9226, "step": 7282 }, { "epoch": 3.299954689623924, "grad_norm": 0.16935862350121222, "learning_rate": 2.5019125653788114e-05, "loss": 0.9298, "step": 7283 }, { "epoch": 3.300407793384685, "grad_norm": 0.15726053411033142, "learning_rate": 2.5007393681689868e-05, "loss": 0.9302, "step": 7284 }, { "epoch": 3.300860897145446, "grad_norm": 0.1510859438826153, "learning_rate": 2.4995663209935863e-05, "loss": 0.9093, "step": 7285 }, { "epoch": 3.3013140009062076, "grad_norm": 0.1469162107000033, "learning_rate": 2.49839342397e-05, "loss": 0.9482, "step": 7286 }, { "epoch": 3.3017671046669688, "grad_norm": 0.1496950360610932, "learning_rate": 2.4972206772156008e-05, "loss": 0.9002, "step": 7287 }, { "epoch": 3.30222020842773, "grad_norm": 0.13683824676578274, "learning_rate": 2.4960480808477493e-05, "loss": 0.9312, "step": 7288 }, { "epoch": 3.302673312188491, "grad_norm": 0.14142979335105132, "learning_rate": 2.4948756349837892e-05, "loss": 0.9514, "step": 7289 }, { "epoch": 3.3031264159492526, "grad_norm": 0.15161845622980724, "learning_rate": 2.4937033397410507e-05, "loss": 0.9232, "step": 7290 }, { "epoch": 3.3035795197100137, "grad_norm": 0.1492600601446208, "learning_rate": 2.4925311952368484e-05, "loss": 0.9227, "step": 7291 }, { "epoch": 3.304032623470775, "grad_norm": 0.143404985691264, "learning_rate": 2.4913592015884795e-05, "loss": 0.9243, "step": 7292 }, { "epoch": 3.304485727231536, "grad_norm": 0.13825047282532799, "learning_rate": 2.490187358913229e-05, "loss": 0.926, "step": 7293 }, { "epoch": 3.304938830992297, "grad_norm": 0.16360651579675403, "learning_rate": 2.4890156673283665e-05, "loss": 0.9326, "step": 7294 }, { "epoch": 3.3053919347530583, "grad_norm": 0.1323787423860414, "learning_rate": 2.487844126951145e-05, "loss": 0.9194, "step": 7295 }, { "epoch": 3.30584503851382, "grad_norm": 0.1407346507091002, "learning_rate": 2.486672737898804e-05, "loss": 0.9317, "step": 7296 }, { "epoch": 3.306298142274581, "grad_norm": 0.14963475259217676, "learning_rate": 2.485501500288568e-05, "loss": 0.9306, "step": 7297 }, { "epoch": 3.306751246035342, "grad_norm": 0.12963596805725122, "learning_rate": 2.4843304142376434e-05, "loss": 0.9289, "step": 7298 }, { "epoch": 3.3072043497961032, "grad_norm": 0.13943286207310027, "learning_rate": 2.4831594798632233e-05, "loss": 0.9386, "step": 7299 }, { "epoch": 3.3076574535568644, "grad_norm": 0.13923005007695194, "learning_rate": 2.4819886972824866e-05, "loss": 0.9136, "step": 7300 }, { "epoch": 3.308110557317626, "grad_norm": 0.12625828674727435, "learning_rate": 2.4808180666125967e-05, "loss": 0.9311, "step": 7301 }, { "epoch": 3.308563661078387, "grad_norm": 0.15091981547771108, "learning_rate": 2.4796475879707003e-05, "loss": 0.9472, "step": 7302 }, { "epoch": 3.309016764839148, "grad_norm": 0.2645170260773624, "learning_rate": 2.4784772614739315e-05, "loss": 0.9391, "step": 7303 }, { "epoch": 3.3094698685999093, "grad_norm": 0.16197212592952115, "learning_rate": 2.477307087239405e-05, "loss": 0.9225, "step": 7304 }, { "epoch": 3.3099229723606705, "grad_norm": 0.1714434655380154, "learning_rate": 2.476137065384225e-05, "loss": 0.9398, "step": 7305 }, { "epoch": 3.3103760761214316, "grad_norm": 0.1306210543326609, "learning_rate": 2.4749671960254766e-05, "loss": 0.9383, "step": 7306 }, { "epoch": 3.310829179882193, "grad_norm": 0.1599664375533918, "learning_rate": 2.4737974792802325e-05, "loss": 0.9277, "step": 7307 }, { "epoch": 3.3112822836429543, "grad_norm": 0.14679332416625293, "learning_rate": 2.4726279152655483e-05, "loss": 0.9234, "step": 7308 }, { "epoch": 3.3117353874037154, "grad_norm": 0.14692081261611833, "learning_rate": 2.4714585040984656e-05, "loss": 0.9397, "step": 7309 }, { "epoch": 3.3121884911644766, "grad_norm": 0.16134874114209913, "learning_rate": 2.470289245896009e-05, "loss": 0.9174, "step": 7310 }, { "epoch": 3.3126415949252377, "grad_norm": 0.14044784727288406, "learning_rate": 2.469120140775191e-05, "loss": 0.9484, "step": 7311 }, { "epoch": 3.3130946986859993, "grad_norm": 0.20432763831539993, "learning_rate": 2.4679511888530033e-05, "loss": 0.9399, "step": 7312 }, { "epoch": 3.3135478024467604, "grad_norm": 0.20594646694307495, "learning_rate": 2.466782390246428e-05, "loss": 0.9383, "step": 7313 }, { "epoch": 3.3140009062075215, "grad_norm": 0.18851742164398486, "learning_rate": 2.4656137450724287e-05, "loss": 0.9172, "step": 7314 }, { "epoch": 3.3144540099682827, "grad_norm": 0.13190784996517876, "learning_rate": 2.464445253447954e-05, "loss": 0.9316, "step": 7315 }, { "epoch": 3.314907113729044, "grad_norm": 0.15836856629722768, "learning_rate": 2.4632769154899386e-05, "loss": 0.9464, "step": 7316 }, { "epoch": 3.315360217489805, "grad_norm": 0.15082501068790372, "learning_rate": 2.4621087313153e-05, "loss": 0.9017, "step": 7317 }, { "epoch": 3.3158133212505665, "grad_norm": 0.17288311782111274, "learning_rate": 2.4609407010409406e-05, "loss": 0.9205, "step": 7318 }, { "epoch": 3.3162664250113276, "grad_norm": 0.13997789347395967, "learning_rate": 2.459772824783749e-05, "loss": 0.9511, "step": 7319 }, { "epoch": 3.3167195287720888, "grad_norm": 0.1859864326506173, "learning_rate": 2.4586051026605966e-05, "loss": 0.9314, "step": 7320 }, { "epoch": 3.31717263253285, "grad_norm": 0.17568263729120157, "learning_rate": 2.4574375347883395e-05, "loss": 0.9153, "step": 7321 }, { "epoch": 3.3176257362936115, "grad_norm": 0.12770416061254056, "learning_rate": 2.4562701212838194e-05, "loss": 0.9057, "step": 7322 }, { "epoch": 3.3180788400543726, "grad_norm": 0.14848270754911655, "learning_rate": 2.4551028622638628e-05, "loss": 0.9516, "step": 7323 }, { "epoch": 3.3185319438151337, "grad_norm": 0.14764898397195209, "learning_rate": 2.4539357578452786e-05, "loss": 0.9535, "step": 7324 }, { "epoch": 3.318985047575895, "grad_norm": 0.15703640558420834, "learning_rate": 2.452768808144862e-05, "loss": 0.929, "step": 7325 }, { "epoch": 3.319438151336656, "grad_norm": 0.16359441857796406, "learning_rate": 2.451602013279393e-05, "loss": 0.952, "step": 7326 }, { "epoch": 3.319891255097417, "grad_norm": 0.17369547774827612, "learning_rate": 2.4504353733656347e-05, "loss": 0.9233, "step": 7327 }, { "epoch": 3.3203443588581787, "grad_norm": 0.15386036614479023, "learning_rate": 2.4492688885203366e-05, "loss": 0.9178, "step": 7328 }, { "epoch": 3.32079746261894, "grad_norm": 0.12828834130331057, "learning_rate": 2.4481025588602297e-05, "loss": 0.9235, "step": 7329 }, { "epoch": 3.321250566379701, "grad_norm": 0.11987638237950317, "learning_rate": 2.4469363845020318e-05, "loss": 0.933, "step": 7330 }, { "epoch": 3.321703670140462, "grad_norm": 0.13773554498503274, "learning_rate": 2.445770365562445e-05, "loss": 0.9578, "step": 7331 }, { "epoch": 3.322156773901223, "grad_norm": 0.12376552213631523, "learning_rate": 2.444604502158156e-05, "loss": 0.9728, "step": 7332 }, { "epoch": 3.322609877661985, "grad_norm": 0.13895298933965652, "learning_rate": 2.443438794405834e-05, "loss": 0.9302, "step": 7333 }, { "epoch": 3.323062981422746, "grad_norm": 0.11921645272901212, "learning_rate": 2.442273242422135e-05, "loss": 0.9331, "step": 7334 }, { "epoch": 3.323516085183507, "grad_norm": 0.13083910581203148, "learning_rate": 2.441107846323698e-05, "loss": 0.9252, "step": 7335 }, { "epoch": 3.323969188944268, "grad_norm": 0.14948067151046743, "learning_rate": 2.439942606227149e-05, "loss": 0.9524, "step": 7336 }, { "epoch": 3.3244222927050293, "grad_norm": 0.12016481108168184, "learning_rate": 2.4387775222490926e-05, "loss": 0.9446, "step": 7337 }, { "epoch": 3.3248753964657904, "grad_norm": 0.14691360578965806, "learning_rate": 2.4376125945061233e-05, "loss": 0.9267, "step": 7338 }, { "epoch": 3.325328500226552, "grad_norm": 0.15573696978582563, "learning_rate": 2.4364478231148176e-05, "loss": 0.9291, "step": 7339 }, { "epoch": 3.325781603987313, "grad_norm": 0.1489126179423759, "learning_rate": 2.435283208191738e-05, "loss": 0.9425, "step": 7340 }, { "epoch": 3.3262347077480743, "grad_norm": 0.14551088630047163, "learning_rate": 2.4341187498534287e-05, "loss": 0.9364, "step": 7341 }, { "epoch": 3.3266878115088354, "grad_norm": 0.15083727153360818, "learning_rate": 2.4329544482164203e-05, "loss": 0.9213, "step": 7342 }, { "epoch": 3.3271409152695965, "grad_norm": 0.1492263770038118, "learning_rate": 2.431790303397227e-05, "loss": 0.9314, "step": 7343 }, { "epoch": 3.327594019030358, "grad_norm": 0.14244245677697218, "learning_rate": 2.4306263155123486e-05, "loss": 0.908, "step": 7344 }, { "epoch": 3.3280471227911192, "grad_norm": 0.13743768855692648, "learning_rate": 2.429462484678266e-05, "loss": 0.9218, "step": 7345 }, { "epoch": 3.3285002265518804, "grad_norm": 0.13991010943957405, "learning_rate": 2.4282988110114475e-05, "loss": 0.9479, "step": 7346 }, { "epoch": 3.3289533303126415, "grad_norm": 0.14469156116167073, "learning_rate": 2.4271352946283445e-05, "loss": 0.921, "step": 7347 }, { "epoch": 3.3294064340734026, "grad_norm": 0.12955613183013504, "learning_rate": 2.425971935645392e-05, "loss": 0.9136, "step": 7348 }, { "epoch": 3.3298595378341638, "grad_norm": 0.13024267878345847, "learning_rate": 2.4248087341790117e-05, "loss": 0.9516, "step": 7349 }, { "epoch": 3.3303126415949253, "grad_norm": 0.1335246665273864, "learning_rate": 2.4236456903456062e-05, "loss": 0.9252, "step": 7350 }, { "epoch": 3.3307657453556865, "grad_norm": 0.1516316396138991, "learning_rate": 2.4224828042615643e-05, "loss": 0.9147, "step": 7351 }, { "epoch": 3.3312188491164476, "grad_norm": 0.1350862627961654, "learning_rate": 2.4213200760432597e-05, "loss": 0.9433, "step": 7352 }, { "epoch": 3.3316719528772087, "grad_norm": 0.16062307792547195, "learning_rate": 2.4201575058070475e-05, "loss": 0.9452, "step": 7353 }, { "epoch": 3.3321250566379703, "grad_norm": 0.13172522298365721, "learning_rate": 2.4189950936692713e-05, "loss": 0.9089, "step": 7354 }, { "epoch": 3.3325781603987314, "grad_norm": 0.16151603327229358, "learning_rate": 2.4178328397462528e-05, "loss": 0.9348, "step": 7355 }, { "epoch": 3.3330312641594926, "grad_norm": 0.17571061177236288, "learning_rate": 2.416670744154305e-05, "loss": 0.9354, "step": 7356 }, { "epoch": 3.3334843679202537, "grad_norm": 0.1615761844647754, "learning_rate": 2.4155088070097186e-05, "loss": 0.9439, "step": 7357 }, { "epoch": 3.333937471681015, "grad_norm": 0.15416542353657978, "learning_rate": 2.4143470284287728e-05, "loss": 0.9403, "step": 7358 }, { "epoch": 3.334390575441776, "grad_norm": 0.14706277786536548, "learning_rate": 2.413185408527728e-05, "loss": 0.9274, "step": 7359 }, { "epoch": 3.3348436792025375, "grad_norm": 0.1808845577269892, "learning_rate": 2.412023947422832e-05, "loss": 0.9608, "step": 7360 }, { "epoch": 3.3352967829632987, "grad_norm": 0.1449443144625884, "learning_rate": 2.4108626452303135e-05, "loss": 0.9202, "step": 7361 }, { "epoch": 3.33574988672406, "grad_norm": 0.1251196926483229, "learning_rate": 2.4097015020663878e-05, "loss": 0.9438, "step": 7362 }, { "epoch": 3.336202990484821, "grad_norm": 0.12254751920673695, "learning_rate": 2.4085405180472528e-05, "loss": 0.9314, "step": 7363 }, { "epoch": 3.336656094245582, "grad_norm": 0.12246318988407134, "learning_rate": 2.407379693289091e-05, "loss": 0.902, "step": 7364 }, { "epoch": 3.3371091980063436, "grad_norm": 0.12456901615620522, "learning_rate": 2.4062190279080674e-05, "loss": 0.936, "step": 7365 }, { "epoch": 3.3375623017671048, "grad_norm": 0.14819218392271935, "learning_rate": 2.4050585220203334e-05, "loss": 0.9472, "step": 7366 }, { "epoch": 3.338015405527866, "grad_norm": 0.1375035136383621, "learning_rate": 2.4038981757420235e-05, "loss": 0.9331, "step": 7367 }, { "epoch": 3.338468509288627, "grad_norm": 0.1384144287550935, "learning_rate": 2.4027379891892563e-05, "loss": 0.9481, "step": 7368 }, { "epoch": 3.338921613049388, "grad_norm": 0.13455951786810144, "learning_rate": 2.4015779624781338e-05, "loss": 0.9257, "step": 7369 }, { "epoch": 3.3393747168101493, "grad_norm": 0.13130420846277513, "learning_rate": 2.4004180957247433e-05, "loss": 0.9358, "step": 7370 }, { "epoch": 3.339827820570911, "grad_norm": 0.1292604317779966, "learning_rate": 2.3992583890451543e-05, "loss": 0.9617, "step": 7371 }, { "epoch": 3.340280924331672, "grad_norm": 0.12564654249677748, "learning_rate": 2.3980988425554232e-05, "loss": 0.9315, "step": 7372 }, { "epoch": 3.340734028092433, "grad_norm": 0.12539251267359772, "learning_rate": 2.3969394563715857e-05, "loss": 0.9376, "step": 7373 }, { "epoch": 3.3411871318531943, "grad_norm": 0.12353699076702346, "learning_rate": 2.3957802306096662e-05, "loss": 0.9264, "step": 7374 }, { "epoch": 3.341640235613956, "grad_norm": 0.13309138336203993, "learning_rate": 2.3946211653856707e-05, "loss": 0.9386, "step": 7375 }, { "epoch": 3.342093339374717, "grad_norm": 0.13225722441944582, "learning_rate": 2.3934622608155883e-05, "loss": 0.9424, "step": 7376 }, { "epoch": 3.342546443135478, "grad_norm": 0.1512335299484691, "learning_rate": 2.392303517015395e-05, "loss": 0.9217, "step": 7377 }, { "epoch": 3.3429995468962392, "grad_norm": 0.1379861890923699, "learning_rate": 2.391144934101048e-05, "loss": 0.9288, "step": 7378 }, { "epoch": 3.3434526506570004, "grad_norm": 0.13053606381115965, "learning_rate": 2.3899865121884892e-05, "loss": 0.9259, "step": 7379 }, { "epoch": 3.3439057544177615, "grad_norm": 0.14398912252964552, "learning_rate": 2.388828251393646e-05, "loss": 0.9265, "step": 7380 }, { "epoch": 3.344358858178523, "grad_norm": 0.14572735214058855, "learning_rate": 2.3876701518324255e-05, "loss": 0.9221, "step": 7381 }, { "epoch": 3.344811961939284, "grad_norm": 0.13783649426127897, "learning_rate": 2.3865122136207228e-05, "loss": 0.9459, "step": 7382 }, { "epoch": 3.3452650657000453, "grad_norm": 0.12950696105462717, "learning_rate": 2.3853544368744164e-05, "loss": 0.9281, "step": 7383 }, { "epoch": 3.3457181694608065, "grad_norm": 0.1455661203990782, "learning_rate": 2.384196821709366e-05, "loss": 0.9377, "step": 7384 }, { "epoch": 3.3461712732215676, "grad_norm": 0.15510627318715603, "learning_rate": 2.3830393682414173e-05, "loss": 0.9499, "step": 7385 }, { "epoch": 3.346624376982329, "grad_norm": 0.14633016455930134, "learning_rate": 2.3818820765863997e-05, "loss": 0.9273, "step": 7386 }, { "epoch": 3.3470774807430903, "grad_norm": 0.14472301619018135, "learning_rate": 2.3807249468601265e-05, "loss": 0.9342, "step": 7387 }, { "epoch": 3.3475305845038514, "grad_norm": 0.1549260466917702, "learning_rate": 2.3795679791783925e-05, "loss": 0.9146, "step": 7388 }, { "epoch": 3.3479836882646126, "grad_norm": 0.1362743961120258, "learning_rate": 2.3784111736569803e-05, "loss": 0.9175, "step": 7389 }, { "epoch": 3.3484367920253737, "grad_norm": 0.14672677365159678, "learning_rate": 2.3772545304116524e-05, "loss": 0.9424, "step": 7390 }, { "epoch": 3.348889895786135, "grad_norm": 0.1462408310553858, "learning_rate": 2.3760980495581573e-05, "loss": 0.9307, "step": 7391 }, { "epoch": 3.3493429995468964, "grad_norm": 0.1298292999216198, "learning_rate": 2.3749417312122263e-05, "loss": 0.9306, "step": 7392 }, { "epoch": 3.3497961033076575, "grad_norm": 0.15207788775077294, "learning_rate": 2.373785575489575e-05, "loss": 0.9249, "step": 7393 }, { "epoch": 3.3502492070684187, "grad_norm": 0.14743789864936324, "learning_rate": 2.3726295825059032e-05, "loss": 0.9346, "step": 7394 }, { "epoch": 3.35070231082918, "grad_norm": 0.1332696633898744, "learning_rate": 2.3714737523768928e-05, "loss": 0.9464, "step": 7395 }, { "epoch": 3.351155414589941, "grad_norm": 0.12038792273271158, "learning_rate": 2.3703180852182114e-05, "loss": 0.9437, "step": 7396 }, { "epoch": 3.3516085183507025, "grad_norm": 0.12878573716244482, "learning_rate": 2.369162581145509e-05, "loss": 0.9467, "step": 7397 }, { "epoch": 3.3520616221114636, "grad_norm": 0.14224085164854447, "learning_rate": 2.3680072402744184e-05, "loss": 0.9257, "step": 7398 }, { "epoch": 3.3525147258722248, "grad_norm": 0.1198914429657462, "learning_rate": 2.3668520627205577e-05, "loss": 0.94, "step": 7399 }, { "epoch": 3.352967829632986, "grad_norm": 0.14110075977414768, "learning_rate": 2.365697048599529e-05, "loss": 0.9302, "step": 7400 }, { "epoch": 3.353420933393747, "grad_norm": 0.14259647988381016, "learning_rate": 2.364542198026916e-05, "loss": 0.928, "step": 7401 }, { "epoch": 3.353874037154508, "grad_norm": 0.13660972268868016, "learning_rate": 2.363387511118288e-05, "loss": 0.9287, "step": 7402 }, { "epoch": 3.3543271409152697, "grad_norm": 0.12666853667484979, "learning_rate": 2.362232987989197e-05, "loss": 0.9321, "step": 7403 }, { "epoch": 3.354780244676031, "grad_norm": 0.15002563984885503, "learning_rate": 2.3610786287551784e-05, "loss": 0.9335, "step": 7404 }, { "epoch": 3.355233348436792, "grad_norm": 0.11553699272469417, "learning_rate": 2.3599244335317528e-05, "loss": 0.899, "step": 7405 }, { "epoch": 3.355686452197553, "grad_norm": 0.1636930045417359, "learning_rate": 2.3587704024344217e-05, "loss": 0.9191, "step": 7406 }, { "epoch": 3.3561395559583147, "grad_norm": 0.16802784825015676, "learning_rate": 2.3576165355786717e-05, "loss": 0.928, "step": 7407 }, { "epoch": 3.356592659719076, "grad_norm": 0.1603383481535576, "learning_rate": 2.3564628330799737e-05, "loss": 0.9144, "step": 7408 }, { "epoch": 3.357045763479837, "grad_norm": 0.12642621295717488, "learning_rate": 2.3553092950537807e-05, "loss": 0.9084, "step": 7409 }, { "epoch": 3.357498867240598, "grad_norm": 0.15189445316787334, "learning_rate": 2.3541559216155303e-05, "loss": 0.9535, "step": 7410 }, { "epoch": 3.3579519710013592, "grad_norm": 0.15342095839695374, "learning_rate": 2.3530027128806444e-05, "loss": 0.9304, "step": 7411 }, { "epoch": 3.3584050747621204, "grad_norm": 0.12533750217530812, "learning_rate": 2.3518496689645245e-05, "loss": 0.9415, "step": 7412 }, { "epoch": 3.358858178522882, "grad_norm": 0.14538374678139301, "learning_rate": 2.3506967899825597e-05, "loss": 0.9461, "step": 7413 }, { "epoch": 3.359311282283643, "grad_norm": 0.13786306707266954, "learning_rate": 2.3495440760501208e-05, "loss": 0.9202, "step": 7414 }, { "epoch": 3.359764386044404, "grad_norm": 0.1358873692198365, "learning_rate": 2.348391527282563e-05, "loss": 0.9407, "step": 7415 }, { "epoch": 3.3602174898051653, "grad_norm": 0.14052951737026256, "learning_rate": 2.3472391437952247e-05, "loss": 0.9264, "step": 7416 }, { "epoch": 3.3606705935659265, "grad_norm": 0.15076323818488033, "learning_rate": 2.346086925703428e-05, "loss": 0.9441, "step": 7417 }, { "epoch": 3.361123697326688, "grad_norm": 0.14983733761250617, "learning_rate": 2.344934873122476e-05, "loss": 0.9294, "step": 7418 }, { "epoch": 3.361576801087449, "grad_norm": 0.12739055239969335, "learning_rate": 2.343782986167659e-05, "loss": 0.9308, "step": 7419 }, { "epoch": 3.3620299048482103, "grad_norm": 0.1446762456745703, "learning_rate": 2.3426312649542478e-05, "loss": 0.9331, "step": 7420 }, { "epoch": 3.3624830086089714, "grad_norm": 0.16699322113761142, "learning_rate": 2.341479709597499e-05, "loss": 0.9326, "step": 7421 }, { "epoch": 3.3629361123697326, "grad_norm": 0.15177404732954491, "learning_rate": 2.3403283202126503e-05, "loss": 0.9154, "step": 7422 }, { "epoch": 3.3633892161304937, "grad_norm": 0.11651847209013973, "learning_rate": 2.3391770969149247e-05, "loss": 0.9393, "step": 7423 }, { "epoch": 3.3638423198912553, "grad_norm": 0.1556507397592214, "learning_rate": 2.338026039819527e-05, "loss": 0.9069, "step": 7424 }, { "epoch": 3.3642954236520164, "grad_norm": 0.1372421080298745, "learning_rate": 2.336875149041647e-05, "loss": 0.9089, "step": 7425 }, { "epoch": 3.3647485274127775, "grad_norm": 0.12814218729474566, "learning_rate": 2.335724424696456e-05, "loss": 0.9544, "step": 7426 }, { "epoch": 3.3652016311735387, "grad_norm": 0.146996546694183, "learning_rate": 2.3345738668991084e-05, "loss": 0.9401, "step": 7427 }, { "epoch": 3.3656547349343, "grad_norm": 0.12126385448475799, "learning_rate": 2.333423475764747e-05, "loss": 0.9194, "step": 7428 }, { "epoch": 3.3661078386950614, "grad_norm": 0.17031768525970234, "learning_rate": 2.3322732514084913e-05, "loss": 0.9404, "step": 7429 }, { "epoch": 3.3665609424558225, "grad_norm": 0.14753472908381482, "learning_rate": 2.3311231939454463e-05, "loss": 0.9626, "step": 7430 }, { "epoch": 3.3670140462165836, "grad_norm": 0.1501636150087672, "learning_rate": 2.3299733034907028e-05, "loss": 0.9538, "step": 7431 }, { "epoch": 3.3674671499773448, "grad_norm": 0.1774949677436266, "learning_rate": 2.3288235801593313e-05, "loss": 0.9191, "step": 7432 }, { "epoch": 3.367920253738106, "grad_norm": 0.1789373017261498, "learning_rate": 2.327674024066389e-05, "loss": 0.9229, "step": 7433 }, { "epoch": 3.368373357498867, "grad_norm": 0.16829622413114176, "learning_rate": 2.3265246353269122e-05, "loss": 0.9341, "step": 7434 }, { "epoch": 3.3688264612596286, "grad_norm": 0.13036203462068077, "learning_rate": 2.3253754140559258e-05, "loss": 0.9201, "step": 7435 }, { "epoch": 3.3692795650203897, "grad_norm": 0.1324760671938356, "learning_rate": 2.324226360368432e-05, "loss": 0.9073, "step": 7436 }, { "epoch": 3.369732668781151, "grad_norm": 0.1563466818933997, "learning_rate": 2.3230774743794217e-05, "loss": 0.9402, "step": 7437 }, { "epoch": 3.370185772541912, "grad_norm": 0.13587476862421377, "learning_rate": 2.3219287562038667e-05, "loss": 0.9313, "step": 7438 }, { "epoch": 3.3706388763026736, "grad_norm": 0.16223146704638633, "learning_rate": 2.3207802059567176e-05, "loss": 0.9356, "step": 7439 }, { "epoch": 3.3710919800634347, "grad_norm": 0.16542625264396058, "learning_rate": 2.319631823752918e-05, "loss": 0.9101, "step": 7440 }, { "epoch": 3.371545083824196, "grad_norm": 0.18223990785567076, "learning_rate": 2.3184836097073845e-05, "loss": 0.918, "step": 7441 }, { "epoch": 3.371998187584957, "grad_norm": 0.16993766868381743, "learning_rate": 2.317335563935025e-05, "loss": 0.9482, "step": 7442 }, { "epoch": 3.372451291345718, "grad_norm": 0.13512035708222647, "learning_rate": 2.3161876865507248e-05, "loss": 0.9377, "step": 7443 }, { "epoch": 3.372904395106479, "grad_norm": 0.16499205206746997, "learning_rate": 2.3150399776693564e-05, "loss": 0.9213, "step": 7444 }, { "epoch": 3.373357498867241, "grad_norm": 0.1648971660302918, "learning_rate": 2.3138924374057727e-05, "loss": 0.9567, "step": 7445 }, { "epoch": 3.373810602628002, "grad_norm": 0.13174844303641559, "learning_rate": 2.3127450658748096e-05, "loss": 0.9516, "step": 7446 }, { "epoch": 3.374263706388763, "grad_norm": 0.17344025831668836, "learning_rate": 2.311597863191289e-05, "loss": 0.945, "step": 7447 }, { "epoch": 3.374716810149524, "grad_norm": 0.18623942214931744, "learning_rate": 2.3104508294700124e-05, "loss": 0.9554, "step": 7448 }, { "epoch": 3.3751699139102853, "grad_norm": 0.1576392813575216, "learning_rate": 2.3093039648257676e-05, "loss": 0.9442, "step": 7449 }, { "epoch": 3.375623017671047, "grad_norm": 0.1461770296508739, "learning_rate": 2.3081572693733224e-05, "loss": 0.9419, "step": 7450 }, { "epoch": 3.376076121431808, "grad_norm": 0.18745988666608157, "learning_rate": 2.3070107432274317e-05, "loss": 0.9458, "step": 7451 }, { "epoch": 3.376529225192569, "grad_norm": 0.18645602210791032, "learning_rate": 2.305864386502828e-05, "loss": 0.9333, "step": 7452 }, { "epoch": 3.3769823289533303, "grad_norm": 0.1292765202844731, "learning_rate": 2.3047181993142332e-05, "loss": 0.9413, "step": 7453 }, { "epoch": 3.3774354327140914, "grad_norm": 0.17945369239686, "learning_rate": 2.303572181776346e-05, "loss": 0.9177, "step": 7454 }, { "epoch": 3.3778885364748525, "grad_norm": 0.21624477759853836, "learning_rate": 2.302426334003851e-05, "loss": 0.9278, "step": 7455 }, { "epoch": 3.378341640235614, "grad_norm": 0.16000438799802394, "learning_rate": 2.301280656111418e-05, "loss": 0.9295, "step": 7456 }, { "epoch": 3.3787947439963752, "grad_norm": 0.1653009283916326, "learning_rate": 2.300135148213695e-05, "loss": 0.9372, "step": 7457 }, { "epoch": 3.3792478477571364, "grad_norm": 0.2003189211387187, "learning_rate": 2.298989810425318e-05, "loss": 0.9349, "step": 7458 }, { "epoch": 3.3797009515178975, "grad_norm": 0.14989359533642022, "learning_rate": 2.297844642860901e-05, "loss": 0.9328, "step": 7459 }, { "epoch": 3.3801540552786586, "grad_norm": 0.150702134008236, "learning_rate": 2.2966996456350467e-05, "loss": 0.9175, "step": 7460 }, { "epoch": 3.38060715903942, "grad_norm": 0.18599174742546942, "learning_rate": 2.2955548188623347e-05, "loss": 0.9413, "step": 7461 }, { "epoch": 3.3810602628001813, "grad_norm": 0.20802528650752639, "learning_rate": 2.2944101626573323e-05, "loss": 0.9301, "step": 7462 }, { "epoch": 3.3815133665609425, "grad_norm": 0.143686896703762, "learning_rate": 2.2932656771345876e-05, "loss": 0.9294, "step": 7463 }, { "epoch": 3.3819664703217036, "grad_norm": 0.2024836164244296, "learning_rate": 2.2921213624086295e-05, "loss": 0.9418, "step": 7464 }, { "epoch": 3.3824195740824647, "grad_norm": 0.23010301276754358, "learning_rate": 2.2909772185939757e-05, "loss": 0.9247, "step": 7465 }, { "epoch": 3.382872677843226, "grad_norm": 0.13760084137104267, "learning_rate": 2.28983324580512e-05, "loss": 0.9428, "step": 7466 }, { "epoch": 3.3833257816039874, "grad_norm": 0.16292227066859896, "learning_rate": 2.288689444156546e-05, "loss": 0.9448, "step": 7467 }, { "epoch": 3.3837788853647486, "grad_norm": 0.20718835026574622, "learning_rate": 2.2875458137627143e-05, "loss": 0.9213, "step": 7468 }, { "epoch": 3.3842319891255097, "grad_norm": 0.1265530679558067, "learning_rate": 2.286402354738071e-05, "loss": 0.947, "step": 7469 }, { "epoch": 3.384685092886271, "grad_norm": 0.18818499680374823, "learning_rate": 2.285259067197043e-05, "loss": 0.9154, "step": 7470 }, { "epoch": 3.3851381966470324, "grad_norm": 0.16060561688460873, "learning_rate": 2.284115951254044e-05, "loss": 0.9144, "step": 7471 }, { "epoch": 3.3855913004077935, "grad_norm": 0.14428900990646729, "learning_rate": 2.2829730070234666e-05, "loss": 0.926, "step": 7472 }, { "epoch": 3.3860444041685547, "grad_norm": 0.1738322476148784, "learning_rate": 2.2818302346196894e-05, "loss": 0.9336, "step": 7473 }, { "epoch": 3.386497507929316, "grad_norm": 0.1898428894650975, "learning_rate": 2.2806876341570723e-05, "loss": 0.9326, "step": 7474 }, { "epoch": 3.386950611690077, "grad_norm": 0.14750463236361594, "learning_rate": 2.279545205749955e-05, "loss": 0.9587, "step": 7475 }, { "epoch": 3.387403715450838, "grad_norm": 0.14818864130071988, "learning_rate": 2.278402949512667e-05, "loss": 0.9097, "step": 7476 }, { "epoch": 3.3878568192115996, "grad_norm": 0.1928925169281741, "learning_rate": 2.2772608655595124e-05, "loss": 0.9223, "step": 7477 }, { "epoch": 3.3883099229723608, "grad_norm": 0.1565880450474125, "learning_rate": 2.276118954004786e-05, "loss": 0.9244, "step": 7478 }, { "epoch": 3.388763026733122, "grad_norm": 0.1276438429541394, "learning_rate": 2.2749772149627578e-05, "loss": 0.9475, "step": 7479 }, { "epoch": 3.389216130493883, "grad_norm": 0.18719306589577095, "learning_rate": 2.2738356485476878e-05, "loss": 0.9466, "step": 7480 }, { "epoch": 3.389669234254644, "grad_norm": 0.18647849935059893, "learning_rate": 2.2726942548738124e-05, "loss": 0.9391, "step": 7481 }, { "epoch": 3.3901223380154057, "grad_norm": 0.14727597954252758, "learning_rate": 2.2715530340553555e-05, "loss": 0.912, "step": 7482 }, { "epoch": 3.390575441776167, "grad_norm": 0.16145377156573312, "learning_rate": 2.2704119862065214e-05, "loss": 0.9245, "step": 7483 }, { "epoch": 3.391028545536928, "grad_norm": 0.17485953796338832, "learning_rate": 2.269271111441494e-05, "loss": 0.9487, "step": 7484 }, { "epoch": 3.391481649297689, "grad_norm": 0.1378851753866253, "learning_rate": 2.2681304098744485e-05, "loss": 0.9117, "step": 7485 }, { "epoch": 3.3919347530584503, "grad_norm": 0.14900475128312915, "learning_rate": 2.266989881619533e-05, "loss": 0.9276, "step": 7486 }, { "epoch": 3.3923878568192114, "grad_norm": 0.17074703987918188, "learning_rate": 2.265849526790886e-05, "loss": 0.9335, "step": 7487 }, { "epoch": 3.392840960579973, "grad_norm": 0.13826937778535262, "learning_rate": 2.2647093455026224e-05, "loss": 0.9424, "step": 7488 }, { "epoch": 3.393294064340734, "grad_norm": 0.15417336342263882, "learning_rate": 2.2635693378688463e-05, "loss": 0.9331, "step": 7489 }, { "epoch": 3.3937471681014952, "grad_norm": 0.13741840476348652, "learning_rate": 2.262429504003639e-05, "loss": 0.9184, "step": 7490 }, { "epoch": 3.3942002718622564, "grad_norm": 0.14931281771417457, "learning_rate": 2.2612898440210645e-05, "loss": 0.9565, "step": 7491 }, { "epoch": 3.3946533756230175, "grad_norm": 0.12639164142804293, "learning_rate": 2.2601503580351744e-05, "loss": 0.9325, "step": 7492 }, { "epoch": 3.395106479383779, "grad_norm": 0.13974086543683428, "learning_rate": 2.259011046159997e-05, "loss": 0.9389, "step": 7493 }, { "epoch": 3.39555958314454, "grad_norm": 0.1352030700374154, "learning_rate": 2.257871908509548e-05, "loss": 0.9343, "step": 7494 }, { "epoch": 3.3960126869053013, "grad_norm": 0.12646158652963582, "learning_rate": 2.256732945197821e-05, "loss": 0.8905, "step": 7495 }, { "epoch": 3.3964657906660625, "grad_norm": 0.1257882552333704, "learning_rate": 2.2555941563387974e-05, "loss": 0.9096, "step": 7496 }, { "epoch": 3.3969188944268236, "grad_norm": 0.12417092012335802, "learning_rate": 2.2544555420464364e-05, "loss": 0.922, "step": 7497 }, { "epoch": 3.3973719981875847, "grad_norm": 0.1269199287279318, "learning_rate": 2.253317102434683e-05, "loss": 0.937, "step": 7498 }, { "epoch": 3.3978251019483463, "grad_norm": 0.11796510932156835, "learning_rate": 2.252178837617464e-05, "loss": 0.9182, "step": 7499 }, { "epoch": 3.3982782057091074, "grad_norm": 0.12408706512294185, "learning_rate": 2.2510407477086845e-05, "loss": 0.9217, "step": 7500 }, { "epoch": 3.3987313094698686, "grad_norm": 0.12589842354972264, "learning_rate": 2.2499028328222403e-05, "loss": 0.9558, "step": 7501 }, { "epoch": 3.3991844132306297, "grad_norm": 0.11774500295623692, "learning_rate": 2.2487650930720014e-05, "loss": 0.9464, "step": 7502 }, { "epoch": 3.3996375169913913, "grad_norm": 0.11419560625248851, "learning_rate": 2.247627528571828e-05, "loss": 0.9253, "step": 7503 }, { "epoch": 3.4000906207521524, "grad_norm": 0.10968253399819335, "learning_rate": 2.2464901394355537e-05, "loss": 0.9285, "step": 7504 }, { "epoch": 3.4005437245129135, "grad_norm": 0.11760894722933271, "learning_rate": 2.245352925777005e-05, "loss": 0.9402, "step": 7505 }, { "epoch": 3.4009968282736747, "grad_norm": 0.1078115563638834, "learning_rate": 2.2442158877099824e-05, "loss": 0.9292, "step": 7506 }, { "epoch": 3.401449932034436, "grad_norm": 0.11401708178024034, "learning_rate": 2.2430790253482712e-05, "loss": 0.9315, "step": 7507 }, { "epoch": 3.401903035795197, "grad_norm": 0.11747656202818393, "learning_rate": 2.241942338805642e-05, "loss": 0.9204, "step": 7508 }, { "epoch": 3.4023561395559585, "grad_norm": 0.11756330714222424, "learning_rate": 2.240805828195843e-05, "loss": 0.9193, "step": 7509 }, { "epoch": 3.4028092433167196, "grad_norm": 0.1233516630728163, "learning_rate": 2.2396694936326106e-05, "loss": 0.9139, "step": 7510 }, { "epoch": 3.4032623470774808, "grad_norm": 0.12685080989988087, "learning_rate": 2.2385333352296576e-05, "loss": 0.9329, "step": 7511 }, { "epoch": 3.403715450838242, "grad_norm": 0.13268659455981588, "learning_rate": 2.237397353100684e-05, "loss": 0.9102, "step": 7512 }, { "epoch": 3.404168554599003, "grad_norm": 0.1389855158079578, "learning_rate": 2.2362615473593684e-05, "loss": 0.9615, "step": 7513 }, { "epoch": 3.4046216583597646, "grad_norm": 0.1273617015699272, "learning_rate": 2.2351259181193753e-05, "loss": 0.9308, "step": 7514 }, { "epoch": 3.4050747621205257, "grad_norm": 0.11601645012713037, "learning_rate": 2.233990465494349e-05, "loss": 0.927, "step": 7515 }, { "epoch": 3.405527865881287, "grad_norm": 0.11772776480786533, "learning_rate": 2.232855189597915e-05, "loss": 0.9368, "step": 7516 }, { "epoch": 3.405980969642048, "grad_norm": 0.1285064083569019, "learning_rate": 2.231720090543686e-05, "loss": 0.9488, "step": 7517 }, { "epoch": 3.406434073402809, "grad_norm": 0.11772368540076275, "learning_rate": 2.2305851684452503e-05, "loss": 0.9409, "step": 7518 }, { "epoch": 3.4068871771635703, "grad_norm": 0.1276222236666031, "learning_rate": 2.2294504234161868e-05, "loss": 0.9402, "step": 7519 }, { "epoch": 3.407340280924332, "grad_norm": 0.1304679161486615, "learning_rate": 2.2283158555700473e-05, "loss": 0.929, "step": 7520 }, { "epoch": 3.407793384685093, "grad_norm": 0.13933287733984662, "learning_rate": 2.2271814650203748e-05, "loss": 0.9379, "step": 7521 }, { "epoch": 3.408246488445854, "grad_norm": 0.11943458800606924, "learning_rate": 2.2260472518806862e-05, "loss": 0.9479, "step": 7522 }, { "epoch": 3.4086995922066152, "grad_norm": 0.12949659974466077, "learning_rate": 2.2249132162644892e-05, "loss": 0.9346, "step": 7523 }, { "epoch": 3.4091526959673764, "grad_norm": 0.1300073706148242, "learning_rate": 2.223779358285266e-05, "loss": 0.9276, "step": 7524 }, { "epoch": 3.409605799728138, "grad_norm": 0.1391021039302358, "learning_rate": 2.2226456780564857e-05, "loss": 0.9247, "step": 7525 }, { "epoch": 3.410058903488899, "grad_norm": 0.10767404881401772, "learning_rate": 2.2215121756915985e-05, "loss": 0.9391, "step": 7526 }, { "epoch": 3.41051200724966, "grad_norm": 0.12422576598891344, "learning_rate": 2.2203788513040337e-05, "loss": 0.9565, "step": 7527 }, { "epoch": 3.4109651110104213, "grad_norm": 0.11109746091227321, "learning_rate": 2.21924570500721e-05, "loss": 0.9369, "step": 7528 }, { "epoch": 3.4114182147711825, "grad_norm": 0.13007613539491902, "learning_rate": 2.2181127369145202e-05, "loss": 0.9147, "step": 7529 }, { "epoch": 3.4118713185319436, "grad_norm": 0.13942109789039767, "learning_rate": 2.2169799471393466e-05, "loss": 0.9346, "step": 7530 }, { "epoch": 3.412324422292705, "grad_norm": 0.14708324643759332, "learning_rate": 2.2158473357950463e-05, "loss": 0.9189, "step": 7531 }, { "epoch": 3.4127775260534663, "grad_norm": 0.12535790983017361, "learning_rate": 2.214714902994965e-05, "loss": 0.9475, "step": 7532 }, { "epoch": 3.4132306298142274, "grad_norm": 0.14920255894592863, "learning_rate": 2.2135826488524264e-05, "loss": 0.9238, "step": 7533 }, { "epoch": 3.4136837335749886, "grad_norm": 0.18328775672651934, "learning_rate": 2.2124505734807392e-05, "loss": 0.9334, "step": 7534 }, { "epoch": 3.41413683733575, "grad_norm": 0.13548848827676035, "learning_rate": 2.211318676993192e-05, "loss": 0.9394, "step": 7535 }, { "epoch": 3.4145899410965113, "grad_norm": 0.12050467781189421, "learning_rate": 2.2101869595030553e-05, "loss": 0.9319, "step": 7536 }, { "epoch": 3.4150430448572724, "grad_norm": 0.17740469914853876, "learning_rate": 2.209055421123584e-05, "loss": 0.9442, "step": 7537 }, { "epoch": 3.4154961486180335, "grad_norm": 0.15275705160091993, "learning_rate": 2.2079240619680115e-05, "loss": 0.9134, "step": 7538 }, { "epoch": 3.4159492523787947, "grad_norm": 0.11886970474634126, "learning_rate": 2.2067928821495598e-05, "loss": 0.9281, "step": 7539 }, { "epoch": 3.416402356139556, "grad_norm": 0.16451354958518719, "learning_rate": 2.2056618817814238e-05, "loss": 0.9627, "step": 7540 }, { "epoch": 3.4168554599003174, "grad_norm": 0.18130179691802484, "learning_rate": 2.204531060976789e-05, "loss": 0.9171, "step": 7541 }, { "epoch": 3.4173085636610785, "grad_norm": 0.11404101442156253, "learning_rate": 2.2034004198488175e-05, "loss": 0.9341, "step": 7542 }, { "epoch": 3.4177616674218396, "grad_norm": 0.155416438285995, "learning_rate": 2.2022699585106544e-05, "loss": 0.9178, "step": 7543 }, { "epoch": 3.4182147711826008, "grad_norm": 0.16287866473317317, "learning_rate": 2.2011396770754296e-05, "loss": 0.9338, "step": 7544 }, { "epoch": 3.418667874943362, "grad_norm": 0.1363071534228113, "learning_rate": 2.2000095756562506e-05, "loss": 0.9269, "step": 7545 }, { "epoch": 3.4191209787041235, "grad_norm": 0.12941680067286904, "learning_rate": 2.1988796543662115e-05, "loss": 0.9375, "step": 7546 }, { "epoch": 3.4195740824648846, "grad_norm": 0.13086085043612822, "learning_rate": 2.1977499133183837e-05, "loss": 0.941, "step": 7547 }, { "epoch": 3.4200271862256457, "grad_norm": 0.11454724476461599, "learning_rate": 2.1966203526258258e-05, "loss": 0.927, "step": 7548 }, { "epoch": 3.420480289986407, "grad_norm": 0.13417975917564987, "learning_rate": 2.1954909724015723e-05, "loss": 0.9257, "step": 7549 }, { "epoch": 3.420933393747168, "grad_norm": 0.12822602436476713, "learning_rate": 2.1943617727586463e-05, "loss": 0.9344, "step": 7550 }, { "epoch": 3.421386497507929, "grad_norm": 0.122400625039134, "learning_rate": 2.1932327538100476e-05, "loss": 0.9436, "step": 7551 }, { "epoch": 3.4218396012686907, "grad_norm": 0.1471800917413598, "learning_rate": 2.1921039156687582e-05, "loss": 0.9355, "step": 7552 }, { "epoch": 3.422292705029452, "grad_norm": 0.12482404392938139, "learning_rate": 2.1909752584477464e-05, "loss": 0.9227, "step": 7553 }, { "epoch": 3.422745808790213, "grad_norm": 0.12473066547667022, "learning_rate": 2.1898467822599567e-05, "loss": 0.9476, "step": 7554 }, { "epoch": 3.423198912550974, "grad_norm": 0.12108263584701862, "learning_rate": 2.188718487218321e-05, "loss": 0.9324, "step": 7555 }, { "epoch": 3.423652016311735, "grad_norm": 0.1258746817043786, "learning_rate": 2.1875903734357477e-05, "loss": 0.9345, "step": 7556 }, { "epoch": 3.424105120072497, "grad_norm": 0.12687791764688575, "learning_rate": 2.1864624410251327e-05, "loss": 0.9121, "step": 7557 }, { "epoch": 3.424558223833258, "grad_norm": 0.14149907254878144, "learning_rate": 2.1853346900993478e-05, "loss": 0.9347, "step": 7558 }, { "epoch": 3.425011327594019, "grad_norm": 0.1118103137038823, "learning_rate": 2.1842071207712516e-05, "loss": 0.9177, "step": 7559 }, { "epoch": 3.42546443135478, "grad_norm": 0.12725074060948716, "learning_rate": 2.183079733153683e-05, "loss": 0.9139, "step": 7560 }, { "epoch": 3.4259175351155413, "grad_norm": 0.13521655770716096, "learning_rate": 2.1819525273594585e-05, "loss": 0.9302, "step": 7561 }, { "epoch": 3.4263706388763024, "grad_norm": 0.13695577907443154, "learning_rate": 2.180825503501385e-05, "loss": 0.9281, "step": 7562 }, { "epoch": 3.426823742637064, "grad_norm": 0.14280836374682113, "learning_rate": 2.1796986616922424e-05, "loss": 0.9212, "step": 7563 }, { "epoch": 3.427276846397825, "grad_norm": 0.1199310842342489, "learning_rate": 2.178572002044799e-05, "loss": 0.9463, "step": 7564 }, { "epoch": 3.4277299501585863, "grad_norm": 0.15079717150724667, "learning_rate": 2.1774455246718004e-05, "loss": 0.9209, "step": 7565 }, { "epoch": 3.4281830539193474, "grad_norm": 0.12455428264825762, "learning_rate": 2.176319229685978e-05, "loss": 0.9315, "step": 7566 }, { "epoch": 3.428636157680109, "grad_norm": 0.1342325135458681, "learning_rate": 2.175193117200041e-05, "loss": 0.9242, "step": 7567 }, { "epoch": 3.42908926144087, "grad_norm": 0.13560341711649584, "learning_rate": 2.174067187326681e-05, "loss": 0.9381, "step": 7568 }, { "epoch": 3.4295423652016312, "grad_norm": 0.13174848273803785, "learning_rate": 2.1729414401785753e-05, "loss": 0.9284, "step": 7569 }, { "epoch": 3.4299954689623924, "grad_norm": 0.12714485491002234, "learning_rate": 2.1718158758683762e-05, "loss": 0.9318, "step": 7570 }, { "epoch": 3.4304485727231535, "grad_norm": 0.14124812528553726, "learning_rate": 2.1706904945087258e-05, "loss": 0.9387, "step": 7571 }, { "epoch": 3.4309016764839146, "grad_norm": 0.11920759107998763, "learning_rate": 2.1695652962122392e-05, "loss": 0.912, "step": 7572 }, { "epoch": 3.431354780244676, "grad_norm": 0.15253628909168487, "learning_rate": 2.1684402810915223e-05, "loss": 0.9458, "step": 7573 }, { "epoch": 3.4318078840054373, "grad_norm": 0.12289478452108314, "learning_rate": 2.167315449259153e-05, "loss": 0.9276, "step": 7574 }, { "epoch": 3.4322609877661985, "grad_norm": 0.13440364738200636, "learning_rate": 2.1661908008277e-05, "loss": 0.9307, "step": 7575 }, { "epoch": 3.4327140915269596, "grad_norm": 0.12068566051420539, "learning_rate": 2.1650663359097078e-05, "loss": 0.9415, "step": 7576 }, { "epoch": 3.4331671952877207, "grad_norm": 0.11992004531837634, "learning_rate": 2.1639420546177022e-05, "loss": 0.918, "step": 7577 }, { "epoch": 3.4336202990484823, "grad_norm": 0.12947650978476863, "learning_rate": 2.162817957064196e-05, "loss": 0.9361, "step": 7578 }, { "epoch": 3.4340734028092434, "grad_norm": 0.13964817209274358, "learning_rate": 2.1616940433616768e-05, "loss": 0.9238, "step": 7579 }, { "epoch": 3.4345265065700046, "grad_norm": 0.1336206930227973, "learning_rate": 2.160570313622621e-05, "loss": 0.9492, "step": 7580 }, { "epoch": 3.4349796103307657, "grad_norm": 0.15125098823759406, "learning_rate": 2.159446767959479e-05, "loss": 0.9495, "step": 7581 }, { "epoch": 3.435432714091527, "grad_norm": 0.13418543112134768, "learning_rate": 2.158323406484691e-05, "loss": 0.9012, "step": 7582 }, { "epoch": 3.435885817852288, "grad_norm": 0.12353808352646647, "learning_rate": 2.157200229310669e-05, "loss": 0.9245, "step": 7583 }, { "epoch": 3.4363389216130495, "grad_norm": 0.12927031581028633, "learning_rate": 2.156077236549816e-05, "loss": 0.9429, "step": 7584 }, { "epoch": 3.4367920253738107, "grad_norm": 0.1181973501735503, "learning_rate": 2.1549544283145094e-05, "loss": 0.9134, "step": 7585 }, { "epoch": 3.437245129134572, "grad_norm": 0.11942914458412755, "learning_rate": 2.1538318047171142e-05, "loss": 0.9176, "step": 7586 }, { "epoch": 3.437698232895333, "grad_norm": 0.11739776015516154, "learning_rate": 2.1527093658699725e-05, "loss": 0.9245, "step": 7587 }, { "epoch": 3.438151336656094, "grad_norm": 0.11598174802334621, "learning_rate": 2.151587111885408e-05, "loss": 0.9257, "step": 7588 }, { "epoch": 3.4386044404168556, "grad_norm": 0.12380629096878587, "learning_rate": 2.1504650428757298e-05, "loss": 0.9155, "step": 7589 }, { "epoch": 3.4390575441776168, "grad_norm": 0.13332571059814785, "learning_rate": 2.1493431589532232e-05, "loss": 0.9374, "step": 7590 }, { "epoch": 3.439510647938378, "grad_norm": 0.1331667725269515, "learning_rate": 2.1482214602301602e-05, "loss": 0.937, "step": 7591 }, { "epoch": 3.439963751699139, "grad_norm": 0.12623426484362368, "learning_rate": 2.1470999468187896e-05, "loss": 0.9292, "step": 7592 }, { "epoch": 3.4404168554599, "grad_norm": 0.1284692459801862, "learning_rate": 2.145978618831346e-05, "loss": 0.9457, "step": 7593 }, { "epoch": 3.4408699592206613, "grad_norm": 0.13639824766824474, "learning_rate": 2.1448574763800405e-05, "loss": 0.9382, "step": 7594 }, { "epoch": 3.441323062981423, "grad_norm": 0.10712412389636244, "learning_rate": 2.1437365195770717e-05, "loss": 0.9192, "step": 7595 }, { "epoch": 3.441776166742184, "grad_norm": 0.1284664377848176, "learning_rate": 2.142615748534615e-05, "loss": 0.9184, "step": 7596 }, { "epoch": 3.442229270502945, "grad_norm": 0.11299916434172591, "learning_rate": 2.1414951633648258e-05, "loss": 0.925, "step": 7597 }, { "epoch": 3.4426823742637063, "grad_norm": 0.12083381415811006, "learning_rate": 2.1403747641798475e-05, "loss": 0.9418, "step": 7598 }, { "epoch": 3.443135478024468, "grad_norm": 0.11439739321991647, "learning_rate": 2.139254551091798e-05, "loss": 0.9373, "step": 7599 }, { "epoch": 3.443588581785229, "grad_norm": 0.1139630165240876, "learning_rate": 2.138134524212782e-05, "loss": 0.9373, "step": 7600 }, { "epoch": 3.44404168554599, "grad_norm": 0.11579699310189104, "learning_rate": 2.137014683654881e-05, "loss": 0.9256, "step": 7601 }, { "epoch": 3.4444947893067512, "grad_norm": 0.137501502846996, "learning_rate": 2.1358950295301625e-05, "loss": 0.9307, "step": 7602 }, { "epoch": 3.4449478930675124, "grad_norm": 0.12200600820970513, "learning_rate": 2.1347755619506714e-05, "loss": 0.94, "step": 7603 }, { "epoch": 3.4454009968282735, "grad_norm": 0.12810086452191963, "learning_rate": 2.133656281028434e-05, "loss": 0.936, "step": 7604 }, { "epoch": 3.445854100589035, "grad_norm": 0.14952194484021827, "learning_rate": 2.132537186875462e-05, "loss": 0.9283, "step": 7605 }, { "epoch": 3.446307204349796, "grad_norm": 0.11502218263296332, "learning_rate": 2.1314182796037424e-05, "loss": 0.9268, "step": 7606 }, { "epoch": 3.4467603081105573, "grad_norm": 0.15948976716779314, "learning_rate": 2.130299559325251e-05, "loss": 0.9043, "step": 7607 }, { "epoch": 3.4472134118713185, "grad_norm": 0.11950417574721443, "learning_rate": 2.1291810261519365e-05, "loss": 0.9198, "step": 7608 }, { "epoch": 3.4476665156320796, "grad_norm": 0.13966649819680116, "learning_rate": 2.1280626801957363e-05, "loss": 0.928, "step": 7609 }, { "epoch": 3.448119619392841, "grad_norm": 0.16977018781073944, "learning_rate": 2.1269445215685637e-05, "loss": 0.9504, "step": 7610 }, { "epoch": 3.4485727231536023, "grad_norm": 0.1412085495111609, "learning_rate": 2.1258265503823166e-05, "loss": 0.9275, "step": 7611 }, { "epoch": 3.4490258269143634, "grad_norm": 0.13615506680425288, "learning_rate": 2.1247087667488735e-05, "loss": 0.9318, "step": 7612 }, { "epoch": 3.4494789306751246, "grad_norm": 0.1484735157846907, "learning_rate": 2.1235911707800908e-05, "loss": 0.935, "step": 7613 }, { "epoch": 3.4499320344358857, "grad_norm": 0.13852513541757427, "learning_rate": 2.122473762587812e-05, "loss": 0.9389, "step": 7614 }, { "epoch": 3.450385138196647, "grad_norm": 0.13106815392446852, "learning_rate": 2.1213565422838563e-05, "loss": 0.9232, "step": 7615 }, { "epoch": 3.4508382419574084, "grad_norm": 0.1395620643665146, "learning_rate": 2.1202395099800287e-05, "loss": 0.9208, "step": 7616 }, { "epoch": 3.4512913457181695, "grad_norm": 0.12967763129092716, "learning_rate": 2.11912266578811e-05, "loss": 0.9461, "step": 7617 }, { "epoch": 3.4517444494789307, "grad_norm": 0.12585509391993954, "learning_rate": 2.1180060098198693e-05, "loss": 0.929, "step": 7618 }, { "epoch": 3.452197553239692, "grad_norm": 0.12249916496244624, "learning_rate": 2.11688954218705e-05, "loss": 0.9474, "step": 7619 }, { "epoch": 3.452650657000453, "grad_norm": 0.43449350644549817, "learning_rate": 2.115773263001381e-05, "loss": 0.9327, "step": 7620 }, { "epoch": 3.4531037607612145, "grad_norm": 0.1335381618844656, "learning_rate": 2.1146571723745705e-05, "loss": 0.9245, "step": 7621 }, { "epoch": 3.4535568645219756, "grad_norm": 0.1156116802785032, "learning_rate": 2.1135412704183067e-05, "loss": 0.9051, "step": 7622 }, { "epoch": 3.4540099682827368, "grad_norm": 0.12900487415250697, "learning_rate": 2.1124255572442634e-05, "loss": 0.9361, "step": 7623 }, { "epoch": 3.454463072043498, "grad_norm": 0.13439678084568824, "learning_rate": 2.11131003296409e-05, "loss": 0.9264, "step": 7624 }, { "epoch": 3.454916175804259, "grad_norm": 0.11740517482576164, "learning_rate": 2.110194697689421e-05, "loss": 0.9258, "step": 7625 }, { "epoch": 3.45536927956502, "grad_norm": 0.13946522316529453, "learning_rate": 2.1090795515318695e-05, "loss": 0.9685, "step": 7626 }, { "epoch": 3.4558223833257817, "grad_norm": 0.13172575328719577, "learning_rate": 2.1079645946030323e-05, "loss": 0.9392, "step": 7627 }, { "epoch": 3.456275487086543, "grad_norm": 0.1283199279200596, "learning_rate": 2.106849827014485e-05, "loss": 0.9234, "step": 7628 }, { "epoch": 3.456728590847304, "grad_norm": 0.12679609859275198, "learning_rate": 2.1057352488777826e-05, "loss": 0.9508, "step": 7629 }, { "epoch": 3.457181694608065, "grad_norm": 0.12754172576274603, "learning_rate": 2.1046208603044677e-05, "loss": 0.9223, "step": 7630 }, { "epoch": 3.4576347983688267, "grad_norm": 0.12999180312650785, "learning_rate": 2.103506661406056e-05, "loss": 0.9243, "step": 7631 }, { "epoch": 3.458087902129588, "grad_norm": 0.47777370531498614, "learning_rate": 2.10239265229405e-05, "loss": 0.9585, "step": 7632 }, { "epoch": 3.458541005890349, "grad_norm": 0.14766593515153897, "learning_rate": 2.10127883307993e-05, "loss": 0.9279, "step": 7633 }, { "epoch": 3.45899410965111, "grad_norm": 0.13528112190431518, "learning_rate": 2.10016520387516e-05, "loss": 0.9538, "step": 7634 }, { "epoch": 3.4594472134118712, "grad_norm": 0.12490437184439508, "learning_rate": 2.099051764791181e-05, "loss": 0.9096, "step": 7635 }, { "epoch": 3.4599003171726324, "grad_norm": 0.12948225476239345, "learning_rate": 2.0979385159394203e-05, "loss": 0.9313, "step": 7636 }, { "epoch": 3.460353420933394, "grad_norm": 0.1429350308568753, "learning_rate": 2.0968254574312815e-05, "loss": 0.9215, "step": 7637 }, { "epoch": 3.460806524694155, "grad_norm": 0.11810238568957246, "learning_rate": 2.0957125893781497e-05, "loss": 0.9422, "step": 7638 }, { "epoch": 3.461259628454916, "grad_norm": 0.14883134270953624, "learning_rate": 2.0945999118913952e-05, "loss": 0.9191, "step": 7639 }, { "epoch": 3.4617127322156773, "grad_norm": 0.13898059135834434, "learning_rate": 2.093487425082364e-05, "loss": 0.9316, "step": 7640 }, { "epoch": 3.4621658359764385, "grad_norm": 0.10717059791364718, "learning_rate": 2.0923751290623862e-05, "loss": 0.9282, "step": 7641 }, { "epoch": 3.4626189397372, "grad_norm": 0.15535048982125355, "learning_rate": 2.0912630239427695e-05, "loss": 0.9324, "step": 7642 }, { "epoch": 3.463072043497961, "grad_norm": 0.13653545258750605, "learning_rate": 2.0901511098348087e-05, "loss": 0.916, "step": 7643 }, { "epoch": 3.4635251472587223, "grad_norm": 0.12218228135951964, "learning_rate": 2.089039386849771e-05, "loss": 0.9345, "step": 7644 }, { "epoch": 3.4639782510194834, "grad_norm": 0.13884764355427026, "learning_rate": 2.0879278550989137e-05, "loss": 0.9446, "step": 7645 }, { "epoch": 3.4644313547802446, "grad_norm": 0.15646688295062097, "learning_rate": 2.0868165146934665e-05, "loss": 0.942, "step": 7646 }, { "epoch": 3.4648844585410057, "grad_norm": 0.16292796361050033, "learning_rate": 2.0857053657446467e-05, "loss": 0.9401, "step": 7647 }, { "epoch": 3.4653375623017673, "grad_norm": 0.13112523075019203, "learning_rate": 2.084594408363648e-05, "loss": 0.9588, "step": 7648 }, { "epoch": 3.4657906660625284, "grad_norm": 0.15807914281960464, "learning_rate": 2.0834836426616456e-05, "loss": 0.9359, "step": 7649 }, { "epoch": 3.4662437698232895, "grad_norm": 0.13749402528466217, "learning_rate": 2.0823730687497985e-05, "loss": 0.9372, "step": 7650 }, { "epoch": 3.4666968735840507, "grad_norm": 0.16118314261235206, "learning_rate": 2.0812626867392423e-05, "loss": 0.93, "step": 7651 }, { "epoch": 3.467149977344812, "grad_norm": 0.13987036834943103, "learning_rate": 2.0801524967410972e-05, "loss": 0.9662, "step": 7652 }, { "epoch": 3.4676030811055734, "grad_norm": 0.12563712902291058, "learning_rate": 2.0790424988664607e-05, "loss": 0.9394, "step": 7653 }, { "epoch": 3.4680561848663345, "grad_norm": 0.1843348651543606, "learning_rate": 2.0779326932264147e-05, "loss": 0.9396, "step": 7654 }, { "epoch": 3.4685092886270956, "grad_norm": 0.1245482210928235, "learning_rate": 2.076823079932018e-05, "loss": 0.92, "step": 7655 }, { "epoch": 3.4689623923878568, "grad_norm": 0.14421153212370214, "learning_rate": 2.0757136590943144e-05, "loss": 0.9459, "step": 7656 }, { "epoch": 3.469415496148618, "grad_norm": 0.13773010937791344, "learning_rate": 2.0746044308243246e-05, "loss": 0.9154, "step": 7657 }, { "epoch": 3.469868599909379, "grad_norm": 0.13341277771641086, "learning_rate": 2.0734953952330512e-05, "loss": 0.9476, "step": 7658 }, { "epoch": 3.4703217036701406, "grad_norm": 0.12529450010644125, "learning_rate": 2.072386552431479e-05, "loss": 0.9297, "step": 7659 }, { "epoch": 3.4707748074309017, "grad_norm": 0.1266329215955637, "learning_rate": 2.0712779025305714e-05, "loss": 0.9443, "step": 7660 }, { "epoch": 3.471227911191663, "grad_norm": 0.150027314832669, "learning_rate": 2.070169445641276e-05, "loss": 0.9148, "step": 7661 }, { "epoch": 3.471681014952424, "grad_norm": 0.12119552696597152, "learning_rate": 2.0690611818745142e-05, "loss": 0.9399, "step": 7662 }, { "epoch": 3.4721341187131856, "grad_norm": 0.1433921509562761, "learning_rate": 2.0679531113411974e-05, "loss": 0.921, "step": 7663 }, { "epoch": 3.4725872224739467, "grad_norm": 0.11721846262426315, "learning_rate": 2.06684523415221e-05, "loss": 0.9456, "step": 7664 }, { "epoch": 3.473040326234708, "grad_norm": 0.13349378203584, "learning_rate": 2.065737550418419e-05, "loss": 0.921, "step": 7665 }, { "epoch": 3.473493429995469, "grad_norm": 0.11651048240648293, "learning_rate": 2.064630060250675e-05, "loss": 0.9281, "step": 7666 }, { "epoch": 3.47394653375623, "grad_norm": 0.14767349893800039, "learning_rate": 2.0635227637598053e-05, "loss": 0.9201, "step": 7667 }, { "epoch": 3.474399637516991, "grad_norm": 0.14620185459854096, "learning_rate": 2.062415661056621e-05, "loss": 0.9127, "step": 7668 }, { "epoch": 3.474852741277753, "grad_norm": 0.14430593206858935, "learning_rate": 2.061308752251911e-05, "loss": 0.9169, "step": 7669 }, { "epoch": 3.475305845038514, "grad_norm": 0.13936416159369583, "learning_rate": 2.0602020374564483e-05, "loss": 0.945, "step": 7670 }, { "epoch": 3.475758948799275, "grad_norm": 0.12623548892488903, "learning_rate": 2.059095516780981e-05, "loss": 0.9109, "step": 7671 }, { "epoch": 3.476212052560036, "grad_norm": 0.1473188756418537, "learning_rate": 2.057989190336245e-05, "loss": 0.9112, "step": 7672 }, { "epoch": 3.4766651563207973, "grad_norm": 0.1547880896174699, "learning_rate": 2.056883058232951e-05, "loss": 0.9439, "step": 7673 }, { "epoch": 3.477118260081559, "grad_norm": 0.13015748602948485, "learning_rate": 2.055777120581791e-05, "loss": 0.936, "step": 7674 }, { "epoch": 3.47757136384232, "grad_norm": 0.12965107268034387, "learning_rate": 2.054671377493441e-05, "loss": 0.9404, "step": 7675 }, { "epoch": 3.478024467603081, "grad_norm": 0.15411785889079124, "learning_rate": 2.0535658290785525e-05, "loss": 0.919, "step": 7676 }, { "epoch": 3.4784775713638423, "grad_norm": 0.11189205850484397, "learning_rate": 2.052460475447764e-05, "loss": 0.9187, "step": 7677 }, { "epoch": 3.4789306751246034, "grad_norm": 0.1551258892607926, "learning_rate": 2.0513553167116864e-05, "loss": 0.9183, "step": 7678 }, { "epoch": 3.4793837788853645, "grad_norm": 0.11286601341823617, "learning_rate": 2.0502503529809195e-05, "loss": 0.9303, "step": 7679 }, { "epoch": 3.479836882646126, "grad_norm": 0.13864705529484084, "learning_rate": 2.049145584366036e-05, "loss": 0.9369, "step": 7680 }, { "epoch": 3.4802899864068872, "grad_norm": 0.13469164823931554, "learning_rate": 2.0480410109775963e-05, "loss": 0.917, "step": 7681 }, { "epoch": 3.4807430901676484, "grad_norm": 0.11511285397485688, "learning_rate": 2.0469366329261352e-05, "loss": 0.9099, "step": 7682 }, { "epoch": 3.4811961939284095, "grad_norm": 0.15488575183805076, "learning_rate": 2.0458324503221697e-05, "loss": 0.9412, "step": 7683 }, { "epoch": 3.4816492976891706, "grad_norm": 0.10985530364098288, "learning_rate": 2.0447284632762e-05, "loss": 0.9185, "step": 7684 }, { "epoch": 3.482102401449932, "grad_norm": 0.14657473762755613, "learning_rate": 2.043624671898702e-05, "loss": 0.9205, "step": 7685 }, { "epoch": 3.4825555052106933, "grad_norm": 0.12146003879862746, "learning_rate": 2.0425210763001376e-05, "loss": 0.9328, "step": 7686 }, { "epoch": 3.4830086089714545, "grad_norm": 0.1313794645164217, "learning_rate": 2.0414176765909433e-05, "loss": 0.9338, "step": 7687 }, { "epoch": 3.4834617127322156, "grad_norm": 0.1325341883349881, "learning_rate": 2.0403144728815417e-05, "loss": 0.9296, "step": 7688 }, { "epoch": 3.4839148164929767, "grad_norm": 0.1163808496143483, "learning_rate": 2.0392114652823308e-05, "loss": 0.9296, "step": 7689 }, { "epoch": 3.484367920253738, "grad_norm": 0.11968417617663245, "learning_rate": 2.0381086539036906e-05, "loss": 0.9393, "step": 7690 }, { "epoch": 3.4848210240144994, "grad_norm": 0.12079126493307372, "learning_rate": 2.037006038855984e-05, "loss": 0.9135, "step": 7691 }, { "epoch": 3.4852741277752606, "grad_norm": 0.13069545123791917, "learning_rate": 2.035903620249551e-05, "loss": 0.9203, "step": 7692 }, { "epoch": 3.4857272315360217, "grad_norm": 0.12643127735996063, "learning_rate": 2.0348013981947137e-05, "loss": 0.9295, "step": 7693 }, { "epoch": 3.486180335296783, "grad_norm": 0.14903465932182605, "learning_rate": 2.0336993728017723e-05, "loss": 0.9366, "step": 7694 }, { "epoch": 3.4866334390575444, "grad_norm": 0.13219548008169246, "learning_rate": 2.0325975441810115e-05, "loss": 0.9366, "step": 7695 }, { "epoch": 3.4870865428183055, "grad_norm": 0.12001680872029087, "learning_rate": 2.031495912442691e-05, "loss": 0.925, "step": 7696 }, { "epoch": 3.4875396465790667, "grad_norm": 0.1361797796922816, "learning_rate": 2.0303944776970587e-05, "loss": 0.9674, "step": 7697 }, { "epoch": 3.487992750339828, "grad_norm": 0.13287501256991624, "learning_rate": 2.029293240054331e-05, "loss": 0.9208, "step": 7698 }, { "epoch": 3.488445854100589, "grad_norm": 0.12161504373312722, "learning_rate": 2.028192199624716e-05, "loss": 0.9302, "step": 7699 }, { "epoch": 3.48889895786135, "grad_norm": 0.18087847655337871, "learning_rate": 2.0270913565183942e-05, "loss": 0.9317, "step": 7700 }, { "epoch": 3.4893520616221116, "grad_norm": 0.11767752107182368, "learning_rate": 2.025990710845532e-05, "loss": 0.9224, "step": 7701 }, { "epoch": 3.4898051653828728, "grad_norm": 0.1800107436551617, "learning_rate": 2.024890262716273e-05, "loss": 0.9243, "step": 7702 }, { "epoch": 3.490258269143634, "grad_norm": 0.16507237728067503, "learning_rate": 2.0237900122407392e-05, "loss": 0.915, "step": 7703 }, { "epoch": 3.490711372904395, "grad_norm": 0.1263277661780718, "learning_rate": 2.0226899595290393e-05, "loss": 0.9367, "step": 7704 }, { "epoch": 3.491164476665156, "grad_norm": 0.17020125060395636, "learning_rate": 2.0215901046912537e-05, "loss": 0.9408, "step": 7705 }, { "epoch": 3.4916175804259177, "grad_norm": 0.14885585812176846, "learning_rate": 2.020490447837451e-05, "loss": 0.9385, "step": 7706 }, { "epoch": 3.492070684186679, "grad_norm": 0.1259471286910491, "learning_rate": 2.019390989077674e-05, "loss": 0.9512, "step": 7707 }, { "epoch": 3.49252378794744, "grad_norm": 0.1589148001147253, "learning_rate": 2.0182917285219493e-05, "loss": 0.9202, "step": 7708 }, { "epoch": 3.492976891708201, "grad_norm": 0.1364128807410265, "learning_rate": 2.0171926662802827e-05, "loss": 0.9385, "step": 7709 }, { "epoch": 3.4934299954689623, "grad_norm": 0.1399598192273102, "learning_rate": 2.0160938024626574e-05, "loss": 0.9114, "step": 7710 }, { "epoch": 3.4938830992297234, "grad_norm": 0.14268056052163242, "learning_rate": 2.014995137179043e-05, "loss": 0.923, "step": 7711 }, { "epoch": 3.494336202990485, "grad_norm": 0.14929639074445245, "learning_rate": 2.0138966705393814e-05, "loss": 0.9327, "step": 7712 }, { "epoch": 3.494789306751246, "grad_norm": 0.13920247204184777, "learning_rate": 2.012798402653602e-05, "loss": 0.9365, "step": 7713 }, { "epoch": 3.4952424105120072, "grad_norm": 0.13513553240078516, "learning_rate": 2.0117003336316084e-05, "loss": 0.9142, "step": 7714 }, { "epoch": 3.4956955142727684, "grad_norm": 0.14568682892809864, "learning_rate": 2.0106024635832898e-05, "loss": 0.9336, "step": 7715 }, { "epoch": 3.4961486180335295, "grad_norm": 0.12244607771053909, "learning_rate": 2.0095047926185095e-05, "loss": 0.9321, "step": 7716 }, { "epoch": 3.496601721794291, "grad_norm": 0.1384904856934077, "learning_rate": 2.0084073208471162e-05, "loss": 0.9348, "step": 7717 }, { "epoch": 3.497054825555052, "grad_norm": 0.12867964209900518, "learning_rate": 2.0073100483789362e-05, "loss": 0.9184, "step": 7718 }, { "epoch": 3.4975079293158133, "grad_norm": 0.11917819044125624, "learning_rate": 2.0062129753237728e-05, "loss": 0.9225, "step": 7719 }, { "epoch": 3.4979610330765745, "grad_norm": 0.12106494165289931, "learning_rate": 2.0051161017914173e-05, "loss": 0.9372, "step": 7720 }, { "epoch": 3.4984141368373356, "grad_norm": 0.11442257528284772, "learning_rate": 2.004019427891633e-05, "loss": 0.9256, "step": 7721 }, { "epoch": 3.4988672405980967, "grad_norm": 0.1203363628678169, "learning_rate": 2.002922953734168e-05, "loss": 0.9234, "step": 7722 }, { "epoch": 3.4993203443588583, "grad_norm": 0.12249983390879397, "learning_rate": 2.0018266794287485e-05, "loss": 0.9309, "step": 7723 }, { "epoch": 3.4997734481196194, "grad_norm": 0.12909758261287735, "learning_rate": 2.0007306050850824e-05, "loss": 0.9419, "step": 7724 }, { "epoch": 3.5002265518803806, "grad_norm": 0.11916951907622869, "learning_rate": 1.9996347308128552e-05, "loss": 0.9242, "step": 7725 }, { "epoch": 3.5006796556411417, "grad_norm": 0.12761118694368825, "learning_rate": 1.998539056721732e-05, "loss": 0.9197, "step": 7726 }, { "epoch": 3.5011327594019033, "grad_norm": 0.12250141147002071, "learning_rate": 1.9974435829213626e-05, "loss": 0.9236, "step": 7727 }, { "epoch": 3.5015858631626644, "grad_norm": 0.11994205251721127, "learning_rate": 1.9963483095213706e-05, "loss": 0.9401, "step": 7728 }, { "epoch": 3.5020389669234255, "grad_norm": 0.13619563869455092, "learning_rate": 1.995253236631366e-05, "loss": 0.9231, "step": 7729 }, { "epoch": 3.5024920706841867, "grad_norm": 0.13229539614345268, "learning_rate": 1.9941583643609307e-05, "loss": 0.9179, "step": 7730 }, { "epoch": 3.502945174444948, "grad_norm": 0.13920076320751062, "learning_rate": 1.993063692819635e-05, "loss": 0.9155, "step": 7731 }, { "epoch": 3.503398278205709, "grad_norm": 0.12812097174952847, "learning_rate": 1.9919692221170223e-05, "loss": 0.9257, "step": 7732 }, { "epoch": 3.50385138196647, "grad_norm": 0.13757276147081, "learning_rate": 1.9908749523626222e-05, "loss": 0.9296, "step": 7733 }, { "epoch": 3.5043044857272316, "grad_norm": 0.14111143358863718, "learning_rate": 1.9897808836659382e-05, "loss": 0.9252, "step": 7734 }, { "epoch": 3.5047575894879928, "grad_norm": 0.13515757383810348, "learning_rate": 1.9886870161364556e-05, "loss": 0.9468, "step": 7735 }, { "epoch": 3.505210693248754, "grad_norm": 0.22472787248737566, "learning_rate": 1.9875933498836423e-05, "loss": 0.9513, "step": 7736 }, { "epoch": 3.505663797009515, "grad_norm": 0.13320453512533817, "learning_rate": 1.986499885016942e-05, "loss": 0.9373, "step": 7737 }, { "epoch": 3.5061169007702766, "grad_norm": 0.1260212666623141, "learning_rate": 1.9854066216457823e-05, "loss": 0.9328, "step": 7738 }, { "epoch": 3.5065700045310377, "grad_norm": 0.1389209154782274, "learning_rate": 1.9843135598795667e-05, "loss": 0.9377, "step": 7739 }, { "epoch": 3.507023108291799, "grad_norm": 0.15062030615393668, "learning_rate": 1.983220699827682e-05, "loss": 0.9376, "step": 7740 }, { "epoch": 3.50747621205256, "grad_norm": 0.12863670680442668, "learning_rate": 1.9821280415994914e-05, "loss": 0.9401, "step": 7741 }, { "epoch": 3.507929315813321, "grad_norm": 0.11412047438976894, "learning_rate": 1.9810355853043418e-05, "loss": 0.9284, "step": 7742 }, { "epoch": 3.5083824195740823, "grad_norm": 0.12539020752811778, "learning_rate": 1.979943331051557e-05, "loss": 0.9502, "step": 7743 }, { "epoch": 3.508835523334844, "grad_norm": 0.11329417276711062, "learning_rate": 1.9788512789504392e-05, "loss": 0.9287, "step": 7744 }, { "epoch": 3.509288627095605, "grad_norm": 0.10857176737694545, "learning_rate": 1.9777594291102756e-05, "loss": 0.8856, "step": 7745 }, { "epoch": 3.509741730856366, "grad_norm": 0.11455773738860041, "learning_rate": 1.9766677816403283e-05, "loss": 0.92, "step": 7746 }, { "epoch": 3.5101948346171272, "grad_norm": 0.1082889155335887, "learning_rate": 1.975576336649842e-05, "loss": 0.9232, "step": 7747 }, { "epoch": 3.510647938377889, "grad_norm": 0.11827277615351751, "learning_rate": 1.974485094248039e-05, "loss": 0.9369, "step": 7748 }, { "epoch": 3.51110104213865, "grad_norm": 0.14388378569972013, "learning_rate": 1.973394054544124e-05, "loss": 0.9322, "step": 7749 }, { "epoch": 3.511554145899411, "grad_norm": 0.12585784742714898, "learning_rate": 1.9723032176472787e-05, "loss": 0.9194, "step": 7750 }, { "epoch": 3.512007249660172, "grad_norm": 0.11985271246436303, "learning_rate": 1.9712125836666645e-05, "loss": 0.9361, "step": 7751 }, { "epoch": 3.5124603534209333, "grad_norm": 0.1778088861412649, "learning_rate": 1.9701221527114262e-05, "loss": 0.9177, "step": 7752 }, { "epoch": 3.5129134571816945, "grad_norm": 0.14373457048585112, "learning_rate": 1.9690319248906834e-05, "loss": 0.9513, "step": 7753 }, { "epoch": 3.5133665609424556, "grad_norm": 0.14380214106753852, "learning_rate": 1.9679419003135413e-05, "loss": 0.9396, "step": 7754 }, { "epoch": 3.513819664703217, "grad_norm": 0.17287469033607764, "learning_rate": 1.9668520790890758e-05, "loss": 0.9184, "step": 7755 }, { "epoch": 3.5142727684639783, "grad_norm": 0.13156562997166135, "learning_rate": 1.9657624613263518e-05, "loss": 0.9321, "step": 7756 }, { "epoch": 3.5147258722247394, "grad_norm": 0.15039421880175574, "learning_rate": 1.9646730471344065e-05, "loss": 0.9091, "step": 7757 }, { "epoch": 3.5151789759855006, "grad_norm": 0.1219417846870854, "learning_rate": 1.9635838366222643e-05, "loss": 0.9262, "step": 7758 }, { "epoch": 3.515632079746262, "grad_norm": 0.13984620379808196, "learning_rate": 1.9624948298989214e-05, "loss": 0.9413, "step": 7759 }, { "epoch": 3.5160851835070233, "grad_norm": 0.11217788091752, "learning_rate": 1.9614060270733593e-05, "loss": 0.9235, "step": 7760 }, { "epoch": 3.5165382872677844, "grad_norm": 0.13850517910268514, "learning_rate": 1.960317428254537e-05, "loss": 0.9166, "step": 7761 }, { "epoch": 3.5169913910285455, "grad_norm": 0.13657230332180745, "learning_rate": 1.9592290335513902e-05, "loss": 0.9406, "step": 7762 }, { "epoch": 3.5174444947893067, "grad_norm": 0.10694616553019211, "learning_rate": 1.958140843072841e-05, "loss": 0.9016, "step": 7763 }, { "epoch": 3.517897598550068, "grad_norm": 0.13741109736492932, "learning_rate": 1.957052856927784e-05, "loss": 0.955, "step": 7764 }, { "epoch": 3.518350702310829, "grad_norm": 0.10911016143309739, "learning_rate": 1.955965075225099e-05, "loss": 0.9272, "step": 7765 }, { "epoch": 3.5188038060715905, "grad_norm": 0.14692716562796762, "learning_rate": 1.9548774980736397e-05, "loss": 0.9356, "step": 7766 }, { "epoch": 3.5192569098323516, "grad_norm": 0.14213189072013302, "learning_rate": 1.9537901255822453e-05, "loss": 0.9168, "step": 7767 }, { "epoch": 3.5197100135931128, "grad_norm": 0.130364760556125, "learning_rate": 1.9527029578597297e-05, "loss": 0.932, "step": 7768 }, { "epoch": 3.520163117353874, "grad_norm": 0.13857635573010005, "learning_rate": 1.9516159950148904e-05, "loss": 0.9284, "step": 7769 }, { "epoch": 3.5206162211146355, "grad_norm": 0.13176185940775567, "learning_rate": 1.9505292371565008e-05, "loss": 0.9124, "step": 7770 }, { "epoch": 3.5210693248753966, "grad_norm": 0.13753006475431312, "learning_rate": 1.949442684393314e-05, "loss": 0.9349, "step": 7771 }, { "epoch": 3.5215224286361577, "grad_norm": 0.13028107271846395, "learning_rate": 1.9483563368340666e-05, "loss": 0.9349, "step": 7772 }, { "epoch": 3.521975532396919, "grad_norm": 0.1394666281025313, "learning_rate": 1.947270194587469e-05, "loss": 0.9444, "step": 7773 }, { "epoch": 3.52242863615768, "grad_norm": 0.11215223748718632, "learning_rate": 1.9461842577622167e-05, "loss": 0.9251, "step": 7774 }, { "epoch": 3.522881739918441, "grad_norm": 0.12423661842944861, "learning_rate": 1.9450985264669793e-05, "loss": 0.939, "step": 7775 }, { "epoch": 3.5233348436792027, "grad_norm": 0.12725091649778442, "learning_rate": 1.944013000810411e-05, "loss": 0.9219, "step": 7776 }, { "epoch": 3.523787947439964, "grad_norm": 0.11634687527381438, "learning_rate": 1.94292768090114e-05, "loss": 0.9462, "step": 7777 }, { "epoch": 3.524241051200725, "grad_norm": 0.1488342624388886, "learning_rate": 1.9418425668477798e-05, "loss": 0.9295, "step": 7778 }, { "epoch": 3.524694154961486, "grad_norm": 0.1252436595869242, "learning_rate": 1.9407576587589183e-05, "loss": 0.9449, "step": 7779 }, { "epoch": 3.5251472587222477, "grad_norm": 0.1399213152246318, "learning_rate": 1.9396729567431237e-05, "loss": 0.9318, "step": 7780 }, { "epoch": 3.525600362483009, "grad_norm": 0.13496870177466047, "learning_rate": 1.9385884609089473e-05, "loss": 0.9184, "step": 7781 }, { "epoch": 3.52605346624377, "grad_norm": 0.15163304709819406, "learning_rate": 1.9375041713649147e-05, "loss": 0.9468, "step": 7782 }, { "epoch": 3.526506570004531, "grad_norm": 0.12438099305358488, "learning_rate": 1.9364200882195355e-05, "loss": 0.931, "step": 7783 }, { "epoch": 3.526959673765292, "grad_norm": 0.1425722144485242, "learning_rate": 1.9353362115812934e-05, "loss": 0.9333, "step": 7784 }, { "epoch": 3.5274127775260533, "grad_norm": 0.13010813023164455, "learning_rate": 1.934252541558658e-05, "loss": 0.9518, "step": 7785 }, { "epoch": 3.5278658812868144, "grad_norm": 0.14719439936637324, "learning_rate": 1.9331690782600722e-05, "loss": 0.944, "step": 7786 }, { "epoch": 3.528318985047576, "grad_norm": 0.173289239234924, "learning_rate": 1.93208582179396e-05, "loss": 0.9085, "step": 7787 }, { "epoch": 3.528772088808337, "grad_norm": 0.11833083985206887, "learning_rate": 1.931002772268728e-05, "loss": 0.9288, "step": 7788 }, { "epoch": 3.5292251925690983, "grad_norm": 0.15429377970527847, "learning_rate": 1.9299199297927563e-05, "loss": 0.925, "step": 7789 }, { "epoch": 3.5296782963298594, "grad_norm": 0.186052230705399, "learning_rate": 1.928837294474411e-05, "loss": 0.9301, "step": 7790 }, { "epoch": 3.530131400090621, "grad_norm": 0.12273265706304368, "learning_rate": 1.9277548664220297e-05, "loss": 0.9287, "step": 7791 }, { "epoch": 3.530584503851382, "grad_norm": 0.1817876379521493, "learning_rate": 1.9266726457439375e-05, "loss": 0.9502, "step": 7792 }, { "epoch": 3.5310376076121432, "grad_norm": 0.19142020142951632, "learning_rate": 1.9255906325484315e-05, "loss": 0.9219, "step": 7793 }, { "epoch": 3.5314907113729044, "grad_norm": 0.12101402690630925, "learning_rate": 1.924508826943794e-05, "loss": 0.9268, "step": 7794 }, { "epoch": 3.5319438151336655, "grad_norm": 0.16603083733975477, "learning_rate": 1.9234272290382827e-05, "loss": 0.9359, "step": 7795 }, { "epoch": 3.5323969188944266, "grad_norm": 0.17471840873989153, "learning_rate": 1.922345838940133e-05, "loss": 0.9383, "step": 7796 }, { "epoch": 3.5328500226551878, "grad_norm": 0.13060205363047125, "learning_rate": 1.9212646567575662e-05, "loss": 0.9033, "step": 7797 }, { "epoch": 3.5333031264159493, "grad_norm": 0.14782254387137886, "learning_rate": 1.920183682598775e-05, "loss": 0.9293, "step": 7798 }, { "epoch": 3.5337562301767105, "grad_norm": 0.1573849769985547, "learning_rate": 1.9191029165719386e-05, "loss": 0.9321, "step": 7799 }, { "epoch": 3.5342093339374716, "grad_norm": 0.12307504297026878, "learning_rate": 1.9180223587852076e-05, "loss": 0.941, "step": 7800 }, { "epoch": 3.5346624376982327, "grad_norm": 0.16101870976271046, "learning_rate": 1.91694200934672e-05, "loss": 0.9234, "step": 7801 }, { "epoch": 3.5351155414589943, "grad_norm": 0.15313621618169865, "learning_rate": 1.915861868364587e-05, "loss": 0.9183, "step": 7802 }, { "epoch": 3.5355686452197554, "grad_norm": 0.12141459197493949, "learning_rate": 1.9147819359468987e-05, "loss": 0.904, "step": 7803 }, { "epoch": 3.5360217489805166, "grad_norm": 0.12726047622249687, "learning_rate": 1.9137022122017305e-05, "loss": 0.9166, "step": 7804 }, { "epoch": 3.5364748527412777, "grad_norm": 0.1287393215047491, "learning_rate": 1.912622697237129e-05, "loss": 0.9394, "step": 7805 }, { "epoch": 3.536927956502039, "grad_norm": 0.12097805840867781, "learning_rate": 1.9115433911611268e-05, "loss": 0.9021, "step": 7806 }, { "epoch": 3.5373810602628, "grad_norm": 0.1392154962015456, "learning_rate": 1.9104642940817294e-05, "loss": 0.9375, "step": 7807 }, { "epoch": 3.5378341640235615, "grad_norm": 0.11062617197330205, "learning_rate": 1.9093854061069278e-05, "loss": 0.9351, "step": 7808 }, { "epoch": 3.5382872677843227, "grad_norm": 0.12517419702852098, "learning_rate": 1.9083067273446857e-05, "loss": 0.9291, "step": 7809 }, { "epoch": 3.538740371545084, "grad_norm": 0.12963006500410848, "learning_rate": 1.9072282579029518e-05, "loss": 0.9351, "step": 7810 }, { "epoch": 3.539193475305845, "grad_norm": 0.12485993079038003, "learning_rate": 1.90614999788965e-05, "loss": 0.9462, "step": 7811 }, { "epoch": 3.5396465790666065, "grad_norm": 0.13805842865197387, "learning_rate": 1.9050719474126837e-05, "loss": 0.9283, "step": 7812 }, { "epoch": 3.5400996828273676, "grad_norm": 0.14159155185133798, "learning_rate": 1.9039941065799338e-05, "loss": 0.9451, "step": 7813 }, { "epoch": 3.5405527865881288, "grad_norm": 0.10603245738044441, "learning_rate": 1.902916475499267e-05, "loss": 0.9203, "step": 7814 }, { "epoch": 3.54100589034889, "grad_norm": 0.14201685693657712, "learning_rate": 1.901839054278521e-05, "loss": 0.9266, "step": 7815 }, { "epoch": 3.541458994109651, "grad_norm": 0.1172024626427806, "learning_rate": 1.900761843025516e-05, "loss": 0.9264, "step": 7816 }, { "epoch": 3.541912097870412, "grad_norm": 0.128791056756598, "learning_rate": 1.8996848418480526e-05, "loss": 0.933, "step": 7817 }, { "epoch": 3.5423652016311733, "grad_norm": 0.14351035804346754, "learning_rate": 1.898608050853906e-05, "loss": 0.9157, "step": 7818 }, { "epoch": 3.542818305391935, "grad_norm": 0.12910081116139013, "learning_rate": 1.8975314701508366e-05, "loss": 0.9318, "step": 7819 }, { "epoch": 3.543271409152696, "grad_norm": 0.12921330111546586, "learning_rate": 1.8964550998465766e-05, "loss": 0.9264, "step": 7820 }, { "epoch": 3.543724512913457, "grad_norm": 0.17405486634796632, "learning_rate": 1.8953789400488448e-05, "loss": 0.9466, "step": 7821 }, { "epoch": 3.5441776166742183, "grad_norm": 0.12422667213267742, "learning_rate": 1.8943029908653328e-05, "loss": 0.934, "step": 7822 }, { "epoch": 3.54463072043498, "grad_norm": 0.13243305849740195, "learning_rate": 1.8932272524037114e-05, "loss": 0.9327, "step": 7823 }, { "epoch": 3.545083824195741, "grad_norm": 0.1287509255361845, "learning_rate": 1.8921517247716362e-05, "loss": 0.9287, "step": 7824 }, { "epoch": 3.545536927956502, "grad_norm": 0.10823035618092784, "learning_rate": 1.8910764080767343e-05, "loss": 0.9273, "step": 7825 }, { "epoch": 3.5459900317172632, "grad_norm": 0.12659003415576747, "learning_rate": 1.8900013024266173e-05, "loss": 0.9194, "step": 7826 }, { "epoch": 3.5464431354780244, "grad_norm": 0.11349869801572554, "learning_rate": 1.888926407928872e-05, "loss": 0.9277, "step": 7827 }, { "epoch": 3.5468962392387855, "grad_norm": 0.11524355276044, "learning_rate": 1.8878517246910672e-05, "loss": 0.9427, "step": 7828 }, { "epoch": 3.5473493429995466, "grad_norm": 0.10862868144520336, "learning_rate": 1.8867772528207465e-05, "loss": 0.929, "step": 7829 }, { "epoch": 3.547802446760308, "grad_norm": 0.11750411211290107, "learning_rate": 1.8857029924254378e-05, "loss": 0.9134, "step": 7830 }, { "epoch": 3.5482555505210693, "grad_norm": 0.10780980576749088, "learning_rate": 1.8846289436126432e-05, "loss": 0.9338, "step": 7831 }, { "epoch": 3.5487086542818305, "grad_norm": 0.1179041841915707, "learning_rate": 1.8835551064898434e-05, "loss": 0.9305, "step": 7832 }, { "epoch": 3.5491617580425916, "grad_norm": 0.10916690767849391, "learning_rate": 1.882481481164503e-05, "loss": 0.9402, "step": 7833 }, { "epoch": 3.549614861803353, "grad_norm": 0.10757170643966915, "learning_rate": 1.881408067744059e-05, "loss": 0.9208, "step": 7834 }, { "epoch": 3.5500679655641143, "grad_norm": 0.11351336142204553, "learning_rate": 1.8803348663359333e-05, "loss": 0.915, "step": 7835 }, { "epoch": 3.5505210693248754, "grad_norm": 0.11382276161769989, "learning_rate": 1.8792618770475212e-05, "loss": 0.9211, "step": 7836 }, { "epoch": 3.5509741730856366, "grad_norm": 0.1078183250777039, "learning_rate": 1.878189099986201e-05, "loss": 0.9281, "step": 7837 }, { "epoch": 3.5514272768463977, "grad_norm": 0.11929405259694498, "learning_rate": 1.8771165352593265e-05, "loss": 0.9089, "step": 7838 }, { "epoch": 3.551880380607159, "grad_norm": 0.11590291737344578, "learning_rate": 1.876044182974233e-05, "loss": 0.9557, "step": 7839 }, { "epoch": 3.5523334843679204, "grad_norm": 0.10424204021375963, "learning_rate": 1.8749720432382333e-05, "loss": 0.9108, "step": 7840 }, { "epoch": 3.5527865881286815, "grad_norm": 0.13105007397408594, "learning_rate": 1.8739001161586162e-05, "loss": 0.9183, "step": 7841 }, { "epoch": 3.5532396918894427, "grad_norm": 0.11926340298716574, "learning_rate": 1.8728284018426556e-05, "loss": 0.9193, "step": 7842 }, { "epoch": 3.553692795650204, "grad_norm": 0.11327131701314565, "learning_rate": 1.8717569003975966e-05, "loss": 0.9382, "step": 7843 }, { "epoch": 3.5541458994109654, "grad_norm": 0.14312188687372915, "learning_rate": 1.8706856119306707e-05, "loss": 0.9193, "step": 7844 }, { "epoch": 3.5545990031717265, "grad_norm": 0.1129223835150224, "learning_rate": 1.8696145365490802e-05, "loss": 0.906, "step": 7845 }, { "epoch": 3.5550521069324876, "grad_norm": 0.1405287867185221, "learning_rate": 1.8685436743600146e-05, "loss": 0.9333, "step": 7846 }, { "epoch": 3.5555052106932488, "grad_norm": 0.15323211081190663, "learning_rate": 1.8674730254706342e-05, "loss": 0.9326, "step": 7847 }, { "epoch": 3.55595831445401, "grad_norm": 0.10848217827569921, "learning_rate": 1.866402589988081e-05, "loss": 0.9237, "step": 7848 }, { "epoch": 3.556411418214771, "grad_norm": 0.1551888644571323, "learning_rate": 1.8653323680194785e-05, "loss": 0.9403, "step": 7849 }, { "epoch": 3.556864521975532, "grad_norm": 0.12209956140338817, "learning_rate": 1.864262359671923e-05, "loss": 0.9351, "step": 7850 }, { "epoch": 3.5573176257362937, "grad_norm": 0.11510143155603257, "learning_rate": 1.863192565052496e-05, "loss": 0.9316, "step": 7851 }, { "epoch": 3.557770729497055, "grad_norm": 0.11791450866511646, "learning_rate": 1.8621229842682518e-05, "loss": 0.9154, "step": 7852 }, { "epoch": 3.558223833257816, "grad_norm": 0.12359268988684358, "learning_rate": 1.8610536174262276e-05, "loss": 0.9322, "step": 7853 }, { "epoch": 3.558676937018577, "grad_norm": 0.10523706268400707, "learning_rate": 1.8599844646334353e-05, "loss": 0.9389, "step": 7854 }, { "epoch": 3.5591300407793387, "grad_norm": 0.11917033834931014, "learning_rate": 1.8589155259968704e-05, "loss": 0.9248, "step": 7855 }, { "epoch": 3.5595831445401, "grad_norm": 0.1245993777530304, "learning_rate": 1.8578468016235014e-05, "loss": 0.9218, "step": 7856 }, { "epoch": 3.560036248300861, "grad_norm": 0.13144538653736063, "learning_rate": 1.8567782916202782e-05, "loss": 0.9493, "step": 7857 }, { "epoch": 3.560489352061622, "grad_norm": 0.11897743839665378, "learning_rate": 1.8557099960941304e-05, "loss": 0.9287, "step": 7858 }, { "epoch": 3.5609424558223832, "grad_norm": 0.12310635808341384, "learning_rate": 1.8546419151519633e-05, "loss": 0.9475, "step": 7859 }, { "epoch": 3.5613955595831444, "grad_norm": 0.12773215877145663, "learning_rate": 1.8535740489006637e-05, "loss": 0.9624, "step": 7860 }, { "epoch": 3.5618486633439055, "grad_norm": 0.1114312185763173, "learning_rate": 1.8525063974470937e-05, "loss": 0.9134, "step": 7861 }, { "epoch": 3.562301767104667, "grad_norm": 0.11624622450693138, "learning_rate": 1.8514389608980975e-05, "loss": 0.9322, "step": 7862 }, { "epoch": 3.562754870865428, "grad_norm": 0.11254539151593179, "learning_rate": 1.8503717393604955e-05, "loss": 0.9227, "step": 7863 }, { "epoch": 3.5632079746261893, "grad_norm": 0.12217409378952826, "learning_rate": 1.8493047329410848e-05, "loss": 0.9349, "step": 7864 }, { "epoch": 3.5636610783869505, "grad_norm": 0.11961282682146679, "learning_rate": 1.8482379417466463e-05, "loss": 0.9393, "step": 7865 }, { "epoch": 3.564114182147712, "grad_norm": 0.10946789045372222, "learning_rate": 1.847171365883933e-05, "loss": 0.9338, "step": 7866 }, { "epoch": 3.564567285908473, "grad_norm": 0.10727595909954471, "learning_rate": 1.846105005459683e-05, "loss": 0.9192, "step": 7867 }, { "epoch": 3.5650203896692343, "grad_norm": 0.11696671088384386, "learning_rate": 1.8450388605806074e-05, "loss": 0.9214, "step": 7868 }, { "epoch": 3.5654734934299954, "grad_norm": 0.12095560909635808, "learning_rate": 1.8439729313533984e-05, "loss": 0.934, "step": 7869 }, { "epoch": 3.5659265971907566, "grad_norm": 0.1304504740094602, "learning_rate": 1.842907217884724e-05, "loss": 0.9216, "step": 7870 }, { "epoch": 3.5663797009515177, "grad_norm": 0.13393862602704407, "learning_rate": 1.841841720281236e-05, "loss": 0.9394, "step": 7871 }, { "epoch": 3.5668328047122793, "grad_norm": 0.13221474704260752, "learning_rate": 1.8407764386495572e-05, "loss": 0.9439, "step": 7872 }, { "epoch": 3.5672859084730404, "grad_norm": 0.11711875490935302, "learning_rate": 1.8397113730962965e-05, "loss": 0.9417, "step": 7873 }, { "epoch": 3.5677390122338015, "grad_norm": 0.13727055696156326, "learning_rate": 1.8386465237280348e-05, "loss": 0.9405, "step": 7874 }, { "epoch": 3.5681921159945627, "grad_norm": 0.11237301259597184, "learning_rate": 1.8375818906513366e-05, "loss": 0.917, "step": 7875 }, { "epoch": 3.5686452197553242, "grad_norm": 0.1269784014834131, "learning_rate": 1.8365174739727402e-05, "loss": 0.9417, "step": 7876 }, { "epoch": 3.5690983235160854, "grad_norm": 0.10918605619233684, "learning_rate": 1.8354532737987633e-05, "loss": 0.9224, "step": 7877 }, { "epoch": 3.5695514272768465, "grad_norm": 0.1390080581992018, "learning_rate": 1.8343892902359056e-05, "loss": 0.9418, "step": 7878 }, { "epoch": 3.5700045310376076, "grad_norm": 0.1263205239456561, "learning_rate": 1.8333255233906393e-05, "loss": 0.9175, "step": 7879 }, { "epoch": 3.5704576347983688, "grad_norm": 0.13874404192870263, "learning_rate": 1.832261973369421e-05, "loss": 0.9228, "step": 7880 }, { "epoch": 3.57091073855913, "grad_norm": 0.13342499386841122, "learning_rate": 1.83119864027868e-05, "loss": 0.9056, "step": 7881 }, { "epoch": 3.571363842319891, "grad_norm": 0.1105485981926602, "learning_rate": 1.830135524224829e-05, "loss": 0.9335, "step": 7882 }, { "epoch": 3.5718169460806526, "grad_norm": 0.1408083246687387, "learning_rate": 1.8290726253142547e-05, "loss": 0.9172, "step": 7883 }, { "epoch": 3.5722700498414137, "grad_norm": 0.12201156427596764, "learning_rate": 1.828009943653323e-05, "loss": 0.9179, "step": 7884 }, { "epoch": 3.572723153602175, "grad_norm": 0.13158851407797975, "learning_rate": 1.8269474793483816e-05, "loss": 0.9333, "step": 7885 }, { "epoch": 3.573176257362936, "grad_norm": 0.15572267495621842, "learning_rate": 1.8258852325057504e-05, "loss": 0.9326, "step": 7886 }, { "epoch": 3.5736293611236976, "grad_norm": 0.12445767660222132, "learning_rate": 1.8248232032317346e-05, "loss": 0.9383, "step": 7887 }, { "epoch": 3.5740824648844587, "grad_norm": 0.144740128233592, "learning_rate": 1.82376139163261e-05, "loss": 0.9312, "step": 7888 }, { "epoch": 3.57453556864522, "grad_norm": 0.11383422449464216, "learning_rate": 1.822699797814638e-05, "loss": 0.9369, "step": 7889 }, { "epoch": 3.574988672405981, "grad_norm": 0.12713881536182473, "learning_rate": 1.821638421884051e-05, "loss": 0.9408, "step": 7890 }, { "epoch": 3.575441776166742, "grad_norm": 0.11247521198364176, "learning_rate": 1.8205772639470676e-05, "loss": 0.9346, "step": 7891 }, { "epoch": 3.575894879927503, "grad_norm": 0.1404170852499858, "learning_rate": 1.8195163241098777e-05, "loss": 0.9698, "step": 7892 }, { "epoch": 3.5763479836882643, "grad_norm": 0.10685076660690503, "learning_rate": 1.8184556024786513e-05, "loss": 0.9121, "step": 7893 }, { "epoch": 3.576801087449026, "grad_norm": 0.12110183596140739, "learning_rate": 1.81739509915954e-05, "loss": 0.9398, "step": 7894 }, { "epoch": 3.577254191209787, "grad_norm": 0.1057411135349822, "learning_rate": 1.8163348142586667e-05, "loss": 0.9211, "step": 7895 }, { "epoch": 3.577707294970548, "grad_norm": 0.11783740619846987, "learning_rate": 1.815274747882141e-05, "loss": 0.9261, "step": 7896 }, { "epoch": 3.5781603987313093, "grad_norm": 0.11153485268103654, "learning_rate": 1.8142149001360424e-05, "loss": 0.9388, "step": 7897 }, { "epoch": 3.578613502492071, "grad_norm": 0.10293128322500356, "learning_rate": 1.8131552711264357e-05, "loss": 0.931, "step": 7898 }, { "epoch": 3.579066606252832, "grad_norm": 0.10751065355302739, "learning_rate": 1.812095860959357e-05, "loss": 0.9318, "step": 7899 }, { "epoch": 3.579519710013593, "grad_norm": 0.10800651331695717, "learning_rate": 1.811036669740827e-05, "loss": 0.926, "step": 7900 }, { "epoch": 3.5799728137743543, "grad_norm": 0.10798642228479849, "learning_rate": 1.8099776975768396e-05, "loss": 0.9171, "step": 7901 }, { "epoch": 3.5804259175351154, "grad_norm": 0.10873669162251408, "learning_rate": 1.8089189445733676e-05, "loss": 0.9282, "step": 7902 }, { "epoch": 3.5808790212958765, "grad_norm": 0.1075327239613845, "learning_rate": 1.8078604108363657e-05, "loss": 0.9307, "step": 7903 }, { "epoch": 3.581332125056638, "grad_norm": 0.1094655393018127, "learning_rate": 1.8068020964717603e-05, "loss": 0.9231, "step": 7904 }, { "epoch": 3.5817852288173992, "grad_norm": 0.11309796916784909, "learning_rate": 1.8057440015854628e-05, "loss": 0.9187, "step": 7905 }, { "epoch": 3.5822383325781604, "grad_norm": 0.10027909303547017, "learning_rate": 1.8046861262833563e-05, "loss": 0.9512, "step": 7906 }, { "epoch": 3.5826914363389215, "grad_norm": 0.12331416986020241, "learning_rate": 1.803628470671307e-05, "loss": 0.925, "step": 7907 }, { "epoch": 3.583144540099683, "grad_norm": 0.12025447902260296, "learning_rate": 1.802571034855156e-05, "loss": 0.9277, "step": 7908 }, { "epoch": 3.583597643860444, "grad_norm": 0.125962906580982, "learning_rate": 1.8015138189407223e-05, "loss": 0.9388, "step": 7909 }, { "epoch": 3.5840507476212053, "grad_norm": 0.1194536130045122, "learning_rate": 1.8004568230338063e-05, "loss": 0.9386, "step": 7910 }, { "epoch": 3.5845038513819665, "grad_norm": 0.11184215156000057, "learning_rate": 1.7994000472401802e-05, "loss": 0.9131, "step": 7911 }, { "epoch": 3.5849569551427276, "grad_norm": 0.11400531573061623, "learning_rate": 1.7983434916656028e-05, "loss": 0.9392, "step": 7912 }, { "epoch": 3.5854100589034887, "grad_norm": 0.10792803793579119, "learning_rate": 1.7972871564158013e-05, "loss": 0.9249, "step": 7913 }, { "epoch": 3.58586316266425, "grad_norm": 0.1132884404923505, "learning_rate": 1.7962310415964893e-05, "loss": 0.9568, "step": 7914 }, { "epoch": 3.5863162664250114, "grad_norm": 0.10936190153975044, "learning_rate": 1.7951751473133518e-05, "loss": 0.9526, "step": 7915 }, { "epoch": 3.5867693701857726, "grad_norm": 0.11879289528998592, "learning_rate": 1.7941194736720574e-05, "loss": 0.9679, "step": 7916 }, { "epoch": 3.5872224739465337, "grad_norm": 0.130032746788634, "learning_rate": 1.7930640207782478e-05, "loss": 0.9349, "step": 7917 }, { "epoch": 3.587675577707295, "grad_norm": 0.11153943904259131, "learning_rate": 1.7920087887375445e-05, "loss": 0.9039, "step": 7918 }, { "epoch": 3.5881286814680564, "grad_norm": 0.11366700994562005, "learning_rate": 1.7909537776555476e-05, "loss": 0.9368, "step": 7919 }, { "epoch": 3.5885817852288175, "grad_norm": 0.11545513860431605, "learning_rate": 1.789898987637834e-05, "loss": 0.9212, "step": 7920 }, { "epoch": 3.5890348889895787, "grad_norm": 0.11526624718740987, "learning_rate": 1.78884441878996e-05, "loss": 0.9372, "step": 7921 }, { "epoch": 3.58948799275034, "grad_norm": 0.10229848016931907, "learning_rate": 1.787790071217457e-05, "loss": 0.9343, "step": 7922 }, { "epoch": 3.589941096511101, "grad_norm": 0.10464557656102647, "learning_rate": 1.7867359450258373e-05, "loss": 0.9303, "step": 7923 }, { "epoch": 3.590394200271862, "grad_norm": 0.10957962234686976, "learning_rate": 1.7856820403205907e-05, "loss": 0.9133, "step": 7924 }, { "epoch": 3.590847304032623, "grad_norm": 0.1121940892933129, "learning_rate": 1.7846283572071815e-05, "loss": 0.9202, "step": 7925 }, { "epoch": 3.5913004077933848, "grad_norm": 0.12053731333236861, "learning_rate": 1.783574895791054e-05, "loss": 0.9445, "step": 7926 }, { "epoch": 3.591753511554146, "grad_norm": 0.10426809972786535, "learning_rate": 1.7825216561776334e-05, "loss": 0.9404, "step": 7927 }, { "epoch": 3.592206615314907, "grad_norm": 0.12148609736678184, "learning_rate": 1.7814686384723173e-05, "loss": 0.9355, "step": 7928 }, { "epoch": 3.592659719075668, "grad_norm": 0.10690258826074182, "learning_rate": 1.7804158427804835e-05, "loss": 0.939, "step": 7929 }, { "epoch": 3.5931128228364297, "grad_norm": 0.11055650722227386, "learning_rate": 1.7793632692074893e-05, "loss": 0.9024, "step": 7930 }, { "epoch": 3.593565926597191, "grad_norm": 0.11948684911998184, "learning_rate": 1.778310917858666e-05, "loss": 0.9289, "step": 7931 }, { "epoch": 3.594019030357952, "grad_norm": 0.11293882368556192, "learning_rate": 1.7772587888393276e-05, "loss": 0.9128, "step": 7932 }, { "epoch": 3.594472134118713, "grad_norm": 0.10915228546911762, "learning_rate": 1.7762068822547603e-05, "loss": 0.9279, "step": 7933 }, { "epoch": 3.5949252378794743, "grad_norm": 0.11781497026167194, "learning_rate": 1.775155198210233e-05, "loss": 0.9326, "step": 7934 }, { "epoch": 3.5953783416402354, "grad_norm": 0.10183213057426249, "learning_rate": 1.774103736810988e-05, "loss": 0.9309, "step": 7935 }, { "epoch": 3.595831445400997, "grad_norm": 0.11974201788630243, "learning_rate": 1.7730524981622496e-05, "loss": 0.9274, "step": 7936 }, { "epoch": 3.596284549161758, "grad_norm": 0.11228891199113794, "learning_rate": 1.772001482369217e-05, "loss": 0.9315, "step": 7937 }, { "epoch": 3.5967376529225192, "grad_norm": 0.10214861468150035, "learning_rate": 1.7709506895370658e-05, "loss": 0.9276, "step": 7938 }, { "epoch": 3.5971907566832804, "grad_norm": 0.10442173325954453, "learning_rate": 1.7699001197709545e-05, "loss": 0.9307, "step": 7939 }, { "epoch": 3.597643860444042, "grad_norm": 0.10655926899194446, "learning_rate": 1.7688497731760122e-05, "loss": 0.9322, "step": 7940 }, { "epoch": 3.598096964204803, "grad_norm": 0.11570606086429193, "learning_rate": 1.7677996498573535e-05, "loss": 0.9129, "step": 7941 }, { "epoch": 3.598550067965564, "grad_norm": 0.12134690872367447, "learning_rate": 1.766749749920063e-05, "loss": 0.9242, "step": 7942 }, { "epoch": 3.5990031717263253, "grad_norm": 0.12024184304188663, "learning_rate": 1.76570007346921e-05, "loss": 0.9242, "step": 7943 }, { "epoch": 3.5994562754870865, "grad_norm": 0.1304624273987997, "learning_rate": 1.7646506206098357e-05, "loss": 0.9226, "step": 7944 }, { "epoch": 3.5999093792478476, "grad_norm": 0.10534523891528952, "learning_rate": 1.7636013914469608e-05, "loss": 0.9295, "step": 7945 }, { "epoch": 3.6003624830086087, "grad_norm": 0.15824441609209816, "learning_rate": 1.7625523860855863e-05, "loss": 0.9338, "step": 7946 }, { "epoch": 3.6008155867693703, "grad_norm": 0.132611508597829, "learning_rate": 1.7615036046306856e-05, "loss": 0.9227, "step": 7947 }, { "epoch": 3.6012686905301314, "grad_norm": 0.11170044250782946, "learning_rate": 1.7604550471872162e-05, "loss": 0.93, "step": 7948 }, { "epoch": 3.6017217942908926, "grad_norm": 0.12082554873868448, "learning_rate": 1.759406713860106e-05, "loss": 0.939, "step": 7949 }, { "epoch": 3.6021748980516537, "grad_norm": 0.1174088360471495, "learning_rate": 1.758358604754268e-05, "loss": 0.9469, "step": 7950 }, { "epoch": 3.6026280018124153, "grad_norm": 0.12531256880758682, "learning_rate": 1.7573107199745848e-05, "loss": 0.9253, "step": 7951 }, { "epoch": 3.6030811055731764, "grad_norm": 0.11364622562919975, "learning_rate": 1.756263059625924e-05, "loss": 0.8969, "step": 7952 }, { "epoch": 3.6035342093339375, "grad_norm": 0.12302003652240963, "learning_rate": 1.755215623813126e-05, "loss": 0.9409, "step": 7953 }, { "epoch": 3.6039873130946987, "grad_norm": 0.13328285141936688, "learning_rate": 1.7541684126410087e-05, "loss": 0.9347, "step": 7954 }, { "epoch": 3.60444041685546, "grad_norm": 0.13498820045115145, "learning_rate": 1.7531214262143718e-05, "loss": 0.9542, "step": 7955 }, { "epoch": 3.604893520616221, "grad_norm": 0.11446238790734675, "learning_rate": 1.752074664637986e-05, "loss": 0.9468, "step": 7956 }, { "epoch": 3.605346624376982, "grad_norm": 0.13623482446022184, "learning_rate": 1.751028128016607e-05, "loss": 0.9495, "step": 7957 }, { "epoch": 3.6057997281377436, "grad_norm": 0.12875101804206573, "learning_rate": 1.749981816454961e-05, "loss": 0.9281, "step": 7958 }, { "epoch": 3.6062528318985048, "grad_norm": 0.1144876468791024, "learning_rate": 1.748935730057757e-05, "loss": 0.9154, "step": 7959 }, { "epoch": 3.606705935659266, "grad_norm": 0.13449242630652566, "learning_rate": 1.7478898689296772e-05, "loss": 0.9316, "step": 7960 }, { "epoch": 3.607159039420027, "grad_norm": 0.11436261709867912, "learning_rate": 1.7468442331753852e-05, "loss": 0.9321, "step": 7961 }, { "epoch": 3.6076121431807886, "grad_norm": 0.13453611964212933, "learning_rate": 1.74579882289952e-05, "loss": 0.9481, "step": 7962 }, { "epoch": 3.6080652469415497, "grad_norm": 0.1330945231643689, "learning_rate": 1.7447536382066957e-05, "loss": 0.9459, "step": 7963 }, { "epoch": 3.608518350702311, "grad_norm": 0.12738447531607977, "learning_rate": 1.7437086792015096e-05, "loss": 0.933, "step": 7964 }, { "epoch": 3.608971454463072, "grad_norm": 0.11918047700044662, "learning_rate": 1.74266394598853e-05, "loss": 0.9385, "step": 7965 }, { "epoch": 3.609424558223833, "grad_norm": 0.12093935188264154, "learning_rate": 1.741619438672309e-05, "loss": 0.9584, "step": 7966 }, { "epoch": 3.6098776619845943, "grad_norm": 0.11548031467090482, "learning_rate": 1.74057515735737e-05, "loss": 0.9289, "step": 7967 }, { "epoch": 3.610330765745356, "grad_norm": 0.1249525990896203, "learning_rate": 1.739531102148219e-05, "loss": 0.9311, "step": 7968 }, { "epoch": 3.610783869506117, "grad_norm": 0.11420719411852719, "learning_rate": 1.7384872731493364e-05, "loss": 0.9495, "step": 7969 }, { "epoch": 3.611236973266878, "grad_norm": 0.11744128042679754, "learning_rate": 1.737443670465178e-05, "loss": 0.9339, "step": 7970 }, { "epoch": 3.6116900770276392, "grad_norm": 0.11077753073609589, "learning_rate": 1.7364002942001828e-05, "loss": 0.9175, "step": 7971 }, { "epoch": 3.612143180788401, "grad_norm": 0.11357322084263279, "learning_rate": 1.735357144458761e-05, "loss": 0.9258, "step": 7972 }, { "epoch": 3.612596284549162, "grad_norm": 0.11333771655317937, "learning_rate": 1.734314221345306e-05, "loss": 0.9442, "step": 7973 }, { "epoch": 3.613049388309923, "grad_norm": 0.11314369356726879, "learning_rate": 1.7332715249641828e-05, "loss": 0.9223, "step": 7974 }, { "epoch": 3.613502492070684, "grad_norm": 0.11200035131840914, "learning_rate": 1.7322290554197385e-05, "loss": 0.9246, "step": 7975 }, { "epoch": 3.6139555958314453, "grad_norm": 0.13022646603414192, "learning_rate": 1.731186812816293e-05, "loss": 0.9386, "step": 7976 }, { "epoch": 3.6144086995922065, "grad_norm": 0.11355874920803642, "learning_rate": 1.730144797258149e-05, "loss": 0.9489, "step": 7977 }, { "epoch": 3.6148618033529676, "grad_norm": 0.12356937461083989, "learning_rate": 1.7291030088495813e-05, "loss": 0.945, "step": 7978 }, { "epoch": 3.615314907113729, "grad_norm": 0.13597673707741664, "learning_rate": 1.7280614476948434e-05, "loss": 0.9362, "step": 7979 }, { "epoch": 3.6157680108744903, "grad_norm": 0.09903708037447216, "learning_rate": 1.7270201138981693e-05, "loss": 0.918, "step": 7980 }, { "epoch": 3.6162211146352514, "grad_norm": 0.11949925270765965, "learning_rate": 1.725979007563764e-05, "loss": 0.9249, "step": 7981 }, { "epoch": 3.6166742183960126, "grad_norm": 0.11131748527171637, "learning_rate": 1.7249381287958187e-05, "loss": 0.9226, "step": 7982 }, { "epoch": 3.617127322156774, "grad_norm": 0.1092133610398351, "learning_rate": 1.7238974776984906e-05, "loss": 0.9417, "step": 7983 }, { "epoch": 3.6175804259175353, "grad_norm": 0.1081229107889297, "learning_rate": 1.722857054375924e-05, "loss": 0.941, "step": 7984 }, { "epoch": 3.6180335296782964, "grad_norm": 0.10526899582769131, "learning_rate": 1.7218168589322336e-05, "loss": 0.9249, "step": 7985 }, { "epoch": 3.6184866334390575, "grad_norm": 0.10925515363636173, "learning_rate": 1.7207768914715166e-05, "loss": 0.938, "step": 7986 }, { "epoch": 3.6189397371998187, "grad_norm": 0.10616207684489959, "learning_rate": 1.719737152097843e-05, "loss": 0.9261, "step": 7987 }, { "epoch": 3.61939284096058, "grad_norm": 0.10700620916118134, "learning_rate": 1.7186976409152642e-05, "loss": 0.9452, "step": 7988 }, { "epoch": 3.619845944721341, "grad_norm": 0.10280815281724356, "learning_rate": 1.717658358027805e-05, "loss": 0.9486, "step": 7989 }, { "epoch": 3.6202990484821025, "grad_norm": 0.12162019372151313, "learning_rate": 1.7166193035394677e-05, "loss": 0.9222, "step": 7990 }, { "epoch": 3.6207521522428636, "grad_norm": 0.10332522302913096, "learning_rate": 1.715580477554235e-05, "loss": 0.9314, "step": 7991 }, { "epoch": 3.6212052560036248, "grad_norm": 0.12648794834764845, "learning_rate": 1.7145418801760625e-05, "loss": 0.9412, "step": 7992 }, { "epoch": 3.621658359764386, "grad_norm": 0.11196212140219532, "learning_rate": 1.7135035115088877e-05, "loss": 0.9187, "step": 7993 }, { "epoch": 3.6221114635251475, "grad_norm": 0.12000154114602248, "learning_rate": 1.7124653716566187e-05, "loss": 0.9261, "step": 7994 }, { "epoch": 3.6225645672859086, "grad_norm": 0.13210436391173622, "learning_rate": 1.711427460723149e-05, "loss": 0.9113, "step": 7995 }, { "epoch": 3.6230176710466697, "grad_norm": 0.12895369682174693, "learning_rate": 1.7103897788123407e-05, "loss": 0.9175, "step": 7996 }, { "epoch": 3.623470774807431, "grad_norm": 0.12224021359786873, "learning_rate": 1.7093523260280402e-05, "loss": 0.9596, "step": 7997 }, { "epoch": 3.623923878568192, "grad_norm": 0.12622013596351672, "learning_rate": 1.7083151024740665e-05, "loss": 0.9176, "step": 7998 }, { "epoch": 3.624376982328953, "grad_norm": 0.1212532898046045, "learning_rate": 1.7072781082542153e-05, "loss": 0.9218, "step": 7999 }, { "epoch": 3.6248300860897147, "grad_norm": 0.11003993614383814, "learning_rate": 1.7062413434722635e-05, "loss": 0.9358, "step": 8000 }, { "epoch": 3.625283189850476, "grad_norm": 0.13179432674942726, "learning_rate": 1.7052048082319602e-05, "loss": 0.9004, "step": 8001 }, { "epoch": 3.625736293611237, "grad_norm": 0.11109125221435269, "learning_rate": 1.7041685026370367e-05, "loss": 0.9144, "step": 8002 }, { "epoch": 3.626189397371998, "grad_norm": 0.13793514965178877, "learning_rate": 1.703132426791195e-05, "loss": 0.9127, "step": 8003 }, { "epoch": 3.6266425011327597, "grad_norm": 0.11843611964711323, "learning_rate": 1.7020965807981208e-05, "loss": 0.9254, "step": 8004 }, { "epoch": 3.627095604893521, "grad_norm": 0.1162226837146443, "learning_rate": 1.7010609647614728e-05, "loss": 0.9406, "step": 8005 }, { "epoch": 3.627548708654282, "grad_norm": 0.12784274566490467, "learning_rate": 1.700025578784885e-05, "loss": 0.9427, "step": 8006 }, { "epoch": 3.628001812415043, "grad_norm": 0.10931552992359302, "learning_rate": 1.6989904229719732e-05, "loss": 0.9238, "step": 8007 }, { "epoch": 3.628454916175804, "grad_norm": 0.11792579385537515, "learning_rate": 1.697955497426327e-05, "loss": 0.9317, "step": 8008 }, { "epoch": 3.6289080199365653, "grad_norm": 0.11650653996954075, "learning_rate": 1.6969208022515147e-05, "loss": 0.9303, "step": 8009 }, { "epoch": 3.6293611236973264, "grad_norm": 0.10751334676944282, "learning_rate": 1.695886337551079e-05, "loss": 0.9207, "step": 8010 }, { "epoch": 3.629814227458088, "grad_norm": 0.12850065311260134, "learning_rate": 1.6948521034285433e-05, "loss": 0.9379, "step": 8011 }, { "epoch": 3.630267331218849, "grad_norm": 0.11952666150083975, "learning_rate": 1.6938180999874027e-05, "loss": 0.9263, "step": 8012 }, { "epoch": 3.6307204349796103, "grad_norm": 0.12854241434096178, "learning_rate": 1.6927843273311357e-05, "loss": 0.94, "step": 8013 }, { "epoch": 3.6311735387403714, "grad_norm": 0.12534965397151457, "learning_rate": 1.691750785563193e-05, "loss": 0.9396, "step": 8014 }, { "epoch": 3.631626642501133, "grad_norm": 0.11984793109117033, "learning_rate": 1.6907174747870013e-05, "loss": 0.929, "step": 8015 }, { "epoch": 3.632079746261894, "grad_norm": 0.12155858233235317, "learning_rate": 1.6896843951059698e-05, "loss": 0.9254, "step": 8016 }, { "epoch": 3.6325328500226552, "grad_norm": 0.11016976342353799, "learning_rate": 1.6886515466234777e-05, "loss": 0.914, "step": 8017 }, { "epoch": 3.6329859537834164, "grad_norm": 0.12013823041310848, "learning_rate": 1.6876189294428882e-05, "loss": 0.9263, "step": 8018 }, { "epoch": 3.6334390575441775, "grad_norm": 0.11666009531111961, "learning_rate": 1.6865865436675345e-05, "loss": 0.938, "step": 8019 }, { "epoch": 3.6338921613049386, "grad_norm": 0.11613014178082642, "learning_rate": 1.685554389400732e-05, "loss": 0.9134, "step": 8020 }, { "epoch": 3.6343452650656998, "grad_norm": 0.12866568578785054, "learning_rate": 1.6845224667457702e-05, "loss": 0.9301, "step": 8021 }, { "epoch": 3.6347983688264613, "grad_norm": 0.11106064328452032, "learning_rate": 1.683490775805914e-05, "loss": 0.9352, "step": 8022 }, { "epoch": 3.6352514725872225, "grad_norm": 0.1320298879661478, "learning_rate": 1.6824593166844097e-05, "loss": 0.9325, "step": 8023 }, { "epoch": 3.6357045763479836, "grad_norm": 0.10660799887212428, "learning_rate": 1.6814280894844753e-05, "loss": 0.9147, "step": 8024 }, { "epoch": 3.6361576801087447, "grad_norm": 0.11967991505038604, "learning_rate": 1.6803970943093112e-05, "loss": 0.9198, "step": 8025 }, { "epoch": 3.6366107838695063, "grad_norm": 0.11389640995403644, "learning_rate": 1.679366331262088e-05, "loss": 0.9674, "step": 8026 }, { "epoch": 3.6370638876302674, "grad_norm": 0.11048157055786818, "learning_rate": 1.67833580044596e-05, "loss": 0.9326, "step": 8027 }, { "epoch": 3.6375169913910286, "grad_norm": 0.11771302292104342, "learning_rate": 1.677305501964051e-05, "loss": 0.9067, "step": 8028 }, { "epoch": 3.6379700951517897, "grad_norm": 0.1165707357015248, "learning_rate": 1.6762754359194685e-05, "loss": 0.9476, "step": 8029 }, { "epoch": 3.638423198912551, "grad_norm": 0.11606312478618415, "learning_rate": 1.6752456024152927e-05, "loss": 0.9165, "step": 8030 }, { "epoch": 3.638876302673312, "grad_norm": 0.1077231113163884, "learning_rate": 1.6742160015545797e-05, "loss": 0.9243, "step": 8031 }, { "epoch": 3.6393294064340735, "grad_norm": 0.12033562091899459, "learning_rate": 1.6731866334403663e-05, "loss": 0.9388, "step": 8032 }, { "epoch": 3.6397825101948347, "grad_norm": 0.12600209384965794, "learning_rate": 1.6721574981756615e-05, "loss": 0.8941, "step": 8033 }, { "epoch": 3.640235613955596, "grad_norm": 0.10818839332326556, "learning_rate": 1.671128595863456e-05, "loss": 0.9458, "step": 8034 }, { "epoch": 3.640688717716357, "grad_norm": 0.12392201884554278, "learning_rate": 1.670099926606711e-05, "loss": 0.9273, "step": 8035 }, { "epoch": 3.6411418214771185, "grad_norm": 0.12017107911162378, "learning_rate": 1.669071490508371e-05, "loss": 0.9158, "step": 8036 }, { "epoch": 3.6415949252378796, "grad_norm": 0.12969191609413291, "learning_rate": 1.668043287671351e-05, "loss": 0.941, "step": 8037 }, { "epoch": 3.6420480289986408, "grad_norm": 0.1322700315947436, "learning_rate": 1.6670153181985486e-05, "loss": 0.9091, "step": 8038 }, { "epoch": 3.642501132759402, "grad_norm": 0.1067471475124351, "learning_rate": 1.665987582192834e-05, "loss": 0.9228, "step": 8039 }, { "epoch": 3.642954236520163, "grad_norm": 0.13009447121686685, "learning_rate": 1.6649600797570544e-05, "loss": 0.9336, "step": 8040 }, { "epoch": 3.643407340280924, "grad_norm": 0.10999730503750511, "learning_rate": 1.663932810994034e-05, "loss": 0.9227, "step": 8041 }, { "epoch": 3.6438604440416853, "grad_norm": 0.10698632302389902, "learning_rate": 1.6629057760065736e-05, "loss": 0.9355, "step": 8042 }, { "epoch": 3.644313547802447, "grad_norm": 0.1350901692199643, "learning_rate": 1.6618789748974532e-05, "loss": 0.9398, "step": 8043 }, { "epoch": 3.644766651563208, "grad_norm": 0.10329713322407275, "learning_rate": 1.6608524077694244e-05, "loss": 0.9464, "step": 8044 }, { "epoch": 3.645219755323969, "grad_norm": 0.13176847889639465, "learning_rate": 1.6598260747252206e-05, "loss": 0.9495, "step": 8045 }, { "epoch": 3.6456728590847303, "grad_norm": 0.13020488287869125, "learning_rate": 1.6587999758675466e-05, "loss": 0.938, "step": 8046 }, { "epoch": 3.646125962845492, "grad_norm": 0.11575734160518132, "learning_rate": 1.6577741112990896e-05, "loss": 0.9511, "step": 8047 }, { "epoch": 3.646579066606253, "grad_norm": 0.1164715561473904, "learning_rate": 1.6567484811225066e-05, "loss": 0.9154, "step": 8048 }, { "epoch": 3.647032170367014, "grad_norm": 0.11088695451926256, "learning_rate": 1.655723085440439e-05, "loss": 0.9246, "step": 8049 }, { "epoch": 3.6474852741277752, "grad_norm": 0.10275867501249822, "learning_rate": 1.6546979243554978e-05, "loss": 0.9388, "step": 8050 }, { "epoch": 3.6479383778885364, "grad_norm": 0.12110422837845813, "learning_rate": 1.6536729979702718e-05, "loss": 0.9317, "step": 8051 }, { "epoch": 3.6483914816492975, "grad_norm": 0.1081501869357378, "learning_rate": 1.6526483063873313e-05, "loss": 0.923, "step": 8052 }, { "epoch": 3.6488445854100586, "grad_norm": 0.1228703173457055, "learning_rate": 1.6516238497092165e-05, "loss": 0.9021, "step": 8053 }, { "epoch": 3.64929768917082, "grad_norm": 0.11649542570771887, "learning_rate": 1.650599628038449e-05, "loss": 0.9416, "step": 8054 }, { "epoch": 3.6497507929315813, "grad_norm": 0.12535068037306915, "learning_rate": 1.649575641477523e-05, "loss": 0.9361, "step": 8055 }, { "epoch": 3.6502038966923425, "grad_norm": 0.11363314918329079, "learning_rate": 1.6485518901289147e-05, "loss": 0.9292, "step": 8056 }, { "epoch": 3.6506570004531036, "grad_norm": 0.11928082084296564, "learning_rate": 1.647528374095069e-05, "loss": 0.924, "step": 8057 }, { "epoch": 3.651110104213865, "grad_norm": 0.12131019419860428, "learning_rate": 1.6465050934784154e-05, "loss": 0.9217, "step": 8058 }, { "epoch": 3.6515632079746263, "grad_norm": 0.10987584009478579, "learning_rate": 1.6454820483813536e-05, "loss": 0.946, "step": 8059 }, { "epoch": 3.6520163117353874, "grad_norm": 0.11272485127785267, "learning_rate": 1.6444592389062616e-05, "loss": 0.9065, "step": 8060 }, { "epoch": 3.6524694154961486, "grad_norm": 0.11121552148232049, "learning_rate": 1.643436665155496e-05, "loss": 0.9127, "step": 8061 }, { "epoch": 3.6529225192569097, "grad_norm": 0.10693071855192054, "learning_rate": 1.6424143272313857e-05, "loss": 0.9153, "step": 8062 }, { "epoch": 3.653375623017671, "grad_norm": 0.10937060887202306, "learning_rate": 1.6413922252362417e-05, "loss": 0.9445, "step": 8063 }, { "epoch": 3.6538287267784324, "grad_norm": 0.10908638020187456, "learning_rate": 1.6403703592723446e-05, "loss": 0.9207, "step": 8064 }, { "epoch": 3.6542818305391935, "grad_norm": 0.1159022686261462, "learning_rate": 1.6393487294419582e-05, "loss": 0.9217, "step": 8065 }, { "epoch": 3.6547349342999547, "grad_norm": 0.10588934979338524, "learning_rate": 1.6383273358473175e-05, "loss": 0.9244, "step": 8066 }, { "epoch": 3.655188038060716, "grad_norm": 0.12816336016744903, "learning_rate": 1.6373061785906344e-05, "loss": 0.9263, "step": 8067 }, { "epoch": 3.6556411418214774, "grad_norm": 0.11295278071951646, "learning_rate": 1.6362852577741018e-05, "loss": 0.9181, "step": 8068 }, { "epoch": 3.6560942455822385, "grad_norm": 0.1130266922011643, "learning_rate": 1.6352645734998816e-05, "loss": 0.9436, "step": 8069 }, { "epoch": 3.6565473493429996, "grad_norm": 0.11112458245765258, "learning_rate": 1.6342441258701196e-05, "loss": 0.9267, "step": 8070 }, { "epoch": 3.6570004531037608, "grad_norm": 0.10837347005506623, "learning_rate": 1.6332239149869313e-05, "loss": 0.9313, "step": 8071 }, { "epoch": 3.657453556864522, "grad_norm": 0.1085838483171437, "learning_rate": 1.6322039409524147e-05, "loss": 0.9486, "step": 8072 }, { "epoch": 3.657906660625283, "grad_norm": 0.10224977104166535, "learning_rate": 1.6311842038686373e-05, "loss": 0.933, "step": 8073 }, { "epoch": 3.658359764386044, "grad_norm": 0.1074074436874957, "learning_rate": 1.6301647038376503e-05, "loss": 0.9368, "step": 8074 }, { "epoch": 3.6588128681468057, "grad_norm": 0.10368757179561107, "learning_rate": 1.6291454409614754e-05, "loss": 0.9288, "step": 8075 }, { "epoch": 3.659265971907567, "grad_norm": 0.10251141385028502, "learning_rate": 1.6281264153421114e-05, "loss": 0.925, "step": 8076 }, { "epoch": 3.659719075668328, "grad_norm": 0.11212827349380575, "learning_rate": 1.6271076270815368e-05, "loss": 0.9479, "step": 8077 }, { "epoch": 3.660172179429089, "grad_norm": 0.11250284795838593, "learning_rate": 1.626089076281702e-05, "loss": 0.9298, "step": 8078 }, { "epoch": 3.6606252831898507, "grad_norm": 0.10600300925726906, "learning_rate": 1.625070763044538e-05, "loss": 0.9262, "step": 8079 }, { "epoch": 3.661078386950612, "grad_norm": 0.11109508372429752, "learning_rate": 1.624052687471947e-05, "loss": 0.9194, "step": 8080 }, { "epoch": 3.661531490711373, "grad_norm": 0.11956220147925234, "learning_rate": 1.6230348496658125e-05, "loss": 0.9602, "step": 8081 }, { "epoch": 3.661984594472134, "grad_norm": 0.10421107919183095, "learning_rate": 1.622017249727992e-05, "loss": 0.9287, "step": 8082 }, { "epoch": 3.6624376982328952, "grad_norm": 0.12162352400448816, "learning_rate": 1.6209998877603156e-05, "loss": 0.9468, "step": 8083 }, { "epoch": 3.6628908019936564, "grad_norm": 0.10203709625028444, "learning_rate": 1.6199827638645974e-05, "loss": 0.9433, "step": 8084 }, { "epoch": 3.6633439057544175, "grad_norm": 0.12588381016150132, "learning_rate": 1.6189658781426197e-05, "loss": 0.9241, "step": 8085 }, { "epoch": 3.663797009515179, "grad_norm": 0.11614913118582947, "learning_rate": 1.6179492306961475e-05, "loss": 0.9353, "step": 8086 }, { "epoch": 3.66425011327594, "grad_norm": 0.10766365935704757, "learning_rate": 1.6169328216269166e-05, "loss": 0.9297, "step": 8087 }, { "epoch": 3.6647032170367013, "grad_norm": 0.12502905664660366, "learning_rate": 1.6159166510366438e-05, "loss": 0.9237, "step": 8088 }, { "epoch": 3.6651563207974625, "grad_norm": 0.11817193568996824, "learning_rate": 1.6149007190270166e-05, "loss": 0.8967, "step": 8089 }, { "epoch": 3.665609424558224, "grad_norm": 0.1199450195140185, "learning_rate": 1.613885025699705e-05, "loss": 0.9223, "step": 8090 }, { "epoch": 3.666062528318985, "grad_norm": 0.11745981129217994, "learning_rate": 1.61286957115635e-05, "loss": 0.9262, "step": 8091 }, { "epoch": 3.6665156320797463, "grad_norm": 0.11548464482554213, "learning_rate": 1.611854355498569e-05, "loss": 0.9401, "step": 8092 }, { "epoch": 3.6669687358405074, "grad_norm": 0.12285981432146938, "learning_rate": 1.61083937882796e-05, "loss": 0.9401, "step": 8093 }, { "epoch": 3.6674218396012686, "grad_norm": 0.09636439136020149, "learning_rate": 1.609824641246091e-05, "loss": 0.9217, "step": 8094 }, { "epoch": 3.6678749433620297, "grad_norm": 0.11535013505896287, "learning_rate": 1.6088101428545114e-05, "loss": 0.9282, "step": 8095 }, { "epoch": 3.6683280471227913, "grad_norm": 0.10724884083983811, "learning_rate": 1.6077958837547445e-05, "loss": 0.9331, "step": 8096 }, { "epoch": 3.6687811508835524, "grad_norm": 0.09598739227288548, "learning_rate": 1.606781864048288e-05, "loss": 0.915, "step": 8097 }, { "epoch": 3.6692342546443135, "grad_norm": 0.11819111790232431, "learning_rate": 1.6057680838366163e-05, "loss": 0.9271, "step": 8098 }, { "epoch": 3.6696873584050747, "grad_norm": 0.09374075560377937, "learning_rate": 1.6047545432211834e-05, "loss": 0.8828, "step": 8099 }, { "epoch": 3.6701404621658362, "grad_norm": 0.12344663520775298, "learning_rate": 1.6037412423034142e-05, "loss": 0.9371, "step": 8100 }, { "epoch": 3.6705935659265974, "grad_norm": 0.11414385979155726, "learning_rate": 1.6027281811847144e-05, "loss": 0.8976, "step": 8101 }, { "epoch": 3.6710466696873585, "grad_norm": 0.10530219981196856, "learning_rate": 1.6017153599664627e-05, "loss": 0.9381, "step": 8102 }, { "epoch": 3.6714997734481196, "grad_norm": 0.09822846214206037, "learning_rate": 1.6007027787500112e-05, "loss": 0.9181, "step": 8103 }, { "epoch": 3.6719528772088808, "grad_norm": 0.11222497163414527, "learning_rate": 1.599690437636696e-05, "loss": 0.9232, "step": 8104 }, { "epoch": 3.672405980969642, "grad_norm": 0.1010342399662205, "learning_rate": 1.598678336727821e-05, "loss": 0.903, "step": 8105 }, { "epoch": 3.672859084730403, "grad_norm": 0.11758779565855629, "learning_rate": 1.5976664761246715e-05, "loss": 0.9213, "step": 8106 }, { "epoch": 3.6733121884911646, "grad_norm": 0.11074569308115774, "learning_rate": 1.5966548559285045e-05, "loss": 0.9184, "step": 8107 }, { "epoch": 3.6737652922519257, "grad_norm": 0.115598359376834, "learning_rate": 1.5956434762405582e-05, "loss": 0.9244, "step": 8108 }, { "epoch": 3.674218396012687, "grad_norm": 0.1129070092014479, "learning_rate": 1.59463233716204e-05, "loss": 0.9508, "step": 8109 }, { "epoch": 3.674671499773448, "grad_norm": 0.11417746160079162, "learning_rate": 1.5936214387941405e-05, "loss": 0.9324, "step": 8110 }, { "epoch": 3.6751246035342096, "grad_norm": 0.10738364981192014, "learning_rate": 1.5926107812380212e-05, "loss": 0.9212, "step": 8111 }, { "epoch": 3.6755777072949707, "grad_norm": 0.10235493825486594, "learning_rate": 1.5916003645948186e-05, "loss": 0.9237, "step": 8112 }, { "epoch": 3.676030811055732, "grad_norm": 0.12452690452137159, "learning_rate": 1.5905901889656515e-05, "loss": 0.9471, "step": 8113 }, { "epoch": 3.676483914816493, "grad_norm": 0.10855169538759678, "learning_rate": 1.5895802544516062e-05, "loss": 0.9133, "step": 8114 }, { "epoch": 3.676937018577254, "grad_norm": 0.11393848238444353, "learning_rate": 1.5885705611537537e-05, "loss": 0.9307, "step": 8115 }, { "epoch": 3.677390122338015, "grad_norm": 0.10253669413264793, "learning_rate": 1.587561109173132e-05, "loss": 0.9298, "step": 8116 }, { "epoch": 3.677843226098777, "grad_norm": 0.11783748145485204, "learning_rate": 1.5865518986107624e-05, "loss": 0.956, "step": 8117 }, { "epoch": 3.678296329859538, "grad_norm": 0.11564670983371704, "learning_rate": 1.5855429295676366e-05, "loss": 0.9334, "step": 8118 }, { "epoch": 3.678749433620299, "grad_norm": 0.11948424776504678, "learning_rate": 1.5845342021447264e-05, "loss": 0.9052, "step": 8119 }, { "epoch": 3.67920253738106, "grad_norm": 0.12424160783525087, "learning_rate": 1.5835257164429773e-05, "loss": 0.9374, "step": 8120 }, { "epoch": 3.6796556411418218, "grad_norm": 0.1227763276641398, "learning_rate": 1.5825174725633083e-05, "loss": 0.9364, "step": 8121 }, { "epoch": 3.680108744902583, "grad_norm": 0.10686703747714284, "learning_rate": 1.58150947060662e-05, "loss": 0.9324, "step": 8122 }, { "epoch": 3.680561848663344, "grad_norm": 0.1368560945687863, "learning_rate": 1.5805017106737822e-05, "loss": 0.9275, "step": 8123 }, { "epoch": 3.681014952424105, "grad_norm": 0.1127022503879809, "learning_rate": 1.5794941928656466e-05, "loss": 0.9301, "step": 8124 }, { "epoch": 3.6814680561848663, "grad_norm": 0.10542268072786859, "learning_rate": 1.5784869172830358e-05, "loss": 0.932, "step": 8125 }, { "epoch": 3.6819211599456274, "grad_norm": 0.12624136810677145, "learning_rate": 1.5774798840267517e-05, "loss": 0.9166, "step": 8126 }, { "epoch": 3.6823742637063885, "grad_norm": 0.10751174746601455, "learning_rate": 1.57647309319757e-05, "loss": 0.9265, "step": 8127 }, { "epoch": 3.68282736746715, "grad_norm": 0.122524225108163, "learning_rate": 1.5754665448962408e-05, "loss": 0.9186, "step": 8128 }, { "epoch": 3.6832804712279112, "grad_norm": 0.1351986312881106, "learning_rate": 1.574460239223494e-05, "loss": 0.9166, "step": 8129 }, { "epoch": 3.6837335749886724, "grad_norm": 0.11373808566606822, "learning_rate": 1.573454176280031e-05, "loss": 0.9549, "step": 8130 }, { "epoch": 3.6841866787494335, "grad_norm": 0.1299423894068666, "learning_rate": 1.5724483561665333e-05, "loss": 0.9327, "step": 8131 }, { "epoch": 3.684639782510195, "grad_norm": 0.1410144614767191, "learning_rate": 1.5714427789836526e-05, "loss": 0.9427, "step": 8132 }, { "epoch": 3.685092886270956, "grad_norm": 0.11055462786424423, "learning_rate": 1.570437444832022e-05, "loss": 0.9262, "step": 8133 }, { "epoch": 3.6855459900317173, "grad_norm": 0.12678342364234327, "learning_rate": 1.569432353812245e-05, "loss": 0.9403, "step": 8134 }, { "epoch": 3.6859990937924785, "grad_norm": 0.12697765139724204, "learning_rate": 1.568427506024906e-05, "loss": 0.9256, "step": 8135 }, { "epoch": 3.6864521975532396, "grad_norm": 0.10977613378594508, "learning_rate": 1.5674229015705616e-05, "loss": 0.9186, "step": 8136 }, { "epoch": 3.6869053013140007, "grad_norm": 0.13931961062858145, "learning_rate": 1.5664185405497424e-05, "loss": 0.9352, "step": 8137 }, { "epoch": 3.687358405074762, "grad_norm": 0.11014959396247144, "learning_rate": 1.565414423062961e-05, "loss": 0.9513, "step": 8138 }, { "epoch": 3.6878115088355234, "grad_norm": 0.10473755422891137, "learning_rate": 1.5644105492106974e-05, "loss": 0.933, "step": 8139 }, { "epoch": 3.6882646125962846, "grad_norm": 0.1072580845131096, "learning_rate": 1.5634069190934164e-05, "loss": 0.9341, "step": 8140 }, { "epoch": 3.6887177163570457, "grad_norm": 0.10412790141391931, "learning_rate": 1.5624035328115496e-05, "loss": 0.916, "step": 8141 }, { "epoch": 3.689170820117807, "grad_norm": 0.11556979536970853, "learning_rate": 1.5614003904655103e-05, "loss": 0.9351, "step": 8142 }, { "epoch": 3.6896239238785684, "grad_norm": 0.11202336044579123, "learning_rate": 1.560397492155685e-05, "loss": 0.9252, "step": 8143 }, { "epoch": 3.6900770276393295, "grad_norm": 0.10661815245779778, "learning_rate": 1.5593948379824344e-05, "loss": 0.9357, "step": 8144 }, { "epoch": 3.6905301314000907, "grad_norm": 0.10511710157436222, "learning_rate": 1.558392428046099e-05, "loss": 0.9018, "step": 8145 }, { "epoch": 3.690983235160852, "grad_norm": 0.10327524932619035, "learning_rate": 1.5573902624469886e-05, "loss": 0.9185, "step": 8146 }, { "epoch": 3.691436338921613, "grad_norm": 0.1019069247790768, "learning_rate": 1.5563883412853967e-05, "loss": 0.9367, "step": 8147 }, { "epoch": 3.691889442682374, "grad_norm": 0.10905254010180408, "learning_rate": 1.5553866646615835e-05, "loss": 0.9372, "step": 8148 }, { "epoch": 3.6923425464431356, "grad_norm": 0.10866531649342524, "learning_rate": 1.554385232675793e-05, "loss": 0.9395, "step": 8149 }, { "epoch": 3.6927956502038968, "grad_norm": 0.11092761199703435, "learning_rate": 1.553384045428237e-05, "loss": 0.9121, "step": 8150 }, { "epoch": 3.693248753964658, "grad_norm": 0.11962821858127189, "learning_rate": 1.55238310301911e-05, "loss": 0.9092, "step": 8151 }, { "epoch": 3.693701857725419, "grad_norm": 0.1223486742431351, "learning_rate": 1.5513824055485774e-05, "loss": 0.9305, "step": 8152 }, { "epoch": 3.6941549614861806, "grad_norm": 0.11096543466861154, "learning_rate": 1.5503819531167806e-05, "loss": 0.9267, "step": 8153 }, { "epoch": 3.6946080652469417, "grad_norm": 0.1051266888516291, "learning_rate": 1.549381745823836e-05, "loss": 0.9337, "step": 8154 }, { "epoch": 3.695061169007703, "grad_norm": 0.11895338524131062, "learning_rate": 1.5483817837698392e-05, "loss": 0.9272, "step": 8155 }, { "epoch": 3.695514272768464, "grad_norm": 0.10525324383981548, "learning_rate": 1.5473820670548587e-05, "loss": 0.9334, "step": 8156 }, { "epoch": 3.695967376529225, "grad_norm": 0.12272191669543807, "learning_rate": 1.5463825957789347e-05, "loss": 0.9266, "step": 8157 }, { "epoch": 3.6964204802899863, "grad_norm": 0.10380522540332923, "learning_rate": 1.545383370042091e-05, "loss": 0.9353, "step": 8158 }, { "epoch": 3.6968735840507474, "grad_norm": 0.12419598430181825, "learning_rate": 1.5443843899443193e-05, "loss": 0.9346, "step": 8159 }, { "epoch": 3.697326687811509, "grad_norm": 0.10656005356795606, "learning_rate": 1.5433856555855923e-05, "loss": 0.913, "step": 8160 }, { "epoch": 3.69777979157227, "grad_norm": 0.1172650071510958, "learning_rate": 1.542387167065853e-05, "loss": 0.9225, "step": 8161 }, { "epoch": 3.6982328953330312, "grad_norm": 0.10584943217001962, "learning_rate": 1.5413889244850254e-05, "loss": 0.9064, "step": 8162 }, { "epoch": 3.6986859990937924, "grad_norm": 0.10372082248747229, "learning_rate": 1.540390927943004e-05, "loss": 0.9419, "step": 8163 }, { "epoch": 3.699139102854554, "grad_norm": 0.10590014741498116, "learning_rate": 1.53939317753966e-05, "loss": 0.9283, "step": 8164 }, { "epoch": 3.699592206615315, "grad_norm": 0.10865258034119826, "learning_rate": 1.5383956733748423e-05, "loss": 0.9313, "step": 8165 }, { "epoch": 3.700045310376076, "grad_norm": 0.10904991408261991, "learning_rate": 1.5373984155483714e-05, "loss": 0.9334, "step": 8166 }, { "epoch": 3.7004984141368373, "grad_norm": 0.12017111063120554, "learning_rate": 1.5364014041600474e-05, "loss": 0.8991, "step": 8167 }, { "epoch": 3.7009515178975985, "grad_norm": 0.11675311471993084, "learning_rate": 1.5354046393096415e-05, "loss": 0.937, "step": 8168 }, { "epoch": 3.7014046216583596, "grad_norm": 0.13508777963026966, "learning_rate": 1.5344081210969045e-05, "loss": 0.9152, "step": 8169 }, { "epoch": 3.7018577254191207, "grad_norm": 0.15022908731395107, "learning_rate": 1.5334118496215572e-05, "loss": 0.9475, "step": 8170 }, { "epoch": 3.7023108291798823, "grad_norm": 0.12243368504317946, "learning_rate": 1.532415824983302e-05, "loss": 0.9391, "step": 8171 }, { "epoch": 3.7027639329406434, "grad_norm": 0.12337820388141697, "learning_rate": 1.531420047281812e-05, "loss": 0.9329, "step": 8172 }, { "epoch": 3.7032170367014046, "grad_norm": 0.12393763249335801, "learning_rate": 1.5304245166167354e-05, "loss": 0.9166, "step": 8173 }, { "epoch": 3.7036701404621657, "grad_norm": 0.1100681779560133, "learning_rate": 1.5294292330876994e-05, "loss": 0.9073, "step": 8174 }, { "epoch": 3.7041232442229273, "grad_norm": 0.10737162343394761, "learning_rate": 1.5284341967943025e-05, "loss": 0.9386, "step": 8175 }, { "epoch": 3.7045763479836884, "grad_norm": 0.10025403647589065, "learning_rate": 1.5274394078361225e-05, "loss": 0.9255, "step": 8176 }, { "epoch": 3.7050294517444495, "grad_norm": 0.11022457231041688, "learning_rate": 1.5264448663127075e-05, "loss": 0.9057, "step": 8177 }, { "epoch": 3.7054825555052107, "grad_norm": 0.11278832481244346, "learning_rate": 1.5254505723235861e-05, "loss": 0.9486, "step": 8178 }, { "epoch": 3.705935659265972, "grad_norm": 0.11359748893046384, "learning_rate": 1.5244565259682574e-05, "loss": 0.9169, "step": 8179 }, { "epoch": 3.706388763026733, "grad_norm": 0.11458097550851183, "learning_rate": 1.5234627273462002e-05, "loss": 0.9288, "step": 8180 }, { "epoch": 3.7068418667874945, "grad_norm": 0.11349080563080341, "learning_rate": 1.522469176556865e-05, "loss": 0.9312, "step": 8181 }, { "epoch": 3.7072949705482556, "grad_norm": 0.10740515210220032, "learning_rate": 1.5214758736996764e-05, "loss": 0.94, "step": 8182 }, { "epoch": 3.7077480743090168, "grad_norm": 0.11513118731803372, "learning_rate": 1.5204828188740407e-05, "loss": 0.9231, "step": 8183 }, { "epoch": 3.708201178069778, "grad_norm": 0.11955905256804107, "learning_rate": 1.5194900121793317e-05, "loss": 0.9475, "step": 8184 }, { "epoch": 3.7086542818305395, "grad_norm": 0.13271592970315008, "learning_rate": 1.5184974537149044e-05, "loss": 0.9175, "step": 8185 }, { "epoch": 3.7091073855913006, "grad_norm": 0.10752574359104954, "learning_rate": 1.5175051435800839e-05, "loss": 0.9093, "step": 8186 }, { "epoch": 3.7095604893520617, "grad_norm": 0.11791904299242406, "learning_rate": 1.5165130818741754e-05, "loss": 0.9202, "step": 8187 }, { "epoch": 3.710013593112823, "grad_norm": 0.11475111346782288, "learning_rate": 1.515521268696456e-05, "loss": 0.9069, "step": 8188 }, { "epoch": 3.710466696873584, "grad_norm": 0.11534439495553313, "learning_rate": 1.5145297041461765e-05, "loss": 0.9302, "step": 8189 }, { "epoch": 3.710919800634345, "grad_norm": 0.11237887722583699, "learning_rate": 1.5135383883225685e-05, "loss": 0.9418, "step": 8190 }, { "epoch": 3.7113729043951063, "grad_norm": 0.11300646777816442, "learning_rate": 1.5125473213248319e-05, "loss": 0.9204, "step": 8191 }, { "epoch": 3.711826008155868, "grad_norm": 0.12250992066563203, "learning_rate": 1.5115565032521482e-05, "loss": 0.9209, "step": 8192 }, { "epoch": 3.712279111916629, "grad_norm": 0.1282556953171212, "learning_rate": 1.5105659342036681e-05, "loss": 0.9083, "step": 8193 }, { "epoch": 3.71273221567739, "grad_norm": 0.11094896483639305, "learning_rate": 1.5095756142785227e-05, "loss": 0.9135, "step": 8194 }, { "epoch": 3.713185319438151, "grad_norm": 0.1331876742854763, "learning_rate": 1.5085855435758125e-05, "loss": 0.9157, "step": 8195 }, { "epoch": 3.713638423198913, "grad_norm": 0.1196058128299587, "learning_rate": 1.5075957221946195e-05, "loss": 0.9355, "step": 8196 }, { "epoch": 3.714091526959674, "grad_norm": 0.10834507951142591, "learning_rate": 1.5066061502339953e-05, "loss": 0.919, "step": 8197 }, { "epoch": 3.714544630720435, "grad_norm": 0.13385692935275634, "learning_rate": 1.505616827792968e-05, "loss": 0.9259, "step": 8198 }, { "epoch": 3.714997734481196, "grad_norm": 0.11119081533960393, "learning_rate": 1.5046277549705437e-05, "loss": 0.9128, "step": 8199 }, { "epoch": 3.7154508382419573, "grad_norm": 0.1260789562172626, "learning_rate": 1.503638931865698e-05, "loss": 0.9319, "step": 8200 }, { "epoch": 3.7159039420027185, "grad_norm": 0.11118341481810776, "learning_rate": 1.5026503585773884e-05, "loss": 0.9672, "step": 8201 }, { "epoch": 3.7163570457634796, "grad_norm": 0.11990283660544439, "learning_rate": 1.5016620352045403e-05, "loss": 0.9264, "step": 8202 }, { "epoch": 3.716810149524241, "grad_norm": 0.12258000828150147, "learning_rate": 1.5006739618460601e-05, "loss": 0.9346, "step": 8203 }, { "epoch": 3.7172632532850023, "grad_norm": 0.11038018587253041, "learning_rate": 1.4996861386008258e-05, "loss": 0.9437, "step": 8204 }, { "epoch": 3.7177163570457634, "grad_norm": 0.14766223488252686, "learning_rate": 1.498698565567689e-05, "loss": 0.9353, "step": 8205 }, { "epoch": 3.7181694608065246, "grad_norm": 0.10581201583328674, "learning_rate": 1.4977112428454813e-05, "loss": 0.9321, "step": 8206 }, { "epoch": 3.718622564567286, "grad_norm": 0.15596478036436504, "learning_rate": 1.4967241705330043e-05, "loss": 0.9316, "step": 8207 }, { "epoch": 3.7190756683280473, "grad_norm": 0.11484661051287148, "learning_rate": 1.4957373487290383e-05, "loss": 0.9089, "step": 8208 }, { "epoch": 3.7195287720888084, "grad_norm": 0.14223723387201948, "learning_rate": 1.494750777532335e-05, "loss": 0.9186, "step": 8209 }, { "epoch": 3.7199818758495695, "grad_norm": 0.12807012047298102, "learning_rate": 1.493764457041626e-05, "loss": 0.9372, "step": 8210 }, { "epoch": 3.7204349796103307, "grad_norm": 0.12703344370469025, "learning_rate": 1.4927783873556098e-05, "loss": 0.9282, "step": 8211 }, { "epoch": 3.720888083371092, "grad_norm": 0.12252854642690743, "learning_rate": 1.4917925685729687e-05, "loss": 0.9432, "step": 8212 }, { "epoch": 3.7213411871318534, "grad_norm": 0.10300549373704086, "learning_rate": 1.4908070007923527e-05, "loss": 0.9443, "step": 8213 }, { "epoch": 3.7217942908926145, "grad_norm": 0.12420871117983623, "learning_rate": 1.4898216841123931e-05, "loss": 0.9294, "step": 8214 }, { "epoch": 3.7222473946533756, "grad_norm": 0.10781836178530965, "learning_rate": 1.4888366186316896e-05, "loss": 0.9077, "step": 8215 }, { "epoch": 3.7227004984141367, "grad_norm": 0.12358425743556459, "learning_rate": 1.4878518044488233e-05, "loss": 0.9206, "step": 8216 }, { "epoch": 3.7231536021748983, "grad_norm": 0.10408457841164394, "learning_rate": 1.4868672416623442e-05, "loss": 0.9144, "step": 8217 }, { "epoch": 3.7236067059356595, "grad_norm": 0.1259743913082584, "learning_rate": 1.4858829303707799e-05, "loss": 0.9452, "step": 8218 }, { "epoch": 3.7240598096964206, "grad_norm": 0.13362597222127293, "learning_rate": 1.484898870672634e-05, "loss": 0.9172, "step": 8219 }, { "epoch": 3.7245129134571817, "grad_norm": 0.12653332008455154, "learning_rate": 1.4839150626663821e-05, "loss": 0.9151, "step": 8220 }, { "epoch": 3.724966017217943, "grad_norm": 0.14100333169479795, "learning_rate": 1.482931506450478e-05, "loss": 0.9253, "step": 8221 }, { "epoch": 3.725419120978704, "grad_norm": 0.14087230107001192, "learning_rate": 1.4819482021233458e-05, "loss": 0.9211, "step": 8222 }, { "epoch": 3.725872224739465, "grad_norm": 0.1534115654021729, "learning_rate": 1.4809651497833897e-05, "loss": 0.9165, "step": 8223 }, { "epoch": 3.7263253285002267, "grad_norm": 0.14591973687810553, "learning_rate": 1.479982349528985e-05, "loss": 0.9504, "step": 8224 }, { "epoch": 3.726778432260988, "grad_norm": 0.12912040378797726, "learning_rate": 1.4789998014584815e-05, "loss": 0.9388, "step": 8225 }, { "epoch": 3.727231536021749, "grad_norm": 0.13650992340620585, "learning_rate": 1.478017505670207e-05, "loss": 0.9439, "step": 8226 }, { "epoch": 3.72768463978251, "grad_norm": 0.12464004108704826, "learning_rate": 1.4770354622624598e-05, "loss": 0.9109, "step": 8227 }, { "epoch": 3.7281377435432717, "grad_norm": 0.1541093048941124, "learning_rate": 1.476053671333518e-05, "loss": 0.9543, "step": 8228 }, { "epoch": 3.728590847304033, "grad_norm": 0.11985032350481933, "learning_rate": 1.4750721329816288e-05, "loss": 0.9053, "step": 8229 }, { "epoch": 3.729043951064794, "grad_norm": 0.13873501558074355, "learning_rate": 1.47409084730502e-05, "loss": 0.9296, "step": 8230 }, { "epoch": 3.729497054825555, "grad_norm": 0.10415575331598684, "learning_rate": 1.4731098144018879e-05, "loss": 0.9372, "step": 8231 }, { "epoch": 3.729950158586316, "grad_norm": 0.1452488544049677, "learning_rate": 1.4721290343704095e-05, "loss": 0.933, "step": 8232 }, { "epoch": 3.7304032623470773, "grad_norm": 0.10245264790696514, "learning_rate": 1.4711485073087327e-05, "loss": 0.9097, "step": 8233 }, { "epoch": 3.7308563661078384, "grad_norm": 0.1355969794073314, "learning_rate": 1.4701682333149796e-05, "loss": 0.9215, "step": 8234 }, { "epoch": 3.7313094698686, "grad_norm": 0.10525215748110632, "learning_rate": 1.4691882124872506e-05, "loss": 0.924, "step": 8235 }, { "epoch": 3.731762573629361, "grad_norm": 0.1314333985092991, "learning_rate": 1.4682084449236164e-05, "loss": 0.9455, "step": 8236 }, { "epoch": 3.7322156773901223, "grad_norm": 0.10159236851655112, "learning_rate": 1.467228930722128e-05, "loss": 0.9233, "step": 8237 }, { "epoch": 3.7326687811508834, "grad_norm": 0.11420358902654083, "learning_rate": 1.4662496699808033e-05, "loss": 0.9217, "step": 8238 }, { "epoch": 3.733121884911645, "grad_norm": 0.10908841650905846, "learning_rate": 1.465270662797643e-05, "loss": 0.9404, "step": 8239 }, { "epoch": 3.733574988672406, "grad_norm": 0.11267747846466657, "learning_rate": 1.4642919092706169e-05, "loss": 0.944, "step": 8240 }, { "epoch": 3.7340280924331672, "grad_norm": 0.11696401228582143, "learning_rate": 1.4633134094976695e-05, "loss": 0.9332, "step": 8241 }, { "epoch": 3.7344811961939284, "grad_norm": 0.11144613959991467, "learning_rate": 1.4623351635767247e-05, "loss": 0.9148, "step": 8242 }, { "epoch": 3.7349342999546895, "grad_norm": 0.11536079728141235, "learning_rate": 1.4613571716056752e-05, "loss": 0.9466, "step": 8243 }, { "epoch": 3.7353874037154506, "grad_norm": 0.11225646665132338, "learning_rate": 1.4603794336823928e-05, "loss": 0.9189, "step": 8244 }, { "epoch": 3.735840507476212, "grad_norm": 0.10436142690869694, "learning_rate": 1.4594019499047196e-05, "loss": 0.9309, "step": 8245 }, { "epoch": 3.7362936112369733, "grad_norm": 0.1202824695006055, "learning_rate": 1.4584247203704775e-05, "loss": 0.9185, "step": 8246 }, { "epoch": 3.7367467149977345, "grad_norm": 0.10974765772478653, "learning_rate": 1.4574477451774578e-05, "loss": 0.9446, "step": 8247 }, { "epoch": 3.7371998187584956, "grad_norm": 0.1146810415473641, "learning_rate": 1.4564710244234302e-05, "loss": 0.9112, "step": 8248 }, { "epoch": 3.737652922519257, "grad_norm": 0.1176077357237517, "learning_rate": 1.4554945582061368e-05, "loss": 0.9353, "step": 8249 }, { "epoch": 3.7381060262800183, "grad_norm": 0.12149824097821248, "learning_rate": 1.4545183466232936e-05, "loss": 0.9356, "step": 8250 }, { "epoch": 3.7385591300407794, "grad_norm": 0.11282984646498287, "learning_rate": 1.4535423897725944e-05, "loss": 0.931, "step": 8251 }, { "epoch": 3.7390122338015406, "grad_norm": 0.12426200999839365, "learning_rate": 1.4525666877517034e-05, "loss": 0.9508, "step": 8252 }, { "epoch": 3.7394653375623017, "grad_norm": 0.11465820741441135, "learning_rate": 1.4515912406582632e-05, "loss": 0.9166, "step": 8253 }, { "epoch": 3.739918441323063, "grad_norm": 0.1158212478481195, "learning_rate": 1.4506160485898875e-05, "loss": 0.9163, "step": 8254 }, { "epoch": 3.740371545083824, "grad_norm": 0.12282035403731179, "learning_rate": 1.449641111644167e-05, "loss": 0.9143, "step": 8255 }, { "epoch": 3.7408246488445855, "grad_norm": 0.12993654093875492, "learning_rate": 1.448666429918665e-05, "loss": 0.9284, "step": 8256 }, { "epoch": 3.7412777526053467, "grad_norm": 0.09860002170213793, "learning_rate": 1.4476920035109214e-05, "loss": 0.9244, "step": 8257 }, { "epoch": 3.741730856366108, "grad_norm": 0.12364633958295973, "learning_rate": 1.4467178325184486e-05, "loss": 0.9413, "step": 8258 }, { "epoch": 3.742183960126869, "grad_norm": 0.11051454223841235, "learning_rate": 1.4457439170387328e-05, "loss": 0.9167, "step": 8259 }, { "epoch": 3.7426370638876305, "grad_norm": 0.11353008813903788, "learning_rate": 1.4447702571692385e-05, "loss": 0.9325, "step": 8260 }, { "epoch": 3.7430901676483916, "grad_norm": 0.11886165161755524, "learning_rate": 1.443796853007399e-05, "loss": 0.9264, "step": 8261 }, { "epoch": 3.7435432714091528, "grad_norm": 0.09784220809302502, "learning_rate": 1.4428237046506284e-05, "loss": 0.9354, "step": 8262 }, { "epoch": 3.743996375169914, "grad_norm": 0.12092245590123468, "learning_rate": 1.4418508121963086e-05, "loss": 0.9395, "step": 8263 }, { "epoch": 3.744449478930675, "grad_norm": 0.09607856119794819, "learning_rate": 1.4408781757418022e-05, "loss": 0.9336, "step": 8264 }, { "epoch": 3.744902582691436, "grad_norm": 0.1101945481068568, "learning_rate": 1.4399057953844415e-05, "loss": 0.9445, "step": 8265 }, { "epoch": 3.7453556864521973, "grad_norm": 0.12306744709726186, "learning_rate": 1.4389336712215341e-05, "loss": 0.907, "step": 8266 }, { "epoch": 3.745808790212959, "grad_norm": 0.09736905273684922, "learning_rate": 1.4379618033503651e-05, "loss": 0.9272, "step": 8267 }, { "epoch": 3.74626189397372, "grad_norm": 0.11960933748152078, "learning_rate": 1.4369901918681896e-05, "loss": 0.9076, "step": 8268 }, { "epoch": 3.746714997734481, "grad_norm": 0.1102505204213376, "learning_rate": 1.4360188368722394e-05, "loss": 0.9384, "step": 8269 }, { "epoch": 3.7471681014952423, "grad_norm": 0.13072761875938205, "learning_rate": 1.4350477384597191e-05, "loss": 0.9407, "step": 8270 }, { "epoch": 3.747621205256004, "grad_norm": 0.11826120380914559, "learning_rate": 1.434076896727811e-05, "loss": 0.947, "step": 8271 }, { "epoch": 3.748074309016765, "grad_norm": 0.10915195780336752, "learning_rate": 1.433106311773667e-05, "loss": 0.9168, "step": 8272 }, { "epoch": 3.748527412777526, "grad_norm": 0.11959621747517107, "learning_rate": 1.4321359836944186e-05, "loss": 0.9203, "step": 8273 }, { "epoch": 3.7489805165382872, "grad_norm": 0.10965010535305868, "learning_rate": 1.4311659125871656e-05, "loss": 0.9296, "step": 8274 }, { "epoch": 3.7494336202990484, "grad_norm": 0.11643418885038184, "learning_rate": 1.4301960985489883e-05, "loss": 0.9339, "step": 8275 }, { "epoch": 3.7498867240598095, "grad_norm": 0.11056836739739018, "learning_rate": 1.4292265416769353e-05, "loss": 0.9478, "step": 8276 }, { "epoch": 3.750339827820571, "grad_norm": 0.11079088571820596, "learning_rate": 1.428257242068035e-05, "loss": 0.9213, "step": 8277 }, { "epoch": 3.750792931581332, "grad_norm": 0.10429725219391897, "learning_rate": 1.4272881998192865e-05, "loss": 0.9123, "step": 8278 }, { "epoch": 3.7512460353420933, "grad_norm": 0.11366872177193271, "learning_rate": 1.4263194150276621e-05, "loss": 0.9307, "step": 8279 }, { "epoch": 3.7516991391028545, "grad_norm": 0.10469028960872713, "learning_rate": 1.425350887790113e-05, "loss": 0.9441, "step": 8280 }, { "epoch": 3.752152242863616, "grad_norm": 0.12106126354374645, "learning_rate": 1.42438261820356e-05, "loss": 0.9297, "step": 8281 }, { "epoch": 3.752605346624377, "grad_norm": 0.10432933600845017, "learning_rate": 1.4234146063649021e-05, "loss": 0.9395, "step": 8282 }, { "epoch": 3.7530584503851383, "grad_norm": 0.13333131853396368, "learning_rate": 1.422446852371008e-05, "loss": 0.9341, "step": 8283 }, { "epoch": 3.7535115541458994, "grad_norm": 0.10791461751949047, "learning_rate": 1.4214793563187254e-05, "loss": 0.9239, "step": 8284 }, { "epoch": 3.7539646579066606, "grad_norm": 0.10153184121309951, "learning_rate": 1.420512118304873e-05, "loss": 0.9113, "step": 8285 }, { "epoch": 3.7544177616674217, "grad_norm": 0.10542597184748727, "learning_rate": 1.4195451384262428e-05, "loss": 0.9333, "step": 8286 }, { "epoch": 3.754870865428183, "grad_norm": 0.10313732324146191, "learning_rate": 1.4185784167796053e-05, "loss": 0.9313, "step": 8287 }, { "epoch": 3.7553239691889444, "grad_norm": 0.09992413007134558, "learning_rate": 1.4176119534617003e-05, "loss": 0.9495, "step": 8288 }, { "epoch": 3.7557770729497055, "grad_norm": 0.09981364462045937, "learning_rate": 1.4166457485692463e-05, "loss": 0.9414, "step": 8289 }, { "epoch": 3.7562301767104667, "grad_norm": 0.10851315477756901, "learning_rate": 1.4156798021989308e-05, "loss": 0.9236, "step": 8290 }, { "epoch": 3.756683280471228, "grad_norm": 0.09769654952482192, "learning_rate": 1.4147141144474215e-05, "loss": 0.9378, "step": 8291 }, { "epoch": 3.7571363842319894, "grad_norm": 0.11547454721355083, "learning_rate": 1.4137486854113536e-05, "loss": 0.9314, "step": 8292 }, { "epoch": 3.7575894879927505, "grad_norm": 0.11279084723788167, "learning_rate": 1.4127835151873432e-05, "loss": 0.9354, "step": 8293 }, { "epoch": 3.7580425917535116, "grad_norm": 0.10695449329615335, "learning_rate": 1.411818603871975e-05, "loss": 0.9198, "step": 8294 }, { "epoch": 3.7584956955142728, "grad_norm": 0.11996325778427742, "learning_rate": 1.410853951561809e-05, "loss": 0.909, "step": 8295 }, { "epoch": 3.758948799275034, "grad_norm": 0.09046192279615486, "learning_rate": 1.409889558353383e-05, "loss": 0.9454, "step": 8296 }, { "epoch": 3.759401903035795, "grad_norm": 0.11715891945887517, "learning_rate": 1.4089254243432024e-05, "loss": 0.9272, "step": 8297 }, { "epoch": 3.759855006796556, "grad_norm": 0.09616628311006475, "learning_rate": 1.4079615496277543e-05, "loss": 0.9353, "step": 8298 }, { "epoch": 3.7603081105573177, "grad_norm": 0.11860404855795385, "learning_rate": 1.4069979343034922e-05, "loss": 0.917, "step": 8299 }, { "epoch": 3.760761214318079, "grad_norm": 0.10468285346823031, "learning_rate": 1.4060345784668501e-05, "loss": 0.917, "step": 8300 }, { "epoch": 3.76121431807884, "grad_norm": 0.10736702033614927, "learning_rate": 1.4050714822142322e-05, "loss": 0.9281, "step": 8301 }, { "epoch": 3.761667421839601, "grad_norm": 0.10161959796253087, "learning_rate": 1.4041086456420158e-05, "loss": 0.9247, "step": 8302 }, { "epoch": 3.7621205256003627, "grad_norm": 0.10872939960656126, "learning_rate": 1.403146068846558e-05, "loss": 0.9198, "step": 8303 }, { "epoch": 3.762573629361124, "grad_norm": 0.10119707674045361, "learning_rate": 1.4021837519241822e-05, "loss": 0.9568, "step": 8304 }, { "epoch": 3.763026733121885, "grad_norm": 0.10475681185739444, "learning_rate": 1.4012216949711923e-05, "loss": 0.9283, "step": 8305 }, { "epoch": 3.763479836882646, "grad_norm": 0.09664182159810565, "learning_rate": 1.4002598980838618e-05, "loss": 0.9391, "step": 8306 }, { "epoch": 3.763932940643407, "grad_norm": 0.11019750446865341, "learning_rate": 1.3992983613584419e-05, "loss": 0.9241, "step": 8307 }, { "epoch": 3.7643860444041684, "grad_norm": 0.12277558724483296, "learning_rate": 1.3983370848911531e-05, "loss": 0.9302, "step": 8308 }, { "epoch": 3.76483914816493, "grad_norm": 0.10460022499084419, "learning_rate": 1.3973760687781956e-05, "loss": 0.9379, "step": 8309 }, { "epoch": 3.765292251925691, "grad_norm": 0.11682560693704959, "learning_rate": 1.396415313115739e-05, "loss": 0.9143, "step": 8310 }, { "epoch": 3.765745355686452, "grad_norm": 0.12184735792080303, "learning_rate": 1.3954548179999266e-05, "loss": 0.9081, "step": 8311 }, { "epoch": 3.7661984594472133, "grad_norm": 0.09707843195740741, "learning_rate": 1.3944945835268802e-05, "loss": 0.9468, "step": 8312 }, { "epoch": 3.766651563207975, "grad_norm": 0.10073675085443101, "learning_rate": 1.3935346097926901e-05, "loss": 0.9093, "step": 8313 }, { "epoch": 3.767104666968736, "grad_norm": 0.10704575531656725, "learning_rate": 1.3925748968934255e-05, "loss": 0.9397, "step": 8314 }, { "epoch": 3.767557770729497, "grad_norm": 0.11261118819481336, "learning_rate": 1.3916154449251247e-05, "loss": 0.9184, "step": 8315 }, { "epoch": 3.7680108744902583, "grad_norm": 0.10248528439415813, "learning_rate": 1.3906562539838047e-05, "loss": 0.9198, "step": 8316 }, { "epoch": 3.7684639782510194, "grad_norm": 0.10755338658093502, "learning_rate": 1.3896973241654514e-05, "loss": 0.9394, "step": 8317 }, { "epoch": 3.7689170820117806, "grad_norm": 0.1016021629900838, "learning_rate": 1.3887386555660297e-05, "loss": 0.9228, "step": 8318 }, { "epoch": 3.7693701857725417, "grad_norm": 0.11466672172427944, "learning_rate": 1.3877802482814745e-05, "loss": 0.9267, "step": 8319 }, { "epoch": 3.7698232895333033, "grad_norm": 0.1136279285131815, "learning_rate": 1.386822102407694e-05, "loss": 0.9431, "step": 8320 }, { "epoch": 3.7702763932940644, "grad_norm": 0.10975563510837703, "learning_rate": 1.3858642180405752e-05, "loss": 0.9212, "step": 8321 }, { "epoch": 3.7707294970548255, "grad_norm": 0.12200979624503608, "learning_rate": 1.384906595275973e-05, "loss": 0.9389, "step": 8322 }, { "epoch": 3.7711826008155866, "grad_norm": 0.11114125908278898, "learning_rate": 1.3839492342097214e-05, "loss": 0.9492, "step": 8323 }, { "epoch": 3.7716357045763482, "grad_norm": 0.11672491190873599, "learning_rate": 1.3829921349376246e-05, "loss": 0.9355, "step": 8324 }, { "epoch": 3.7720888083371094, "grad_norm": 0.1071919597262888, "learning_rate": 1.3820352975554618e-05, "loss": 0.9411, "step": 8325 }, { "epoch": 3.7725419120978705, "grad_norm": 0.10798432239098557, "learning_rate": 1.3810787221589839e-05, "loss": 0.9408, "step": 8326 }, { "epoch": 3.7729950158586316, "grad_norm": 0.11622820149724888, "learning_rate": 1.3801224088439211e-05, "loss": 0.9449, "step": 8327 }, { "epoch": 3.7734481196193927, "grad_norm": 0.09784988451649286, "learning_rate": 1.3791663577059704e-05, "loss": 0.9472, "step": 8328 }, { "epoch": 3.773901223380154, "grad_norm": 0.11677006371602543, "learning_rate": 1.378210568840809e-05, "loss": 0.9393, "step": 8329 }, { "epoch": 3.774354327140915, "grad_norm": 0.09597506824864316, "learning_rate": 1.3772550423440838e-05, "loss": 0.9363, "step": 8330 }, { "epoch": 3.7748074309016766, "grad_norm": 0.1202559418715665, "learning_rate": 1.3762997783114149e-05, "loss": 0.9403, "step": 8331 }, { "epoch": 3.7752605346624377, "grad_norm": 0.10122037684577574, "learning_rate": 1.3753447768384005e-05, "loss": 0.9345, "step": 8332 }, { "epoch": 3.775713638423199, "grad_norm": 0.10467131128966195, "learning_rate": 1.3743900380206072e-05, "loss": 0.9377, "step": 8333 }, { "epoch": 3.77616674218396, "grad_norm": 0.09894113386260478, "learning_rate": 1.3734355619535804e-05, "loss": 0.9387, "step": 8334 }, { "epoch": 3.7766198459447216, "grad_norm": 0.10127174007187015, "learning_rate": 1.372481348732834e-05, "loss": 0.9087, "step": 8335 }, { "epoch": 3.7770729497054827, "grad_norm": 0.0955880105050447, "learning_rate": 1.371527398453861e-05, "loss": 0.9174, "step": 8336 }, { "epoch": 3.777526053466244, "grad_norm": 0.11189918599080295, "learning_rate": 1.3705737112121224e-05, "loss": 0.9145, "step": 8337 }, { "epoch": 3.777979157227005, "grad_norm": 0.10128926819430376, "learning_rate": 1.3696202871030595e-05, "loss": 0.9191, "step": 8338 }, { "epoch": 3.778432260987766, "grad_norm": 0.10430048582183789, "learning_rate": 1.3686671262220812e-05, "loss": 0.921, "step": 8339 }, { "epoch": 3.778885364748527, "grad_norm": 0.09923891058433712, "learning_rate": 1.3677142286645717e-05, "loss": 0.9449, "step": 8340 }, { "epoch": 3.779338468509289, "grad_norm": 0.10102271140552502, "learning_rate": 1.366761594525892e-05, "loss": 0.9398, "step": 8341 }, { "epoch": 3.77979157227005, "grad_norm": 0.10680029243574606, "learning_rate": 1.3658092239013718e-05, "loss": 0.9314, "step": 8342 }, { "epoch": 3.780244676030811, "grad_norm": 0.10120249566280391, "learning_rate": 1.36485711688632e-05, "loss": 0.9286, "step": 8343 }, { "epoch": 3.780697779791572, "grad_norm": 0.10011419551861699, "learning_rate": 1.363905273576013e-05, "loss": 0.9353, "step": 8344 }, { "epoch": 3.7811508835523338, "grad_norm": 0.096851911691175, "learning_rate": 1.3629536940657068e-05, "loss": 0.9405, "step": 8345 }, { "epoch": 3.781603987313095, "grad_norm": 0.11071295668596075, "learning_rate": 1.3620023784506269e-05, "loss": 0.9163, "step": 8346 }, { "epoch": 3.782057091073856, "grad_norm": 0.09928003452953747, "learning_rate": 1.3610513268259724e-05, "loss": 0.9245, "step": 8347 }, { "epoch": 3.782510194834617, "grad_norm": 0.1179854064198568, "learning_rate": 1.3601005392869193e-05, "loss": 0.9275, "step": 8348 }, { "epoch": 3.7829632985953783, "grad_norm": 0.095910024112715, "learning_rate": 1.3591500159286125e-05, "loss": 0.9256, "step": 8349 }, { "epoch": 3.7834164023561394, "grad_norm": 0.11656204981194469, "learning_rate": 1.358199756846176e-05, "loss": 0.9369, "step": 8350 }, { "epoch": 3.7838695061169005, "grad_norm": 0.09488454622168771, "learning_rate": 1.357249762134702e-05, "loss": 0.9356, "step": 8351 }, { "epoch": 3.784322609877662, "grad_norm": 0.11189361251908285, "learning_rate": 1.3563000318892608e-05, "loss": 0.9307, "step": 8352 }, { "epoch": 3.7847757136384232, "grad_norm": 0.10696945419535497, "learning_rate": 1.3553505662048916e-05, "loss": 0.9384, "step": 8353 }, { "epoch": 3.7852288173991844, "grad_norm": 0.09506069062730911, "learning_rate": 1.3544013651766118e-05, "loss": 0.911, "step": 8354 }, { "epoch": 3.7856819211599455, "grad_norm": 0.11843117644080381, "learning_rate": 1.3534524288994097e-05, "loss": 0.9309, "step": 8355 }, { "epoch": 3.786135024920707, "grad_norm": 0.0988930610388978, "learning_rate": 1.3525037574682456e-05, "loss": 0.9453, "step": 8356 }, { "epoch": 3.786588128681468, "grad_norm": 0.10398866541894827, "learning_rate": 1.3515553509780577e-05, "loss": 0.9341, "step": 8357 }, { "epoch": 3.7870412324422293, "grad_norm": 0.09718608891438799, "learning_rate": 1.3506072095237533e-05, "loss": 0.9279, "step": 8358 }, { "epoch": 3.7874943362029905, "grad_norm": 0.1059751182790431, "learning_rate": 1.3496593332002163e-05, "loss": 0.9402, "step": 8359 }, { "epoch": 3.7879474399637516, "grad_norm": 0.09355927430083316, "learning_rate": 1.3487117221023019e-05, "loss": 0.9185, "step": 8360 }, { "epoch": 3.7884005437245127, "grad_norm": 0.10543378414394204, "learning_rate": 1.3477643763248414e-05, "loss": 0.9248, "step": 8361 }, { "epoch": 3.788853647485274, "grad_norm": 0.09813135483139114, "learning_rate": 1.3468172959626365e-05, "loss": 0.9331, "step": 8362 }, { "epoch": 3.7893067512460354, "grad_norm": 0.10441714739899549, "learning_rate": 1.3458704811104623e-05, "loss": 0.9211, "step": 8363 }, { "epoch": 3.7897598550067966, "grad_norm": 0.09936289991608256, "learning_rate": 1.3449239318630718e-05, "loss": 0.9255, "step": 8364 }, { "epoch": 3.7902129587675577, "grad_norm": 0.10175260871878564, "learning_rate": 1.3439776483151858e-05, "loss": 0.937, "step": 8365 }, { "epoch": 3.790666062528319, "grad_norm": 0.11585190072735878, "learning_rate": 1.343031630561503e-05, "loss": 0.9433, "step": 8366 }, { "epoch": 3.7911191662890804, "grad_norm": 0.10069802479237741, "learning_rate": 1.3420858786966914e-05, "loss": 0.9286, "step": 8367 }, { "epoch": 3.7915722700498415, "grad_norm": 0.10775594862811287, "learning_rate": 1.341140392815397e-05, "loss": 0.9455, "step": 8368 }, { "epoch": 3.7920253738106027, "grad_norm": 0.10165989580844749, "learning_rate": 1.3401951730122345e-05, "loss": 0.9383, "step": 8369 }, { "epoch": 3.792478477571364, "grad_norm": 0.10395025416340085, "learning_rate": 1.3392502193817962e-05, "loss": 0.9242, "step": 8370 }, { "epoch": 3.792931581332125, "grad_norm": 0.1086771852611872, "learning_rate": 1.3383055320186444e-05, "loss": 0.9358, "step": 8371 }, { "epoch": 3.793384685092886, "grad_norm": 0.10133447478848569, "learning_rate": 1.3373611110173159e-05, "loss": 0.9252, "step": 8372 }, { "epoch": 3.7938377888536476, "grad_norm": 0.10585459851690912, "learning_rate": 1.3364169564723222e-05, "loss": 0.9312, "step": 8373 }, { "epoch": 3.7942908926144088, "grad_norm": 0.10297829726448657, "learning_rate": 1.3354730684781453e-05, "loss": 0.9202, "step": 8374 }, { "epoch": 3.79474399637517, "grad_norm": 0.1040716509014085, "learning_rate": 1.3345294471292452e-05, "loss": 0.9155, "step": 8375 }, { "epoch": 3.795197100135931, "grad_norm": 0.10112132105030054, "learning_rate": 1.3335860925200486e-05, "loss": 0.9349, "step": 8376 }, { "epoch": 3.7956502038966926, "grad_norm": 0.12155223741726459, "learning_rate": 1.3326430047449619e-05, "loss": 0.908, "step": 8377 }, { "epoch": 3.7961033076574537, "grad_norm": 0.10821667823956077, "learning_rate": 1.3317001838983599e-05, "loss": 0.9157, "step": 8378 }, { "epoch": 3.796556411418215, "grad_norm": 0.107257119643552, "learning_rate": 1.3307576300745955e-05, "loss": 0.9138, "step": 8379 }, { "epoch": 3.797009515178976, "grad_norm": 0.12832103238227277, "learning_rate": 1.3298153433679906e-05, "loss": 0.933, "step": 8380 }, { "epoch": 3.797462618939737, "grad_norm": 0.09280742483742177, "learning_rate": 1.3288733238728409e-05, "loss": 0.9301, "step": 8381 }, { "epoch": 3.7979157227004983, "grad_norm": 0.13470812607925206, "learning_rate": 1.3279315716834198e-05, "loss": 0.921, "step": 8382 }, { "epoch": 3.7983688264612594, "grad_norm": 0.09896769817386308, "learning_rate": 1.3269900868939662e-05, "loss": 0.9108, "step": 8383 }, { "epoch": 3.798821930222021, "grad_norm": 0.1270118811398199, "learning_rate": 1.3260488695987e-05, "loss": 0.922, "step": 8384 }, { "epoch": 3.799275033982782, "grad_norm": 0.1224835149854089, "learning_rate": 1.3251079198918081e-05, "loss": 0.9303, "step": 8385 }, { "epoch": 3.7997281377435432, "grad_norm": 0.09604309722108106, "learning_rate": 1.3241672378674565e-05, "loss": 0.9069, "step": 8386 }, { "epoch": 3.8001812415043044, "grad_norm": 0.12470354029501282, "learning_rate": 1.3232268236197791e-05, "loss": 0.9429, "step": 8387 }, { "epoch": 3.800634345265066, "grad_norm": 0.11103692043731325, "learning_rate": 1.3222866772428872e-05, "loss": 0.9267, "step": 8388 }, { "epoch": 3.801087449025827, "grad_norm": 0.12881899632296331, "learning_rate": 1.3213467988308613e-05, "loss": 0.9387, "step": 8389 }, { "epoch": 3.801540552786588, "grad_norm": 0.11964710028648495, "learning_rate": 1.3204071884777591e-05, "loss": 0.9475, "step": 8390 }, { "epoch": 3.8019936565473493, "grad_norm": 0.1302680362704043, "learning_rate": 1.3194678462776094e-05, "loss": 0.9339, "step": 8391 }, { "epoch": 3.8024467603081105, "grad_norm": 0.12398912042005081, "learning_rate": 1.3185287723244122e-05, "loss": 0.9257, "step": 8392 }, { "epoch": 3.8028998640688716, "grad_norm": 0.1263021380449312, "learning_rate": 1.317589966712145e-05, "loss": 0.9361, "step": 8393 }, { "epoch": 3.8033529678296327, "grad_norm": 0.13963463151529734, "learning_rate": 1.3166514295347548e-05, "loss": 0.9298, "step": 8394 }, { "epoch": 3.8038060715903943, "grad_norm": 0.11216077476167376, "learning_rate": 1.3157131608861646e-05, "loss": 0.9207, "step": 8395 }, { "epoch": 3.8042591753511554, "grad_norm": 0.13525097446523904, "learning_rate": 1.314775160860267e-05, "loss": 0.9272, "step": 8396 }, { "epoch": 3.8047122791119166, "grad_norm": 0.1098525586453982, "learning_rate": 1.3138374295509326e-05, "loss": 0.9462, "step": 8397 }, { "epoch": 3.8051653828726777, "grad_norm": 0.12932724108805804, "learning_rate": 1.3128999670520007e-05, "loss": 0.9497, "step": 8398 }, { "epoch": 3.8056184866334393, "grad_norm": 0.12527521413486187, "learning_rate": 1.311962773457284e-05, "loss": 0.9408, "step": 8399 }, { "epoch": 3.8060715903942004, "grad_norm": 0.11474398322251927, "learning_rate": 1.311025848860572e-05, "loss": 0.9204, "step": 8400 }, { "epoch": 3.8065246941549615, "grad_norm": 0.12554774647749303, "learning_rate": 1.3100891933556228e-05, "loss": 0.9197, "step": 8401 }, { "epoch": 3.8069777979157227, "grad_norm": 0.12631323509671466, "learning_rate": 1.3091528070361719e-05, "loss": 0.9137, "step": 8402 }, { "epoch": 3.807430901676484, "grad_norm": 0.10655197530316426, "learning_rate": 1.308216689995923e-05, "loss": 0.9212, "step": 8403 }, { "epoch": 3.807884005437245, "grad_norm": 0.137768181440768, "learning_rate": 1.3072808423285586e-05, "loss": 0.9455, "step": 8404 }, { "epoch": 3.8083371091980065, "grad_norm": 0.09935004686421212, "learning_rate": 1.3063452641277277e-05, "loss": 0.9163, "step": 8405 }, { "epoch": 3.8087902129587676, "grad_norm": 0.12444871543481213, "learning_rate": 1.3054099554870585e-05, "loss": 0.9346, "step": 8406 }, { "epoch": 3.8092433167195288, "grad_norm": 0.10845586085267657, "learning_rate": 1.3044749165001488e-05, "loss": 0.9516, "step": 8407 }, { "epoch": 3.80969642048029, "grad_norm": 0.12592063741374185, "learning_rate": 1.3035401472605682e-05, "loss": 0.9518, "step": 8408 }, { "epoch": 3.8101495242410515, "grad_norm": 0.11393979665740998, "learning_rate": 1.3026056478618637e-05, "loss": 0.9426, "step": 8409 }, { "epoch": 3.8106026280018126, "grad_norm": 0.1176257893462647, "learning_rate": 1.3016714183975507e-05, "loss": 0.8911, "step": 8410 }, { "epoch": 3.8110557317625737, "grad_norm": 0.11469592164528504, "learning_rate": 1.300737458961122e-05, "loss": 0.9386, "step": 8411 }, { "epoch": 3.811508835523335, "grad_norm": 0.1405643267197601, "learning_rate": 1.2998037696460379e-05, "loss": 0.9239, "step": 8412 }, { "epoch": 3.811961939284096, "grad_norm": 0.09214945705447264, "learning_rate": 1.2988703505457387e-05, "loss": 0.933, "step": 8413 }, { "epoch": 3.812415043044857, "grad_norm": 0.1495775179395374, "learning_rate": 1.29793720175363e-05, "loss": 0.932, "step": 8414 }, { "epoch": 3.8128681468056183, "grad_norm": 0.1037648002224768, "learning_rate": 1.2970043233630971e-05, "loss": 0.9256, "step": 8415 }, { "epoch": 3.81332125056638, "grad_norm": 0.13045843498819482, "learning_rate": 1.2960717154674943e-05, "loss": 0.9563, "step": 8416 }, { "epoch": 3.813774354327141, "grad_norm": 0.10344075268671372, "learning_rate": 1.295139378160148e-05, "loss": 0.9317, "step": 8417 }, { "epoch": 3.814227458087902, "grad_norm": 0.13220586944254448, "learning_rate": 1.2942073115343621e-05, "loss": 0.9323, "step": 8418 }, { "epoch": 3.814680561848663, "grad_norm": 0.11027743458777507, "learning_rate": 1.293275515683408e-05, "loss": 0.9397, "step": 8419 }, { "epoch": 3.815133665609425, "grad_norm": 0.12082285293191829, "learning_rate": 1.2923439907005352e-05, "loss": 0.9425, "step": 8420 }, { "epoch": 3.815586769370186, "grad_norm": 0.1256082761830293, "learning_rate": 1.2914127366789613e-05, "loss": 0.9165, "step": 8421 }, { "epoch": 3.816039873130947, "grad_norm": 0.09384835519005764, "learning_rate": 1.2904817537118812e-05, "loss": 0.9508, "step": 8422 }, { "epoch": 3.816492976891708, "grad_norm": 0.12818307824906497, "learning_rate": 1.2895510418924592e-05, "loss": 0.9097, "step": 8423 }, { "epoch": 3.8169460806524693, "grad_norm": 0.09861669812007895, "learning_rate": 1.2886206013138325e-05, "loss": 0.9509, "step": 8424 }, { "epoch": 3.8173991844132305, "grad_norm": 0.11506653778470757, "learning_rate": 1.2876904320691153e-05, "loss": 0.9283, "step": 8425 }, { "epoch": 3.8178522881739916, "grad_norm": 0.11061156186099058, "learning_rate": 1.2867605342513887e-05, "loss": 0.9026, "step": 8426 }, { "epoch": 3.818305391934753, "grad_norm": 0.10989480446683393, "learning_rate": 1.2858309079537125e-05, "loss": 0.929, "step": 8427 }, { "epoch": 3.8187584956955143, "grad_norm": 0.12802090697190044, "learning_rate": 1.2849015532691143e-05, "loss": 0.933, "step": 8428 }, { "epoch": 3.8192115994562754, "grad_norm": 0.1183708700598628, "learning_rate": 1.2839724702905984e-05, "loss": 0.9528, "step": 8429 }, { "epoch": 3.8196647032170365, "grad_norm": 0.1323119698512776, "learning_rate": 1.2830436591111389e-05, "loss": 0.9375, "step": 8430 }, { "epoch": 3.820117806977798, "grad_norm": 0.11936271278929095, "learning_rate": 1.2821151198236855e-05, "loss": 0.9456, "step": 8431 }, { "epoch": 3.8205709107385593, "grad_norm": 0.09990053217010256, "learning_rate": 1.2811868525211582e-05, "loss": 0.9318, "step": 8432 }, { "epoch": 3.8210240144993204, "grad_norm": 0.11711209157579948, "learning_rate": 1.2802588572964499e-05, "loss": 0.923, "step": 8433 }, { "epoch": 3.8214771182600815, "grad_norm": 0.10400943847685472, "learning_rate": 1.2793311342424297e-05, "loss": 0.9338, "step": 8434 }, { "epoch": 3.8219302220208426, "grad_norm": 0.11131930563396975, "learning_rate": 1.2784036834519342e-05, "loss": 0.9361, "step": 8435 }, { "epoch": 3.822383325781604, "grad_norm": 0.1083927876402194, "learning_rate": 1.277476505017778e-05, "loss": 0.9233, "step": 8436 }, { "epoch": 3.8228364295423654, "grad_norm": 0.11284571919398911, "learning_rate": 1.2765495990327437e-05, "loss": 0.9483, "step": 8437 }, { "epoch": 3.8232895333031265, "grad_norm": 0.10978206784986341, "learning_rate": 1.2756229655895913e-05, "loss": 0.9286, "step": 8438 }, { "epoch": 3.8237426370638876, "grad_norm": 0.10982994103507746, "learning_rate": 1.2746966047810499e-05, "loss": 0.9431, "step": 8439 }, { "epoch": 3.8241957408246487, "grad_norm": 0.10446042483559288, "learning_rate": 1.2737705166998224e-05, "loss": 0.9144, "step": 8440 }, { "epoch": 3.8246488445854103, "grad_norm": 0.10774806887571721, "learning_rate": 1.2728447014385833e-05, "loss": 0.931, "step": 8441 }, { "epoch": 3.8251019483461715, "grad_norm": 0.10890578509405362, "learning_rate": 1.2719191590899839e-05, "loss": 0.9326, "step": 8442 }, { "epoch": 3.8255550521069326, "grad_norm": 0.09956011876284793, "learning_rate": 1.2709938897466438e-05, "loss": 0.9203, "step": 8443 }, { "epoch": 3.8260081558676937, "grad_norm": 0.10839449935968551, "learning_rate": 1.2700688935011551e-05, "loss": 0.9305, "step": 8444 }, { "epoch": 3.826461259628455, "grad_norm": 0.09938775555546657, "learning_rate": 1.2691441704460879e-05, "loss": 0.9274, "step": 8445 }, { "epoch": 3.826914363389216, "grad_norm": 0.10621976202073347, "learning_rate": 1.2682197206739777e-05, "loss": 0.9324, "step": 8446 }, { "epoch": 3.827367467149977, "grad_norm": 0.09152234337314687, "learning_rate": 1.267295544277339e-05, "loss": 0.9493, "step": 8447 }, { "epoch": 3.8278205709107387, "grad_norm": 0.11059488612237182, "learning_rate": 1.2663716413486547e-05, "loss": 0.936, "step": 8448 }, { "epoch": 3.8282736746715, "grad_norm": 0.11562040272558341, "learning_rate": 1.2654480119803831e-05, "loss": 0.9185, "step": 8449 }, { "epoch": 3.828726778432261, "grad_norm": 0.1067171360805408, "learning_rate": 1.2645246562649524e-05, "loss": 0.9405, "step": 8450 }, { "epoch": 3.829179882193022, "grad_norm": 0.12065853870220004, "learning_rate": 1.2636015742947665e-05, "loss": 0.9315, "step": 8451 }, { "epoch": 3.8296329859537837, "grad_norm": 0.11705130223450827, "learning_rate": 1.2626787661621997e-05, "loss": 0.9363, "step": 8452 }, { "epoch": 3.830086089714545, "grad_norm": 0.13158921238613097, "learning_rate": 1.2617562319595984e-05, "loss": 0.8999, "step": 8453 }, { "epoch": 3.830539193475306, "grad_norm": 0.12304282425027464, "learning_rate": 1.2608339717792846e-05, "loss": 0.9132, "step": 8454 }, { "epoch": 3.830992297236067, "grad_norm": 0.11559837028631334, "learning_rate": 1.2599119857135489e-05, "loss": 0.93, "step": 8455 }, { "epoch": 3.831445400996828, "grad_norm": 0.12043689237542637, "learning_rate": 1.2589902738546589e-05, "loss": 0.9332, "step": 8456 }, { "epoch": 3.8318985047575893, "grad_norm": 0.11047756923563005, "learning_rate": 1.2580688362948497e-05, "loss": 0.9623, "step": 8457 }, { "epoch": 3.8323516085183504, "grad_norm": 0.12673091599712086, "learning_rate": 1.2571476731263346e-05, "loss": 0.9414, "step": 8458 }, { "epoch": 3.832804712279112, "grad_norm": 0.09931079214292399, "learning_rate": 1.256226784441295e-05, "loss": 0.9529, "step": 8459 }, { "epoch": 3.833257816039873, "grad_norm": 0.11313268934108182, "learning_rate": 1.2553061703318851e-05, "loss": 0.9195, "step": 8460 }, { "epoch": 3.8337109198006343, "grad_norm": 0.11827431155888017, "learning_rate": 1.2543858308902355e-05, "loss": 0.9428, "step": 8461 }, { "epoch": 3.8341640235613954, "grad_norm": 0.10820606635563577, "learning_rate": 1.2534657662084438e-05, "loss": 0.9327, "step": 8462 }, { "epoch": 3.834617127322157, "grad_norm": 0.12213701621388544, "learning_rate": 1.2525459763785851e-05, "loss": 0.9223, "step": 8463 }, { "epoch": 3.835070231082918, "grad_norm": 0.1087217069314131, "learning_rate": 1.2516264614927037e-05, "loss": 0.9274, "step": 8464 }, { "epoch": 3.8355233348436792, "grad_norm": 0.11315776195250647, "learning_rate": 1.2507072216428191e-05, "loss": 0.9495, "step": 8465 }, { "epoch": 3.8359764386044404, "grad_norm": 0.11189236106342774, "learning_rate": 1.2497882569209195e-05, "loss": 0.9407, "step": 8466 }, { "epoch": 3.8364295423652015, "grad_norm": 0.11369908589781866, "learning_rate": 1.2488695674189702e-05, "loss": 0.9271, "step": 8467 }, { "epoch": 3.8368826461259626, "grad_norm": 0.112689925024548, "learning_rate": 1.2479511532289048e-05, "loss": 0.9454, "step": 8468 }, { "epoch": 3.837335749886724, "grad_norm": 0.10432820529287598, "learning_rate": 1.2470330144426307e-05, "loss": 0.937, "step": 8469 }, { "epoch": 3.8377888536474853, "grad_norm": 0.10888524739350534, "learning_rate": 1.2461151511520306e-05, "loss": 0.9492, "step": 8470 }, { "epoch": 3.8382419574082465, "grad_norm": 0.21781218973962577, "learning_rate": 1.245197563448954e-05, "loss": 0.9498, "step": 8471 }, { "epoch": 3.8386950611690076, "grad_norm": 0.11340087978117601, "learning_rate": 1.244280251425229e-05, "loss": 0.9207, "step": 8472 }, { "epoch": 3.839148164929769, "grad_norm": 0.10609803589602695, "learning_rate": 1.24336321517265e-05, "loss": 0.9229, "step": 8473 }, { "epoch": 3.8396012686905303, "grad_norm": 0.09722995585124043, "learning_rate": 1.2424464547829901e-05, "loss": 0.9259, "step": 8474 }, { "epoch": 3.8400543724512914, "grad_norm": 0.10700865581271096, "learning_rate": 1.241529970347989e-05, "loss": 0.9264, "step": 8475 }, { "epoch": 3.8405074762120526, "grad_norm": 0.10546305372349314, "learning_rate": 1.240613761959364e-05, "loss": 0.9527, "step": 8476 }, { "epoch": 3.8409605799728137, "grad_norm": 0.11552840591876107, "learning_rate": 1.2396978297088e-05, "loss": 0.9033, "step": 8477 }, { "epoch": 3.841413683733575, "grad_norm": 0.0892815201347556, "learning_rate": 1.2387821736879561e-05, "loss": 0.9141, "step": 8478 }, { "epoch": 3.841866787494336, "grad_norm": 0.11533637598053964, "learning_rate": 1.2378667939884665e-05, "loss": 0.9313, "step": 8479 }, { "epoch": 3.8423198912550975, "grad_norm": 0.09045442524748037, "learning_rate": 1.236951690701932e-05, "loss": 0.9193, "step": 8480 }, { "epoch": 3.8427729950158587, "grad_norm": 0.10697869096793056, "learning_rate": 1.236036863919933e-05, "loss": 0.9375, "step": 8481 }, { "epoch": 3.84322609877662, "grad_norm": 0.0951053149649538, "learning_rate": 1.2351223137340146e-05, "loss": 0.9435, "step": 8482 }, { "epoch": 3.843679202537381, "grad_norm": 0.09086870137667695, "learning_rate": 1.2342080402357008e-05, "loss": 0.9138, "step": 8483 }, { "epoch": 3.8441323062981425, "grad_norm": 0.09487892877281262, "learning_rate": 1.2332940435164838e-05, "loss": 0.9334, "step": 8484 }, { "epoch": 3.8445854100589036, "grad_norm": 0.10617717545203945, "learning_rate": 1.2323803236678282e-05, "loss": 0.938, "step": 8485 }, { "epoch": 3.8450385138196648, "grad_norm": 0.09682906217791584, "learning_rate": 1.231466880781174e-05, "loss": 0.9008, "step": 8486 }, { "epoch": 3.845491617580426, "grad_norm": 0.10324437208823335, "learning_rate": 1.2305537149479298e-05, "loss": 0.9263, "step": 8487 }, { "epoch": 3.845944721341187, "grad_norm": 0.09704566842092674, "learning_rate": 1.2296408262594803e-05, "loss": 0.9063, "step": 8488 }, { "epoch": 3.846397825101948, "grad_norm": 0.10487471450649442, "learning_rate": 1.2287282148071777e-05, "loss": 0.9203, "step": 8489 }, { "epoch": 3.8468509288627093, "grad_norm": 0.0985798742858029, "learning_rate": 1.2278158806823516e-05, "loss": 0.9044, "step": 8490 }, { "epoch": 3.847304032623471, "grad_norm": 0.09982362393646162, "learning_rate": 1.2269038239762994e-05, "loss": 0.9404, "step": 8491 }, { "epoch": 3.847757136384232, "grad_norm": 0.10618914813534383, "learning_rate": 1.2259920447802943e-05, "loss": 0.9319, "step": 8492 }, { "epoch": 3.848210240144993, "grad_norm": 0.10053860086731269, "learning_rate": 1.22508054318558e-05, "loss": 0.8985, "step": 8493 }, { "epoch": 3.8486633439057543, "grad_norm": 0.10656876029976511, "learning_rate": 1.2241693192833699e-05, "loss": 0.9335, "step": 8494 }, { "epoch": 3.849116447666516, "grad_norm": 0.10389487697316405, "learning_rate": 1.223258373164856e-05, "loss": 0.9331, "step": 8495 }, { "epoch": 3.849569551427277, "grad_norm": 0.10031348875762754, "learning_rate": 1.2223477049211972e-05, "loss": 0.9221, "step": 8496 }, { "epoch": 3.850022655188038, "grad_norm": 0.0939492286297308, "learning_rate": 1.2214373146435255e-05, "loss": 0.9268, "step": 8497 }, { "epoch": 3.8504757589487992, "grad_norm": 0.12036221349393836, "learning_rate": 1.2205272024229449e-05, "loss": 0.9425, "step": 8498 }, { "epoch": 3.8509288627095604, "grad_norm": 0.09681721198912063, "learning_rate": 1.219617368350535e-05, "loss": 0.9311, "step": 8499 }, { "epoch": 3.8513819664703215, "grad_norm": 0.10236955352307804, "learning_rate": 1.2187078125173427e-05, "loss": 0.9314, "step": 8500 }, { "epoch": 3.851835070231083, "grad_norm": 0.09534707920520666, "learning_rate": 1.2177985350143913e-05, "loss": 0.9431, "step": 8501 }, { "epoch": 3.852288173991844, "grad_norm": 0.09637935043926558, "learning_rate": 1.2168895359326719e-05, "loss": 0.9329, "step": 8502 }, { "epoch": 3.8527412777526053, "grad_norm": 0.10496821585625928, "learning_rate": 1.2159808153631528e-05, "loss": 0.9468, "step": 8503 }, { "epoch": 3.8531943815133665, "grad_norm": 0.09404955237598893, "learning_rate": 1.2150723733967701e-05, "loss": 0.9312, "step": 8504 }, { "epoch": 3.853647485274128, "grad_norm": 0.0989815211721202, "learning_rate": 1.214164210124433e-05, "loss": 0.9256, "step": 8505 }, { "epoch": 3.854100589034889, "grad_norm": 0.09566546450055335, "learning_rate": 1.2132563256370253e-05, "loss": 0.9312, "step": 8506 }, { "epoch": 3.8545536927956503, "grad_norm": 0.09607368888865482, "learning_rate": 1.2123487200253994e-05, "loss": 0.9264, "step": 8507 }, { "epoch": 3.8550067965564114, "grad_norm": 0.10710696693717307, "learning_rate": 1.2114413933803828e-05, "loss": 0.9347, "step": 8508 }, { "epoch": 3.8554599003171726, "grad_norm": 0.09920217474196305, "learning_rate": 1.2105343457927723e-05, "loss": 0.9437, "step": 8509 }, { "epoch": 3.8559130040779337, "grad_norm": 0.10938820458180823, "learning_rate": 1.2096275773533406e-05, "loss": 0.9275, "step": 8510 }, { "epoch": 3.856366107838695, "grad_norm": 0.09329862716703793, "learning_rate": 1.208721088152827e-05, "loss": 0.9228, "step": 8511 }, { "epoch": 3.8568192115994564, "grad_norm": 0.09515623846153416, "learning_rate": 1.207814878281949e-05, "loss": 0.9408, "step": 8512 }, { "epoch": 3.8572723153602175, "grad_norm": 0.11078649372322519, "learning_rate": 1.206908947831391e-05, "loss": 0.9335, "step": 8513 }, { "epoch": 3.8577254191209787, "grad_norm": 0.09649026959718916, "learning_rate": 1.2060032968918116e-05, "loss": 0.9182, "step": 8514 }, { "epoch": 3.85817852288174, "grad_norm": 0.09822273372439898, "learning_rate": 1.2050979255538428e-05, "loss": 0.9056, "step": 8515 }, { "epoch": 3.8586316266425014, "grad_norm": 0.09174025444480403, "learning_rate": 1.2041928339080849e-05, "loss": 0.9376, "step": 8516 }, { "epoch": 3.8590847304032625, "grad_norm": 0.10439362074837479, "learning_rate": 1.2032880220451148e-05, "loss": 0.9373, "step": 8517 }, { "epoch": 3.8595378341640236, "grad_norm": 0.10080844857201846, "learning_rate": 1.2023834900554773e-05, "loss": 0.9264, "step": 8518 }, { "epoch": 3.8599909379247848, "grad_norm": 0.10923865037578703, "learning_rate": 1.201479238029693e-05, "loss": 0.9416, "step": 8519 }, { "epoch": 3.860444041685546, "grad_norm": 0.10754243271473775, "learning_rate": 1.2005752660582508e-05, "loss": 0.9366, "step": 8520 }, { "epoch": 3.860897145446307, "grad_norm": 0.09464024668252458, "learning_rate": 1.199671574231613e-05, "loss": 0.929, "step": 8521 }, { "epoch": 3.861350249207068, "grad_norm": 0.11070147078940912, "learning_rate": 1.1987681626402154e-05, "loss": 0.9359, "step": 8522 }, { "epoch": 3.8618033529678297, "grad_norm": 0.09992481775163778, "learning_rate": 1.1978650313744628e-05, "loss": 0.9407, "step": 8523 }, { "epoch": 3.862256456728591, "grad_norm": 0.1074346366124702, "learning_rate": 1.1969621805247362e-05, "loss": 0.9426, "step": 8524 }, { "epoch": 3.862709560489352, "grad_norm": 0.09816998110456848, "learning_rate": 1.1960596101813829e-05, "loss": 0.9065, "step": 8525 }, { "epoch": 3.863162664250113, "grad_norm": 0.10128917728622494, "learning_rate": 1.1951573204347278e-05, "loss": 0.9157, "step": 8526 }, { "epoch": 3.8636157680108747, "grad_norm": 0.09611066536629602, "learning_rate": 1.1942553113750624e-05, "loss": 0.9498, "step": 8527 }, { "epoch": 3.864068871771636, "grad_norm": 0.09577364714289709, "learning_rate": 1.1933535830926557e-05, "loss": 0.9197, "step": 8528 }, { "epoch": 3.864521975532397, "grad_norm": 0.10411989606443728, "learning_rate": 1.1924521356777445e-05, "loss": 0.92, "step": 8529 }, { "epoch": 3.864975079293158, "grad_norm": 0.09079896434664188, "learning_rate": 1.1915509692205371e-05, "loss": 0.9297, "step": 8530 }, { "epoch": 3.865428183053919, "grad_norm": 0.0995871290116558, "learning_rate": 1.1906500838112183e-05, "loss": 0.9323, "step": 8531 }, { "epoch": 3.8658812868146804, "grad_norm": 0.09392387431981648, "learning_rate": 1.1897494795399385e-05, "loss": 0.903, "step": 8532 }, { "epoch": 3.866334390575442, "grad_norm": 0.09683928753395271, "learning_rate": 1.188849156496826e-05, "loss": 0.9305, "step": 8533 }, { "epoch": 3.866787494336203, "grad_norm": 0.09702066181211176, "learning_rate": 1.1879491147719762e-05, "loss": 0.9214, "step": 8534 }, { "epoch": 3.867240598096964, "grad_norm": 0.09011847683682234, "learning_rate": 1.1870493544554602e-05, "loss": 0.9225, "step": 8535 }, { "epoch": 3.8676937018577253, "grad_norm": 0.09603586337248841, "learning_rate": 1.1861498756373164e-05, "loss": 0.9412, "step": 8536 }, { "epoch": 3.868146805618487, "grad_norm": 0.08982495079154094, "learning_rate": 1.1852506784075613e-05, "loss": 0.9123, "step": 8537 }, { "epoch": 3.868599909379248, "grad_norm": 0.10525837999272687, "learning_rate": 1.1843517628561769e-05, "loss": 0.9422, "step": 8538 }, { "epoch": 3.869053013140009, "grad_norm": 0.09928549752848959, "learning_rate": 1.1834531290731194e-05, "loss": 0.9384, "step": 8539 }, { "epoch": 3.8695061169007703, "grad_norm": 0.10273974472197572, "learning_rate": 1.1825547771483196e-05, "loss": 0.9286, "step": 8540 }, { "epoch": 3.8699592206615314, "grad_norm": 0.11914717034400824, "learning_rate": 1.1816567071716754e-05, "loss": 0.9277, "step": 8541 }, { "epoch": 3.8704123244222925, "grad_norm": 0.10990403600282071, "learning_rate": 1.18075891923306e-05, "loss": 0.9107, "step": 8542 }, { "epoch": 3.8708654281830537, "grad_norm": 0.12242576153869578, "learning_rate": 1.1798614134223154e-05, "loss": 0.9186, "step": 8543 }, { "epoch": 3.8713185319438153, "grad_norm": 0.10560602764044225, "learning_rate": 1.1789641898292601e-05, "loss": 0.9434, "step": 8544 }, { "epoch": 3.8717716357045764, "grad_norm": 0.11414703760030198, "learning_rate": 1.178067248543679e-05, "loss": 0.9414, "step": 8545 }, { "epoch": 3.8722247394653375, "grad_norm": 0.1214920127136553, "learning_rate": 1.1771705896553302e-05, "loss": 0.9218, "step": 8546 }, { "epoch": 3.8726778432260986, "grad_norm": 0.09912117271869059, "learning_rate": 1.1762742132539469e-05, "loss": 0.9373, "step": 8547 }, { "epoch": 3.8731309469868602, "grad_norm": 0.11433434231940504, "learning_rate": 1.1753781194292296e-05, "loss": 0.9211, "step": 8548 }, { "epoch": 3.8735840507476214, "grad_norm": 0.09896569481590088, "learning_rate": 1.1744823082708537e-05, "loss": 0.943, "step": 8549 }, { "epoch": 3.8740371545083825, "grad_norm": 0.10041908997005872, "learning_rate": 1.1735867798684639e-05, "loss": 0.9081, "step": 8550 }, { "epoch": 3.8744902582691436, "grad_norm": 0.09863685039323354, "learning_rate": 1.1726915343116793e-05, "loss": 0.8909, "step": 8551 }, { "epoch": 3.8749433620299047, "grad_norm": 0.09844057515486651, "learning_rate": 1.171796571690087e-05, "loss": 0.9296, "step": 8552 }, { "epoch": 3.875396465790666, "grad_norm": 0.11537459580564874, "learning_rate": 1.170901892093252e-05, "loss": 0.9269, "step": 8553 }, { "epoch": 3.875849569551427, "grad_norm": 0.09675345287717993, "learning_rate": 1.170007495610701e-05, "loss": 0.9336, "step": 8554 }, { "epoch": 3.8763026733121886, "grad_norm": 0.11197732674939885, "learning_rate": 1.1691133823319434e-05, "loss": 0.922, "step": 8555 }, { "epoch": 3.8767557770729497, "grad_norm": 0.09937848263744439, "learning_rate": 1.1682195523464514e-05, "loss": 0.9328, "step": 8556 }, { "epoch": 3.877208880833711, "grad_norm": 0.1108280688933509, "learning_rate": 1.167326005743676e-05, "loss": 0.9222, "step": 8557 }, { "epoch": 3.877661984594472, "grad_norm": 0.11514562701559634, "learning_rate": 1.1664327426130346e-05, "loss": 0.9508, "step": 8558 }, { "epoch": 3.8781150883552336, "grad_norm": 0.10256831569448678, "learning_rate": 1.1655397630439174e-05, "loss": 0.9212, "step": 8559 }, { "epoch": 3.8785681921159947, "grad_norm": 0.13004828667203067, "learning_rate": 1.1646470671256886e-05, "loss": 0.9275, "step": 8560 }, { "epoch": 3.879021295876756, "grad_norm": 0.09912350716555014, "learning_rate": 1.1637546549476806e-05, "loss": 0.9144, "step": 8561 }, { "epoch": 3.879474399637517, "grad_norm": 0.15162270428983118, "learning_rate": 1.1628625265992009e-05, "loss": 0.9242, "step": 8562 }, { "epoch": 3.879927503398278, "grad_norm": 0.09986293937042467, "learning_rate": 1.1619706821695247e-05, "loss": 0.934, "step": 8563 }, { "epoch": 3.880380607159039, "grad_norm": 0.15516574338633, "learning_rate": 1.1610791217479037e-05, "loss": 0.9474, "step": 8564 }, { "epoch": 3.880833710919801, "grad_norm": 0.12539735892253714, "learning_rate": 1.1601878454235566e-05, "loss": 0.9449, "step": 8565 }, { "epoch": 3.881286814680562, "grad_norm": 0.14075140454193494, "learning_rate": 1.1592968532856745e-05, "loss": 0.9261, "step": 8566 }, { "epoch": 3.881739918441323, "grad_norm": 0.1203421924305679, "learning_rate": 1.1584061454234235e-05, "loss": 0.9329, "step": 8567 }, { "epoch": 3.882193022202084, "grad_norm": 0.12199235269678274, "learning_rate": 1.157515721925936e-05, "loss": 0.9263, "step": 8568 }, { "epoch": 3.8826461259628458, "grad_norm": 0.10846530063561545, "learning_rate": 1.1566255828823216e-05, "loss": 0.9437, "step": 8569 }, { "epoch": 3.883099229723607, "grad_norm": 0.12845435414701742, "learning_rate": 1.1557357283816559e-05, "loss": 0.9359, "step": 8570 }, { "epoch": 3.883552333484368, "grad_norm": 0.11310025617718382, "learning_rate": 1.1548461585129909e-05, "loss": 0.9368, "step": 8571 }, { "epoch": 3.884005437245129, "grad_norm": 0.1245272742317992, "learning_rate": 1.1539568733653455e-05, "loss": 0.9536, "step": 8572 }, { "epoch": 3.8844585410058903, "grad_norm": 0.11385031715821829, "learning_rate": 1.1530678730277151e-05, "loss": 0.9292, "step": 8573 }, { "epoch": 3.8849116447666514, "grad_norm": 0.1226573414096497, "learning_rate": 1.152179157589063e-05, "loss": 0.9358, "step": 8574 }, { "epoch": 3.8853647485274125, "grad_norm": 0.10338962097563102, "learning_rate": 1.1512907271383234e-05, "loss": 0.9248, "step": 8575 }, { "epoch": 3.885817852288174, "grad_norm": 0.11281131359156084, "learning_rate": 1.1504025817644057e-05, "loss": 0.9277, "step": 8576 }, { "epoch": 3.8862709560489352, "grad_norm": 0.13230357693361805, "learning_rate": 1.1495147215561872e-05, "loss": 0.93, "step": 8577 }, { "epoch": 3.8867240598096964, "grad_norm": 0.09668688146800933, "learning_rate": 1.1486271466025194e-05, "loss": 0.9275, "step": 8578 }, { "epoch": 3.8871771635704575, "grad_norm": 0.13726587770239687, "learning_rate": 1.1477398569922223e-05, "loss": 0.9541, "step": 8579 }, { "epoch": 3.887630267331219, "grad_norm": 0.09033756837523958, "learning_rate": 1.1468528528140906e-05, "loss": 0.9206, "step": 8580 }, { "epoch": 3.88808337109198, "grad_norm": 0.1101361617666884, "learning_rate": 1.1459661341568883e-05, "loss": 0.9216, "step": 8581 }, { "epoch": 3.8885364748527413, "grad_norm": 0.10141290946516339, "learning_rate": 1.1450797011093501e-05, "loss": 0.9681, "step": 8582 }, { "epoch": 3.8889895786135025, "grad_norm": 0.11966230587879129, "learning_rate": 1.1441935537601858e-05, "loss": 0.9265, "step": 8583 }, { "epoch": 3.8894426823742636, "grad_norm": 0.10029540836300188, "learning_rate": 1.1433076921980715e-05, "loss": 0.9279, "step": 8584 }, { "epoch": 3.8898957861350247, "grad_norm": 0.10067635397497239, "learning_rate": 1.1424221165116602e-05, "loss": 0.9291, "step": 8585 }, { "epoch": 3.890348889895786, "grad_norm": 0.09379274739331682, "learning_rate": 1.1415368267895706e-05, "loss": 0.8944, "step": 8586 }, { "epoch": 3.8908019936565474, "grad_norm": 0.11015824981509852, "learning_rate": 1.1406518231203983e-05, "loss": 0.9247, "step": 8587 }, { "epoch": 3.8912550974173086, "grad_norm": 0.11619601001643723, "learning_rate": 1.139767105592705e-05, "loss": 0.9291, "step": 8588 }, { "epoch": 3.8917082011780697, "grad_norm": 0.11928290682985243, "learning_rate": 1.138882674295029e-05, "loss": 0.9491, "step": 8589 }, { "epoch": 3.892161304938831, "grad_norm": 0.10301419691294139, "learning_rate": 1.1379985293158762e-05, "loss": 0.9277, "step": 8590 }, { "epoch": 3.8926144086995924, "grad_norm": 0.11486792733887423, "learning_rate": 1.1371146707437238e-05, "loss": 0.9466, "step": 8591 }, { "epoch": 3.8930675124603535, "grad_norm": 0.10010716987969782, "learning_rate": 1.136231098667024e-05, "loss": 0.9272, "step": 8592 }, { "epoch": 3.8935206162211147, "grad_norm": 0.11313524017271187, "learning_rate": 1.135347813174195e-05, "loss": 0.925, "step": 8593 }, { "epoch": 3.893973719981876, "grad_norm": 0.10823526887845673, "learning_rate": 1.134464814353632e-05, "loss": 0.9491, "step": 8594 }, { "epoch": 3.894426823742637, "grad_norm": 0.11545001050372877, "learning_rate": 1.1335821022936964e-05, "loss": 0.9182, "step": 8595 }, { "epoch": 3.894879927503398, "grad_norm": 0.11446884134036718, "learning_rate": 1.1326996770827257e-05, "loss": 0.9333, "step": 8596 }, { "epoch": 3.8953330312641596, "grad_norm": 0.11639965417340363, "learning_rate": 1.1318175388090231e-05, "loss": 0.9297, "step": 8597 }, { "epoch": 3.8957861350249208, "grad_norm": 0.10128505742010199, "learning_rate": 1.1309356875608693e-05, "loss": 0.9439, "step": 8598 }, { "epoch": 3.896239238785682, "grad_norm": 0.10485076951155518, "learning_rate": 1.1300541234265112e-05, "loss": 0.9354, "step": 8599 }, { "epoch": 3.896692342546443, "grad_norm": 0.10660298514308743, "learning_rate": 1.1291728464941688e-05, "loss": 0.9408, "step": 8600 }, { "epoch": 3.8971454463072046, "grad_norm": 0.10403740988270538, "learning_rate": 1.1282918568520351e-05, "loss": 0.9126, "step": 8601 }, { "epoch": 3.8975985500679657, "grad_norm": 0.09632164307720928, "learning_rate": 1.1274111545882702e-05, "loss": 0.9129, "step": 8602 }, { "epoch": 3.898051653828727, "grad_norm": 0.10994372385131214, "learning_rate": 1.1265307397910115e-05, "loss": 0.9237, "step": 8603 }, { "epoch": 3.898504757589488, "grad_norm": 0.10006142316781025, "learning_rate": 1.1256506125483604e-05, "loss": 0.9179, "step": 8604 }, { "epoch": 3.898957861350249, "grad_norm": 0.11084770994562153, "learning_rate": 1.1247707729483963e-05, "loss": 0.8984, "step": 8605 }, { "epoch": 3.8994109651110103, "grad_norm": 0.09775947339455798, "learning_rate": 1.1238912210791653e-05, "loss": 0.9056, "step": 8606 }, { "epoch": 3.8998640688717714, "grad_norm": 0.12600978982812136, "learning_rate": 1.1230119570286853e-05, "loss": 0.9398, "step": 8607 }, { "epoch": 3.900317172632533, "grad_norm": 0.09793615491371524, "learning_rate": 1.1221329808849481e-05, "loss": 0.933, "step": 8608 }, { "epoch": 3.900770276393294, "grad_norm": 0.12559801541126903, "learning_rate": 1.121254292735913e-05, "loss": 0.9211, "step": 8609 }, { "epoch": 3.9012233801540552, "grad_norm": 0.10603193532839943, "learning_rate": 1.1203758926695159e-05, "loss": 0.9362, "step": 8610 }, { "epoch": 3.9016764839148164, "grad_norm": 0.10475874748409199, "learning_rate": 1.1194977807736552e-05, "loss": 0.9046, "step": 8611 }, { "epoch": 3.902129587675578, "grad_norm": 0.09561267138485706, "learning_rate": 1.1186199571362089e-05, "loss": 0.9352, "step": 8612 }, { "epoch": 3.902582691436339, "grad_norm": 0.10933479699330158, "learning_rate": 1.117742421845021e-05, "loss": 0.9097, "step": 8613 }, { "epoch": 3.9030357951971, "grad_norm": 0.10567472824464673, "learning_rate": 1.1168651749879106e-05, "loss": 0.9363, "step": 8614 }, { "epoch": 3.9034888989578613, "grad_norm": 0.11238023648173813, "learning_rate": 1.1159882166526637e-05, "loss": 0.9251, "step": 8615 }, { "epoch": 3.9039420027186225, "grad_norm": 0.098132062608964, "learning_rate": 1.1151115469270412e-05, "loss": 0.9557, "step": 8616 }, { "epoch": 3.9043951064793836, "grad_norm": 0.12024841734714478, "learning_rate": 1.114235165898773e-05, "loss": 0.9165, "step": 8617 }, { "epoch": 3.9048482102401447, "grad_norm": 0.09721805996622564, "learning_rate": 1.1133590736555586e-05, "loss": 0.9243, "step": 8618 }, { "epoch": 3.9053013140009063, "grad_norm": 0.10913707282423528, "learning_rate": 1.1124832702850736e-05, "loss": 0.9093, "step": 8619 }, { "epoch": 3.9057544177616674, "grad_norm": 0.09538851882017285, "learning_rate": 1.1116077558749589e-05, "loss": 0.9079, "step": 8620 }, { "epoch": 3.9062075215224286, "grad_norm": 0.10188286574081858, "learning_rate": 1.110732530512832e-05, "loss": 0.9213, "step": 8621 }, { "epoch": 3.9066606252831897, "grad_norm": 0.10336933390434311, "learning_rate": 1.1098575942862757e-05, "loss": 0.9355, "step": 8622 }, { "epoch": 3.9071137290439513, "grad_norm": 0.10066633648524616, "learning_rate": 1.1089829472828497e-05, "loss": 0.9309, "step": 8623 }, { "epoch": 3.9075668328047124, "grad_norm": 0.1112880901701866, "learning_rate": 1.1081085895900792e-05, "loss": 0.9407, "step": 8624 }, { "epoch": 3.9080199365654735, "grad_norm": 0.09872109707590578, "learning_rate": 1.1072345212954661e-05, "loss": 0.9315, "step": 8625 }, { "epoch": 3.9084730403262347, "grad_norm": 0.11267007555200105, "learning_rate": 1.1063607424864787e-05, "loss": 0.9551, "step": 8626 }, { "epoch": 3.908926144086996, "grad_norm": 0.10613112481772198, "learning_rate": 1.1054872532505567e-05, "loss": 0.9207, "step": 8627 }, { "epoch": 3.909379247847757, "grad_norm": 0.1135199145518665, "learning_rate": 1.1046140536751148e-05, "loss": 0.9403, "step": 8628 }, { "epoch": 3.9098323516085185, "grad_norm": 0.10782980137888838, "learning_rate": 1.1037411438475344e-05, "loss": 0.9438, "step": 8629 }, { "epoch": 3.9102854553692796, "grad_norm": 0.45909642545360557, "learning_rate": 1.1028685238551704e-05, "loss": 0.9399, "step": 8630 }, { "epoch": 3.9107385591300408, "grad_norm": 0.11071661472251397, "learning_rate": 1.1019961937853468e-05, "loss": 0.9319, "step": 8631 }, { "epoch": 3.911191662890802, "grad_norm": 0.1264478198459965, "learning_rate": 1.1011241537253614e-05, "loss": 0.9356, "step": 8632 }, { "epoch": 3.9116447666515635, "grad_norm": 0.10200174899212089, "learning_rate": 1.1002524037624793e-05, "loss": 0.9053, "step": 8633 }, { "epoch": 3.9120978704123246, "grad_norm": 0.12444438803085138, "learning_rate": 1.0993809439839404e-05, "loss": 0.9188, "step": 8634 }, { "epoch": 3.9125509741730857, "grad_norm": 0.0952336568779052, "learning_rate": 1.0985097744769529e-05, "loss": 0.9233, "step": 8635 }, { "epoch": 3.913004077933847, "grad_norm": 0.11735763012092289, "learning_rate": 1.0976388953286952e-05, "loss": 0.929, "step": 8636 }, { "epoch": 3.913457181694608, "grad_norm": 0.09750035854441406, "learning_rate": 1.0967683066263208e-05, "loss": 0.932, "step": 8637 }, { "epoch": 3.913910285455369, "grad_norm": 0.1047199494536005, "learning_rate": 1.095898008456949e-05, "loss": 0.9258, "step": 8638 }, { "epoch": 3.9143633892161303, "grad_norm": 0.09626491739086321, "learning_rate": 1.0950280009076745e-05, "loss": 0.9332, "step": 8639 }, { "epoch": 3.914816492976892, "grad_norm": 0.10459914110191543, "learning_rate": 1.0941582840655593e-05, "loss": 0.9214, "step": 8640 }, { "epoch": 3.915269596737653, "grad_norm": 0.089730485415963, "learning_rate": 1.09328885801764e-05, "loss": 0.9176, "step": 8641 }, { "epoch": 3.915722700498414, "grad_norm": 0.10457509441850382, "learning_rate": 1.0924197228509206e-05, "loss": 0.9227, "step": 8642 }, { "epoch": 3.916175804259175, "grad_norm": 0.09445554196712704, "learning_rate": 1.0915508786523766e-05, "loss": 0.9291, "step": 8643 }, { "epoch": 3.916628908019937, "grad_norm": 0.10977468622204839, "learning_rate": 1.0906823255089574e-05, "loss": 0.9248, "step": 8644 }, { "epoch": 3.917082011780698, "grad_norm": 0.09601133559232206, "learning_rate": 1.0898140635075789e-05, "loss": 0.9231, "step": 8645 }, { "epoch": 3.917535115541459, "grad_norm": 0.10462201079088568, "learning_rate": 1.088946092735132e-05, "loss": 0.9145, "step": 8646 }, { "epoch": 3.91798821930222, "grad_norm": 0.10912308293123835, "learning_rate": 1.0880784132784744e-05, "loss": 0.9251, "step": 8647 }, { "epoch": 3.9184413230629813, "grad_norm": 0.09721662311354504, "learning_rate": 1.0872110252244395e-05, "loss": 0.9444, "step": 8648 }, { "epoch": 3.9188944268237424, "grad_norm": 0.10020836043238245, "learning_rate": 1.0863439286598258e-05, "loss": 0.927, "step": 8649 }, { "epoch": 3.9193475305845036, "grad_norm": 0.09267925491331185, "learning_rate": 1.0854771236714083e-05, "loss": 0.9386, "step": 8650 }, { "epoch": 3.919800634345265, "grad_norm": 0.10480583391474646, "learning_rate": 1.084610610345929e-05, "loss": 0.9158, "step": 8651 }, { "epoch": 3.9202537381060263, "grad_norm": 0.09426136888903543, "learning_rate": 1.083744388770101e-05, "loss": 0.9049, "step": 8652 }, { "epoch": 3.9207068418667874, "grad_norm": 0.10054076714737148, "learning_rate": 1.0828784590306106e-05, "loss": 0.9269, "step": 8653 }, { "epoch": 3.9211599456275485, "grad_norm": 0.09368310959070275, "learning_rate": 1.0820128212141116e-05, "loss": 0.9323, "step": 8654 }, { "epoch": 3.92161304938831, "grad_norm": 0.09889046343592124, "learning_rate": 1.0811474754072323e-05, "loss": 0.9291, "step": 8655 }, { "epoch": 3.9220661531490713, "grad_norm": 0.09437979961174796, "learning_rate": 1.0802824216965684e-05, "loss": 0.9141, "step": 8656 }, { "epoch": 3.9225192569098324, "grad_norm": 0.1059931481442803, "learning_rate": 1.079417660168689e-05, "loss": 0.9264, "step": 8657 }, { "epoch": 3.9229723606705935, "grad_norm": 0.09278852175364129, "learning_rate": 1.0785531909101321e-05, "loss": 0.9399, "step": 8658 }, { "epoch": 3.9234254644313546, "grad_norm": 0.09868389991704943, "learning_rate": 1.0776890140074058e-05, "loss": 0.9242, "step": 8659 }, { "epoch": 3.923878568192116, "grad_norm": 0.09910901128398243, "learning_rate": 1.0768251295469926e-05, "loss": 0.9397, "step": 8660 }, { "epoch": 3.9243316719528774, "grad_norm": 0.09766048839788002, "learning_rate": 1.0759615376153412e-05, "loss": 0.9465, "step": 8661 }, { "epoch": 3.9247847757136385, "grad_norm": 0.1003781374944822, "learning_rate": 1.075098238298876e-05, "loss": 0.9135, "step": 8662 }, { "epoch": 3.9252378794743996, "grad_norm": 0.10161840977966717, "learning_rate": 1.0742352316839856e-05, "loss": 0.9401, "step": 8663 }, { "epoch": 3.9256909832351607, "grad_norm": 0.1077690517494735, "learning_rate": 1.0733725178570368e-05, "loss": 0.9326, "step": 8664 }, { "epoch": 3.9261440869959223, "grad_norm": 0.09331500594434948, "learning_rate": 1.0725100969043605e-05, "loss": 0.9453, "step": 8665 }, { "epoch": 3.9265971907566835, "grad_norm": 0.0937109728980805, "learning_rate": 1.071647968912263e-05, "loss": 0.9106, "step": 8666 }, { "epoch": 3.9270502945174446, "grad_norm": 0.09550818241834383, "learning_rate": 1.070786133967019e-05, "loss": 0.9187, "step": 8667 }, { "epoch": 3.9275033982782057, "grad_norm": 0.09221573170177538, "learning_rate": 1.0699245921548735e-05, "loss": 0.9311, "step": 8668 }, { "epoch": 3.927956502038967, "grad_norm": 0.09153860976126159, "learning_rate": 1.0690633435620428e-05, "loss": 0.9429, "step": 8669 }, { "epoch": 3.928409605799728, "grad_norm": 0.09276580701868582, "learning_rate": 1.0682023882747155e-05, "loss": 0.9252, "step": 8670 }, { "epoch": 3.928862709560489, "grad_norm": 0.09012558442985445, "learning_rate": 1.067341726379048e-05, "loss": 0.9226, "step": 8671 }, { "epoch": 3.9293158133212507, "grad_norm": 0.09476684229819878, "learning_rate": 1.0664813579611683e-05, "loss": 0.9152, "step": 8672 }, { "epoch": 3.929768917082012, "grad_norm": 0.0877326951628061, "learning_rate": 1.0656212831071774e-05, "loss": 0.9325, "step": 8673 }, { "epoch": 3.930222020842773, "grad_norm": 0.1006148373853965, "learning_rate": 1.064761501903143e-05, "loss": 0.9182, "step": 8674 }, { "epoch": 3.930675124603534, "grad_norm": 0.08800996276219748, "learning_rate": 1.0639020144351071e-05, "loss": 0.9454, "step": 8675 }, { "epoch": 3.9311282283642957, "grad_norm": 0.10487936688482453, "learning_rate": 1.0630428207890783e-05, "loss": 0.927, "step": 8676 }, { "epoch": 3.931581332125057, "grad_norm": 0.0933394082127611, "learning_rate": 1.0621839210510405e-05, "loss": 0.9273, "step": 8677 }, { "epoch": 3.932034435885818, "grad_norm": 0.10459290477459753, "learning_rate": 1.0613253153069452e-05, "loss": 0.9124, "step": 8678 }, { "epoch": 3.932487539646579, "grad_norm": 0.09735952085201018, "learning_rate": 1.060467003642713e-05, "loss": 0.9285, "step": 8679 }, { "epoch": 3.93294064340734, "grad_norm": 0.10847411768906512, "learning_rate": 1.0596089861442396e-05, "loss": 0.9371, "step": 8680 }, { "epoch": 3.9333937471681013, "grad_norm": 0.09815504502091663, "learning_rate": 1.0587512628973865e-05, "loss": 0.9299, "step": 8681 }, { "epoch": 3.9338468509288624, "grad_norm": 0.1051399342216594, "learning_rate": 1.0578938339879907e-05, "loss": 0.9281, "step": 8682 }, { "epoch": 3.934299954689624, "grad_norm": 0.09159754182984794, "learning_rate": 1.0570366995018544e-05, "loss": 0.9263, "step": 8683 }, { "epoch": 3.934753058450385, "grad_norm": 0.1091438582423798, "learning_rate": 1.0561798595247552e-05, "loss": 0.9304, "step": 8684 }, { "epoch": 3.9352061622111463, "grad_norm": 0.0984866191475586, "learning_rate": 1.0553233141424371e-05, "loss": 0.9143, "step": 8685 }, { "epoch": 3.9356592659719074, "grad_norm": 0.1011484830926263, "learning_rate": 1.0544670634406189e-05, "loss": 0.9253, "step": 8686 }, { "epoch": 3.936112369732669, "grad_norm": 0.09354945285068018, "learning_rate": 1.0536111075049856e-05, "loss": 0.9179, "step": 8687 }, { "epoch": 3.93656547349343, "grad_norm": 0.1099735908794546, "learning_rate": 1.0527554464211943e-05, "loss": 0.9292, "step": 8688 }, { "epoch": 3.9370185772541912, "grad_norm": 0.10061916409671107, "learning_rate": 1.051900080274875e-05, "loss": 0.912, "step": 8689 }, { "epoch": 3.9374716810149524, "grad_norm": 0.08796985514773609, "learning_rate": 1.0510450091516242e-05, "loss": 0.9019, "step": 8690 }, { "epoch": 3.9379247847757135, "grad_norm": 0.08538961949013761, "learning_rate": 1.0501902331370127e-05, "loss": 0.9193, "step": 8691 }, { "epoch": 3.9383778885364746, "grad_norm": 0.08711297153606547, "learning_rate": 1.0493357523165772e-05, "loss": 0.9202, "step": 8692 }, { "epoch": 3.938830992297236, "grad_norm": 0.08881906295414317, "learning_rate": 1.0484815667758306e-05, "loss": 0.9506, "step": 8693 }, { "epoch": 3.9392840960579973, "grad_norm": 0.08550824977878381, "learning_rate": 1.0476276766002504e-05, "loss": 0.8887, "step": 8694 }, { "epoch": 3.9397371998187585, "grad_norm": 0.09394476952753299, "learning_rate": 1.0467740818752898e-05, "loss": 0.9329, "step": 8695 }, { "epoch": 3.9401903035795196, "grad_norm": 0.08999271978014307, "learning_rate": 1.0459207826863692e-05, "loss": 0.9369, "step": 8696 }, { "epoch": 3.940643407340281, "grad_norm": 0.09384778210169639, "learning_rate": 1.0450677791188783e-05, "loss": 0.9287, "step": 8697 }, { "epoch": 3.9410965111010423, "grad_norm": 0.09159107051981882, "learning_rate": 1.044215071258182e-05, "loss": 0.9392, "step": 8698 }, { "epoch": 3.9415496148618034, "grad_norm": 0.08849634177756167, "learning_rate": 1.0433626591896102e-05, "loss": 0.9004, "step": 8699 }, { "epoch": 3.9420027186225646, "grad_norm": 0.08986918860182234, "learning_rate": 1.0425105429984681e-05, "loss": 0.9176, "step": 8700 }, { "epoch": 3.9424558223833257, "grad_norm": 0.09064344948882226, "learning_rate": 1.0416587227700261e-05, "loss": 0.9388, "step": 8701 }, { "epoch": 3.942908926144087, "grad_norm": 0.09570077836128259, "learning_rate": 1.0408071985895312e-05, "loss": 0.9355, "step": 8702 }, { "epoch": 3.943362029904848, "grad_norm": 0.08675935809612788, "learning_rate": 1.0399559705421951e-05, "loss": 0.9151, "step": 8703 }, { "epoch": 3.9438151336656095, "grad_norm": 0.08589937851903753, "learning_rate": 1.0391050387132014e-05, "loss": 0.9245, "step": 8704 }, { "epoch": 3.9442682374263707, "grad_norm": 0.0945763506853331, "learning_rate": 1.0382544031877072e-05, "loss": 0.9217, "step": 8705 }, { "epoch": 3.944721341187132, "grad_norm": 0.09748224604160675, "learning_rate": 1.037404064050835e-05, "loss": 0.934, "step": 8706 }, { "epoch": 3.945174444947893, "grad_norm": 0.09567102417175087, "learning_rate": 1.0365540213876826e-05, "loss": 0.9304, "step": 8707 }, { "epoch": 3.9456275487086545, "grad_norm": 0.0927256711724287, "learning_rate": 1.0357042752833135e-05, "loss": 0.9375, "step": 8708 }, { "epoch": 3.9460806524694156, "grad_norm": 0.09544759116459514, "learning_rate": 1.034854825822766e-05, "loss": 0.9348, "step": 8709 }, { "epoch": 3.9465337562301768, "grad_norm": 0.09489152222543332, "learning_rate": 1.0340056730910435e-05, "loss": 0.9183, "step": 8710 }, { "epoch": 3.946986859990938, "grad_norm": 0.09978967851450059, "learning_rate": 1.0331568171731261e-05, "loss": 0.9483, "step": 8711 }, { "epoch": 3.947439963751699, "grad_norm": 0.09413258163211204, "learning_rate": 1.0323082581539591e-05, "loss": 0.9282, "step": 8712 }, { "epoch": 3.94789306751246, "grad_norm": 0.10004603883701198, "learning_rate": 1.0314599961184579e-05, "loss": 0.9465, "step": 8713 }, { "epoch": 3.9483461712732213, "grad_norm": 0.106073794498443, "learning_rate": 1.0306120311515136e-05, "loss": 0.9374, "step": 8714 }, { "epoch": 3.948799275033983, "grad_norm": 0.10095665589717953, "learning_rate": 1.0297643633379803e-05, "loss": 0.9456, "step": 8715 }, { "epoch": 3.949252378794744, "grad_norm": 0.10313217179973318, "learning_rate": 1.0289169927626897e-05, "loss": 0.9137, "step": 8716 }, { "epoch": 3.949705482555505, "grad_norm": 0.11102650214502867, "learning_rate": 1.0280699195104367e-05, "loss": 0.895, "step": 8717 }, { "epoch": 3.9501585863162663, "grad_norm": 0.09959329956888435, "learning_rate": 1.0272231436659927e-05, "loss": 0.9392, "step": 8718 }, { "epoch": 3.950611690077028, "grad_norm": 0.11016253252251203, "learning_rate": 1.0263766653140955e-05, "loss": 0.9275, "step": 8719 }, { "epoch": 3.951064793837789, "grad_norm": 0.09748847708779278, "learning_rate": 1.025530484539452e-05, "loss": 0.9324, "step": 8720 }, { "epoch": 3.95151789759855, "grad_norm": 0.10166814491972812, "learning_rate": 1.024684601426745e-05, "loss": 0.9325, "step": 8721 }, { "epoch": 3.9519710013593112, "grad_norm": 0.09789065970163145, "learning_rate": 1.0238390160606211e-05, "loss": 0.9106, "step": 8722 }, { "epoch": 3.9524241051200724, "grad_norm": 0.10255486545385019, "learning_rate": 1.0229937285257022e-05, "loss": 0.9265, "step": 8723 }, { "epoch": 3.9528772088808335, "grad_norm": 0.09075008592765611, "learning_rate": 1.0221487389065765e-05, "loss": 0.9307, "step": 8724 }, { "epoch": 3.953330312641595, "grad_norm": 0.10956764244346748, "learning_rate": 1.021304047287805e-05, "loss": 0.9256, "step": 8725 }, { "epoch": 3.953783416402356, "grad_norm": 0.10833073300290352, "learning_rate": 1.0204596537539163e-05, "loss": 0.9325, "step": 8726 }, { "epoch": 3.9542365201631173, "grad_norm": 0.11088102174229006, "learning_rate": 1.019615558389413e-05, "loss": 0.9336, "step": 8727 }, { "epoch": 3.9546896239238785, "grad_norm": 0.11577969361885267, "learning_rate": 1.0187717612787633e-05, "loss": 0.9284, "step": 8728 }, { "epoch": 3.95514272768464, "grad_norm": 0.10872938774059518, "learning_rate": 1.0179282625064108e-05, "loss": 0.9345, "step": 8729 }, { "epoch": 3.955595831445401, "grad_norm": 0.12276416116143547, "learning_rate": 1.0170850621567631e-05, "loss": 0.9462, "step": 8730 }, { "epoch": 3.9560489352061623, "grad_norm": 0.0920620025853044, "learning_rate": 1.0162421603142043e-05, "loss": 0.9278, "step": 8731 }, { "epoch": 3.9565020389669234, "grad_norm": 0.11563016488370333, "learning_rate": 1.015399557063084e-05, "loss": 0.9355, "step": 8732 }, { "epoch": 3.9569551427276846, "grad_norm": 0.08518922395782813, "learning_rate": 1.0145572524877223e-05, "loss": 0.9233, "step": 8733 }, { "epoch": 3.9574082464884457, "grad_norm": 0.11032498289760806, "learning_rate": 1.0137152466724128e-05, "loss": 0.9316, "step": 8734 }, { "epoch": 3.957861350249207, "grad_norm": 0.09573120030710797, "learning_rate": 1.0128735397014142e-05, "loss": 0.9203, "step": 8735 }, { "epoch": 3.9583144540099684, "grad_norm": 0.11493864283779419, "learning_rate": 1.0120321316589612e-05, "loss": 0.9236, "step": 8736 }, { "epoch": 3.9587675577707295, "grad_norm": 0.0930587095580976, "learning_rate": 1.0111910226292529e-05, "loss": 0.9459, "step": 8737 }, { "epoch": 3.9592206615314907, "grad_norm": 0.09749103886177413, "learning_rate": 1.0103502126964631e-05, "loss": 0.9238, "step": 8738 }, { "epoch": 3.959673765292252, "grad_norm": 0.09359777002943095, "learning_rate": 1.0095097019447322e-05, "loss": 0.9479, "step": 8739 }, { "epoch": 3.9601268690530134, "grad_norm": 0.09561150682091564, "learning_rate": 1.0086694904581714e-05, "loss": 0.9313, "step": 8740 }, { "epoch": 3.9605799728137745, "grad_norm": 0.08558035005100391, "learning_rate": 1.0078295783208643e-05, "loss": 0.9443, "step": 8741 }, { "epoch": 3.9610330765745356, "grad_norm": 0.09981538336083115, "learning_rate": 1.0069899656168607e-05, "loss": 0.9193, "step": 8742 }, { "epoch": 3.9614861803352968, "grad_norm": 0.09596967185817407, "learning_rate": 1.006150652430185e-05, "loss": 0.9277, "step": 8743 }, { "epoch": 3.961939284096058, "grad_norm": 0.09471586316736641, "learning_rate": 1.0053116388448272e-05, "loss": 0.9295, "step": 8744 }, { "epoch": 3.962392387856819, "grad_norm": 0.08519847331331881, "learning_rate": 1.004472924944751e-05, "loss": 0.9247, "step": 8745 }, { "epoch": 3.96284549161758, "grad_norm": 0.10292252421249423, "learning_rate": 1.0036345108138862e-05, "loss": 0.9312, "step": 8746 }, { "epoch": 3.9632985953783417, "grad_norm": 0.08988409505958105, "learning_rate": 1.0027963965361373e-05, "loss": 0.9419, "step": 8747 }, { "epoch": 3.963751699139103, "grad_norm": 0.09758928107766991, "learning_rate": 1.0019585821953752e-05, "loss": 0.9321, "step": 8748 }, { "epoch": 3.964204802899864, "grad_norm": 0.09051593624289349, "learning_rate": 1.0011210678754403e-05, "loss": 0.9256, "step": 8749 }, { "epoch": 3.964657906660625, "grad_norm": 0.09184430586256889, "learning_rate": 1.000283853660148e-05, "loss": 0.9329, "step": 8750 }, { "epoch": 3.9651110104213867, "grad_norm": 0.09204828083867184, "learning_rate": 9.994469396332765e-06, "loss": 0.9103, "step": 8751 }, { "epoch": 3.965564114182148, "grad_norm": 0.08931584244566484, "learning_rate": 9.986103258785808e-06, "loss": 0.9238, "step": 8752 }, { "epoch": 3.966017217942909, "grad_norm": 0.09415711926981243, "learning_rate": 9.977740124797801e-06, "loss": 0.9387, "step": 8753 }, { "epoch": 3.96647032170367, "grad_norm": 0.08669987993898706, "learning_rate": 9.96937999520569e-06, "loss": 0.9309, "step": 8754 }, { "epoch": 3.966923425464431, "grad_norm": 0.09258861200277362, "learning_rate": 9.961022870846064e-06, "loss": 0.9348, "step": 8755 }, { "epoch": 3.9673765292251923, "grad_norm": 0.09520520312092731, "learning_rate": 9.952668752555263e-06, "loss": 0.9174, "step": 8756 }, { "epoch": 3.967829632985954, "grad_norm": 0.1020128273554053, "learning_rate": 9.9443176411693e-06, "loss": 0.9323, "step": 8757 }, { "epoch": 3.968282736746715, "grad_norm": 0.10625199668242533, "learning_rate": 9.935969537523862e-06, "loss": 0.9289, "step": 8758 }, { "epoch": 3.968735840507476, "grad_norm": 0.09880655348230757, "learning_rate": 9.927624442454396e-06, "loss": 0.9464, "step": 8759 }, { "epoch": 3.9691889442682373, "grad_norm": 0.09838823122804405, "learning_rate": 9.919282356795992e-06, "loss": 0.9339, "step": 8760 }, { "epoch": 3.969642048028999, "grad_norm": 0.10387536412962095, "learning_rate": 9.91094328138348e-06, "loss": 0.9187, "step": 8761 }, { "epoch": 3.97009515178976, "grad_norm": 0.091621198736673, "learning_rate": 9.90260721705135e-06, "loss": 0.9462, "step": 8762 }, { "epoch": 3.970548255550521, "grad_norm": 0.11219960520387537, "learning_rate": 9.894274164633835e-06, "loss": 0.94, "step": 8763 }, { "epoch": 3.9710013593112823, "grad_norm": 0.09632651974265949, "learning_rate": 9.885944124964832e-06, "loss": 0.9191, "step": 8764 }, { "epoch": 3.9714544630720434, "grad_norm": 0.107563131655795, "learning_rate": 9.87761709887793e-06, "loss": 0.9339, "step": 8765 }, { "epoch": 3.9719075668328045, "grad_norm": 0.10690403365208971, "learning_rate": 9.86929308720646e-06, "loss": 0.9083, "step": 8766 }, { "epoch": 3.9723606705935657, "grad_norm": 0.1018104969966924, "learning_rate": 9.860972090783405e-06, "loss": 0.9057, "step": 8767 }, { "epoch": 3.9728137743543273, "grad_norm": 0.11081515977300305, "learning_rate": 9.852654110441482e-06, "loss": 0.9177, "step": 8768 }, { "epoch": 3.9732668781150884, "grad_norm": 0.09178807440324026, "learning_rate": 9.84433914701307e-06, "loss": 0.9346, "step": 8769 }, { "epoch": 3.9737199818758495, "grad_norm": 0.10317361412178232, "learning_rate": 9.836027201330291e-06, "loss": 0.9123, "step": 8770 }, { "epoch": 3.9741730856366106, "grad_norm": 0.10939675761212748, "learning_rate": 9.82771827422492e-06, "loss": 0.921, "step": 8771 }, { "epoch": 3.9746261893973722, "grad_norm": 0.09349762819412713, "learning_rate": 9.819412366528467e-06, "loss": 0.8997, "step": 8772 }, { "epoch": 3.9750792931581334, "grad_norm": 0.10463402599100996, "learning_rate": 9.811109479072112e-06, "loss": 0.918, "step": 8773 }, { "epoch": 3.9755323969188945, "grad_norm": 0.10108201720481429, "learning_rate": 9.802809612686737e-06, "loss": 0.9233, "step": 8774 }, { "epoch": 3.9759855006796556, "grad_norm": 0.08627064696021501, "learning_rate": 9.794512768202949e-06, "loss": 0.949, "step": 8775 }, { "epoch": 3.9764386044404167, "grad_norm": 0.10619094455259966, "learning_rate": 9.786218946451007e-06, "loss": 0.9245, "step": 8776 }, { "epoch": 3.976891708201178, "grad_norm": 0.09234688707261875, "learning_rate": 9.77792814826092e-06, "loss": 0.9261, "step": 8777 }, { "epoch": 3.977344811961939, "grad_norm": 0.09976067084953069, "learning_rate": 9.769640374462344e-06, "loss": 0.9305, "step": 8778 }, { "epoch": 3.9777979157227006, "grad_norm": 0.09308007940311301, "learning_rate": 9.761355625884672e-06, "loss": 0.9125, "step": 8779 }, { "epoch": 3.9782510194834617, "grad_norm": 0.09113034104792496, "learning_rate": 9.753073903356972e-06, "loss": 0.9195, "step": 8780 }, { "epoch": 3.978704123244223, "grad_norm": 0.10561227567883713, "learning_rate": 9.744795207708013e-06, "loss": 0.9262, "step": 8781 }, { "epoch": 3.979157227004984, "grad_norm": 0.08910754160668892, "learning_rate": 9.736519539766251e-06, "loss": 0.9619, "step": 8782 }, { "epoch": 3.9796103307657456, "grad_norm": 0.11101892528808101, "learning_rate": 9.728246900359877e-06, "loss": 0.9146, "step": 8783 }, { "epoch": 3.9800634345265067, "grad_norm": 0.09426003780152116, "learning_rate": 9.719977290316738e-06, "loss": 0.93, "step": 8784 }, { "epoch": 3.980516538287268, "grad_norm": 0.1001416244221306, "learning_rate": 9.711710710464381e-06, "loss": 0.923, "step": 8785 }, { "epoch": 3.980969642048029, "grad_norm": 0.0934750376308233, "learning_rate": 9.703447161630088e-06, "loss": 0.9224, "step": 8786 }, { "epoch": 3.98142274580879, "grad_norm": 0.09514803806061371, "learning_rate": 9.695186644640784e-06, "loss": 0.9552, "step": 8787 }, { "epoch": 3.981875849569551, "grad_norm": 0.0967103829599888, "learning_rate": 9.686929160323143e-06, "loss": 0.931, "step": 8788 }, { "epoch": 3.982328953330313, "grad_norm": 0.09732702543807675, "learning_rate": 9.67867470950349e-06, "loss": 0.944, "step": 8789 }, { "epoch": 3.982782057091074, "grad_norm": 0.1092152756123379, "learning_rate": 9.670423293007883e-06, "loss": 0.9387, "step": 8790 }, { "epoch": 3.983235160851835, "grad_norm": 0.10417531645550411, "learning_rate": 9.662174911662042e-06, "loss": 0.9223, "step": 8791 }, { "epoch": 3.983688264612596, "grad_norm": 0.10552420908418594, "learning_rate": 9.653929566291417e-06, "loss": 0.9326, "step": 8792 }, { "epoch": 3.9841413683733577, "grad_norm": 0.11081928752261301, "learning_rate": 9.645687257721139e-06, "loss": 0.9222, "step": 8793 }, { "epoch": 3.984594472134119, "grad_norm": 0.09789424162307021, "learning_rate": 9.637447986776016e-06, "loss": 0.9274, "step": 8794 }, { "epoch": 3.98504757589488, "grad_norm": 0.10646614038840542, "learning_rate": 9.629211754280594e-06, "loss": 0.9207, "step": 8795 }, { "epoch": 3.985500679655641, "grad_norm": 0.10077924536478501, "learning_rate": 9.620978561059067e-06, "loss": 0.9301, "step": 8796 }, { "epoch": 3.9859537834164023, "grad_norm": 0.10847985637316122, "learning_rate": 9.612748407935375e-06, "loss": 0.9378, "step": 8797 }, { "epoch": 3.9864068871771634, "grad_norm": 0.10290004284606555, "learning_rate": 9.604521295733105e-06, "loss": 0.8958, "step": 8798 }, { "epoch": 3.9868599909379245, "grad_norm": 0.0957452002030907, "learning_rate": 9.596297225275588e-06, "loss": 0.9079, "step": 8799 }, { "epoch": 3.987313094698686, "grad_norm": 0.10752077431311692, "learning_rate": 9.588076197385807e-06, "loss": 0.9306, "step": 8800 }, { "epoch": 3.9877661984594472, "grad_norm": 0.09765784725010936, "learning_rate": 9.579858212886456e-06, "loss": 0.9441, "step": 8801 }, { "epoch": 3.9882193022202084, "grad_norm": 0.09754949844313798, "learning_rate": 9.57164327259994e-06, "loss": 0.9202, "step": 8802 }, { "epoch": 3.9886724059809695, "grad_norm": 0.10491435701673738, "learning_rate": 9.563431377348338e-06, "loss": 0.9255, "step": 8803 }, { "epoch": 3.989125509741731, "grad_norm": 0.09331950903288945, "learning_rate": 9.55522252795344e-06, "loss": 0.9427, "step": 8804 }, { "epoch": 3.989578613502492, "grad_norm": 0.10472844312894458, "learning_rate": 9.547016725236715e-06, "loss": 0.939, "step": 8805 }, { "epoch": 3.9900317172632533, "grad_norm": 0.10406315241272097, "learning_rate": 9.53881397001935e-06, "loss": 0.918, "step": 8806 }, { "epoch": 3.9904848210240145, "grad_norm": 0.09939279315857881, "learning_rate": 9.530614263122198e-06, "loss": 0.9215, "step": 8807 }, { "epoch": 3.9909379247847756, "grad_norm": 0.10950316969697113, "learning_rate": 9.522417605365839e-06, "loss": 0.9084, "step": 8808 }, { "epoch": 3.9913910285455367, "grad_norm": 0.08987551593445825, "learning_rate": 9.51422399757052e-06, "loss": 0.9197, "step": 8809 }, { "epoch": 3.991844132306298, "grad_norm": 0.0925110300837039, "learning_rate": 9.506033440556188e-06, "loss": 0.9181, "step": 8810 }, { "epoch": 3.9922972360670594, "grad_norm": 0.08759383263282704, "learning_rate": 9.49784593514251e-06, "loss": 0.9164, "step": 8811 }, { "epoch": 3.9927503398278206, "grad_norm": 0.08299751814812785, "learning_rate": 9.4896614821488e-06, "loss": 0.9189, "step": 8812 }, { "epoch": 3.9932034435885817, "grad_norm": 0.09491569215638479, "learning_rate": 9.48148008239413e-06, "loss": 0.9263, "step": 8813 }, { "epoch": 3.993656547349343, "grad_norm": 0.08367274980352299, "learning_rate": 9.473301736697196e-06, "loss": 0.9305, "step": 8814 }, { "epoch": 3.9941096511101044, "grad_norm": 0.10004893442214968, "learning_rate": 9.465126445876458e-06, "loss": 0.9244, "step": 8815 }, { "epoch": 3.9945627548708655, "grad_norm": 0.10471921344970893, "learning_rate": 9.456954210750005e-06, "loss": 0.951, "step": 8816 }, { "epoch": 3.9950158586316267, "grad_norm": 0.09303740578190237, "learning_rate": 9.448785032135675e-06, "loss": 0.9261, "step": 8817 }, { "epoch": 3.995468962392388, "grad_norm": 0.10039743610832556, "learning_rate": 9.440618910850964e-06, "loss": 0.936, "step": 8818 }, { "epoch": 3.995922066153149, "grad_norm": 0.09781300772592202, "learning_rate": 9.43245584771307e-06, "loss": 0.9057, "step": 8819 }, { "epoch": 3.99637516991391, "grad_norm": 0.10923251246288246, "learning_rate": 9.42429584353891e-06, "loss": 0.9219, "step": 8820 }, { "epoch": 3.9968282736746716, "grad_norm": 0.104030275979634, "learning_rate": 9.416138899145042e-06, "loss": 0.9305, "step": 8821 }, { "epoch": 3.9972813774354328, "grad_norm": 0.09288299478180284, "learning_rate": 9.407985015347782e-06, "loss": 0.9336, "step": 8822 }, { "epoch": 3.997734481196194, "grad_norm": 0.10498825628375912, "learning_rate": 9.399834192963081e-06, "loss": 0.9251, "step": 8823 }, { "epoch": 3.998187584956955, "grad_norm": 0.09374917195176258, "learning_rate": 9.391686432806635e-06, "loss": 0.9343, "step": 8824 }, { "epoch": 3.9986406887177166, "grad_norm": 0.1012888994921123, "learning_rate": 9.383541735693798e-06, "loss": 0.9393, "step": 8825 }, { "epoch": 3.9990937924784777, "grad_norm": 0.10434286619044526, "learning_rate": 9.375400102439616e-06, "loss": 0.9057, "step": 8826 }, { "epoch": 3.999546896239239, "grad_norm": 0.0975134839576523, "learning_rate": 9.367261533858865e-06, "loss": 0.9396, "step": 8827 }, { "epoch": 4.0, "grad_norm": 0.10004469493838847, "learning_rate": 9.359126030765964e-06, "loss": 0.9165, "step": 8828 }, { "epoch": 4.000453103760761, "grad_norm": 0.10364839673626258, "learning_rate": 9.350993593975075e-06, "loss": 0.9101, "step": 8829 }, { "epoch": 4.000906207521522, "grad_norm": 0.09429453706009129, "learning_rate": 9.342864224300006e-06, "loss": 0.9308, "step": 8830 }, { "epoch": 4.001359311282283, "grad_norm": 0.09833342657688163, "learning_rate": 9.334737922554308e-06, "loss": 0.9155, "step": 8831 }, { "epoch": 4.0018124150430445, "grad_norm": 0.10048541456716958, "learning_rate": 9.326614689551174e-06, "loss": 0.9377, "step": 8832 }, { "epoch": 4.002265518803806, "grad_norm": 0.09192901007505283, "learning_rate": 9.318494526103538e-06, "loss": 0.9299, "step": 8833 }, { "epoch": 4.002718622564568, "grad_norm": 0.10080488118687564, "learning_rate": 9.310377433023992e-06, "loss": 0.9574, "step": 8834 }, { "epoch": 4.003171726325329, "grad_norm": 0.10685025354158188, "learning_rate": 9.302263411124816e-06, "loss": 0.9305, "step": 8835 }, { "epoch": 4.00362483008609, "grad_norm": 0.09027502128279463, "learning_rate": 9.294152461218027e-06, "loss": 0.9348, "step": 8836 }, { "epoch": 4.004077933846851, "grad_norm": 0.10097241884925187, "learning_rate": 9.286044584115278e-06, "loss": 0.9231, "step": 8837 }, { "epoch": 4.004531037607612, "grad_norm": 0.099174608594514, "learning_rate": 9.277939780627983e-06, "loss": 0.942, "step": 8838 }, { "epoch": 4.004984141368373, "grad_norm": 0.10880527975232837, "learning_rate": 9.269838051567159e-06, "loss": 0.8998, "step": 8839 }, { "epoch": 4.0054372451291345, "grad_norm": 0.09349893586592609, "learning_rate": 9.2617393977436e-06, "loss": 0.9242, "step": 8840 }, { "epoch": 4.005890348889896, "grad_norm": 0.10432218325041294, "learning_rate": 9.25364381996773e-06, "loss": 0.9233, "step": 8841 }, { "epoch": 4.006343452650657, "grad_norm": 0.09752358067991995, "learning_rate": 9.245551319049721e-06, "loss": 0.9093, "step": 8842 }, { "epoch": 4.006796556411418, "grad_norm": 0.08975454164359739, "learning_rate": 9.23746189579938e-06, "loss": 0.9269, "step": 8843 }, { "epoch": 4.00724966017218, "grad_norm": 0.11684474130823833, "learning_rate": 9.229375551026262e-06, "loss": 0.8984, "step": 8844 }, { "epoch": 4.007702763932941, "grad_norm": 0.09501178706729393, "learning_rate": 9.221292285539568e-06, "loss": 0.9091, "step": 8845 }, { "epoch": 4.008155867693702, "grad_norm": 0.11140383928705766, "learning_rate": 9.213212100148202e-06, "loss": 0.9407, "step": 8846 }, { "epoch": 4.008608971454463, "grad_norm": 0.10689753015006286, "learning_rate": 9.20513499566079e-06, "loss": 0.9287, "step": 8847 }, { "epoch": 4.009062075215224, "grad_norm": 0.08788255802838059, "learning_rate": 9.197060972885602e-06, "loss": 0.9216, "step": 8848 }, { "epoch": 4.0095151789759855, "grad_norm": 0.10074468136780595, "learning_rate": 9.188990032630642e-06, "loss": 0.9201, "step": 8849 }, { "epoch": 4.009968282736747, "grad_norm": 0.10566663228022981, "learning_rate": 9.18092217570357e-06, "loss": 0.9166, "step": 8850 }, { "epoch": 4.010421386497508, "grad_norm": 0.09046390341722099, "learning_rate": 9.17285740291178e-06, "loss": 0.9314, "step": 8851 }, { "epoch": 4.010874490258269, "grad_norm": 0.10156596281214957, "learning_rate": 9.164795715062302e-06, "loss": 0.9154, "step": 8852 }, { "epoch": 4.01132759401903, "grad_norm": 0.09301437054120447, "learning_rate": 9.156737112961913e-06, "loss": 0.9321, "step": 8853 }, { "epoch": 4.011780697779791, "grad_norm": 0.0845423266198243, "learning_rate": 9.148681597417046e-06, "loss": 0.9134, "step": 8854 }, { "epoch": 4.012233801540553, "grad_norm": 0.09978116394061871, "learning_rate": 9.140629169233826e-06, "loss": 0.9333, "step": 8855 }, { "epoch": 4.012686905301314, "grad_norm": 0.08898468541283187, "learning_rate": 9.132579829218095e-06, "loss": 0.9349, "step": 8856 }, { "epoch": 4.0131400090620755, "grad_norm": 0.08788749226384923, "learning_rate": 9.124533578175346e-06, "loss": 0.92, "step": 8857 }, { "epoch": 4.013593112822837, "grad_norm": 0.08664001888565583, "learning_rate": 9.116490416910815e-06, "loss": 0.9113, "step": 8858 }, { "epoch": 4.014046216583598, "grad_norm": 0.0887784292483501, "learning_rate": 9.108450346229367e-06, "loss": 0.9223, "step": 8859 }, { "epoch": 4.014499320344359, "grad_norm": 0.09143793901580416, "learning_rate": 9.100413366935626e-06, "loss": 0.9312, "step": 8860 }, { "epoch": 4.01495242410512, "grad_norm": 0.08404992050581986, "learning_rate": 9.092379479833843e-06, "loss": 0.926, "step": 8861 }, { "epoch": 4.015405527865881, "grad_norm": 0.10147295601669978, "learning_rate": 9.084348685727993e-06, "loss": 0.9267, "step": 8862 }, { "epoch": 4.015858631626642, "grad_norm": 0.08985351994112584, "learning_rate": 9.076320985421745e-06, "loss": 0.9288, "step": 8863 }, { "epoch": 4.016311735387403, "grad_norm": 0.0977610042030221, "learning_rate": 9.068296379718431e-06, "loss": 0.9279, "step": 8864 }, { "epoch": 4.0167648391481645, "grad_norm": 0.09896803589168227, "learning_rate": 9.06027486942112e-06, "loss": 0.9141, "step": 8865 }, { "epoch": 4.0172179429089265, "grad_norm": 0.09738166613213098, "learning_rate": 9.052256455332511e-06, "loss": 0.9205, "step": 8866 }, { "epoch": 4.017671046669688, "grad_norm": 0.09658178546123497, "learning_rate": 9.044241138255051e-06, "loss": 0.9275, "step": 8867 }, { "epoch": 4.018124150430449, "grad_norm": 0.09737290367612891, "learning_rate": 9.03622891899083e-06, "loss": 0.9305, "step": 8868 }, { "epoch": 4.01857725419121, "grad_norm": 0.08891724365771916, "learning_rate": 9.028219798341675e-06, "loss": 0.9204, "step": 8869 }, { "epoch": 4.019030357951971, "grad_norm": 0.09793904994426847, "learning_rate": 9.020213777109057e-06, "loss": 0.9206, "step": 8870 }, { "epoch": 4.019483461712732, "grad_norm": 0.08971825593757277, "learning_rate": 9.012210856094152e-06, "loss": 0.9167, "step": 8871 }, { "epoch": 4.019936565473493, "grad_norm": 0.0936769059514767, "learning_rate": 9.00421103609785e-06, "loss": 0.9293, "step": 8872 }, { "epoch": 4.0203896692342544, "grad_norm": 0.09341354355449591, "learning_rate": 8.996214317920691e-06, "loss": 0.9396, "step": 8873 }, { "epoch": 4.020842772995016, "grad_norm": 0.0966375815030351, "learning_rate": 8.988220702362941e-06, "loss": 0.9216, "step": 8874 }, { "epoch": 4.021295876755777, "grad_norm": 0.09432380452822169, "learning_rate": 8.980230190224528e-06, "loss": 0.9115, "step": 8875 }, { "epoch": 4.021748980516539, "grad_norm": 0.09430321115907107, "learning_rate": 8.972242782305093e-06, "loss": 0.9463, "step": 8876 }, { "epoch": 4.0222020842773, "grad_norm": 0.0968438914066803, "learning_rate": 8.964258479403951e-06, "loss": 0.9206, "step": 8877 }, { "epoch": 4.022655188038061, "grad_norm": 0.08499548659743804, "learning_rate": 8.956277282320091e-06, "loss": 0.9223, "step": 8878 }, { "epoch": 4.023108291798822, "grad_norm": 0.09937431954818354, "learning_rate": 8.948299191852237e-06, "loss": 0.943, "step": 8879 }, { "epoch": 4.023561395559583, "grad_norm": 0.08795084437495959, "learning_rate": 8.940324208798748e-06, "loss": 0.9332, "step": 8880 }, { "epoch": 4.024014499320344, "grad_norm": 0.09112507107659717, "learning_rate": 8.932352333957723e-06, "loss": 0.9392, "step": 8881 }, { "epoch": 4.0244676030811055, "grad_norm": 0.09295259685578508, "learning_rate": 8.924383568126904e-06, "loss": 0.913, "step": 8882 }, { "epoch": 4.024920706841867, "grad_norm": 0.08994314284296946, "learning_rate": 8.916417912103768e-06, "loss": 0.9394, "step": 8883 }, { "epoch": 4.025373810602628, "grad_norm": 0.09724109995077111, "learning_rate": 8.908455366685431e-06, "loss": 0.9209, "step": 8884 }, { "epoch": 4.025826914363389, "grad_norm": 0.09239228070682894, "learning_rate": 8.900495932668742e-06, "loss": 0.9159, "step": 8885 }, { "epoch": 4.02628001812415, "grad_norm": 0.11091750223687065, "learning_rate": 8.892539610850216e-06, "loss": 0.9368, "step": 8886 }, { "epoch": 4.026733121884912, "grad_norm": 0.0859311247871569, "learning_rate": 8.884586402026043e-06, "loss": 0.9317, "step": 8887 }, { "epoch": 4.027186225645673, "grad_norm": 0.10142433781099311, "learning_rate": 8.876636306992146e-06, "loss": 0.9144, "step": 8888 }, { "epoch": 4.027639329406434, "grad_norm": 0.09317072578684847, "learning_rate": 8.868689326544081e-06, "loss": 0.9351, "step": 8889 }, { "epoch": 4.0280924331671955, "grad_norm": 0.10297590332293972, "learning_rate": 8.860745461477149e-06, "loss": 0.9229, "step": 8890 }, { "epoch": 4.028545536927957, "grad_norm": 0.0965912620041598, "learning_rate": 8.85280471258629e-06, "loss": 0.9323, "step": 8891 }, { "epoch": 4.028998640688718, "grad_norm": 0.0954748988709227, "learning_rate": 8.844867080666163e-06, "loss": 0.9259, "step": 8892 }, { "epoch": 4.029451744449479, "grad_norm": 0.0943241430899144, "learning_rate": 8.83693256651109e-06, "loss": 0.9206, "step": 8893 }, { "epoch": 4.02990484821024, "grad_norm": 0.08980691276958509, "learning_rate": 8.829001170915124e-06, "loss": 0.9241, "step": 8894 }, { "epoch": 4.030357951971001, "grad_norm": 0.09428338878488285, "learning_rate": 8.821072894671957e-06, "loss": 0.9108, "step": 8895 }, { "epoch": 4.030811055731762, "grad_norm": 0.09112734272582043, "learning_rate": 8.813147738574992e-06, "loss": 0.9174, "step": 8896 }, { "epoch": 4.031264159492523, "grad_norm": 0.09343068852055048, "learning_rate": 8.805225703417322e-06, "loss": 0.9026, "step": 8897 }, { "epoch": 4.031717263253285, "grad_norm": 0.0986891590593075, "learning_rate": 8.797306789991707e-06, "loss": 0.9119, "step": 8898 }, { "epoch": 4.0321703670140465, "grad_norm": 0.09414458955972625, "learning_rate": 8.789390999090637e-06, "loss": 0.9364, "step": 8899 }, { "epoch": 4.032623470774808, "grad_norm": 0.08997222399154663, "learning_rate": 8.78147833150624e-06, "loss": 0.9292, "step": 8900 }, { "epoch": 4.033076574535569, "grad_norm": 0.0929735484226183, "learning_rate": 8.77356878803037e-06, "loss": 0.9076, "step": 8901 }, { "epoch": 4.03352967829633, "grad_norm": 0.0901891499320868, "learning_rate": 8.765662369454539e-06, "loss": 0.926, "step": 8902 }, { "epoch": 4.033982782057091, "grad_norm": 0.08959006389178964, "learning_rate": 8.757759076569985e-06, "loss": 0.9198, "step": 8903 }, { "epoch": 4.034435885817852, "grad_norm": 0.09296208208803328, "learning_rate": 8.749858910167575e-06, "loss": 0.9119, "step": 8904 }, { "epoch": 4.034888989578613, "grad_norm": 0.08665420098008692, "learning_rate": 8.741961871037929e-06, "loss": 0.9222, "step": 8905 }, { "epoch": 4.035342093339374, "grad_norm": 0.08525541692340276, "learning_rate": 8.734067959971306e-06, "loss": 0.9359, "step": 8906 }, { "epoch": 4.035795197100136, "grad_norm": 0.09509322720929793, "learning_rate": 8.726177177757655e-06, "loss": 0.9237, "step": 8907 }, { "epoch": 4.036248300860898, "grad_norm": 0.08402715926443506, "learning_rate": 8.718289525186652e-06, "loss": 0.9298, "step": 8908 }, { "epoch": 4.036701404621659, "grad_norm": 0.09603136652353278, "learning_rate": 8.710405003047607e-06, "loss": 0.9352, "step": 8909 }, { "epoch": 4.03715450838242, "grad_norm": 0.08688721718154371, "learning_rate": 8.702523612129564e-06, "loss": 0.9199, "step": 8910 }, { "epoch": 4.037607612143181, "grad_norm": 0.10839495525574454, "learning_rate": 8.694645353221207e-06, "loss": 0.9224, "step": 8911 }, { "epoch": 4.038060715903942, "grad_norm": 0.08449757683438792, "learning_rate": 8.68677022711096e-06, "loss": 0.9027, "step": 8912 }, { "epoch": 4.038513819664703, "grad_norm": 0.10187364632464364, "learning_rate": 8.67889823458688e-06, "loss": 0.9289, "step": 8913 }, { "epoch": 4.038966923425464, "grad_norm": 0.08951119325256561, "learning_rate": 8.671029376436753e-06, "loss": 0.9357, "step": 8914 }, { "epoch": 4.0394200271862255, "grad_norm": 0.11190399644677765, "learning_rate": 8.663163653448028e-06, "loss": 0.927, "step": 8915 }, { "epoch": 4.039873130946987, "grad_norm": 0.08434361783457948, "learning_rate": 8.655301066407831e-06, "loss": 0.9145, "step": 8916 }, { "epoch": 4.040326234707748, "grad_norm": 0.10080701350408336, "learning_rate": 8.647441616103016e-06, "loss": 0.9324, "step": 8917 }, { "epoch": 4.040779338468509, "grad_norm": 0.08963179213442095, "learning_rate": 8.639585303320071e-06, "loss": 0.9294, "step": 8918 }, { "epoch": 4.041232442229271, "grad_norm": 0.0965349767345931, "learning_rate": 8.631732128845214e-06, "loss": 0.9328, "step": 8919 }, { "epoch": 4.041685545990032, "grad_norm": 0.09231364653483042, "learning_rate": 8.62388209346431e-06, "loss": 0.9389, "step": 8920 }, { "epoch": 4.042138649750793, "grad_norm": 0.09764477788956229, "learning_rate": 8.616035197962956e-06, "loss": 0.9076, "step": 8921 }, { "epoch": 4.042591753511554, "grad_norm": 0.0972535691310007, "learning_rate": 8.608191443126395e-06, "loss": 0.9123, "step": 8922 }, { "epoch": 4.043044857272315, "grad_norm": 0.0897203786816313, "learning_rate": 8.600350829739556e-06, "loss": 0.9129, "step": 8923 }, { "epoch": 4.043497961033077, "grad_norm": 0.09149313162703382, "learning_rate": 8.592513358587094e-06, "loss": 0.9317, "step": 8924 }, { "epoch": 4.043951064793838, "grad_norm": 0.09299214849757906, "learning_rate": 8.584679030453293e-06, "loss": 0.91, "step": 8925 }, { "epoch": 4.044404168554599, "grad_norm": 0.09153721938225368, "learning_rate": 8.576847846122183e-06, "loss": 0.9346, "step": 8926 }, { "epoch": 4.04485727231536, "grad_norm": 0.10174940493285448, "learning_rate": 8.569019806377418e-06, "loss": 0.923, "step": 8927 }, { "epoch": 4.045310376076121, "grad_norm": 0.09399621891615233, "learning_rate": 8.561194912002398e-06, "loss": 0.9095, "step": 8928 }, { "epoch": 4.045763479836882, "grad_norm": 0.08876059067536693, "learning_rate": 8.553373163780146e-06, "loss": 0.9153, "step": 8929 }, { "epoch": 4.046216583597644, "grad_norm": 0.10219882700499926, "learning_rate": 8.545554562493437e-06, "loss": 0.9459, "step": 8930 }, { "epoch": 4.046669687358405, "grad_norm": 0.09797922364049087, "learning_rate": 8.537739108924672e-06, "loss": 0.926, "step": 8931 }, { "epoch": 4.0471227911191665, "grad_norm": 0.09558900528395395, "learning_rate": 8.529926803855958e-06, "loss": 0.9244, "step": 8932 }, { "epoch": 4.047575894879928, "grad_norm": 0.09200484545326411, "learning_rate": 8.522117648069113e-06, "loss": 0.9134, "step": 8933 }, { "epoch": 4.048028998640689, "grad_norm": 0.09046766467956098, "learning_rate": 8.514311642345592e-06, "loss": 0.9264, "step": 8934 }, { "epoch": 4.04848210240145, "grad_norm": 0.09089951850398835, "learning_rate": 8.506508787466581e-06, "loss": 0.9137, "step": 8935 }, { "epoch": 4.048935206162211, "grad_norm": 0.09475002478064451, "learning_rate": 8.498709084212914e-06, "loss": 0.9091, "step": 8936 }, { "epoch": 4.049388309922972, "grad_norm": 0.08373166000272765, "learning_rate": 8.490912533365137e-06, "loss": 0.9409, "step": 8937 }, { "epoch": 4.049841413683733, "grad_norm": 0.10970373329861786, "learning_rate": 8.483119135703459e-06, "loss": 0.9187, "step": 8938 }, { "epoch": 4.050294517444494, "grad_norm": 0.08506332774311809, "learning_rate": 8.475328892007781e-06, "loss": 0.9456, "step": 8939 }, { "epoch": 4.050747621205256, "grad_norm": 0.09414708406539948, "learning_rate": 8.467541803057706e-06, "loss": 0.9359, "step": 8940 }, { "epoch": 4.051200724966018, "grad_norm": 0.08982681415259872, "learning_rate": 8.459757869632481e-06, "loss": 0.9182, "step": 8941 }, { "epoch": 4.051653828726779, "grad_norm": 0.08838359358192567, "learning_rate": 8.451977092511092e-06, "loss": 0.9175, "step": 8942 }, { "epoch": 4.05210693248754, "grad_norm": 0.0905713983009871, "learning_rate": 8.444199472472152e-06, "loss": 0.9254, "step": 8943 }, { "epoch": 4.052560036248301, "grad_norm": 0.09577912811516562, "learning_rate": 8.436425010294008e-06, "loss": 0.9311, "step": 8944 }, { "epoch": 4.053013140009062, "grad_norm": 0.09160542151806977, "learning_rate": 8.428653706754644e-06, "loss": 0.9232, "step": 8945 }, { "epoch": 4.053466243769823, "grad_norm": 0.09554358217837046, "learning_rate": 8.420885562631777e-06, "loss": 0.9134, "step": 8946 }, { "epoch": 4.053919347530584, "grad_norm": 0.0925422708092054, "learning_rate": 8.41312057870277e-06, "loss": 0.9104, "step": 8947 }, { "epoch": 4.0543724512913455, "grad_norm": 0.10498407331350532, "learning_rate": 8.405358755744672e-06, "loss": 0.9255, "step": 8948 }, { "epoch": 4.054825555052107, "grad_norm": 0.09272512302555026, "learning_rate": 8.397600094534248e-06, "loss": 0.9319, "step": 8949 }, { "epoch": 4.055278658812868, "grad_norm": 0.11688118992465732, "learning_rate": 8.389844595847908e-06, "loss": 0.9313, "step": 8950 }, { "epoch": 4.05573176257363, "grad_norm": 0.09275536322303594, "learning_rate": 8.382092260461778e-06, "loss": 0.9109, "step": 8951 }, { "epoch": 4.056184866334391, "grad_norm": 0.09575282435705186, "learning_rate": 8.374343089151642e-06, "loss": 0.9236, "step": 8952 }, { "epoch": 4.056637970095152, "grad_norm": 0.08351771530179315, "learning_rate": 8.366597082692976e-06, "loss": 0.9052, "step": 8953 }, { "epoch": 4.057091073855913, "grad_norm": 0.1110450258398474, "learning_rate": 8.358854241860932e-06, "loss": 0.9464, "step": 8954 }, { "epoch": 4.057544177616674, "grad_norm": 0.08561053179248447, "learning_rate": 8.35111456743038e-06, "loss": 0.905, "step": 8955 }, { "epoch": 4.057997281377435, "grad_norm": 0.08957729194519522, "learning_rate": 8.343378060175818e-06, "loss": 0.9205, "step": 8956 }, { "epoch": 4.058450385138197, "grad_norm": 0.09760137034916554, "learning_rate": 8.335644720871477e-06, "loss": 0.9238, "step": 8957 }, { "epoch": 4.058903488898958, "grad_norm": 0.0902184803276101, "learning_rate": 8.327914550291245e-06, "loss": 0.9241, "step": 8958 }, { "epoch": 4.059356592659719, "grad_norm": 0.10216148700281052, "learning_rate": 8.320187549208682e-06, "loss": 0.897, "step": 8959 }, { "epoch": 4.05980969642048, "grad_norm": 0.08913251489969105, "learning_rate": 8.312463718397068e-06, "loss": 0.9172, "step": 8960 }, { "epoch": 4.060262800181241, "grad_norm": 0.10777873109178239, "learning_rate": 8.30474305862933e-06, "loss": 0.9348, "step": 8961 }, { "epoch": 4.060715903942003, "grad_norm": 0.08294460557605254, "learning_rate": 8.297025570678103e-06, "loss": 0.9104, "step": 8962 }, { "epoch": 4.061169007702764, "grad_norm": 0.09892854403119473, "learning_rate": 8.289311255315682e-06, "loss": 0.9405, "step": 8963 }, { "epoch": 4.061622111463525, "grad_norm": 0.08823415684701187, "learning_rate": 8.281600113314069e-06, "loss": 0.9308, "step": 8964 }, { "epoch": 4.0620752152242865, "grad_norm": 0.09991496150432853, "learning_rate": 8.273892145444922e-06, "loss": 0.9286, "step": 8965 }, { "epoch": 4.062528318985048, "grad_norm": 0.09555391746440825, "learning_rate": 8.266187352479612e-06, "loss": 0.9151, "step": 8966 }, { "epoch": 4.062981422745809, "grad_norm": 0.10069087512965587, "learning_rate": 8.258485735189161e-06, "loss": 0.905, "step": 8967 }, { "epoch": 4.06343452650657, "grad_norm": 0.09264584041405038, "learning_rate": 8.25078729434428e-06, "loss": 0.9214, "step": 8968 }, { "epoch": 4.063887630267331, "grad_norm": 0.09740108250355933, "learning_rate": 8.243092030715396e-06, "loss": 0.9054, "step": 8969 }, { "epoch": 4.064340734028092, "grad_norm": 0.09338753130936805, "learning_rate": 8.235399945072564e-06, "loss": 0.9132, "step": 8970 }, { "epoch": 4.064793837788853, "grad_norm": 0.09556331470804393, "learning_rate": 8.22771103818557e-06, "loss": 0.905, "step": 8971 }, { "epoch": 4.065246941549615, "grad_norm": 0.10491680130837636, "learning_rate": 8.220025310823842e-06, "loss": 0.9291, "step": 8972 }, { "epoch": 4.065700045310376, "grad_norm": 0.08931623386690905, "learning_rate": 8.212342763756526e-06, "loss": 0.9335, "step": 8973 }, { "epoch": 4.066153149071138, "grad_norm": 0.09399851110041893, "learning_rate": 8.204663397752415e-06, "loss": 0.9223, "step": 8974 }, { "epoch": 4.066606252831899, "grad_norm": 0.10135700690220882, "learning_rate": 8.196987213580017e-06, "loss": 0.9322, "step": 8975 }, { "epoch": 4.06705935659266, "grad_norm": 0.09801088755281732, "learning_rate": 8.1893142120075e-06, "loss": 0.9262, "step": 8976 }, { "epoch": 4.067512460353421, "grad_norm": 0.09822761602038169, "learning_rate": 8.181644393802704e-06, "loss": 0.929, "step": 8977 }, { "epoch": 4.067965564114182, "grad_norm": 0.10211881658799696, "learning_rate": 8.173977759733187e-06, "loss": 0.9238, "step": 8978 }, { "epoch": 4.068418667874943, "grad_norm": 0.09264860332039543, "learning_rate": 8.166314310566146e-06, "loss": 0.9171, "step": 8979 }, { "epoch": 4.068871771635704, "grad_norm": 0.10077587091797148, "learning_rate": 8.158654047068504e-06, "loss": 0.917, "step": 8980 }, { "epoch": 4.0693248753964655, "grad_norm": 0.0937319639168501, "learning_rate": 8.150996970006813e-06, "loss": 0.9278, "step": 8981 }, { "epoch": 4.069777979157227, "grad_norm": 0.10149312930800289, "learning_rate": 8.143343080147362e-06, "loss": 0.9236, "step": 8982 }, { "epoch": 4.070231082917989, "grad_norm": 0.09305606890341894, "learning_rate": 8.13569237825608e-06, "loss": 0.9014, "step": 8983 }, { "epoch": 4.07068418667875, "grad_norm": 0.08871255404077655, "learning_rate": 8.128044865098577e-06, "loss": 0.9502, "step": 8984 }, { "epoch": 4.071137290439511, "grad_norm": 0.0956413051215034, "learning_rate": 8.120400541440183e-06, "loss": 0.9375, "step": 8985 }, { "epoch": 4.071590394200272, "grad_norm": 0.09818994843395235, "learning_rate": 8.112759408045856e-06, "loss": 0.9199, "step": 8986 }, { "epoch": 4.072043497961033, "grad_norm": 0.09675463665476616, "learning_rate": 8.10512146568029e-06, "loss": 0.9318, "step": 8987 }, { "epoch": 4.072496601721794, "grad_norm": 0.10166904268649606, "learning_rate": 8.097486715107803e-06, "loss": 0.9614, "step": 8988 }, { "epoch": 4.072949705482555, "grad_norm": 0.1060565383631579, "learning_rate": 8.089855157092445e-06, "loss": 0.9151, "step": 8989 }, { "epoch": 4.0734028092433165, "grad_norm": 0.08845940626063216, "learning_rate": 8.082226792397904e-06, "loss": 0.9399, "step": 8990 }, { "epoch": 4.073855913004078, "grad_norm": 0.10679196418834382, "learning_rate": 8.074601621787593e-06, "loss": 0.927, "step": 8991 }, { "epoch": 4.074309016764839, "grad_norm": 0.0897989694769874, "learning_rate": 8.06697964602456e-06, "loss": 0.9424, "step": 8992 }, { "epoch": 4.0747621205256, "grad_norm": 0.09858075322156751, "learning_rate": 8.059360865871553e-06, "loss": 0.9423, "step": 8993 }, { "epoch": 4.075215224286362, "grad_norm": 0.0965514055600343, "learning_rate": 8.05174528209101e-06, "loss": 0.925, "step": 8994 }, { "epoch": 4.075668328047123, "grad_norm": 0.09152016241715741, "learning_rate": 8.04413289544503e-06, "loss": 0.9176, "step": 8995 }, { "epoch": 4.076121431807884, "grad_norm": 0.08950390987430905, "learning_rate": 8.036523706695419e-06, "loss": 0.9192, "step": 8996 }, { "epoch": 4.076574535568645, "grad_norm": 0.08815283437635305, "learning_rate": 8.028917716603621e-06, "loss": 0.9059, "step": 8997 }, { "epoch": 4.0770276393294065, "grad_norm": 0.09116078402024096, "learning_rate": 8.021314925930807e-06, "loss": 0.9112, "step": 8998 }, { "epoch": 4.077480743090168, "grad_norm": 0.0908631736549905, "learning_rate": 8.0137153354378e-06, "loss": 0.9319, "step": 8999 }, { "epoch": 4.077933846850929, "grad_norm": 0.09328004719852494, "learning_rate": 8.006118945885096e-06, "loss": 0.9112, "step": 9000 }, { "epoch": 4.07838695061169, "grad_norm": 0.08363302589678183, "learning_rate": 7.9985257580329e-06, "loss": 0.9378, "step": 9001 }, { "epoch": 4.078840054372451, "grad_norm": 0.0960455884945005, "learning_rate": 7.990935772641065e-06, "loss": 0.9219, "step": 9002 }, { "epoch": 4.079293158133212, "grad_norm": 0.0944209646014018, "learning_rate": 7.983348990469152e-06, "loss": 0.942, "step": 9003 }, { "epoch": 4.079746261893973, "grad_norm": 0.09859263797676263, "learning_rate": 7.975765412276368e-06, "loss": 0.9202, "step": 9004 }, { "epoch": 4.080199365654735, "grad_norm": 0.08841443083239223, "learning_rate": 7.968185038821646e-06, "loss": 0.9103, "step": 9005 }, { "epoch": 4.080652469415496, "grad_norm": 0.09136774766991743, "learning_rate": 7.960607870863545e-06, "loss": 0.9086, "step": 9006 }, { "epoch": 4.0811055731762576, "grad_norm": 0.09013690903633474, "learning_rate": 7.953033909160353e-06, "loss": 0.9298, "step": 9007 }, { "epoch": 4.081558676937019, "grad_norm": 0.08999248593919253, "learning_rate": 7.945463154469996e-06, "loss": 0.923, "step": 9008 }, { "epoch": 4.08201178069778, "grad_norm": 0.08788829414791248, "learning_rate": 7.937895607550108e-06, "loss": 0.9343, "step": 9009 }, { "epoch": 4.082464884458541, "grad_norm": 0.08575649295878338, "learning_rate": 7.930331269157974e-06, "loss": 0.942, "step": 9010 }, { "epoch": 4.082917988219302, "grad_norm": 0.08917672918502469, "learning_rate": 7.922770140050594e-06, "loss": 0.9421, "step": 9011 }, { "epoch": 4.083371091980063, "grad_norm": 0.08719736915549361, "learning_rate": 7.915212220984619e-06, "loss": 0.91, "step": 9012 }, { "epoch": 4.083824195740824, "grad_norm": 0.09244745025115476, "learning_rate": 7.907657512716378e-06, "loss": 0.9047, "step": 9013 }, { "epoch": 4.0842772995015855, "grad_norm": 0.08445650198577653, "learning_rate": 7.900106016001906e-06, "loss": 0.9137, "step": 9014 }, { "epoch": 4.0847304032623475, "grad_norm": 0.08528431469296963, "learning_rate": 7.892557731596882e-06, "loss": 0.9223, "step": 9015 }, { "epoch": 4.085183507023109, "grad_norm": 0.09489979423657136, "learning_rate": 7.885012660256696e-06, "loss": 0.9063, "step": 9016 }, { "epoch": 4.08563661078387, "grad_norm": 0.08811393278670553, "learning_rate": 7.87747080273638e-06, "loss": 0.9259, "step": 9017 }, { "epoch": 4.086089714544631, "grad_norm": 0.09604189271513551, "learning_rate": 7.869932159790687e-06, "loss": 0.9311, "step": 9018 }, { "epoch": 4.086542818305392, "grad_norm": 0.09170560065996043, "learning_rate": 7.862396732174016e-06, "loss": 0.9137, "step": 9019 }, { "epoch": 4.086995922066153, "grad_norm": 0.08474508578128107, "learning_rate": 7.854864520640446e-06, "loss": 0.9033, "step": 9020 }, { "epoch": 4.087449025826914, "grad_norm": 0.0963411968569839, "learning_rate": 7.847335525943763e-06, "loss": 0.9288, "step": 9021 }, { "epoch": 4.087902129587675, "grad_norm": 0.09028484555423733, "learning_rate": 7.839809748837384e-06, "loss": 0.9382, "step": 9022 }, { "epoch": 4.0883552333484365, "grad_norm": 0.09199858231281789, "learning_rate": 7.832287190074455e-06, "loss": 0.9154, "step": 9023 }, { "epoch": 4.088808337109198, "grad_norm": 0.09314679368378338, "learning_rate": 7.824767850407755e-06, "loss": 0.9311, "step": 9024 }, { "epoch": 4.089261440869959, "grad_norm": 0.09038010723653618, "learning_rate": 7.817251730589786e-06, "loss": 0.9187, "step": 9025 }, { "epoch": 4.089714544630721, "grad_norm": 0.1032373328153095, "learning_rate": 7.809738831372677e-06, "loss": 0.9329, "step": 9026 }, { "epoch": 4.090167648391482, "grad_norm": 0.08796225004801037, "learning_rate": 7.802229153508283e-06, "loss": 0.9225, "step": 9027 }, { "epoch": 4.090620752152243, "grad_norm": 0.08867488910309719, "learning_rate": 7.794722697748107e-06, "loss": 0.9074, "step": 9028 }, { "epoch": 4.091073855913004, "grad_norm": 0.0967766240974072, "learning_rate": 7.787219464843324e-06, "loss": 0.9329, "step": 9029 }, { "epoch": 4.091526959673765, "grad_norm": 0.08412739439675436, "learning_rate": 7.779719455544819e-06, "loss": 0.9229, "step": 9030 }, { "epoch": 4.0919800634345265, "grad_norm": 0.0980755289514933, "learning_rate": 7.772222670603118e-06, "loss": 0.9356, "step": 9031 }, { "epoch": 4.092433167195288, "grad_norm": 0.09283321573480784, "learning_rate": 7.764729110768462e-06, "loss": 0.943, "step": 9032 }, { "epoch": 4.092886270956049, "grad_norm": 0.09330168367343443, "learning_rate": 7.757238776790722e-06, "loss": 0.9221, "step": 9033 }, { "epoch": 4.09333937471681, "grad_norm": 0.09476960156771491, "learning_rate": 7.7497516694195e-06, "loss": 0.9293, "step": 9034 }, { "epoch": 4.093792478477571, "grad_norm": 0.08825383048109214, "learning_rate": 7.742267789404034e-06, "loss": 0.9317, "step": 9035 }, { "epoch": 4.094245582238333, "grad_norm": 0.09179436779253415, "learning_rate": 7.734787137493245e-06, "loss": 0.922, "step": 9036 }, { "epoch": 4.094698685999094, "grad_norm": 0.08786660640603583, "learning_rate": 7.727309714435756e-06, "loss": 0.9239, "step": 9037 }, { "epoch": 4.095151789759855, "grad_norm": 0.09032604638706121, "learning_rate": 7.719835520979831e-06, "loss": 0.9281, "step": 9038 }, { "epoch": 4.095604893520616, "grad_norm": 0.100230785805892, "learning_rate": 7.712364557873454e-06, "loss": 0.9186, "step": 9039 }, { "epoch": 4.0960579972813775, "grad_norm": 0.09044768824586898, "learning_rate": 7.704896825864234e-06, "loss": 0.9172, "step": 9040 }, { "epoch": 4.096511101042139, "grad_norm": 0.0963171389570885, "learning_rate": 7.697432325699509e-06, "loss": 0.9409, "step": 9041 }, { "epoch": 4.0969642048029, "grad_norm": 0.0975486449223849, "learning_rate": 7.689971058126247e-06, "loss": 0.9272, "step": 9042 }, { "epoch": 4.097417308563661, "grad_norm": 0.09635232735094781, "learning_rate": 7.682513023891132e-06, "loss": 0.9038, "step": 9043 }, { "epoch": 4.097870412324422, "grad_norm": 0.09832872357922409, "learning_rate": 7.675058223740501e-06, "loss": 0.9299, "step": 9044 }, { "epoch": 4.098323516085183, "grad_norm": 0.10613094866039903, "learning_rate": 7.667606658420359e-06, "loss": 0.9256, "step": 9045 }, { "epoch": 4.098776619845944, "grad_norm": 0.08867774374436613, "learning_rate": 7.660158328676424e-06, "loss": 0.9318, "step": 9046 }, { "epoch": 4.099229723606706, "grad_norm": 0.10138410582061812, "learning_rate": 7.65271323525404e-06, "loss": 0.9265, "step": 9047 }, { "epoch": 4.0996828273674675, "grad_norm": 0.10471657457660051, "learning_rate": 7.645271378898287e-06, "loss": 0.9411, "step": 9048 }, { "epoch": 4.100135931128229, "grad_norm": 0.10256077588570728, "learning_rate": 7.637832760353858e-06, "loss": 0.9232, "step": 9049 }, { "epoch": 4.10058903488899, "grad_norm": 0.10220982361289824, "learning_rate": 7.630397380365177e-06, "loss": 0.9384, "step": 9050 }, { "epoch": 4.101042138649751, "grad_norm": 0.09246660755589207, "learning_rate": 7.6229652396762985e-06, "loss": 0.9, "step": 9051 }, { "epoch": 4.101495242410512, "grad_norm": 0.09493932701623775, "learning_rate": 7.6155363390309955e-06, "loss": 0.9402, "step": 9052 }, { "epoch": 4.101948346171273, "grad_norm": 0.09445593570100638, "learning_rate": 7.608110679172682e-06, "loss": 0.9486, "step": 9053 }, { "epoch": 4.102401449932034, "grad_norm": 0.09785071242922383, "learning_rate": 7.600688260844449e-06, "loss": 0.9393, "step": 9054 }, { "epoch": 4.102854553692795, "grad_norm": 0.10055456254978047, "learning_rate": 7.593269084789101e-06, "loss": 0.9368, "step": 9055 }, { "epoch": 4.1033076574535565, "grad_norm": 0.08202774087525448, "learning_rate": 7.585853151749067e-06, "loss": 0.9181, "step": 9056 }, { "epoch": 4.103760761214318, "grad_norm": 0.0953033832712986, "learning_rate": 7.578440462466496e-06, "loss": 0.9471, "step": 9057 }, { "epoch": 4.10421386497508, "grad_norm": 0.0924699606046603, "learning_rate": 7.571031017683177e-06, "loss": 0.9123, "step": 9058 }, { "epoch": 4.104666968735841, "grad_norm": 0.09115730301424313, "learning_rate": 7.563624818140609e-06, "loss": 0.938, "step": 9059 }, { "epoch": 4.105120072496602, "grad_norm": 0.0940933008672301, "learning_rate": 7.556221864579933e-06, "loss": 0.9176, "step": 9060 }, { "epoch": 4.105573176257363, "grad_norm": 0.0894667487972616, "learning_rate": 7.548822157741971e-06, "loss": 0.9355, "step": 9061 }, { "epoch": 4.106026280018124, "grad_norm": 0.08745343445241158, "learning_rate": 7.541425698367253e-06, "loss": 0.9274, "step": 9062 }, { "epoch": 4.106479383778885, "grad_norm": 0.0847691162330356, "learning_rate": 7.534032487195935e-06, "loss": 0.9097, "step": 9063 }, { "epoch": 4.1069324875396465, "grad_norm": 0.08454147807255163, "learning_rate": 7.526642524967895e-06, "loss": 0.9123, "step": 9064 }, { "epoch": 4.107385591300408, "grad_norm": 0.08732266154740526, "learning_rate": 7.519255812422641e-06, "loss": 0.9449, "step": 9065 }, { "epoch": 4.107838695061169, "grad_norm": 0.07915504808561334, "learning_rate": 7.511872350299407e-06, "loss": 0.9144, "step": 9066 }, { "epoch": 4.10829179882193, "grad_norm": 0.08435569803015301, "learning_rate": 7.504492139337038e-06, "loss": 0.9208, "step": 9067 }, { "epoch": 4.108744902582692, "grad_norm": 0.08601899777907653, "learning_rate": 7.497115180274113e-06, "loss": 0.9046, "step": 9068 }, { "epoch": 4.109198006343453, "grad_norm": 0.08482980817072015, "learning_rate": 7.4897414738488435e-06, "loss": 0.9217, "step": 9069 }, { "epoch": 4.109651110104214, "grad_norm": 0.08868378334751584, "learning_rate": 7.482371020799157e-06, "loss": 0.9355, "step": 9070 }, { "epoch": 4.110104213864975, "grad_norm": 0.08882070887226319, "learning_rate": 7.4750038218626054e-06, "loss": 0.9188, "step": 9071 }, { "epoch": 4.110557317625736, "grad_norm": 0.08930546601809985, "learning_rate": 7.46763987777646e-06, "loss": 0.927, "step": 9072 }, { "epoch": 4.1110104213864975, "grad_norm": 0.09162969460065197, "learning_rate": 7.460279189277648e-06, "loss": 0.9344, "step": 9073 }, { "epoch": 4.111463525147259, "grad_norm": 0.08740045427668648, "learning_rate": 7.452921757102749e-06, "loss": 0.8988, "step": 9074 }, { "epoch": 4.11191662890802, "grad_norm": 0.08989563114241307, "learning_rate": 7.445567581988063e-06, "loss": 0.9279, "step": 9075 }, { "epoch": 4.112369732668781, "grad_norm": 0.09655592130788172, "learning_rate": 7.438216664669516e-06, "loss": 0.9314, "step": 9076 }, { "epoch": 4.112822836429542, "grad_norm": 0.08996152175814617, "learning_rate": 7.4308690058827544e-06, "loss": 0.9383, "step": 9077 }, { "epoch": 4.113275940190303, "grad_norm": 0.08762397074684446, "learning_rate": 7.423524606363051e-06, "loss": 0.9342, "step": 9078 }, { "epoch": 4.113729043951065, "grad_norm": 0.09297883065192956, "learning_rate": 7.416183466845406e-06, "loss": 0.9312, "step": 9079 }, { "epoch": 4.114182147711826, "grad_norm": 0.09063504469003628, "learning_rate": 7.408845588064441e-06, "loss": 0.9382, "step": 9080 }, { "epoch": 4.1146352514725875, "grad_norm": 0.09411947029433129, "learning_rate": 7.401510970754473e-06, "loss": 0.9108, "step": 9081 }, { "epoch": 4.115088355233349, "grad_norm": 0.08684995247146184, "learning_rate": 7.394179615649513e-06, "loss": 0.9157, "step": 9082 }, { "epoch": 4.11554145899411, "grad_norm": 0.09478877232172046, "learning_rate": 7.386851523483205e-06, "loss": 0.9292, "step": 9083 }, { "epoch": 4.115994562754871, "grad_norm": 0.09651437547058762, "learning_rate": 7.379526694988906e-06, "loss": 0.9255, "step": 9084 }, { "epoch": 4.116447666515632, "grad_norm": 0.09287569067172954, "learning_rate": 7.37220513089961e-06, "loss": 0.9244, "step": 9085 }, { "epoch": 4.116900770276393, "grad_norm": 0.09738471069393152, "learning_rate": 7.364886831948022e-06, "loss": 0.9284, "step": 9086 }, { "epoch": 4.117353874037154, "grad_norm": 0.09178251208898455, "learning_rate": 7.3575717988664826e-06, "loss": 0.9283, "step": 9087 }, { "epoch": 4.117806977797915, "grad_norm": 0.09800096110113946, "learning_rate": 7.35026003238704e-06, "loss": 0.9101, "step": 9088 }, { "epoch": 4.1182600815586765, "grad_norm": 0.08911636340980615, "learning_rate": 7.342951533241396e-06, "loss": 0.9462, "step": 9089 }, { "epoch": 4.1187131853194385, "grad_norm": 0.09425998203666887, "learning_rate": 7.335646302160913e-06, "loss": 0.9371, "step": 9090 }, { "epoch": 4.1191662890802, "grad_norm": 0.0838656891488178, "learning_rate": 7.328344339876663e-06, "loss": 0.9104, "step": 9091 }, { "epoch": 4.119619392840961, "grad_norm": 0.09389239338814674, "learning_rate": 7.321045647119351e-06, "loss": 0.9254, "step": 9092 }, { "epoch": 4.120072496601722, "grad_norm": 0.08674415118025307, "learning_rate": 7.313750224619394e-06, "loss": 0.9212, "step": 9093 }, { "epoch": 4.120525600362483, "grad_norm": 0.08490157068086512, "learning_rate": 7.3064580731068415e-06, "loss": 0.9149, "step": 9094 }, { "epoch": 4.120978704123244, "grad_norm": 0.09207131139780463, "learning_rate": 7.299169193311453e-06, "loss": 0.9188, "step": 9095 }, { "epoch": 4.121431807884005, "grad_norm": 0.08422756968293688, "learning_rate": 7.29188358596264e-06, "loss": 0.9291, "step": 9096 }, { "epoch": 4.1218849116447664, "grad_norm": 0.0807729772076309, "learning_rate": 7.2846012517894695e-06, "loss": 0.9367, "step": 9097 }, { "epoch": 4.122338015405528, "grad_norm": 0.08568765174990095, "learning_rate": 7.277322191520731e-06, "loss": 0.9388, "step": 9098 }, { "epoch": 4.122791119166289, "grad_norm": 0.08223182116124883, "learning_rate": 7.270046405884832e-06, "loss": 0.9383, "step": 9099 }, { "epoch": 4.123244222927051, "grad_norm": 0.09294442041017807, "learning_rate": 7.2627738956098985e-06, "loss": 0.9186, "step": 9100 }, { "epoch": 4.123697326687812, "grad_norm": 0.09280754293727966, "learning_rate": 7.255504661423684e-06, "loss": 0.9303, "step": 9101 }, { "epoch": 4.124150430448573, "grad_norm": 0.09013054316403948, "learning_rate": 7.248238704053663e-06, "loss": 0.8986, "step": 9102 }, { "epoch": 4.124603534209334, "grad_norm": 0.09000230863618258, "learning_rate": 7.240976024226931e-06, "loss": 0.913, "step": 9103 }, { "epoch": 4.125056637970095, "grad_norm": 0.08247126142363916, "learning_rate": 7.233716622670304e-06, "loss": 0.9099, "step": 9104 }, { "epoch": 4.125509741730856, "grad_norm": 0.08603154548161579, "learning_rate": 7.226460500110235e-06, "loss": 0.9239, "step": 9105 }, { "epoch": 4.1259628454916175, "grad_norm": 0.08277739541400213, "learning_rate": 7.219207657272851e-06, "loss": 0.9206, "step": 9106 }, { "epoch": 4.126415949252379, "grad_norm": 0.10323145608699001, "learning_rate": 7.211958094883984e-06, "loss": 0.9366, "step": 9107 }, { "epoch": 4.12686905301314, "grad_norm": 0.0893584517462771, "learning_rate": 7.204711813669085e-06, "loss": 0.9134, "step": 9108 }, { "epoch": 4.127322156773901, "grad_norm": 0.10187878529420061, "learning_rate": 7.197468814353339e-06, "loss": 0.9206, "step": 9109 }, { "epoch": 4.127775260534662, "grad_norm": 0.08998639862551014, "learning_rate": 7.190229097661539e-06, "loss": 0.9089, "step": 9110 }, { "epoch": 4.128228364295424, "grad_norm": 0.09299752729435412, "learning_rate": 7.1829926643182025e-06, "loss": 0.9221, "step": 9111 }, { "epoch": 4.128681468056185, "grad_norm": 0.08356848506167144, "learning_rate": 7.17575951504748e-06, "loss": 0.9118, "step": 9112 }, { "epoch": 4.129134571816946, "grad_norm": 0.091916251517685, "learning_rate": 7.1685296505732234e-06, "loss": 0.9156, "step": 9113 }, { "epoch": 4.1295876755777075, "grad_norm": 0.09699293301265133, "learning_rate": 7.161303071618935e-06, "loss": 0.9406, "step": 9114 }, { "epoch": 4.130040779338469, "grad_norm": 0.09379512130801697, "learning_rate": 7.1540797789077855e-06, "loss": 0.9297, "step": 9115 }, { "epoch": 4.13049388309923, "grad_norm": 0.10500235402726861, "learning_rate": 7.146859773162642e-06, "loss": 0.9417, "step": 9116 }, { "epoch": 4.130946986859991, "grad_norm": 0.09357797118112034, "learning_rate": 7.1396430551060115e-06, "loss": 0.9284, "step": 9117 }, { "epoch": 4.131400090620752, "grad_norm": 0.09666416979539336, "learning_rate": 7.1324296254601065e-06, "loss": 0.9248, "step": 9118 }, { "epoch": 4.131853194381513, "grad_norm": 0.08408060353795402, "learning_rate": 7.1252194849467685e-06, "loss": 0.9255, "step": 9119 }, { "epoch": 4.132306298142274, "grad_norm": 0.09683864666874051, "learning_rate": 7.118012634287557e-06, "loss": 0.9373, "step": 9120 }, { "epoch": 4.132759401903035, "grad_norm": 0.08505360472703012, "learning_rate": 7.110809074203664e-06, "loss": 0.8859, "step": 9121 }, { "epoch": 4.133212505663797, "grad_norm": 0.09023194721153324, "learning_rate": 7.103608805415962e-06, "loss": 0.936, "step": 9122 }, { "epoch": 4.1336656094245585, "grad_norm": 0.0875434954633943, "learning_rate": 7.096411828645009e-06, "loss": 0.912, "step": 9123 }, { "epoch": 4.13411871318532, "grad_norm": 0.09075287721167566, "learning_rate": 7.089218144611023e-06, "loss": 0.9332, "step": 9124 }, { "epoch": 4.134571816946081, "grad_norm": 0.08447086186524522, "learning_rate": 7.082027754033882e-06, "loss": 0.9259, "step": 9125 }, { "epoch": 4.135024920706842, "grad_norm": 0.0911459544193304, "learning_rate": 7.0748406576331444e-06, "loss": 0.9255, "step": 9126 }, { "epoch": 4.135478024467603, "grad_norm": 0.08985878202355611, "learning_rate": 7.067656856128056e-06, "loss": 0.9288, "step": 9127 }, { "epoch": 4.135931128228364, "grad_norm": 0.08616887878027683, "learning_rate": 7.060476350237491e-06, "loss": 0.9305, "step": 9128 }, { "epoch": 4.136384231989125, "grad_norm": 0.0955753005457424, "learning_rate": 7.0532991406800435e-06, "loss": 0.9195, "step": 9129 }, { "epoch": 4.136837335749886, "grad_norm": 0.08794665773091075, "learning_rate": 7.046125228173935e-06, "loss": 0.9142, "step": 9130 }, { "epoch": 4.137290439510648, "grad_norm": 0.08772989447790787, "learning_rate": 7.038954613437092e-06, "loss": 0.9293, "step": 9131 }, { "epoch": 4.137743543271409, "grad_norm": 0.09597114922481892, "learning_rate": 7.031787297187076e-06, "loss": 0.9223, "step": 9132 }, { "epoch": 4.138196647032171, "grad_norm": 0.08417241271349009, "learning_rate": 7.0246232801411566e-06, "loss": 0.9278, "step": 9133 }, { "epoch": 4.138649750792932, "grad_norm": 0.08808152876725915, "learning_rate": 7.017462563016244e-06, "loss": 0.9297, "step": 9134 }, { "epoch": 4.139102854553693, "grad_norm": 0.09052054771512795, "learning_rate": 7.010305146528913e-06, "loss": 0.9383, "step": 9135 }, { "epoch": 4.139555958314454, "grad_norm": 0.09215496900821676, "learning_rate": 7.0031510313954476e-06, "loss": 0.9169, "step": 9136 }, { "epoch": 4.140009062075215, "grad_norm": 0.09565654512257435, "learning_rate": 6.996000218331751e-06, "loss": 0.9219, "step": 9137 }, { "epoch": 4.140462165835976, "grad_norm": 0.09175938553979518, "learning_rate": 6.98885270805345e-06, "loss": 0.9653, "step": 9138 }, { "epoch": 4.1409152695967375, "grad_norm": 0.09727215272506663, "learning_rate": 6.981708501275784e-06, "loss": 0.9243, "step": 9139 }, { "epoch": 4.141368373357499, "grad_norm": 0.09251674042154827, "learning_rate": 6.9745675987137156e-06, "loss": 0.9264, "step": 9140 }, { "epoch": 4.14182147711826, "grad_norm": 0.08737212948001544, "learning_rate": 6.967430001081834e-06, "loss": 0.9258, "step": 9141 }, { "epoch": 4.142274580879021, "grad_norm": 0.09478434989270192, "learning_rate": 6.960295709094413e-06, "loss": 0.916, "step": 9142 }, { "epoch": 4.142727684639783, "grad_norm": 0.0859229434221616, "learning_rate": 6.953164723465416e-06, "loss": 0.9357, "step": 9143 }, { "epoch": 4.143180788400544, "grad_norm": 0.08794834708126746, "learning_rate": 6.946037044908434e-06, "loss": 0.9207, "step": 9144 }, { "epoch": 4.143633892161305, "grad_norm": 0.09198879610974207, "learning_rate": 6.9389126741367685e-06, "loss": 0.9096, "step": 9145 }, { "epoch": 4.144086995922066, "grad_norm": 0.0945643507555747, "learning_rate": 6.931791611863352e-06, "loss": 0.9274, "step": 9146 }, { "epoch": 4.144540099682827, "grad_norm": 0.0883492091437255, "learning_rate": 6.924673858800832e-06, "loss": 0.9134, "step": 9147 }, { "epoch": 4.144993203443589, "grad_norm": 0.08991674501535875, "learning_rate": 6.91755941566147e-06, "loss": 0.9193, "step": 9148 }, { "epoch": 4.14544630720435, "grad_norm": 0.09132133209704837, "learning_rate": 6.9104482831572515e-06, "loss": 0.9365, "step": 9149 }, { "epoch": 4.145899410965111, "grad_norm": 0.09077785981129231, "learning_rate": 6.90334046199979e-06, "loss": 0.9187, "step": 9150 }, { "epoch": 4.146352514725872, "grad_norm": 0.09677065899407894, "learning_rate": 6.89623595290037e-06, "loss": 0.9104, "step": 9151 }, { "epoch": 4.146805618486633, "grad_norm": 0.09660760672893821, "learning_rate": 6.889134756569977e-06, "loss": 0.9434, "step": 9152 }, { "epoch": 4.147258722247394, "grad_norm": 0.09003499869326621, "learning_rate": 6.8820368737192264e-06, "loss": 0.9325, "step": 9153 }, { "epoch": 4.147711826008156, "grad_norm": 0.1010493109305365, "learning_rate": 6.874942305058434e-06, "loss": 0.9482, "step": 9154 }, { "epoch": 4.148164929768917, "grad_norm": 0.09560005842919551, "learning_rate": 6.867851051297556e-06, "loss": 0.9195, "step": 9155 }, { "epoch": 4.1486180335296785, "grad_norm": 0.11630888225789422, "learning_rate": 6.860763113146247e-06, "loss": 0.9562, "step": 9156 }, { "epoch": 4.14907113729044, "grad_norm": 0.10054866408137973, "learning_rate": 6.853678491313806e-06, "loss": 0.9042, "step": 9157 }, { "epoch": 4.149524241051201, "grad_norm": 0.08990135326204494, "learning_rate": 6.846597186509192e-06, "loss": 0.9206, "step": 9158 }, { "epoch": 4.149977344811962, "grad_norm": 0.11169221617462957, "learning_rate": 6.839519199441067e-06, "loss": 0.9247, "step": 9159 }, { "epoch": 4.150430448572723, "grad_norm": 0.0902849433541655, "learning_rate": 6.832444530817728e-06, "loss": 0.9462, "step": 9160 }, { "epoch": 4.150883552333484, "grad_norm": 0.09066224509564254, "learning_rate": 6.825373181347168e-06, "loss": 0.9045, "step": 9161 }, { "epoch": 4.151336656094245, "grad_norm": 0.09413137897322767, "learning_rate": 6.818305151737017e-06, "loss": 0.9006, "step": 9162 }, { "epoch": 4.151789759855006, "grad_norm": 0.09457071412305773, "learning_rate": 6.811240442694602e-06, "loss": 0.9316, "step": 9163 }, { "epoch": 4.152242863615768, "grad_norm": 0.091560478749308, "learning_rate": 6.804179054926892e-06, "loss": 0.9256, "step": 9164 }, { "epoch": 4.15269596737653, "grad_norm": 0.09727280716911754, "learning_rate": 6.797120989140551e-06, "loss": 0.9248, "step": 9165 }, { "epoch": 4.153149071137291, "grad_norm": 0.08829175759088005, "learning_rate": 6.790066246041891e-06, "loss": 0.9403, "step": 9166 }, { "epoch": 4.153602174898052, "grad_norm": 0.0872376487857295, "learning_rate": 6.78301482633688e-06, "loss": 0.9108, "step": 9167 }, { "epoch": 4.154055278658813, "grad_norm": 0.09342776408163574, "learning_rate": 6.775966730731194e-06, "loss": 0.9153, "step": 9168 }, { "epoch": 4.154508382419574, "grad_norm": 0.09281982675793117, "learning_rate": 6.768921959930126e-06, "loss": 0.9357, "step": 9169 }, { "epoch": 4.154961486180335, "grad_norm": 0.08329888458635878, "learning_rate": 6.76188051463869e-06, "loss": 0.9384, "step": 9170 }, { "epoch": 4.155414589941096, "grad_norm": 0.08440957476993663, "learning_rate": 6.754842395561518e-06, "loss": 0.9397, "step": 9171 }, { "epoch": 4.1558676937018575, "grad_norm": 0.08180424628763261, "learning_rate": 6.747807603402945e-06, "loss": 0.9171, "step": 9172 }, { "epoch": 4.156320797462619, "grad_norm": 0.08004977264759267, "learning_rate": 6.7407761388669445e-06, "loss": 0.9232, "step": 9173 }, { "epoch": 4.15677390122338, "grad_norm": 0.07767591777967374, "learning_rate": 6.73374800265719e-06, "loss": 0.923, "step": 9174 }, { "epoch": 4.157227004984142, "grad_norm": 0.08970573179224114, "learning_rate": 6.7267231954769895e-06, "loss": 0.9154, "step": 9175 }, { "epoch": 4.157680108744903, "grad_norm": 0.08901796056943391, "learning_rate": 6.719701718029323e-06, "loss": 0.9289, "step": 9176 }, { "epoch": 4.158133212505664, "grad_norm": 0.08739410066063699, "learning_rate": 6.7126835710168694e-06, "loss": 0.9096, "step": 9177 }, { "epoch": 4.158586316266425, "grad_norm": 0.08796902082680584, "learning_rate": 6.7056687551419274e-06, "loss": 0.9225, "step": 9178 }, { "epoch": 4.159039420027186, "grad_norm": 0.08530026098001232, "learning_rate": 6.6986572711065055e-06, "loss": 0.9285, "step": 9179 }, { "epoch": 4.159492523787947, "grad_norm": 0.08283159180816402, "learning_rate": 6.691649119612243e-06, "loss": 0.9199, "step": 9180 }, { "epoch": 4.159945627548709, "grad_norm": 0.08626949591176625, "learning_rate": 6.684644301360484e-06, "loss": 0.9215, "step": 9181 }, { "epoch": 4.16039873130947, "grad_norm": 0.09281920325324287, "learning_rate": 6.677642817052183e-06, "loss": 0.9136, "step": 9182 }, { "epoch": 4.160851835070231, "grad_norm": 0.08534847038179037, "learning_rate": 6.670644667388022e-06, "loss": 0.9297, "step": 9183 }, { "epoch": 4.161304938830992, "grad_norm": 0.08354730682821819, "learning_rate": 6.663649853068302e-06, "loss": 0.9047, "step": 9184 }, { "epoch": 4.161758042591753, "grad_norm": 0.08475533980978012, "learning_rate": 6.6566583747930305e-06, "loss": 0.933, "step": 9185 }, { "epoch": 4.162211146352515, "grad_norm": 0.08829481122747623, "learning_rate": 6.649670233261845e-06, "loss": 0.9384, "step": 9186 }, { "epoch": 4.162664250113276, "grad_norm": 0.08057898698387954, "learning_rate": 6.642685429174065e-06, "loss": 0.9226, "step": 9187 }, { "epoch": 4.163117353874037, "grad_norm": 0.08996037403580866, "learning_rate": 6.635703963228692e-06, "loss": 0.9313, "step": 9188 }, { "epoch": 4.1635704576347985, "grad_norm": 0.0851349338580073, "learning_rate": 6.62872583612435e-06, "loss": 0.9289, "step": 9189 }, { "epoch": 4.16402356139556, "grad_norm": 0.09044292878510937, "learning_rate": 6.621751048559386e-06, "loss": 0.9283, "step": 9190 }, { "epoch": 4.164476665156321, "grad_norm": 0.08802897960892546, "learning_rate": 6.614779601231758e-06, "loss": 0.9215, "step": 9191 }, { "epoch": 4.164929768917082, "grad_norm": 0.08138497953581257, "learning_rate": 6.607811494839138e-06, "loss": 0.9347, "step": 9192 }, { "epoch": 4.165382872677843, "grad_norm": 0.09189388125646979, "learning_rate": 6.600846730078813e-06, "loss": 0.9431, "step": 9193 }, { "epoch": 4.165835976438604, "grad_norm": 0.08553261598635507, "learning_rate": 6.593885307647791e-06, "loss": 0.9337, "step": 9194 }, { "epoch": 4.166289080199365, "grad_norm": 0.09211449302412207, "learning_rate": 6.586927228242701e-06, "loss": 0.9164, "step": 9195 }, { "epoch": 4.166742183960127, "grad_norm": 0.09223541665969698, "learning_rate": 6.5799724925598475e-06, "loss": 0.9212, "step": 9196 }, { "epoch": 4.167195287720888, "grad_norm": 0.08575039639092509, "learning_rate": 6.57302110129523e-06, "loss": 0.9161, "step": 9197 }, { "epoch": 4.16764839148165, "grad_norm": 0.09144573236686145, "learning_rate": 6.56607305514446e-06, "loss": 0.9173, "step": 9198 }, { "epoch": 4.168101495242411, "grad_norm": 0.09353431361361826, "learning_rate": 6.559128354802871e-06, "loss": 0.9161, "step": 9199 }, { "epoch": 4.168554599003172, "grad_norm": 0.08812315118254846, "learning_rate": 6.552187000965418e-06, "loss": 0.9253, "step": 9200 }, { "epoch": 4.169007702763933, "grad_norm": 0.09064386979413631, "learning_rate": 6.545248994326754e-06, "loss": 0.9319, "step": 9201 }, { "epoch": 4.169460806524694, "grad_norm": 0.10006477090630142, "learning_rate": 6.5383143355811685e-06, "loss": 0.926, "step": 9202 }, { "epoch": 4.169913910285455, "grad_norm": 0.08787878149406841, "learning_rate": 6.5313830254226216e-06, "loss": 0.9275, "step": 9203 }, { "epoch": 4.170367014046216, "grad_norm": 0.09381712844693427, "learning_rate": 6.524455064544768e-06, "loss": 0.9003, "step": 9204 }, { "epoch": 4.1708201178069775, "grad_norm": 0.0807315194749406, "learning_rate": 6.517530453640879e-06, "loss": 0.9066, "step": 9205 }, { "epoch": 4.171273221567739, "grad_norm": 0.08505900017989783, "learning_rate": 6.510609193403939e-06, "loss": 0.9235, "step": 9206 }, { "epoch": 4.171726325328501, "grad_norm": 0.09345454444527455, "learning_rate": 6.5036912845265566e-06, "loss": 0.9156, "step": 9207 }, { "epoch": 4.172179429089262, "grad_norm": 0.09403030343665608, "learning_rate": 6.496776727701038e-06, "loss": 0.9443, "step": 9208 }, { "epoch": 4.172632532850023, "grad_norm": 0.09864257080360944, "learning_rate": 6.489865523619321e-06, "loss": 0.9226, "step": 9209 }, { "epoch": 4.173085636610784, "grad_norm": 0.08649989545713001, "learning_rate": 6.4829576729730446e-06, "loss": 0.9272, "step": 9210 }, { "epoch": 4.173538740371545, "grad_norm": 0.09712047304754963, "learning_rate": 6.476053176453487e-06, "loss": 0.9203, "step": 9211 }, { "epoch": 4.173991844132306, "grad_norm": 0.09263620143713577, "learning_rate": 6.46915203475158e-06, "loss": 0.908, "step": 9212 }, { "epoch": 4.174444947893067, "grad_norm": 0.08511029111043135, "learning_rate": 6.462254248557962e-06, "loss": 0.9282, "step": 9213 }, { "epoch": 4.1748980516538285, "grad_norm": 0.0957336004065988, "learning_rate": 6.45535981856289e-06, "loss": 0.8859, "step": 9214 }, { "epoch": 4.17535115541459, "grad_norm": 0.08227577295219882, "learning_rate": 6.448468745456322e-06, "loss": 0.906, "step": 9215 }, { "epoch": 4.175804259175351, "grad_norm": 0.08709217989869063, "learning_rate": 6.44158102992785e-06, "loss": 0.9358, "step": 9216 }, { "epoch": 4.176257362936113, "grad_norm": 0.07805824968380232, "learning_rate": 6.434696672666754e-06, "loss": 0.9143, "step": 9217 }, { "epoch": 4.176710466696874, "grad_norm": 0.08456356085137182, "learning_rate": 6.427815674361961e-06, "loss": 0.9147, "step": 9218 }, { "epoch": 4.177163570457635, "grad_norm": 0.09487974351107473, "learning_rate": 6.42093803570206e-06, "loss": 0.9183, "step": 9219 }, { "epoch": 4.177616674218396, "grad_norm": 0.08568634757515758, "learning_rate": 6.41406375737533e-06, "loss": 0.9135, "step": 9220 }, { "epoch": 4.178069777979157, "grad_norm": 0.08418153235947697, "learning_rate": 6.40719284006968e-06, "loss": 0.9173, "step": 9221 }, { "epoch": 4.1785228817399185, "grad_norm": 0.09095766826073129, "learning_rate": 6.400325284472715e-06, "loss": 0.9434, "step": 9222 }, { "epoch": 4.17897598550068, "grad_norm": 0.0907236157984442, "learning_rate": 6.393461091271667e-06, "loss": 0.9314, "step": 9223 }, { "epoch": 4.179429089261441, "grad_norm": 0.08170301456220362, "learning_rate": 6.386600261153466e-06, "loss": 0.9437, "step": 9224 }, { "epoch": 4.179882193022202, "grad_norm": 0.08416712581300087, "learning_rate": 6.37974279480468e-06, "loss": 0.9286, "step": 9225 }, { "epoch": 4.180335296782963, "grad_norm": 0.08297457637136381, "learning_rate": 6.372888692911567e-06, "loss": 0.9151, "step": 9226 }, { "epoch": 4.180788400543724, "grad_norm": 0.08354261241004259, "learning_rate": 6.366037956160025e-06, "loss": 0.9337, "step": 9227 }, { "epoch": 4.181241504304486, "grad_norm": 0.07956261052192562, "learning_rate": 6.359190585235611e-06, "loss": 0.911, "step": 9228 }, { "epoch": 4.181694608065247, "grad_norm": 0.08116728925405352, "learning_rate": 6.352346580823572e-06, "loss": 0.9269, "step": 9229 }, { "epoch": 4.182147711826008, "grad_norm": 0.0847741007068157, "learning_rate": 6.345505943608792e-06, "loss": 0.9226, "step": 9230 }, { "epoch": 4.1826008155867695, "grad_norm": 0.08492192523170405, "learning_rate": 6.338668674275848e-06, "loss": 0.9476, "step": 9231 }, { "epoch": 4.183053919347531, "grad_norm": 0.08568697455454588, "learning_rate": 6.331834773508934e-06, "loss": 0.9058, "step": 9232 }, { "epoch": 4.183507023108292, "grad_norm": 0.07872156908846015, "learning_rate": 6.325004241991957e-06, "loss": 0.9219, "step": 9233 }, { "epoch": 4.183960126869053, "grad_norm": 0.0841423050886388, "learning_rate": 6.31817708040845e-06, "loss": 0.9348, "step": 9234 }, { "epoch": 4.184413230629814, "grad_norm": 0.08966802875913372, "learning_rate": 6.311353289441635e-06, "loss": 0.946, "step": 9235 }, { "epoch": 4.184866334390575, "grad_norm": 0.0865360304942227, "learning_rate": 6.304532869774375e-06, "loss": 0.931, "step": 9236 }, { "epoch": 4.185319438151336, "grad_norm": 0.09028913824826207, "learning_rate": 6.297715822089201e-06, "loss": 0.944, "step": 9237 }, { "epoch": 4.1857725419120975, "grad_norm": 0.08031259877846089, "learning_rate": 6.290902147068334e-06, "loss": 0.9103, "step": 9238 }, { "epoch": 4.1862256456728595, "grad_norm": 0.09347357730297066, "learning_rate": 6.284091845393594e-06, "loss": 0.9281, "step": 9239 }, { "epoch": 4.186678749433621, "grad_norm": 0.08698660569746382, "learning_rate": 6.277284917746538e-06, "loss": 0.9283, "step": 9240 }, { "epoch": 4.187131853194382, "grad_norm": 0.09135304467875541, "learning_rate": 6.270481364808332e-06, "loss": 0.9121, "step": 9241 }, { "epoch": 4.187584956955143, "grad_norm": 0.08356869900352207, "learning_rate": 6.263681187259836e-06, "loss": 0.9393, "step": 9242 }, { "epoch": 4.188038060715904, "grad_norm": 0.08977531313830775, "learning_rate": 6.256884385781541e-06, "loss": 0.9039, "step": 9243 }, { "epoch": 4.188491164476665, "grad_norm": 0.08559763997708165, "learning_rate": 6.250090961053645e-06, "loss": 0.8969, "step": 9244 }, { "epoch": 4.188944268237426, "grad_norm": 0.08769532722479113, "learning_rate": 6.243300913755952e-06, "loss": 0.9388, "step": 9245 }, { "epoch": 4.189397371998187, "grad_norm": 0.09407742670988041, "learning_rate": 6.236514244567984e-06, "loss": 0.9379, "step": 9246 }, { "epoch": 4.1898504757589485, "grad_norm": 0.08606479594111818, "learning_rate": 6.229730954168887e-06, "loss": 0.919, "step": 9247 }, { "epoch": 4.19030357951971, "grad_norm": 0.08119090105459789, "learning_rate": 6.222951043237468e-06, "loss": 0.9125, "step": 9248 }, { "epoch": 4.190756683280471, "grad_norm": 0.08631206114501783, "learning_rate": 6.216174512452236e-06, "loss": 0.9051, "step": 9249 }, { "epoch": 4.191209787041233, "grad_norm": 0.08174604921590704, "learning_rate": 6.209401362491303e-06, "loss": 0.9299, "step": 9250 }, { "epoch": 4.191662890801994, "grad_norm": 0.07984209736058073, "learning_rate": 6.202631594032498e-06, "loss": 0.9044, "step": 9251 }, { "epoch": 4.192115994562755, "grad_norm": 0.08257070804279983, "learning_rate": 6.195865207753269e-06, "loss": 0.9319, "step": 9252 }, { "epoch": 4.192569098323516, "grad_norm": 0.08716848011642901, "learning_rate": 6.189102204330759e-06, "loss": 0.9413, "step": 9253 }, { "epoch": 4.193022202084277, "grad_norm": 0.08426436984600101, "learning_rate": 6.182342584441752e-06, "loss": 0.9241, "step": 9254 }, { "epoch": 4.1934753058450385, "grad_norm": 0.09096782382382745, "learning_rate": 6.175586348762688e-06, "loss": 0.9356, "step": 9255 }, { "epoch": 4.1939284096058, "grad_norm": 0.08394723933553686, "learning_rate": 6.168833497969698e-06, "loss": 0.9217, "step": 9256 }, { "epoch": 4.194381513366561, "grad_norm": 0.08845538837161476, "learning_rate": 6.1620840327385335e-06, "loss": 0.9604, "step": 9257 }, { "epoch": 4.194834617127322, "grad_norm": 0.08810743600335169, "learning_rate": 6.155337953744651e-06, "loss": 0.9375, "step": 9258 }, { "epoch": 4.195287720888083, "grad_norm": 0.08311278754488645, "learning_rate": 6.148595261663128e-06, "loss": 0.9172, "step": 9259 }, { "epoch": 4.195740824648845, "grad_norm": 0.08987876718970846, "learning_rate": 6.141855957168736e-06, "loss": 0.9256, "step": 9260 }, { "epoch": 4.196193928409606, "grad_norm": 0.08673590864306344, "learning_rate": 6.135120040935874e-06, "loss": 0.9152, "step": 9261 }, { "epoch": 4.196647032170367, "grad_norm": 0.09027163253317025, "learning_rate": 6.12838751363864e-06, "loss": 0.9129, "step": 9262 }, { "epoch": 4.197100135931128, "grad_norm": 0.08477579710729075, "learning_rate": 6.1216583759507696e-06, "loss": 0.8925, "step": 9263 }, { "epoch": 4.1975532396918895, "grad_norm": 0.08448683602430929, "learning_rate": 6.114932628545647e-06, "loss": 0.9354, "step": 9264 }, { "epoch": 4.198006343452651, "grad_norm": 0.08058483957077224, "learning_rate": 6.108210272096351e-06, "loss": 0.9347, "step": 9265 }, { "epoch": 4.198459447213412, "grad_norm": 0.08210942491820308, "learning_rate": 6.1014913072755846e-06, "loss": 0.9303, "step": 9266 }, { "epoch": 4.198912550974173, "grad_norm": 0.09290451416111235, "learning_rate": 6.094775734755756e-06, "loss": 0.9363, "step": 9267 }, { "epoch": 4.199365654734934, "grad_norm": 0.0773759479733759, "learning_rate": 6.0880635552088786e-06, "loss": 0.9351, "step": 9268 }, { "epoch": 4.199818758495695, "grad_norm": 0.09075986804912779, "learning_rate": 6.0813547693066865e-06, "loss": 0.9397, "step": 9269 }, { "epoch": 4.200271862256456, "grad_norm": 0.09176780850386843, "learning_rate": 6.074649377720509e-06, "loss": 0.9277, "step": 9270 }, { "epoch": 4.200724966017218, "grad_norm": 0.08549284889420172, "learning_rate": 6.067947381121406e-06, "loss": 0.9279, "step": 9271 }, { "epoch": 4.2011780697779795, "grad_norm": 0.09257654378385032, "learning_rate": 6.061248780180036e-06, "loss": 0.9204, "step": 9272 }, { "epoch": 4.201631173538741, "grad_norm": 0.08189316212959116, "learning_rate": 6.054553575566746e-06, "loss": 0.9151, "step": 9273 }, { "epoch": 4.202084277299502, "grad_norm": 0.08254384473394087, "learning_rate": 6.047861767951553e-06, "loss": 0.9076, "step": 9274 }, { "epoch": 4.202537381060263, "grad_norm": 0.09215885629735679, "learning_rate": 6.041173358004102e-06, "loss": 0.9045, "step": 9275 }, { "epoch": 4.202990484821024, "grad_norm": 0.08999814847907713, "learning_rate": 6.0344883463937385e-06, "loss": 0.9232, "step": 9276 }, { "epoch": 4.203443588581785, "grad_norm": 0.085039975260093, "learning_rate": 6.0278067337894295e-06, "loss": 0.9175, "step": 9277 }, { "epoch": 4.203896692342546, "grad_norm": 0.0887182534936085, "learning_rate": 6.021128520859836e-06, "loss": 0.9268, "step": 9278 }, { "epoch": 4.204349796103307, "grad_norm": 0.09152169659340528, "learning_rate": 6.014453708273249e-06, "loss": 0.9155, "step": 9279 }, { "epoch": 4.2048028998640685, "grad_norm": 0.08639868774030486, "learning_rate": 6.007782296697629e-06, "loss": 0.9399, "step": 9280 }, { "epoch": 4.2052560036248305, "grad_norm": 0.08539629806133828, "learning_rate": 6.001114286800618e-06, "loss": 0.9602, "step": 9281 }, { "epoch": 4.205709107385592, "grad_norm": 0.07920665378342764, "learning_rate": 5.994449679249474e-06, "loss": 0.9247, "step": 9282 }, { "epoch": 4.206162211146353, "grad_norm": 0.08080282144766482, "learning_rate": 5.987788474711163e-06, "loss": 0.9313, "step": 9283 }, { "epoch": 4.206615314907114, "grad_norm": 0.09236923850859396, "learning_rate": 5.981130673852269e-06, "loss": 0.9232, "step": 9284 }, { "epoch": 4.207068418667875, "grad_norm": 0.08679609726399885, "learning_rate": 5.974476277339069e-06, "loss": 0.9184, "step": 9285 }, { "epoch": 4.207521522428636, "grad_norm": 0.08335443840138454, "learning_rate": 5.967825285837467e-06, "loss": 0.9299, "step": 9286 }, { "epoch": 4.207974626189397, "grad_norm": 0.0891464964695746, "learning_rate": 5.9611777000130585e-06, "loss": 0.9157, "step": 9287 }, { "epoch": 4.2084277299501585, "grad_norm": 0.09129758179347737, "learning_rate": 5.954533520531076e-06, "loss": 0.9127, "step": 9288 }, { "epoch": 4.20888083371092, "grad_norm": 0.08545350229321326, "learning_rate": 5.94789274805641e-06, "loss": 0.9198, "step": 9289 }, { "epoch": 4.209333937471681, "grad_norm": 0.09937994021774499, "learning_rate": 5.94125538325363e-06, "loss": 0.9205, "step": 9290 }, { "epoch": 4.209787041232442, "grad_norm": 0.08347848780914094, "learning_rate": 5.93462142678694e-06, "loss": 0.9009, "step": 9291 }, { "epoch": 4.210240144993204, "grad_norm": 0.08698773805941405, "learning_rate": 5.927990879320229e-06, "loss": 0.9366, "step": 9292 }, { "epoch": 4.210693248753965, "grad_norm": 0.08853435534911741, "learning_rate": 5.9213637415170165e-06, "loss": 0.9176, "step": 9293 }, { "epoch": 4.211146352514726, "grad_norm": 0.0833880925850425, "learning_rate": 5.914740014040514e-06, "loss": 0.9195, "step": 9294 }, { "epoch": 4.211599456275487, "grad_norm": 0.08754721568628808, "learning_rate": 5.908119697553561e-06, "loss": 0.8975, "step": 9295 }, { "epoch": 4.212052560036248, "grad_norm": 0.09413405185487567, "learning_rate": 5.901502792718665e-06, "loss": 0.928, "step": 9296 }, { "epoch": 4.2125056637970095, "grad_norm": 0.08784130484572683, "learning_rate": 5.894889300197992e-06, "loss": 0.9236, "step": 9297 }, { "epoch": 4.212958767557771, "grad_norm": 0.09386536943110699, "learning_rate": 5.888279220653381e-06, "loss": 0.935, "step": 9298 }, { "epoch": 4.213411871318532, "grad_norm": 0.10030121604572925, "learning_rate": 5.881672554746316e-06, "loss": 0.9202, "step": 9299 }, { "epoch": 4.213864975079293, "grad_norm": 0.08927219104258062, "learning_rate": 5.875069303137926e-06, "loss": 0.9046, "step": 9300 }, { "epoch": 4.214318078840054, "grad_norm": 0.094405791907022, "learning_rate": 5.8684694664890326e-06, "loss": 0.9155, "step": 9301 }, { "epoch": 4.214771182600815, "grad_norm": 0.08398483617592142, "learning_rate": 5.861873045460082e-06, "loss": 0.9313, "step": 9302 }, { "epoch": 4.215224286361577, "grad_norm": 0.0924560597633328, "learning_rate": 5.855280040711209e-06, "loss": 0.9095, "step": 9303 }, { "epoch": 4.215677390122338, "grad_norm": 0.07929230176210464, "learning_rate": 5.848690452902168e-06, "loss": 0.9168, "step": 9304 }, { "epoch": 4.2161304938830995, "grad_norm": 0.08721680283250358, "learning_rate": 5.84210428269242e-06, "loss": 0.9295, "step": 9305 }, { "epoch": 4.216583597643861, "grad_norm": 0.09216895633888264, "learning_rate": 5.835521530741033e-06, "loss": 0.9519, "step": 9306 }, { "epoch": 4.217036701404622, "grad_norm": 0.08275411774986859, "learning_rate": 5.828942197706782e-06, "loss": 0.9159, "step": 9307 }, { "epoch": 4.217489805165383, "grad_norm": 0.07708804865718244, "learning_rate": 5.822366284248064e-06, "loss": 0.9065, "step": 9308 }, { "epoch": 4.217942908926144, "grad_norm": 0.09046276085605397, "learning_rate": 5.815793791022937e-06, "loss": 0.9409, "step": 9309 }, { "epoch": 4.218396012686905, "grad_norm": 0.08772436496012628, "learning_rate": 5.8092247186891395e-06, "loss": 0.9395, "step": 9310 }, { "epoch": 4.218849116447666, "grad_norm": 0.09491320212151114, "learning_rate": 5.802659067904039e-06, "loss": 0.9104, "step": 9311 }, { "epoch": 4.219302220208427, "grad_norm": 0.08741969565929573, "learning_rate": 5.796096839324699e-06, "loss": 0.9519, "step": 9312 }, { "epoch": 4.2197553239691885, "grad_norm": 0.08632855632596705, "learning_rate": 5.789538033607786e-06, "loss": 0.9204, "step": 9313 }, { "epoch": 4.2202084277299505, "grad_norm": 0.08005638867457773, "learning_rate": 5.7829826514096855e-06, "loss": 0.9214, "step": 9314 }, { "epoch": 4.220661531490712, "grad_norm": 0.08200041554209787, "learning_rate": 5.77643069338639e-06, "loss": 0.9283, "step": 9315 }, { "epoch": 4.221114635251473, "grad_norm": 0.07984251563700138, "learning_rate": 5.769882160193563e-06, "loss": 0.9236, "step": 9316 }, { "epoch": 4.221567739012234, "grad_norm": 0.08178476357695456, "learning_rate": 5.763337052486555e-06, "loss": 0.9032, "step": 9317 }, { "epoch": 4.222020842772995, "grad_norm": 0.08797975674377727, "learning_rate": 5.756795370920323e-06, "loss": 0.9199, "step": 9318 }, { "epoch": 4.222473946533756, "grad_norm": 0.07955141673651854, "learning_rate": 5.750257116149533e-06, "loss": 0.915, "step": 9319 }, { "epoch": 4.222927050294517, "grad_norm": 0.07872561342375059, "learning_rate": 5.7437222888284636e-06, "loss": 0.9109, "step": 9320 }, { "epoch": 4.2233801540552784, "grad_norm": 0.08755977831485932, "learning_rate": 5.737190889611084e-06, "loss": 0.9169, "step": 9321 }, { "epoch": 4.22383325781604, "grad_norm": 0.08402741216800674, "learning_rate": 5.730662919150991e-06, "loss": 0.9325, "step": 9322 }, { "epoch": 4.224286361576801, "grad_norm": 0.08087911828823433, "learning_rate": 5.724138378101471e-06, "loss": 0.9195, "step": 9323 }, { "epoch": 4.224739465337563, "grad_norm": 0.08386553172513596, "learning_rate": 5.717617267115438e-06, "loss": 0.914, "step": 9324 }, { "epoch": 4.225192569098324, "grad_norm": 0.08236828097899918, "learning_rate": 5.711099586845472e-06, "loss": 0.9102, "step": 9325 }, { "epoch": 4.225645672859085, "grad_norm": 0.09068199016734414, "learning_rate": 5.7045853379438285e-06, "loss": 0.9342, "step": 9326 }, { "epoch": 4.226098776619846, "grad_norm": 0.07801539188542592, "learning_rate": 5.698074521062378e-06, "loss": 0.9171, "step": 9327 }, { "epoch": 4.226551880380607, "grad_norm": 0.08402276774330485, "learning_rate": 5.691567136852696e-06, "loss": 0.9235, "step": 9328 }, { "epoch": 4.227004984141368, "grad_norm": 0.07868316072936123, "learning_rate": 5.685063185965977e-06, "loss": 0.9245, "step": 9329 }, { "epoch": 4.2274580879021295, "grad_norm": 0.0775357651689144, "learning_rate": 5.678562669053098e-06, "loss": 0.9127, "step": 9330 }, { "epoch": 4.227911191662891, "grad_norm": 0.07979151743232789, "learning_rate": 5.672065586764564e-06, "loss": 0.9135, "step": 9331 }, { "epoch": 4.228364295423652, "grad_norm": 0.08128744653062667, "learning_rate": 5.665571939750569e-06, "loss": 0.9366, "step": 9332 }, { "epoch": 4.228817399184413, "grad_norm": 0.07701863354613424, "learning_rate": 5.659081728660947e-06, "loss": 0.9254, "step": 9333 }, { "epoch": 4.229270502945174, "grad_norm": 0.08160789951661777, "learning_rate": 5.652594954145167e-06, "loss": 0.9291, "step": 9334 }, { "epoch": 4.229723606705936, "grad_norm": 0.08080539316360354, "learning_rate": 5.646111616852401e-06, "loss": 0.9059, "step": 9335 }, { "epoch": 4.230176710466697, "grad_norm": 0.08548098943069486, "learning_rate": 5.639631717431435e-06, "loss": 0.9255, "step": 9336 }, { "epoch": 4.230629814227458, "grad_norm": 0.08518455750383032, "learning_rate": 5.633155256530738e-06, "loss": 0.9051, "step": 9337 }, { "epoch": 4.2310829179882194, "grad_norm": 0.07997811152002748, "learning_rate": 5.626682234798409e-06, "loss": 0.9079, "step": 9338 }, { "epoch": 4.231536021748981, "grad_norm": 0.07829708549486863, "learning_rate": 5.62021265288224e-06, "loss": 0.9354, "step": 9339 }, { "epoch": 4.231989125509742, "grad_norm": 0.08400099756609941, "learning_rate": 5.613746511429643e-06, "loss": 0.9148, "step": 9340 }, { "epoch": 4.232442229270503, "grad_norm": 0.07968310035850301, "learning_rate": 5.607283811087691e-06, "loss": 0.9131, "step": 9341 }, { "epoch": 4.232895333031264, "grad_norm": 0.08096179345638198, "learning_rate": 5.600824552503139e-06, "loss": 0.917, "step": 9342 }, { "epoch": 4.233348436792025, "grad_norm": 0.08409662724661289, "learning_rate": 5.594368736322367e-06, "loss": 0.9267, "step": 9343 }, { "epoch": 4.233801540552786, "grad_norm": 0.08254021517747483, "learning_rate": 5.587916363191435e-06, "loss": 0.9285, "step": 9344 }, { "epoch": 4.234254644313548, "grad_norm": 0.0817399354609636, "learning_rate": 5.581467433756032e-06, "loss": 0.9176, "step": 9345 }, { "epoch": 4.234707748074309, "grad_norm": 0.08030756764464958, "learning_rate": 5.575021948661534e-06, "loss": 0.9268, "step": 9346 }, { "epoch": 4.2351608518350705, "grad_norm": 0.08283918145704458, "learning_rate": 5.568579908552934e-06, "loss": 0.9186, "step": 9347 }, { "epoch": 4.235613955595832, "grad_norm": 0.08147744040552035, "learning_rate": 5.562141314074927e-06, "loss": 0.9243, "step": 9348 }, { "epoch": 4.236067059356593, "grad_norm": 0.08265918866004687, "learning_rate": 5.555706165871822e-06, "loss": 0.9445, "step": 9349 }, { "epoch": 4.236520163117354, "grad_norm": 0.08182324473387334, "learning_rate": 5.549274464587595e-06, "loss": 0.9272, "step": 9350 }, { "epoch": 4.236973266878115, "grad_norm": 0.08220857254241083, "learning_rate": 5.542846210865893e-06, "loss": 0.9388, "step": 9351 }, { "epoch": 4.237426370638876, "grad_norm": 0.07778365585694623, "learning_rate": 5.5364214053500055e-06, "loss": 0.9303, "step": 9352 }, { "epoch": 4.237879474399637, "grad_norm": 0.08586828759887592, "learning_rate": 5.530000048682867e-06, "loss": 0.9186, "step": 9353 }, { "epoch": 4.238332578160398, "grad_norm": 0.08460921036691246, "learning_rate": 5.523582141507078e-06, "loss": 0.9217, "step": 9354 }, { "epoch": 4.23878568192116, "grad_norm": 0.0868277679442277, "learning_rate": 5.517167684464904e-06, "loss": 0.9097, "step": 9355 }, { "epoch": 4.239238785681922, "grad_norm": 0.08199898227278646, "learning_rate": 5.510756678198243e-06, "loss": 0.9239, "step": 9356 }, { "epoch": 4.239691889442683, "grad_norm": 0.0905517465091708, "learning_rate": 5.504349123348669e-06, "loss": 0.9472, "step": 9357 }, { "epoch": 4.240144993203444, "grad_norm": 0.08756431332609177, "learning_rate": 5.4979450205573915e-06, "loss": 0.9359, "step": 9358 }, { "epoch": 4.240598096964205, "grad_norm": 0.09649326105185374, "learning_rate": 5.4915443704652935e-06, "loss": 0.9358, "step": 9359 }, { "epoch": 4.241051200724966, "grad_norm": 0.08030211066782601, "learning_rate": 5.485147173712894e-06, "loss": 0.9028, "step": 9360 }, { "epoch": 4.241504304485727, "grad_norm": 0.0872307452318914, "learning_rate": 5.478753430940376e-06, "loss": 0.9182, "step": 9361 }, { "epoch": 4.241957408246488, "grad_norm": 0.09552509453168398, "learning_rate": 5.47236314278758e-06, "loss": 0.9357, "step": 9362 }, { "epoch": 4.2424105120072495, "grad_norm": 0.08497921134674738, "learning_rate": 5.465976309893992e-06, "loss": 0.9146, "step": 9363 }, { "epoch": 4.242863615768011, "grad_norm": 0.09011119668835806, "learning_rate": 5.459592932898768e-06, "loss": 0.9085, "step": 9364 }, { "epoch": 4.243316719528772, "grad_norm": 0.0876289659708046, "learning_rate": 5.453213012440683e-06, "loss": 0.9121, "step": 9365 }, { "epoch": 4.243769823289533, "grad_norm": 0.08754057909474885, "learning_rate": 5.446836549158221e-06, "loss": 0.9234, "step": 9366 }, { "epoch": 4.244222927050295, "grad_norm": 0.07950734174151253, "learning_rate": 5.440463543689465e-06, "loss": 0.9102, "step": 9367 }, { "epoch": 4.244676030811056, "grad_norm": 0.09066149957458787, "learning_rate": 5.434093996672189e-06, "loss": 0.9114, "step": 9368 }, { "epoch": 4.245129134571817, "grad_norm": 0.07895393172997443, "learning_rate": 5.42772790874381e-06, "loss": 0.9254, "step": 9369 }, { "epoch": 4.245582238332578, "grad_norm": 0.08420414996078392, "learning_rate": 5.42136528054138e-06, "loss": 0.9187, "step": 9370 }, { "epoch": 4.246035342093339, "grad_norm": 0.08561299801878841, "learning_rate": 5.41500611270164e-06, "loss": 0.9498, "step": 9371 }, { "epoch": 4.246488445854101, "grad_norm": 0.07974284427405519, "learning_rate": 5.408650405860951e-06, "loss": 0.9217, "step": 9372 }, { "epoch": 4.246941549614862, "grad_norm": 0.0825102720275998, "learning_rate": 5.402298160655361e-06, "loss": 0.9127, "step": 9373 }, { "epoch": 4.247394653375623, "grad_norm": 0.08536106090352548, "learning_rate": 5.3959493777205355e-06, "loss": 0.9184, "step": 9374 }, { "epoch": 4.247847757136384, "grad_norm": 0.08357828639037261, "learning_rate": 5.389604057691827e-06, "loss": 0.9267, "step": 9375 }, { "epoch": 4.248300860897145, "grad_norm": 0.08113797500835962, "learning_rate": 5.38326220120422e-06, "loss": 0.9006, "step": 9376 }, { "epoch": 4.248753964657906, "grad_norm": 0.08390709186905684, "learning_rate": 5.376923808892352e-06, "loss": 0.9138, "step": 9377 }, { "epoch": 4.249207068418668, "grad_norm": 0.08345321125787185, "learning_rate": 5.37058888139053e-06, "loss": 0.9203, "step": 9378 }, { "epoch": 4.249660172179429, "grad_norm": 0.07919945632389765, "learning_rate": 5.3642574193326945e-06, "loss": 0.9105, "step": 9379 }, { "epoch": 4.2501132759401905, "grad_norm": 0.07777663692825532, "learning_rate": 5.357929423352466e-06, "loss": 0.9254, "step": 9380 }, { "epoch": 4.250566379700952, "grad_norm": 0.08726675672573309, "learning_rate": 5.35160489408308e-06, "loss": 0.9418, "step": 9381 }, { "epoch": 4.251019483461713, "grad_norm": 0.08221285119679447, "learning_rate": 5.345283832157466e-06, "loss": 0.9221, "step": 9382 }, { "epoch": 4.251472587222474, "grad_norm": 0.08949232072317989, "learning_rate": 5.338966238208167e-06, "loss": 0.9235, "step": 9383 }, { "epoch": 4.251925690983235, "grad_norm": 0.0917037068807348, "learning_rate": 5.332652112867425e-06, "loss": 0.9375, "step": 9384 }, { "epoch": 4.252378794743996, "grad_norm": 0.08156607246297728, "learning_rate": 5.326341456767092e-06, "loss": 0.9341, "step": 9385 }, { "epoch": 4.252831898504757, "grad_norm": 0.0907256899407845, "learning_rate": 5.320034270538683e-06, "loss": 0.9309, "step": 9386 }, { "epoch": 4.253285002265518, "grad_norm": 0.090432211601808, "learning_rate": 5.31373055481339e-06, "loss": 0.9079, "step": 9387 }, { "epoch": 4.2537381060262796, "grad_norm": 0.0880633005911474, "learning_rate": 5.307430310222024e-06, "loss": 0.9099, "step": 9388 }, { "epoch": 4.254191209787042, "grad_norm": 0.09001822662630436, "learning_rate": 5.30113353739508e-06, "loss": 0.9308, "step": 9389 }, { "epoch": 4.254644313547803, "grad_norm": 0.08085106669084484, "learning_rate": 5.294840236962673e-06, "loss": 0.9146, "step": 9390 }, { "epoch": 4.255097417308564, "grad_norm": 0.08400391657800954, "learning_rate": 5.288550409554609e-06, "loss": 0.9297, "step": 9391 }, { "epoch": 4.255550521069325, "grad_norm": 0.08674953289006893, "learning_rate": 5.2822640558003055e-06, "loss": 0.9243, "step": 9392 }, { "epoch": 4.256003624830086, "grad_norm": 0.08776018688231722, "learning_rate": 5.2759811763288685e-06, "loss": 0.9111, "step": 9393 }, { "epoch": 4.256456728590847, "grad_norm": 0.08292864989209359, "learning_rate": 5.269701771769033e-06, "loss": 0.9365, "step": 9394 }, { "epoch": 4.256909832351608, "grad_norm": 0.08067248694705492, "learning_rate": 5.263425842749188e-06, "loss": 0.9049, "step": 9395 }, { "epoch": 4.2573629361123695, "grad_norm": 0.08852117398326262, "learning_rate": 5.2571533898973895e-06, "loss": 0.9216, "step": 9396 }, { "epoch": 4.257816039873131, "grad_norm": 0.09196018052478823, "learning_rate": 5.2508844138413215e-06, "loss": 0.9108, "step": 9397 }, { "epoch": 4.258269143633892, "grad_norm": 0.07686864654455139, "learning_rate": 5.244618915208355e-06, "loss": 0.9221, "step": 9398 }, { "epoch": 4.258722247394654, "grad_norm": 0.10077883928946742, "learning_rate": 5.238356894625476e-06, "loss": 0.9475, "step": 9399 }, { "epoch": 4.259175351155415, "grad_norm": 0.09207335330525088, "learning_rate": 5.232098352719352e-06, "loss": 0.9421, "step": 9400 }, { "epoch": 4.259628454916176, "grad_norm": 0.08193367960711272, "learning_rate": 5.225843290116283e-06, "loss": 0.9028, "step": 9401 }, { "epoch": 4.260081558676937, "grad_norm": 0.08060541154318135, "learning_rate": 5.219591707442218e-06, "loss": 0.9097, "step": 9402 }, { "epoch": 4.260534662437698, "grad_norm": 0.09095728672083524, "learning_rate": 5.213343605322787e-06, "loss": 0.9404, "step": 9403 }, { "epoch": 4.260987766198459, "grad_norm": 0.09196240451213936, "learning_rate": 5.20709898438323e-06, "loss": 0.9226, "step": 9404 }, { "epoch": 4.261440869959221, "grad_norm": 0.08279166358077107, "learning_rate": 5.200857845248481e-06, "loss": 0.9441, "step": 9405 }, { "epoch": 4.261893973719982, "grad_norm": 0.08147871648855516, "learning_rate": 5.1946201885430914e-06, "loss": 0.9362, "step": 9406 }, { "epoch": 4.262347077480743, "grad_norm": 0.08653126180136671, "learning_rate": 5.188386014891285e-06, "loss": 0.9212, "step": 9407 }, { "epoch": 4.262800181241504, "grad_norm": 0.08454698628156827, "learning_rate": 5.182155324916922e-06, "loss": 0.9129, "step": 9408 }, { "epoch": 4.263253285002266, "grad_norm": 0.07940754842897099, "learning_rate": 5.175928119243545e-06, "loss": 0.9071, "step": 9409 }, { "epoch": 4.263706388763027, "grad_norm": 0.0812719681727881, "learning_rate": 5.169704398494286e-06, "loss": 0.9157, "step": 9410 }, { "epoch": 4.264159492523788, "grad_norm": 0.07743662030173905, "learning_rate": 5.163484163291994e-06, "loss": 0.9401, "step": 9411 }, { "epoch": 4.264612596284549, "grad_norm": 0.08479927514625046, "learning_rate": 5.1572674142591306e-06, "loss": 0.951, "step": 9412 }, { "epoch": 4.2650657000453105, "grad_norm": 0.07801163406856168, "learning_rate": 5.1510541520178294e-06, "loss": 0.9384, "step": 9413 }, { "epoch": 4.265518803806072, "grad_norm": 0.0804840618963616, "learning_rate": 5.144844377189864e-06, "loss": 0.9277, "step": 9414 }, { "epoch": 4.265971907566833, "grad_norm": 0.08357676373036194, "learning_rate": 5.138638090396648e-06, "loss": 0.9038, "step": 9415 }, { "epoch": 4.266425011327594, "grad_norm": 0.08222342506178934, "learning_rate": 5.132435292259277e-06, "loss": 0.9091, "step": 9416 }, { "epoch": 4.266878115088355, "grad_norm": 0.08284388453381096, "learning_rate": 5.126235983398462e-06, "loss": 0.9238, "step": 9417 }, { "epoch": 4.267331218849116, "grad_norm": 0.08898962562950169, "learning_rate": 5.120040164434596e-06, "loss": 0.9367, "step": 9418 }, { "epoch": 4.267784322609877, "grad_norm": 0.08455519972890158, "learning_rate": 5.1138478359877e-06, "loss": 0.9303, "step": 9419 }, { "epoch": 4.268237426370639, "grad_norm": 0.09177075279123684, "learning_rate": 5.107658998677463e-06, "loss": 0.9238, "step": 9420 }, { "epoch": 4.2686905301314, "grad_norm": 0.0979425420469692, "learning_rate": 5.101473653123212e-06, "loss": 0.9066, "step": 9421 }, { "epoch": 4.269143633892162, "grad_norm": 0.08274289193855014, "learning_rate": 5.0952917999439155e-06, "loss": 0.9302, "step": 9422 }, { "epoch": 4.269596737652923, "grad_norm": 0.09539937512424275, "learning_rate": 5.089113439758229e-06, "loss": 0.9178, "step": 9423 }, { "epoch": 4.270049841413684, "grad_norm": 0.09017406079823241, "learning_rate": 5.082938573184413e-06, "loss": 0.945, "step": 9424 }, { "epoch": 4.270502945174445, "grad_norm": 0.08842097202514615, "learning_rate": 5.0767672008404175e-06, "loss": 0.9278, "step": 9425 }, { "epoch": 4.270956048935206, "grad_norm": 0.09224173852607916, "learning_rate": 5.070599323343812e-06, "loss": 0.9161, "step": 9426 }, { "epoch": 4.271409152695967, "grad_norm": 0.08114541954107014, "learning_rate": 5.064434941311844e-06, "loss": 0.9379, "step": 9427 }, { "epoch": 4.271862256456728, "grad_norm": 0.12382991989259172, "learning_rate": 5.058274055361385e-06, "loss": 0.9405, "step": 9428 }, { "epoch": 4.2723153602174895, "grad_norm": 0.08483526015194864, "learning_rate": 5.052116666108981e-06, "loss": 0.9145, "step": 9429 }, { "epoch": 4.272768463978251, "grad_norm": 0.08767497831394253, "learning_rate": 5.045962774170811e-06, "loss": 0.9329, "step": 9430 }, { "epoch": 4.273221567739013, "grad_norm": 0.0905528463572058, "learning_rate": 5.039812380162694e-06, "loss": 0.9259, "step": 9431 }, { "epoch": 4.273674671499774, "grad_norm": 0.08133014632820192, "learning_rate": 5.033665484700141e-06, "loss": 0.9186, "step": 9432 }, { "epoch": 4.274127775260535, "grad_norm": 0.08617115361127665, "learning_rate": 5.02752208839826e-06, "loss": 0.952, "step": 9433 }, { "epoch": 4.274580879021296, "grad_norm": 0.08874621076645199, "learning_rate": 5.021382191871857e-06, "loss": 0.9337, "step": 9434 }, { "epoch": 4.275033982782057, "grad_norm": 0.08782694168841992, "learning_rate": 5.015245795735344e-06, "loss": 0.9312, "step": 9435 }, { "epoch": 4.275487086542818, "grad_norm": 0.08855647263557391, "learning_rate": 5.009112900602824e-06, "loss": 0.9362, "step": 9436 }, { "epoch": 4.275940190303579, "grad_norm": 0.08896478168727305, "learning_rate": 5.002983507088024e-06, "loss": 0.9391, "step": 9437 }, { "epoch": 4.2763932940643405, "grad_norm": 0.08821201412294415, "learning_rate": 4.996857615804312e-06, "loss": 0.9177, "step": 9438 }, { "epoch": 4.276846397825102, "grad_norm": 0.10164546418088738, "learning_rate": 4.990735227364738e-06, "loss": 0.9204, "step": 9439 }, { "epoch": 4.277299501585863, "grad_norm": 0.08366605784775213, "learning_rate": 4.984616342381969e-06, "loss": 0.9339, "step": 9440 }, { "epoch": 4.277752605346624, "grad_norm": 0.08276593946170933, "learning_rate": 4.978500961468355e-06, "loss": 0.9553, "step": 9441 }, { "epoch": 4.278205709107386, "grad_norm": 0.09679543449969467, "learning_rate": 4.972389085235851e-06, "loss": 0.9098, "step": 9442 }, { "epoch": 4.278658812868147, "grad_norm": 0.08759363606032608, "learning_rate": 4.96628071429611e-06, "loss": 0.9297, "step": 9443 }, { "epoch": 4.279111916628908, "grad_norm": 0.08648742384914554, "learning_rate": 4.9601758492603935e-06, "loss": 0.9351, "step": 9444 }, { "epoch": 4.279565020389669, "grad_norm": 0.09727153529582389, "learning_rate": 4.9540744907396395e-06, "loss": 0.9064, "step": 9445 }, { "epoch": 4.2800181241504305, "grad_norm": 0.08168251615373567, "learning_rate": 4.947976639344428e-06, "loss": 0.9335, "step": 9446 }, { "epoch": 4.280471227911192, "grad_norm": 0.0844238935130322, "learning_rate": 4.941882295684966e-06, "loss": 0.9191, "step": 9447 }, { "epoch": 4.280924331671953, "grad_norm": 0.08612127303936089, "learning_rate": 4.935791460371149e-06, "loss": 0.9124, "step": 9448 }, { "epoch": 4.281377435432714, "grad_norm": 0.08267051596911883, "learning_rate": 4.929704134012485e-06, "loss": 0.9235, "step": 9449 }, { "epoch": 4.281830539193475, "grad_norm": 0.08562282985385518, "learning_rate": 4.923620317218158e-06, "loss": 0.9332, "step": 9450 }, { "epoch": 4.282283642954236, "grad_norm": 0.08091824402316758, "learning_rate": 4.9175400105969795e-06, "loss": 0.9202, "step": 9451 }, { "epoch": 4.282736746714997, "grad_norm": 0.08162766947944117, "learning_rate": 4.911463214757435e-06, "loss": 0.9421, "step": 9452 }, { "epoch": 4.283189850475759, "grad_norm": 0.09255188541508078, "learning_rate": 4.905389930307625e-06, "loss": 0.9194, "step": 9453 }, { "epoch": 4.28364295423652, "grad_norm": 0.08325957615978047, "learning_rate": 4.899320157855334e-06, "loss": 0.9255, "step": 9454 }, { "epoch": 4.2840960579972815, "grad_norm": 0.08532918769152395, "learning_rate": 4.89325389800797e-06, "loss": 0.935, "step": 9455 }, { "epoch": 4.284549161758043, "grad_norm": 0.08664300066147707, "learning_rate": 4.887191151372585e-06, "loss": 0.9214, "step": 9456 }, { "epoch": 4.285002265518804, "grad_norm": 0.07885464345853656, "learning_rate": 4.881131918555917e-06, "loss": 0.9322, "step": 9457 }, { "epoch": 4.285455369279565, "grad_norm": 0.0946619655012351, "learning_rate": 4.875076200164302e-06, "loss": 0.9259, "step": 9458 }, { "epoch": 4.285908473040326, "grad_norm": 0.08017491938440306, "learning_rate": 4.8690239968037704e-06, "loss": 0.909, "step": 9459 }, { "epoch": 4.286361576801087, "grad_norm": 0.08101227997814944, "learning_rate": 4.862975309079967e-06, "loss": 0.9273, "step": 9460 }, { "epoch": 4.286814680561848, "grad_norm": 0.09286779817442589, "learning_rate": 4.856930137598204e-06, "loss": 0.9308, "step": 9461 }, { "epoch": 4.2872677843226095, "grad_norm": 0.08086950725267578, "learning_rate": 4.850888482963436e-06, "loss": 0.9218, "step": 9462 }, { "epoch": 4.2877208880833715, "grad_norm": 0.07908723000235886, "learning_rate": 4.844850345780257e-06, "loss": 0.9254, "step": 9463 }, { "epoch": 4.288173991844133, "grad_norm": 0.08543162009120754, "learning_rate": 4.838815726652923e-06, "loss": 0.9532, "step": 9464 }, { "epoch": 4.288627095604894, "grad_norm": 0.08127089433436023, "learning_rate": 4.832784626185332e-06, "loss": 0.9123, "step": 9465 }, { "epoch": 4.289080199365655, "grad_norm": 0.08009783551779988, "learning_rate": 4.8267570449810434e-06, "loss": 0.9576, "step": 9466 }, { "epoch": 4.289533303126416, "grad_norm": 0.08469858262594215, "learning_rate": 4.820732983643219e-06, "loss": 0.9229, "step": 9467 }, { "epoch": 4.289986406887177, "grad_norm": 0.08232066818869649, "learning_rate": 4.814712442774725e-06, "loss": 0.9193, "step": 9468 }, { "epoch": 4.290439510647938, "grad_norm": 0.084274261049214, "learning_rate": 4.808695422978038e-06, "loss": 0.9345, "step": 9469 }, { "epoch": 4.290892614408699, "grad_norm": 0.0801964343119507, "learning_rate": 4.8026819248553125e-06, "loss": 0.9154, "step": 9470 }, { "epoch": 4.2913457181694605, "grad_norm": 0.08558285328131697, "learning_rate": 4.796671949008307e-06, "loss": 0.9222, "step": 9471 }, { "epoch": 4.291798821930222, "grad_norm": 0.07751710596198505, "learning_rate": 4.790665496038482e-06, "loss": 0.9279, "step": 9472 }, { "epoch": 4.292251925690984, "grad_norm": 0.09185635914906742, "learning_rate": 4.7846625665469004e-06, "loss": 0.916, "step": 9473 }, { "epoch": 4.292705029451745, "grad_norm": 0.08423300610430763, "learning_rate": 4.77866316113428e-06, "loss": 0.9138, "step": 9474 }, { "epoch": 4.293158133212506, "grad_norm": 0.08549668373773883, "learning_rate": 4.7726672804010175e-06, "loss": 0.9442, "step": 9475 }, { "epoch": 4.293611236973267, "grad_norm": 0.07901540695225946, "learning_rate": 4.766674924947116e-06, "loss": 0.9094, "step": 9476 }, { "epoch": 4.294064340734028, "grad_norm": 0.08208206088460926, "learning_rate": 4.7606860953722575e-06, "loss": 0.9329, "step": 9477 }, { "epoch": 4.294517444494789, "grad_norm": 0.07984988350188281, "learning_rate": 4.754700792275744e-06, "loss": 0.9052, "step": 9478 }, { "epoch": 4.2949705482555505, "grad_norm": 0.07652389202391825, "learning_rate": 4.748719016256558e-06, "loss": 0.9036, "step": 9479 }, { "epoch": 4.295423652016312, "grad_norm": 0.14824157875298505, "learning_rate": 4.742740767913283e-06, "loss": 0.9339, "step": 9480 }, { "epoch": 4.295876755777073, "grad_norm": 0.08166796393531876, "learning_rate": 4.736766047844206e-06, "loss": 0.9193, "step": 9481 }, { "epoch": 4.296329859537834, "grad_norm": 0.08160657879993775, "learning_rate": 4.730794856647207e-06, "loss": 0.9202, "step": 9482 }, { "epoch": 4.296782963298595, "grad_norm": 0.08003311581830427, "learning_rate": 4.724827194919841e-06, "loss": 0.9324, "step": 9483 }, { "epoch": 4.297236067059357, "grad_norm": 0.07530339755387983, "learning_rate": 4.718863063259318e-06, "loss": 0.9066, "step": 9484 }, { "epoch": 4.297689170820118, "grad_norm": 0.07540834053798795, "learning_rate": 4.712902462262468e-06, "loss": 0.9043, "step": 9485 }, { "epoch": 4.298142274580879, "grad_norm": 0.07801505252656911, "learning_rate": 4.70694539252579e-06, "loss": 0.9205, "step": 9486 }, { "epoch": 4.29859537834164, "grad_norm": 0.08093396718197733, "learning_rate": 4.7009918546454135e-06, "loss": 0.9391, "step": 9487 }, { "epoch": 4.2990484821024015, "grad_norm": 0.07666768337653639, "learning_rate": 4.6950418492171365e-06, "loss": 0.9281, "step": 9488 }, { "epoch": 4.299501585863163, "grad_norm": 0.08034447259599414, "learning_rate": 4.689095376836376e-06, "loss": 0.9211, "step": 9489 }, { "epoch": 4.299954689623924, "grad_norm": 0.07770275940088638, "learning_rate": 4.683152438098218e-06, "loss": 0.9284, "step": 9490 }, { "epoch": 4.300407793384685, "grad_norm": 0.07894808790369136, "learning_rate": 4.677213033597383e-06, "loss": 0.9168, "step": 9491 }, { "epoch": 4.300860897145446, "grad_norm": 0.0860156072739337, "learning_rate": 4.671277163928234e-06, "loss": 0.9439, "step": 9492 }, { "epoch": 4.301314000906207, "grad_norm": 0.07667753752009882, "learning_rate": 4.6653448296848015e-06, "loss": 0.926, "step": 9493 }, { "epoch": 4.301767104666968, "grad_norm": 0.07681572039267442, "learning_rate": 4.659416031460726e-06, "loss": 0.9412, "step": 9494 }, { "epoch": 4.30222020842773, "grad_norm": 0.08086262632880607, "learning_rate": 4.653490769849342e-06, "loss": 0.9289, "step": 9495 }, { "epoch": 4.3026733121884915, "grad_norm": 0.08434155088233473, "learning_rate": 4.647569045443581e-06, "loss": 0.8974, "step": 9496 }, { "epoch": 4.303126415949253, "grad_norm": 0.07848026536122463, "learning_rate": 4.641650858836064e-06, "loss": 0.9376, "step": 9497 }, { "epoch": 4.303579519710014, "grad_norm": 0.08305149518609925, "learning_rate": 4.635736210619022e-06, "loss": 0.9406, "step": 9498 }, { "epoch": 4.304032623470775, "grad_norm": 0.07795568625774921, "learning_rate": 4.629825101384348e-06, "loss": 0.933, "step": 9499 }, { "epoch": 4.304485727231536, "grad_norm": 0.08893249802734353, "learning_rate": 4.62391753172359e-06, "loss": 0.9358, "step": 9500 }, { "epoch": 4.304938830992297, "grad_norm": 0.08364321825776089, "learning_rate": 4.61801350222792e-06, "loss": 0.9194, "step": 9501 }, { "epoch": 4.305391934753058, "grad_norm": 0.0819363106604072, "learning_rate": 4.61211301348818e-06, "loss": 0.895, "step": 9502 }, { "epoch": 4.305845038513819, "grad_norm": 0.08312320967880125, "learning_rate": 4.606216066094833e-06, "loss": 0.9099, "step": 9503 }, { "epoch": 4.3062981422745805, "grad_norm": 0.08116970940513815, "learning_rate": 4.600322660638013e-06, "loss": 0.9092, "step": 9504 }, { "epoch": 4.306751246035342, "grad_norm": 0.09031539995507086, "learning_rate": 4.594432797707469e-06, "loss": 0.9237, "step": 9505 }, { "epoch": 4.307204349796104, "grad_norm": 0.07936095094232788, "learning_rate": 4.588546477892632e-06, "loss": 0.902, "step": 9506 }, { "epoch": 4.307657453556865, "grad_norm": 0.0806718506277223, "learning_rate": 4.582663701782557e-06, "loss": 0.9331, "step": 9507 }, { "epoch": 4.308110557317626, "grad_norm": 0.09049993285456125, "learning_rate": 4.576784469965927e-06, "loss": 0.9278, "step": 9508 }, { "epoch": 4.308563661078387, "grad_norm": 0.08069369021000938, "learning_rate": 4.570908783031116e-06, "loss": 0.9158, "step": 9509 }, { "epoch": 4.309016764839148, "grad_norm": 0.07805719060461656, "learning_rate": 4.565036641566094e-06, "loss": 0.923, "step": 9510 }, { "epoch": 4.309469868599909, "grad_norm": 0.08882196760993723, "learning_rate": 4.5591680461585195e-06, "loss": 0.9227, "step": 9511 }, { "epoch": 4.3099229723606705, "grad_norm": 0.08756668838780439, "learning_rate": 4.553302997395661e-06, "loss": 0.936, "step": 9512 }, { "epoch": 4.310376076121432, "grad_norm": 0.09081938319257232, "learning_rate": 4.547441495864462e-06, "loss": 0.9285, "step": 9513 }, { "epoch": 4.310829179882193, "grad_norm": 0.09767273992751223, "learning_rate": 4.541583542151484e-06, "loss": 0.9267, "step": 9514 }, { "epoch": 4.311282283642954, "grad_norm": 0.08442747860869247, "learning_rate": 4.535729136842944e-06, "loss": 0.9208, "step": 9515 }, { "epoch": 4.311735387403715, "grad_norm": 0.08430994115126973, "learning_rate": 4.529878280524722e-06, "loss": 0.9221, "step": 9516 }, { "epoch": 4.312188491164477, "grad_norm": 0.09288697942363301, "learning_rate": 4.524030973782308e-06, "loss": 0.9323, "step": 9517 }, { "epoch": 4.312641594925238, "grad_norm": 0.07921796074766146, "learning_rate": 4.518187217200867e-06, "loss": 0.9247, "step": 9518 }, { "epoch": 4.313094698685999, "grad_norm": 0.07871565405018019, "learning_rate": 4.51234701136519e-06, "loss": 0.9168, "step": 9519 }, { "epoch": 4.31354780244676, "grad_norm": 0.08277377019605046, "learning_rate": 4.50651035685973e-06, "loss": 0.917, "step": 9520 }, { "epoch": 4.3140009062075215, "grad_norm": 0.07444541230344809, "learning_rate": 4.5006772542685575e-06, "loss": 0.934, "step": 9521 }, { "epoch": 4.314454009968283, "grad_norm": 0.07921748305033062, "learning_rate": 4.494847704175422e-06, "loss": 0.9489, "step": 9522 }, { "epoch": 4.314907113729044, "grad_norm": 0.0795733027743604, "learning_rate": 4.489021707163694e-06, "loss": 0.9223, "step": 9523 }, { "epoch": 4.315360217489805, "grad_norm": 0.07382806008891546, "learning_rate": 4.483199263816392e-06, "loss": 0.9219, "step": 9524 }, { "epoch": 4.315813321250566, "grad_norm": 0.07671064099479048, "learning_rate": 4.477380374716171e-06, "loss": 0.9354, "step": 9525 }, { "epoch": 4.316266425011327, "grad_norm": 0.07946097393746071, "learning_rate": 4.471565040445365e-06, "loss": 0.9497, "step": 9526 }, { "epoch": 4.316719528772089, "grad_norm": 0.07418431580297781, "learning_rate": 4.46575326158591e-06, "loss": 0.9104, "step": 9527 }, { "epoch": 4.31717263253285, "grad_norm": 0.08081758959761084, "learning_rate": 4.4599450387194e-06, "loss": 0.925, "step": 9528 }, { "epoch": 4.3176257362936115, "grad_norm": 0.08538262733434426, "learning_rate": 4.454140372427089e-06, "loss": 0.9318, "step": 9529 }, { "epoch": 4.318078840054373, "grad_norm": 0.0773557789717134, "learning_rate": 4.448339263289856e-06, "loss": 0.9041, "step": 9530 }, { "epoch": 4.318531943815134, "grad_norm": 0.0891845009677287, "learning_rate": 4.44254171188824e-06, "loss": 0.9385, "step": 9531 }, { "epoch": 4.318985047575895, "grad_norm": 0.08181982284151433, "learning_rate": 4.4367477188024035e-06, "loss": 0.9544, "step": 9532 }, { "epoch": 4.319438151336656, "grad_norm": 0.07949151656362582, "learning_rate": 4.430957284612176e-06, "loss": 0.9064, "step": 9533 }, { "epoch": 4.319891255097417, "grad_norm": 0.07339830981985025, "learning_rate": 4.425170409897015e-06, "loss": 0.9055, "step": 9534 }, { "epoch": 4.320344358858178, "grad_norm": 0.07875816551645815, "learning_rate": 4.419387095236016e-06, "loss": 0.9193, "step": 9535 }, { "epoch": 4.320797462618939, "grad_norm": 0.07591247625600107, "learning_rate": 4.413607341207944e-06, "loss": 0.9343, "step": 9536 }, { "epoch": 4.321250566379701, "grad_norm": 0.07468256463789054, "learning_rate": 4.40783114839118e-06, "loss": 0.9119, "step": 9537 }, { "epoch": 4.3217036701404625, "grad_norm": 0.08336197972714161, "learning_rate": 4.402058517363772e-06, "loss": 0.9162, "step": 9538 }, { "epoch": 4.322156773901224, "grad_norm": 0.07566485940896703, "learning_rate": 4.3962894487033884e-06, "loss": 0.9235, "step": 9539 }, { "epoch": 4.322609877661985, "grad_norm": 0.07740116223341252, "learning_rate": 4.390523942987371e-06, "loss": 0.911, "step": 9540 }, { "epoch": 4.323062981422746, "grad_norm": 0.08706285191429086, "learning_rate": 4.384762000792662e-06, "loss": 0.9133, "step": 9541 }, { "epoch": 4.323516085183507, "grad_norm": 0.07857317908022075, "learning_rate": 4.379003622695894e-06, "loss": 0.9194, "step": 9542 }, { "epoch": 4.323969188944268, "grad_norm": 0.09090472098995868, "learning_rate": 4.373248809273314e-06, "loss": 0.9086, "step": 9543 }, { "epoch": 4.324422292705029, "grad_norm": 0.0816929416300493, "learning_rate": 4.3674975611008105e-06, "loss": 0.914, "step": 9544 }, { "epoch": 4.3248753964657904, "grad_norm": 0.08339832103547476, "learning_rate": 4.361749878753938e-06, "loss": 0.9182, "step": 9545 }, { "epoch": 4.325328500226552, "grad_norm": 0.08445633540721416, "learning_rate": 4.356005762807867e-06, "loss": 0.9092, "step": 9546 }, { "epoch": 4.325781603987313, "grad_norm": 0.0811717112559621, "learning_rate": 4.350265213837435e-06, "loss": 0.937, "step": 9547 }, { "epoch": 4.326234707748075, "grad_norm": 0.07676659798285457, "learning_rate": 4.344528232417102e-06, "loss": 0.93, "step": 9548 }, { "epoch": 4.326687811508836, "grad_norm": 0.08370617276120657, "learning_rate": 4.338794819120998e-06, "loss": 0.9044, "step": 9549 }, { "epoch": 4.327140915269597, "grad_norm": 0.0863268119036863, "learning_rate": 4.333064974522852e-06, "loss": 0.916, "step": 9550 }, { "epoch": 4.327594019030358, "grad_norm": 0.08070302951353472, "learning_rate": 4.327338699196091e-06, "loss": 0.9373, "step": 9551 }, { "epoch": 4.328047122791119, "grad_norm": 0.08009588493678907, "learning_rate": 4.321615993713737e-06, "loss": 0.9401, "step": 9552 }, { "epoch": 4.32850022655188, "grad_norm": 0.08056477405032399, "learning_rate": 4.315896858648478e-06, "loss": 0.9072, "step": 9553 }, { "epoch": 4.3289533303126415, "grad_norm": 0.08280596246013247, "learning_rate": 4.3101812945726484e-06, "loss": 0.9282, "step": 9554 }, { "epoch": 4.329406434073403, "grad_norm": 0.08123107349828687, "learning_rate": 4.304469302058203e-06, "loss": 0.922, "step": 9555 }, { "epoch": 4.329859537834164, "grad_norm": 0.07808396167679903, "learning_rate": 4.2987608816767735e-06, "loss": 0.8949, "step": 9556 }, { "epoch": 4.330312641594925, "grad_norm": 0.09305702850042247, "learning_rate": 4.293056033999596e-06, "loss": 0.9133, "step": 9557 }, { "epoch": 4.330765745355686, "grad_norm": 0.08679609494558149, "learning_rate": 4.287354759597584e-06, "loss": 0.9335, "step": 9558 }, { "epoch": 4.331218849116448, "grad_norm": 0.07941693528426906, "learning_rate": 4.281657059041267e-06, "loss": 0.9013, "step": 9559 }, { "epoch": 4.331671952877209, "grad_norm": 0.08690270048738402, "learning_rate": 4.275962932900824e-06, "loss": 0.9529, "step": 9560 }, { "epoch": 4.33212505663797, "grad_norm": 0.08289813824737988, "learning_rate": 4.270272381746088e-06, "loss": 0.8959, "step": 9561 }, { "epoch": 4.3325781603987314, "grad_norm": 0.08097272507700903, "learning_rate": 4.264585406146515e-06, "loss": 0.9276, "step": 9562 }, { "epoch": 4.333031264159493, "grad_norm": 0.07888007273159676, "learning_rate": 4.258902006671229e-06, "loss": 0.906, "step": 9563 }, { "epoch": 4.333484367920254, "grad_norm": 0.07837569572601756, "learning_rate": 4.2532221838889635e-06, "loss": 0.9073, "step": 9564 }, { "epoch": 4.333937471681015, "grad_norm": 0.08501723884033165, "learning_rate": 4.247545938368123e-06, "loss": 0.915, "step": 9565 }, { "epoch": 4.334390575441776, "grad_norm": 0.07753699162746266, "learning_rate": 4.241873270676738e-06, "loss": 0.902, "step": 9566 }, { "epoch": 4.334843679202537, "grad_norm": 0.08150814661144694, "learning_rate": 4.236204181382486e-06, "loss": 0.9322, "step": 9567 }, { "epoch": 4.335296782963298, "grad_norm": 0.07928218995936105, "learning_rate": 4.230538671052693e-06, "loss": 0.9241, "step": 9568 }, { "epoch": 4.335749886724059, "grad_norm": 0.07703022310363522, "learning_rate": 4.2248767402543e-06, "loss": 0.9262, "step": 9569 }, { "epoch": 4.336202990484821, "grad_norm": 0.08030342312184863, "learning_rate": 4.21921838955393e-06, "loss": 0.9347, "step": 9570 }, { "epoch": 4.3366560942455825, "grad_norm": 0.07955998434948858, "learning_rate": 4.213563619517809e-06, "loss": 0.9273, "step": 9571 }, { "epoch": 4.337109198006344, "grad_norm": 0.08148386450551132, "learning_rate": 4.20791243071184e-06, "loss": 0.9335, "step": 9572 }, { "epoch": 4.337562301767105, "grad_norm": 0.08206928642077155, "learning_rate": 4.202264823701536e-06, "loss": 0.9209, "step": 9573 }, { "epoch": 4.338015405527866, "grad_norm": 0.07978035058111708, "learning_rate": 4.196620799052084e-06, "loss": 0.9144, "step": 9574 }, { "epoch": 4.338468509288627, "grad_norm": 0.08145483637917753, "learning_rate": 4.190980357328278e-06, "loss": 0.9127, "step": 9575 }, { "epoch": 4.338921613049388, "grad_norm": 0.07696917267621205, "learning_rate": 4.185343499094568e-06, "loss": 0.9157, "step": 9576 }, { "epoch": 4.339374716810149, "grad_norm": 0.07829796719564222, "learning_rate": 4.179710224915061e-06, "loss": 0.9302, "step": 9577 }, { "epoch": 4.33982782057091, "grad_norm": 0.08536093367594634, "learning_rate": 4.174080535353478e-06, "loss": 0.9441, "step": 9578 }, { "epoch": 4.340280924331672, "grad_norm": 0.08871297638719655, "learning_rate": 4.1684544309732104e-06, "loss": 0.9179, "step": 9579 }, { "epoch": 4.340734028092434, "grad_norm": 0.08289591613080764, "learning_rate": 4.162831912337262e-06, "loss": 0.9051, "step": 9580 }, { "epoch": 4.341187131853195, "grad_norm": 0.08203919657580976, "learning_rate": 4.157212980008298e-06, "loss": 0.917, "step": 9581 }, { "epoch": 4.341640235613956, "grad_norm": 0.0851918671461655, "learning_rate": 4.151597634548608e-06, "loss": 0.9157, "step": 9582 }, { "epoch": 4.342093339374717, "grad_norm": 0.09387167890118675, "learning_rate": 4.145985876520144e-06, "loss": 0.9202, "step": 9583 }, { "epoch": 4.342546443135478, "grad_norm": 0.07774820773694098, "learning_rate": 4.140377706484478e-06, "loss": 0.9409, "step": 9584 }, { "epoch": 4.342999546896239, "grad_norm": 0.0906705480804487, "learning_rate": 4.134773125002842e-06, "loss": 0.8924, "step": 9585 }, { "epoch": 4.343452650657, "grad_norm": 0.08765687740269165, "learning_rate": 4.129172132636088e-06, "loss": 0.9396, "step": 9586 }, { "epoch": 4.3439057544177615, "grad_norm": 0.07993478020206987, "learning_rate": 4.123574729944735e-06, "loss": 0.9249, "step": 9587 }, { "epoch": 4.344358858178523, "grad_norm": 0.08782485189715476, "learning_rate": 4.117980917488917e-06, "loss": 0.9182, "step": 9588 }, { "epoch": 4.344811961939284, "grad_norm": 0.08321297235241157, "learning_rate": 4.112390695828414e-06, "loss": 0.9198, "step": 9589 }, { "epoch": 4.345265065700046, "grad_norm": 0.07988296890673066, "learning_rate": 4.106804065522672e-06, "loss": 0.9322, "step": 9590 }, { "epoch": 4.345718169460807, "grad_norm": 0.08836899653603525, "learning_rate": 4.101221027130731e-06, "loss": 0.9049, "step": 9591 }, { "epoch": 4.346171273221568, "grad_norm": 0.08719589574860293, "learning_rate": 4.095641581211327e-06, "loss": 0.933, "step": 9592 }, { "epoch": 4.346624376982329, "grad_norm": 0.08101703564388427, "learning_rate": 4.090065728322783e-06, "loss": 0.9441, "step": 9593 }, { "epoch": 4.34707748074309, "grad_norm": 0.08454545332124554, "learning_rate": 4.084493469023105e-06, "loss": 0.9311, "step": 9594 }, { "epoch": 4.347530584503851, "grad_norm": 0.08364054786408948, "learning_rate": 4.078924803869919e-06, "loss": 0.9358, "step": 9595 }, { "epoch": 4.347983688264613, "grad_norm": 0.08109716885537772, "learning_rate": 4.073359733420481e-06, "loss": 0.9422, "step": 9596 }, { "epoch": 4.348436792025374, "grad_norm": 0.07682694123560103, "learning_rate": 4.067798258231714e-06, "loss": 0.9138, "step": 9597 }, { "epoch": 4.348889895786135, "grad_norm": 0.07572881937641501, "learning_rate": 4.0622403788601564e-06, "loss": 0.9327, "step": 9598 }, { "epoch": 4.349342999546896, "grad_norm": 0.08560605345711986, "learning_rate": 4.056686095862015e-06, "loss": 0.9191, "step": 9599 }, { "epoch": 4.349796103307657, "grad_norm": 0.08091479248643065, "learning_rate": 4.051135409793099e-06, "loss": 0.9262, "step": 9600 }, { "epoch": 4.350249207068419, "grad_norm": 0.078732056979989, "learning_rate": 4.045588321208893e-06, "loss": 0.9095, "step": 9601 }, { "epoch": 4.35070231082918, "grad_norm": 0.07985581020558931, "learning_rate": 4.040044830664495e-06, "loss": 0.9356, "step": 9602 }, { "epoch": 4.351155414589941, "grad_norm": 0.07952417605971886, "learning_rate": 4.034504938714671e-06, "loss": 0.9415, "step": 9603 }, { "epoch": 4.3516085183507025, "grad_norm": 0.08166096633496833, "learning_rate": 4.0289686459138e-06, "loss": 0.9119, "step": 9604 }, { "epoch": 4.352061622111464, "grad_norm": 0.07316899006046348, "learning_rate": 4.023435952815904e-06, "loss": 0.9289, "step": 9605 }, { "epoch": 4.352514725872225, "grad_norm": 0.08648837155563688, "learning_rate": 4.017906859974665e-06, "loss": 0.913, "step": 9606 }, { "epoch": 4.352967829632986, "grad_norm": 0.08207148226455596, "learning_rate": 4.0123813679433835e-06, "loss": 0.9007, "step": 9607 }, { "epoch": 4.353420933393747, "grad_norm": 0.08134225236127604, "learning_rate": 4.0068594772750206e-06, "loss": 0.9471, "step": 9608 }, { "epoch": 4.353874037154508, "grad_norm": 0.08782354768791446, "learning_rate": 4.0013411885221476e-06, "loss": 0.9313, "step": 9609 }, { "epoch": 4.354327140915269, "grad_norm": 0.08314644053865125, "learning_rate": 3.995826502237008e-06, "loss": 0.9243, "step": 9610 }, { "epoch": 4.35478024467603, "grad_norm": 0.07945397671522228, "learning_rate": 3.990315418971453e-06, "loss": 0.9069, "step": 9611 }, { "epoch": 4.355233348436792, "grad_norm": 0.07772255897146213, "learning_rate": 3.9848079392770065e-06, "loss": 0.9291, "step": 9612 }, { "epoch": 4.355686452197554, "grad_norm": 0.08591943672702242, "learning_rate": 3.979304063704805e-06, "loss": 0.9254, "step": 9613 }, { "epoch": 4.356139555958315, "grad_norm": 0.08615654257662198, "learning_rate": 3.973803792805626e-06, "loss": 0.9562, "step": 9614 }, { "epoch": 4.356592659719076, "grad_norm": 0.08628725366528284, "learning_rate": 3.9683071271299136e-06, "loss": 0.9385, "step": 9615 }, { "epoch": 4.357045763479837, "grad_norm": 0.08841651671542072, "learning_rate": 3.962814067227707e-06, "loss": 0.9059, "step": 9616 }, { "epoch": 4.357498867240598, "grad_norm": 0.0801844087432536, "learning_rate": 3.957324613648736e-06, "loss": 0.9393, "step": 9617 }, { "epoch": 4.357951971001359, "grad_norm": 0.08920829744521662, "learning_rate": 3.951838766942322e-06, "loss": 0.9241, "step": 9618 }, { "epoch": 4.35840507476212, "grad_norm": 0.08018356625806668, "learning_rate": 3.94635652765746e-06, "loss": 0.9205, "step": 9619 }, { "epoch": 4.3588581785228815, "grad_norm": 0.07334887846960211, "learning_rate": 3.9408778963427655e-06, "loss": 0.9303, "step": 9620 }, { "epoch": 4.359311282283643, "grad_norm": 0.07778365277525891, "learning_rate": 3.935402873546488e-06, "loss": 0.9198, "step": 9621 }, { "epoch": 4.359764386044404, "grad_norm": 0.0789684525989924, "learning_rate": 3.9299314598165365e-06, "loss": 0.9326, "step": 9622 }, { "epoch": 4.360217489805166, "grad_norm": 0.0771568303148584, "learning_rate": 3.9244636557004415e-06, "loss": 0.913, "step": 9623 }, { "epoch": 4.360670593565927, "grad_norm": 0.07795497446678681, "learning_rate": 3.918999461745392e-06, "loss": 0.9219, "step": 9624 }, { "epoch": 4.361123697326688, "grad_norm": 0.09087626149242298, "learning_rate": 3.913538878498182e-06, "loss": 0.9347, "step": 9625 }, { "epoch": 4.361576801087449, "grad_norm": 0.08525823045870491, "learning_rate": 3.908081906505281e-06, "loss": 0.9452, "step": 9626 }, { "epoch": 4.36202990484821, "grad_norm": 0.0785155104458813, "learning_rate": 3.9026285463127725e-06, "loss": 0.9138, "step": 9627 }, { "epoch": 4.362483008608971, "grad_norm": 0.07392760480360713, "learning_rate": 3.8971787984663925e-06, "loss": 0.9347, "step": 9628 }, { "epoch": 4.3629361123697326, "grad_norm": 0.07675355233053911, "learning_rate": 3.8917326635115045e-06, "loss": 0.9018, "step": 9629 }, { "epoch": 4.363389216130494, "grad_norm": 0.07797843390025977, "learning_rate": 3.88629014199311e-06, "loss": 0.9395, "step": 9630 }, { "epoch": 4.363842319891255, "grad_norm": 0.07882741108122467, "learning_rate": 3.880851234455869e-06, "loss": 0.9376, "step": 9631 }, { "epoch": 4.364295423652016, "grad_norm": 0.08518804747295836, "learning_rate": 3.875415941444054e-06, "loss": 0.9305, "step": 9632 }, { "epoch": 4.364748527412777, "grad_norm": 0.07997910480602202, "learning_rate": 3.869984263501594e-06, "loss": 0.9553, "step": 9633 }, { "epoch": 4.365201631173539, "grad_norm": 0.07548047118650696, "learning_rate": 3.864556201172041e-06, "loss": 0.9265, "step": 9634 }, { "epoch": 4.3656547349343, "grad_norm": 0.07805984978445796, "learning_rate": 3.8591317549986085e-06, "loss": 0.9101, "step": 9635 }, { "epoch": 4.366107838695061, "grad_norm": 0.082698535893654, "learning_rate": 3.853710925524121e-06, "loss": 0.9397, "step": 9636 }, { "epoch": 4.3665609424558225, "grad_norm": 0.07729403953831705, "learning_rate": 3.848293713291056e-06, "loss": 0.8934, "step": 9637 }, { "epoch": 4.367014046216584, "grad_norm": 0.07397624251842908, "learning_rate": 3.842880118841516e-06, "loss": 0.916, "step": 9638 }, { "epoch": 4.367467149977345, "grad_norm": 0.07686044671434959, "learning_rate": 3.837470142717275e-06, "loss": 0.9198, "step": 9639 }, { "epoch": 4.367920253738106, "grad_norm": 0.08162102842719947, "learning_rate": 3.8320637854597055e-06, "loss": 0.9358, "step": 9640 }, { "epoch": 4.368373357498867, "grad_norm": 0.08142329647140106, "learning_rate": 3.82666104760983e-06, "loss": 0.9265, "step": 9641 }, { "epoch": 4.368826461259628, "grad_norm": 0.07867437563610971, "learning_rate": 3.821261929708327e-06, "loss": 0.9522, "step": 9642 }, { "epoch": 4.369279565020389, "grad_norm": 0.0768380460702517, "learning_rate": 3.8158664322954785e-06, "loss": 0.9183, "step": 9643 }, { "epoch": 4.369732668781151, "grad_norm": 0.07984473938603422, "learning_rate": 3.81047455591125e-06, "loss": 0.9333, "step": 9644 }, { "epoch": 4.370185772541912, "grad_norm": 0.07900465594175227, "learning_rate": 3.8050863010952e-06, "loss": 0.9164, "step": 9645 }, { "epoch": 4.370638876302674, "grad_norm": 0.07969030551439393, "learning_rate": 3.7997016683865506e-06, "loss": 0.9304, "step": 9646 }, { "epoch": 4.371091980063435, "grad_norm": 0.07885935371527812, "learning_rate": 3.794320658324151e-06, "loss": 0.9033, "step": 9647 }, { "epoch": 4.371545083824196, "grad_norm": 0.07733200649432383, "learning_rate": 3.7889432714464992e-06, "loss": 0.9054, "step": 9648 }, { "epoch": 4.371998187584957, "grad_norm": 0.07742084738515137, "learning_rate": 3.7835695082917156e-06, "loss": 0.9497, "step": 9649 }, { "epoch": 4.372451291345718, "grad_norm": 0.08246532574111773, "learning_rate": 3.7781993693975573e-06, "loss": 0.933, "step": 9650 }, { "epoch": 4.372904395106479, "grad_norm": 0.07927905382111537, "learning_rate": 3.7728328553014425e-06, "loss": 0.9257, "step": 9651 }, { "epoch": 4.37335749886724, "grad_norm": 0.07533079384935187, "learning_rate": 3.7674699665403956e-06, "loss": 0.9142, "step": 9652 }, { "epoch": 4.3738106026280015, "grad_norm": 0.07902243736080207, "learning_rate": 3.7621107036511074e-06, "loss": 0.9179, "step": 9653 }, { "epoch": 4.3742637063887635, "grad_norm": 0.07870138306557889, "learning_rate": 3.756755067169877e-06, "loss": 0.9226, "step": 9654 }, { "epoch": 4.374716810149525, "grad_norm": 0.07471632406628173, "learning_rate": 3.7514030576326677e-06, "loss": 0.916, "step": 9655 }, { "epoch": 4.375169913910286, "grad_norm": 0.07128288777822789, "learning_rate": 3.746054675575064e-06, "loss": 0.8956, "step": 9656 }, { "epoch": 4.375623017671047, "grad_norm": 0.08684056509407517, "learning_rate": 3.7407099215322818e-06, "loss": 0.9182, "step": 9657 }, { "epoch": 4.376076121431808, "grad_norm": 0.0832426775789872, "learning_rate": 3.735368796039196e-06, "loss": 0.914, "step": 9658 }, { "epoch": 4.376529225192569, "grad_norm": 0.0789132445119734, "learning_rate": 3.73003129963029e-06, "loss": 0.9032, "step": 9659 }, { "epoch": 4.37698232895333, "grad_norm": 0.0816703404294499, "learning_rate": 3.724697432839719e-06, "loss": 0.9366, "step": 9660 }, { "epoch": 4.377435432714091, "grad_norm": 0.08133964112841753, "learning_rate": 3.7193671962012334e-06, "loss": 0.9165, "step": 9661 }, { "epoch": 4.3778885364748525, "grad_norm": 0.07585064053230406, "learning_rate": 3.7140405902482603e-06, "loss": 0.9345, "step": 9662 }, { "epoch": 4.378341640235614, "grad_norm": 0.07461823980485857, "learning_rate": 3.708717615513835e-06, "loss": 0.9019, "step": 9663 }, { "epoch": 4.378794743996375, "grad_norm": 0.08040078187357341, "learning_rate": 3.703398272530647e-06, "loss": 0.9355, "step": 9664 }, { "epoch": 4.379247847757137, "grad_norm": 0.07599975602453937, "learning_rate": 3.698082561831009e-06, "loss": 0.9216, "step": 9665 }, { "epoch": 4.379700951517898, "grad_norm": 0.07819181757743568, "learning_rate": 3.6927704839468725e-06, "loss": 0.9144, "step": 9666 }, { "epoch": 4.380154055278659, "grad_norm": 0.08111861680920995, "learning_rate": 3.6874620394098436e-06, "loss": 0.9236, "step": 9667 }, { "epoch": 4.38060715903942, "grad_norm": 0.07579021880183257, "learning_rate": 3.682157228751129e-06, "loss": 0.9264, "step": 9668 }, { "epoch": 4.381060262800181, "grad_norm": 0.07626736588901335, "learning_rate": 3.6768560525016186e-06, "loss": 0.9134, "step": 9669 }, { "epoch": 4.3815133665609425, "grad_norm": 0.07507081481056299, "learning_rate": 3.671558511191791e-06, "loss": 0.9409, "step": 9670 }, { "epoch": 4.381966470321704, "grad_norm": 0.08102844378406934, "learning_rate": 3.6662646053517992e-06, "loss": 0.9202, "step": 9671 }, { "epoch": 4.382419574082465, "grad_norm": 0.07209365345796281, "learning_rate": 3.660974335511407e-06, "loss": 0.9271, "step": 9672 }, { "epoch": 4.382872677843226, "grad_norm": 0.07913702036978214, "learning_rate": 3.6556877022000214e-06, "loss": 0.9172, "step": 9673 }, { "epoch": 4.383325781603987, "grad_norm": 0.07871813584317405, "learning_rate": 3.650404705946704e-06, "loss": 0.9159, "step": 9674 }, { "epoch": 4.383778885364748, "grad_norm": 0.07790853444312333, "learning_rate": 3.645125347280112e-06, "loss": 0.9158, "step": 9675 }, { "epoch": 4.38423198912551, "grad_norm": 0.07788390214997822, "learning_rate": 3.6398496267285865e-06, "loss": 0.9247, "step": 9676 }, { "epoch": 4.384685092886271, "grad_norm": 0.08315036118307613, "learning_rate": 3.6345775448200615e-06, "loss": 0.946, "step": 9677 }, { "epoch": 4.385138196647032, "grad_norm": 0.07959007259374548, "learning_rate": 3.6293091020821413e-06, "loss": 0.9326, "step": 9678 }, { "epoch": 4.3855913004077935, "grad_norm": 0.08695504167729394, "learning_rate": 3.6240442990420397e-06, "loss": 0.9181, "step": 9679 }, { "epoch": 4.386044404168555, "grad_norm": 0.07511036234684296, "learning_rate": 3.618783136226629e-06, "loss": 0.8995, "step": 9680 }, { "epoch": 4.386497507929316, "grad_norm": 0.07654825597321159, "learning_rate": 3.613525614162399e-06, "loss": 0.9236, "step": 9681 }, { "epoch": 4.386950611690077, "grad_norm": 0.08057429595648549, "learning_rate": 3.6082717333754746e-06, "loss": 0.9123, "step": 9682 }, { "epoch": 4.387403715450838, "grad_norm": 0.07812998040036942, "learning_rate": 3.603021494391641e-06, "loss": 0.8967, "step": 9683 }, { "epoch": 4.387856819211599, "grad_norm": 0.07240624701635441, "learning_rate": 3.5977748977362813e-06, "loss": 0.9062, "step": 9684 }, { "epoch": 4.38830992297236, "grad_norm": 0.08040615456958876, "learning_rate": 3.592531943934452e-06, "loss": 0.9204, "step": 9685 }, { "epoch": 4.3887630267331215, "grad_norm": 0.07615994379252934, "learning_rate": 3.587292633510817e-06, "loss": 0.9298, "step": 9686 }, { "epoch": 4.3892161304938835, "grad_norm": 0.0749820775725777, "learning_rate": 3.582056966989691e-06, "loss": 0.9243, "step": 9687 }, { "epoch": 4.389669234254645, "grad_norm": 0.07948192072392277, "learning_rate": 3.5768249448950145e-06, "loss": 0.9103, "step": 9688 }, { "epoch": 4.390122338015406, "grad_norm": 0.08194448315853153, "learning_rate": 3.5715965677503773e-06, "loss": 0.9377, "step": 9689 }, { "epoch": 4.390575441776167, "grad_norm": 0.08215942570912989, "learning_rate": 3.5663718360789877e-06, "loss": 0.9281, "step": 9690 }, { "epoch": 4.391028545536928, "grad_norm": 0.07952354730695521, "learning_rate": 3.561150750403686e-06, "loss": 0.9312, "step": 9691 }, { "epoch": 4.391481649297689, "grad_norm": 0.07551056087141085, "learning_rate": 3.5559333112469816e-06, "loss": 0.9327, "step": 9692 }, { "epoch": 4.39193475305845, "grad_norm": 0.08011565299030247, "learning_rate": 3.550719519130974e-06, "loss": 0.9227, "step": 9693 }, { "epoch": 4.392387856819211, "grad_norm": 0.08138349021556676, "learning_rate": 3.545509374577445e-06, "loss": 0.9407, "step": 9694 }, { "epoch": 4.3928409605799725, "grad_norm": 0.0785425457953518, "learning_rate": 3.540302878107751e-06, "loss": 0.9299, "step": 9695 }, { "epoch": 4.393294064340734, "grad_norm": 0.07938147677767284, "learning_rate": 3.535100030242946e-06, "loss": 0.9209, "step": 9696 }, { "epoch": 4.393747168101495, "grad_norm": 0.07654047445262088, "learning_rate": 3.529900831503672e-06, "loss": 0.9238, "step": 9697 }, { "epoch": 4.394200271862257, "grad_norm": 0.08016188230867623, "learning_rate": 3.524705282410241e-06, "loss": 0.9206, "step": 9698 }, { "epoch": 4.394653375623018, "grad_norm": 0.0801998686421263, "learning_rate": 3.519513383482567e-06, "loss": 0.9136, "step": 9699 }, { "epoch": 4.395106479383779, "grad_norm": 0.07779853388321006, "learning_rate": 3.514325135240233e-06, "loss": 0.8979, "step": 9700 }, { "epoch": 4.39555958314454, "grad_norm": 0.07973430227905938, "learning_rate": 3.5091405382024335e-06, "loss": 0.941, "step": 9701 }, { "epoch": 4.396012686905301, "grad_norm": 0.07580741452938353, "learning_rate": 3.503959592887989e-06, "loss": 0.9115, "step": 9702 }, { "epoch": 4.3964657906660625, "grad_norm": 0.07572768371169208, "learning_rate": 3.498782299815382e-06, "loss": 0.8976, "step": 9703 }, { "epoch": 4.396918894426824, "grad_norm": 0.08053072643398745, "learning_rate": 3.4936086595027095e-06, "loss": 0.9365, "step": 9704 }, { "epoch": 4.397371998187585, "grad_norm": 0.07260414227733969, "learning_rate": 3.4884386724677223e-06, "loss": 0.923, "step": 9705 }, { "epoch": 4.397825101948346, "grad_norm": 0.08163671111844137, "learning_rate": 3.48327233922777e-06, "loss": 0.9269, "step": 9706 }, { "epoch": 4.398278205709107, "grad_norm": 0.07560689480861167, "learning_rate": 3.4781096602998845e-06, "loss": 0.9452, "step": 9707 }, { "epoch": 4.398731309469869, "grad_norm": 0.0712896122310142, "learning_rate": 3.472950636200687e-06, "loss": 0.9231, "step": 9708 }, { "epoch": 4.39918441323063, "grad_norm": 0.08280376735799452, "learning_rate": 3.4677952674464676e-06, "loss": 0.9265, "step": 9709 }, { "epoch": 4.399637516991391, "grad_norm": 0.07521310875870038, "learning_rate": 3.462643554553129e-06, "loss": 0.9359, "step": 9710 }, { "epoch": 4.400090620752152, "grad_norm": 0.07542895356658226, "learning_rate": 3.457495498036205e-06, "loss": 0.9081, "step": 9711 }, { "epoch": 4.4005437245129135, "grad_norm": 0.07722179597634843, "learning_rate": 3.452351098410893e-06, "loss": 0.9297, "step": 9712 }, { "epoch": 4.400996828273675, "grad_norm": 0.07621029219746983, "learning_rate": 3.4472103561919856e-06, "loss": 0.9336, "step": 9713 }, { "epoch": 4.401449932034436, "grad_norm": 0.07532900773387619, "learning_rate": 3.4420732718939465e-06, "loss": 0.8845, "step": 9714 }, { "epoch": 4.401903035795197, "grad_norm": 0.08735779378461211, "learning_rate": 3.43693984603084e-06, "loss": 0.9118, "step": 9715 }, { "epoch": 4.402356139555958, "grad_norm": 0.07664073091983989, "learning_rate": 3.4318100791163935e-06, "loss": 0.9253, "step": 9716 }, { "epoch": 4.402809243316719, "grad_norm": 0.07335908413333701, "learning_rate": 3.4266839716639465e-06, "loss": 0.9124, "step": 9717 }, { "epoch": 4.403262347077481, "grad_norm": 0.08584735542521986, "learning_rate": 3.421561524186472e-06, "loss": 0.9174, "step": 9718 }, { "epoch": 4.403715450838242, "grad_norm": 0.07720533921088832, "learning_rate": 3.4164427371966035e-06, "loss": 0.932, "step": 9719 }, { "epoch": 4.4041685545990035, "grad_norm": 0.07574015897327455, "learning_rate": 3.411327611206572e-06, "loss": 0.9196, "step": 9720 }, { "epoch": 4.404621658359765, "grad_norm": 0.07758892706940082, "learning_rate": 3.4062161467282785e-06, "loss": 0.9234, "step": 9721 }, { "epoch": 4.405074762120526, "grad_norm": 0.07306294829554114, "learning_rate": 3.401108344273216e-06, "loss": 0.9187, "step": 9722 }, { "epoch": 4.405527865881287, "grad_norm": 0.0740131802993205, "learning_rate": 3.3960042043525586e-06, "loss": 0.9326, "step": 9723 }, { "epoch": 4.405980969642048, "grad_norm": 0.07689039110532443, "learning_rate": 3.3909037274770664e-06, "loss": 0.9394, "step": 9724 }, { "epoch": 4.406434073402809, "grad_norm": 0.07367020435604431, "learning_rate": 3.3858069141571747e-06, "loss": 0.9038, "step": 9725 }, { "epoch": 4.40688717716357, "grad_norm": 0.0775979587797961, "learning_rate": 3.3807137649029253e-06, "loss": 0.9263, "step": 9726 }, { "epoch": 4.407340280924331, "grad_norm": 0.07463723310251612, "learning_rate": 3.375624280223995e-06, "loss": 0.9059, "step": 9727 }, { "epoch": 4.4077933846850925, "grad_norm": 0.07468360236258147, "learning_rate": 3.3705384606297088e-06, "loss": 0.9407, "step": 9728 }, { "epoch": 4.4082464884458545, "grad_norm": 0.07424499808719887, "learning_rate": 3.365456306629007e-06, "loss": 0.9332, "step": 9729 }, { "epoch": 4.408699592206616, "grad_norm": 0.08047758559719084, "learning_rate": 3.360377818730487e-06, "loss": 0.9274, "step": 9730 }, { "epoch": 4.409152695967377, "grad_norm": 0.07399061146376314, "learning_rate": 3.355302997442351e-06, "loss": 0.9093, "step": 9731 }, { "epoch": 4.409605799728138, "grad_norm": 0.07201714618547422, "learning_rate": 3.3502318432724555e-06, "loss": 0.9269, "step": 9732 }, { "epoch": 4.410058903488899, "grad_norm": 0.08244260704022104, "learning_rate": 3.3451643567282832e-06, "loss": 0.9144, "step": 9733 }, { "epoch": 4.41051200724966, "grad_norm": 0.07597905008458045, "learning_rate": 3.3401005383169395e-06, "loss": 0.9059, "step": 9734 }, { "epoch": 4.410965111010421, "grad_norm": 0.08487091508839979, "learning_rate": 3.3350403885451832e-06, "loss": 0.9478, "step": 9735 }, { "epoch": 4.4114182147711825, "grad_norm": 0.0801689169631301, "learning_rate": 3.3299839079193874e-06, "loss": 0.9271, "step": 9736 }, { "epoch": 4.411871318531944, "grad_norm": 0.07531874900176098, "learning_rate": 3.324931096945574e-06, "loss": 0.9159, "step": 9737 }, { "epoch": 4.412324422292705, "grad_norm": 0.08045061890947909, "learning_rate": 3.3198819561293736e-06, "loss": 0.9182, "step": 9738 }, { "epoch": 4.412777526053466, "grad_norm": 0.08012038972289862, "learning_rate": 3.314836485976085e-06, "loss": 0.9088, "step": 9739 }, { "epoch": 4.413230629814228, "grad_norm": 0.07721717544764105, "learning_rate": 3.309794686990606e-06, "loss": 0.9229, "step": 9740 }, { "epoch": 4.413683733574989, "grad_norm": 0.08069920827602953, "learning_rate": 3.304756559677489e-06, "loss": 0.9354, "step": 9741 }, { "epoch": 4.41413683733575, "grad_norm": 0.08771613894877159, "learning_rate": 3.2997221045409077e-06, "loss": 0.923, "step": 9742 }, { "epoch": 4.414589941096511, "grad_norm": 0.07623404913745772, "learning_rate": 3.2946913220846644e-06, "loss": 0.9448, "step": 9743 }, { "epoch": 4.415043044857272, "grad_norm": 0.08673505831141758, "learning_rate": 3.2896642128122138e-06, "loss": 0.9301, "step": 9744 }, { "epoch": 4.4154961486180335, "grad_norm": 0.08321773037152874, "learning_rate": 3.2846407772266197e-06, "loss": 0.9386, "step": 9745 }, { "epoch": 4.415949252378795, "grad_norm": 0.08021932518239204, "learning_rate": 3.2796210158305963e-06, "loss": 0.9409, "step": 9746 }, { "epoch": 4.416402356139556, "grad_norm": 0.07544733029989338, "learning_rate": 3.2746049291264747e-06, "loss": 0.9252, "step": 9747 }, { "epoch": 4.416855459900317, "grad_norm": 0.08612691659063508, "learning_rate": 3.269592517616236e-06, "loss": 0.9321, "step": 9748 }, { "epoch": 4.417308563661078, "grad_norm": 0.07950962675232313, "learning_rate": 3.2645837818014736e-06, "loss": 0.9185, "step": 9749 }, { "epoch": 4.417761667421839, "grad_norm": 0.08088279918487226, "learning_rate": 3.259578722183432e-06, "loss": 0.9455, "step": 9750 }, { "epoch": 4.418214771182601, "grad_norm": 0.08078245668151532, "learning_rate": 3.2545773392629764e-06, "loss": 0.9228, "step": 9751 }, { "epoch": 4.418667874943362, "grad_norm": 0.08905500456891799, "learning_rate": 3.2495796335406047e-06, "loss": 0.922, "step": 9752 }, { "epoch": 4.4191209787041235, "grad_norm": 0.07943645782475527, "learning_rate": 3.2445856055164504e-06, "loss": 0.9394, "step": 9753 }, { "epoch": 4.419574082464885, "grad_norm": 0.07930622990129235, "learning_rate": 3.2395952556902642e-06, "loss": 0.9463, "step": 9754 }, { "epoch": 4.420027186225646, "grad_norm": 0.0850639169105489, "learning_rate": 3.234608584561465e-06, "loss": 0.9163, "step": 9755 }, { "epoch": 4.420480289986407, "grad_norm": 0.07601008521574587, "learning_rate": 3.2296255926290575e-06, "loss": 0.9509, "step": 9756 }, { "epoch": 4.420933393747168, "grad_norm": 0.07545870213268335, "learning_rate": 3.2246462803917233e-06, "loss": 0.9305, "step": 9757 }, { "epoch": 4.421386497507929, "grad_norm": 0.07101621121847593, "learning_rate": 3.2196706483477304e-06, "loss": 0.9123, "step": 9758 }, { "epoch": 4.42183960126869, "grad_norm": 0.07906940710107119, "learning_rate": 3.214698696995027e-06, "loss": 0.946, "step": 9759 }, { "epoch": 4.422292705029451, "grad_norm": 0.0727866987765445, "learning_rate": 3.2097304268311392e-06, "loss": 0.915, "step": 9760 }, { "epoch": 4.4227458087902125, "grad_norm": 0.07371240753629404, "learning_rate": 3.2047658383532832e-06, "loss": 0.9445, "step": 9761 }, { "epoch": 4.4231989125509745, "grad_norm": 0.07164098926458712, "learning_rate": 3.1998049320582568e-06, "loss": 0.9328, "step": 9762 }, { "epoch": 4.423652016311736, "grad_norm": 0.07497042174586929, "learning_rate": 3.194847708442508e-06, "loss": 0.8954, "step": 9763 }, { "epoch": 4.424105120072497, "grad_norm": 0.07060333075082591, "learning_rate": 3.189894168002128e-06, "loss": 0.928, "step": 9764 }, { "epoch": 4.424558223833258, "grad_norm": 0.08067596827851467, "learning_rate": 3.184944311232818e-06, "loss": 0.9223, "step": 9765 }, { "epoch": 4.425011327594019, "grad_norm": 0.07152171349514805, "learning_rate": 3.1799981386299383e-06, "loss": 0.9189, "step": 9766 }, { "epoch": 4.42546443135478, "grad_norm": 0.07148915379622159, "learning_rate": 3.1750556506884436e-06, "loss": 0.9089, "step": 9767 }, { "epoch": 4.425917535115541, "grad_norm": 0.08095536100934314, "learning_rate": 3.1701168479029555e-06, "loss": 0.8999, "step": 9768 }, { "epoch": 4.426370638876302, "grad_norm": 0.07284997959943514, "learning_rate": 3.1651817307677013e-06, "loss": 0.9262, "step": 9769 }, { "epoch": 4.426823742637064, "grad_norm": 0.07483077260578971, "learning_rate": 3.160250299776557e-06, "loss": 0.9102, "step": 9770 }, { "epoch": 4.427276846397825, "grad_norm": 0.07702442881999215, "learning_rate": 3.155322555423017e-06, "loss": 0.9205, "step": 9771 }, { "epoch": 4.427729950158587, "grad_norm": 0.07510670010425866, "learning_rate": 3.1503984982002066e-06, "loss": 0.9414, "step": 9772 }, { "epoch": 4.428183053919348, "grad_norm": 0.07199046666180034, "learning_rate": 3.145478128600905e-06, "loss": 0.9516, "step": 9773 }, { "epoch": 4.428636157680109, "grad_norm": 0.07143941615446081, "learning_rate": 3.1405614471174827e-06, "loss": 0.9128, "step": 9774 }, { "epoch": 4.42908926144087, "grad_norm": 0.07642793932733706, "learning_rate": 3.135648454241986e-06, "loss": 0.9341, "step": 9775 }, { "epoch": 4.429542365201631, "grad_norm": 0.07202027978765549, "learning_rate": 3.1307391504660446e-06, "loss": 0.9235, "step": 9776 }, { "epoch": 4.429995468962392, "grad_norm": 0.07798855418078918, "learning_rate": 3.1258335362809666e-06, "loss": 0.9063, "step": 9777 }, { "epoch": 4.4304485727231535, "grad_norm": 0.07528145853377852, "learning_rate": 3.120931612177658e-06, "loss": 0.9262, "step": 9778 }, { "epoch": 4.430901676483915, "grad_norm": 0.07428458319460896, "learning_rate": 3.116033378646659e-06, "loss": 0.9394, "step": 9779 }, { "epoch": 4.431354780244676, "grad_norm": 0.07111286124819977, "learning_rate": 3.1111388361781603e-06, "loss": 0.9173, "step": 9780 }, { "epoch": 4.431807884005437, "grad_norm": 0.07401509589240458, "learning_rate": 3.1062479852619563e-06, "loss": 0.926, "step": 9781 }, { "epoch": 4.432260987766199, "grad_norm": 0.07528998920514932, "learning_rate": 3.1013608263875007e-06, "loss": 0.939, "step": 9782 }, { "epoch": 4.43271409152696, "grad_norm": 0.07322606314089965, "learning_rate": 3.0964773600438458e-06, "loss": 0.9472, "step": 9783 }, { "epoch": 4.433167195287721, "grad_norm": 0.07203168279548106, "learning_rate": 3.0915975867197032e-06, "loss": 0.9351, "step": 9784 }, { "epoch": 4.433620299048482, "grad_norm": 0.0764777377954361, "learning_rate": 3.0867215069033987e-06, "loss": 0.9237, "step": 9785 }, { "epoch": 4.4340734028092434, "grad_norm": 0.07975844962957626, "learning_rate": 3.0818491210828962e-06, "loss": 0.9378, "step": 9786 }, { "epoch": 4.434526506570005, "grad_norm": 0.07625756484854451, "learning_rate": 3.076980429745784e-06, "loss": 0.9313, "step": 9787 }, { "epoch": 4.434979610330766, "grad_norm": 0.07789263966330122, "learning_rate": 3.072115433379277e-06, "loss": 0.9317, "step": 9788 }, { "epoch": 4.435432714091527, "grad_norm": 0.07776587805515406, "learning_rate": 3.0672541324702387e-06, "loss": 0.9395, "step": 9789 }, { "epoch": 4.435885817852288, "grad_norm": 0.07621293591547662, "learning_rate": 3.062396527505138e-06, "loss": 0.9291, "step": 9790 }, { "epoch": 4.436338921613049, "grad_norm": 0.07626530406044463, "learning_rate": 3.057542618970102e-06, "loss": 0.9514, "step": 9791 }, { "epoch": 4.43679202537381, "grad_norm": 0.07243403912324557, "learning_rate": 3.052692407350852e-06, "loss": 0.9149, "step": 9792 }, { "epoch": 4.437245129134572, "grad_norm": 0.07118997287894263, "learning_rate": 3.0478458931327794e-06, "loss": 0.9211, "step": 9793 }, { "epoch": 4.437698232895333, "grad_norm": 0.0761272158684118, "learning_rate": 3.0430030768008765e-06, "loss": 0.9372, "step": 9794 }, { "epoch": 4.4381513366560945, "grad_norm": 0.07106940564565255, "learning_rate": 3.038163958839766e-06, "loss": 0.9161, "step": 9795 }, { "epoch": 4.438604440416856, "grad_norm": 0.0788294312980866, "learning_rate": 3.033328539733731e-06, "loss": 0.9201, "step": 9796 }, { "epoch": 4.439057544177617, "grad_norm": 0.07271941111160249, "learning_rate": 3.028496819966642e-06, "loss": 0.9025, "step": 9797 }, { "epoch": 4.439510647938378, "grad_norm": 0.07825338920696566, "learning_rate": 3.023668800022037e-06, "loss": 0.9134, "step": 9798 }, { "epoch": 4.439963751699139, "grad_norm": 0.08058191086431435, "learning_rate": 3.0188444803830497e-06, "loss": 0.9036, "step": 9799 }, { "epoch": 4.4404168554599, "grad_norm": 0.06905782842141865, "learning_rate": 3.01402386153248e-06, "loss": 0.9067, "step": 9800 }, { "epoch": 4.440869959220661, "grad_norm": 0.07833859779155067, "learning_rate": 3.0092069439527204e-06, "loss": 0.935, "step": 9801 }, { "epoch": 4.441323062981422, "grad_norm": 0.07528227493404424, "learning_rate": 3.004393728125825e-06, "loss": 0.9292, "step": 9802 }, { "epoch": 4.441776166742184, "grad_norm": 0.07558848889857407, "learning_rate": 2.9995842145334573e-06, "loss": 0.9336, "step": 9803 }, { "epoch": 4.442229270502946, "grad_norm": 0.07115091997055849, "learning_rate": 2.994778403656908e-06, "loss": 0.9235, "step": 9804 }, { "epoch": 4.442682374263707, "grad_norm": 0.07738570544955475, "learning_rate": 2.989976295977122e-06, "loss": 0.922, "step": 9805 }, { "epoch": 4.443135478024468, "grad_norm": 0.07435078722346887, "learning_rate": 2.985177891974642e-06, "loss": 0.9328, "step": 9806 }, { "epoch": 4.443588581785229, "grad_norm": 0.07417295898210205, "learning_rate": 2.980383192129668e-06, "loss": 0.9139, "step": 9807 }, { "epoch": 4.44404168554599, "grad_norm": 0.07483774019990294, "learning_rate": 2.97559219692201e-06, "loss": 0.9265, "step": 9808 }, { "epoch": 4.444494789306751, "grad_norm": 0.07602002612533003, "learning_rate": 2.9708049068311174e-06, "loss": 0.9307, "step": 9809 }, { "epoch": 4.444947893067512, "grad_norm": 0.07394382806549309, "learning_rate": 2.9660213223360545e-06, "loss": 0.9295, "step": 9810 }, { "epoch": 4.4454009968282735, "grad_norm": 0.08412175726247435, "learning_rate": 2.961241443915537e-06, "loss": 0.9164, "step": 9811 }, { "epoch": 4.445854100589035, "grad_norm": 0.07738360901237729, "learning_rate": 2.956465272047888e-06, "loss": 0.9204, "step": 9812 }, { "epoch": 4.446307204349796, "grad_norm": 0.0728832606939303, "learning_rate": 2.951692807211086e-06, "loss": 0.9165, "step": 9813 }, { "epoch": 4.446760308110557, "grad_norm": 0.07877413688725782, "learning_rate": 2.946924049882709e-06, "loss": 0.9132, "step": 9814 }, { "epoch": 4.447213411871319, "grad_norm": 0.07918222722698427, "learning_rate": 2.942159000539975e-06, "loss": 0.9326, "step": 9815 }, { "epoch": 4.44766651563208, "grad_norm": 0.07269938779678212, "learning_rate": 2.9373976596597465e-06, "loss": 0.9257, "step": 9816 }, { "epoch": 4.448119619392841, "grad_norm": 0.07414857427313015, "learning_rate": 2.9326400277184874e-06, "loss": 0.9406, "step": 9817 }, { "epoch": 4.448572723153602, "grad_norm": 0.07249581180002783, "learning_rate": 2.927886105192319e-06, "loss": 0.9035, "step": 9818 }, { "epoch": 4.449025826914363, "grad_norm": 0.0727733733750445, "learning_rate": 2.923135892556963e-06, "loss": 0.9319, "step": 9819 }, { "epoch": 4.449478930675125, "grad_norm": 0.07720486481919855, "learning_rate": 2.918389390287799e-06, "loss": 0.9211, "step": 9820 }, { "epoch": 4.449932034435886, "grad_norm": 0.0687106456408616, "learning_rate": 2.913646598859803e-06, "loss": 0.921, "step": 9821 }, { "epoch": 4.450385138196647, "grad_norm": 0.07181884313949725, "learning_rate": 2.908907518747617e-06, "loss": 0.9345, "step": 9822 }, { "epoch": 4.450838241957408, "grad_norm": 0.07652390607816816, "learning_rate": 2.9041721504254795e-06, "loss": 0.9403, "step": 9823 }, { "epoch": 4.451291345718169, "grad_norm": 0.07299532189039802, "learning_rate": 2.8994404943672606e-06, "loss": 0.8906, "step": 9824 }, { "epoch": 4.45174444947893, "grad_norm": 0.07316859243327346, "learning_rate": 2.894712551046488e-06, "loss": 0.9402, "step": 9825 }, { "epoch": 4.452197553239692, "grad_norm": 0.07412875116936375, "learning_rate": 2.8899883209362813e-06, "loss": 0.9254, "step": 9826 }, { "epoch": 4.452650657000453, "grad_norm": 0.06982780312101029, "learning_rate": 2.8852678045094173e-06, "loss": 0.9379, "step": 9827 }, { "epoch": 4.4531037607612145, "grad_norm": 0.07534081832429357, "learning_rate": 2.8805510022382788e-06, "loss": 0.9227, "step": 9828 }, { "epoch": 4.453556864521976, "grad_norm": 0.07905315692541925, "learning_rate": 2.875837914594892e-06, "loss": 0.9207, "step": 9829 }, { "epoch": 4.454009968282737, "grad_norm": 0.07486138337133816, "learning_rate": 2.8711285420509028e-06, "loss": 0.9017, "step": 9830 }, { "epoch": 4.454463072043498, "grad_norm": 0.07366426110167777, "learning_rate": 2.8664228850775998e-06, "loss": 0.9124, "step": 9831 }, { "epoch": 4.454916175804259, "grad_norm": 0.07842324961103712, "learning_rate": 2.8617209441458784e-06, "loss": 0.9532, "step": 9832 }, { "epoch": 4.45536927956502, "grad_norm": 0.07373435467205476, "learning_rate": 2.857022719726268e-06, "loss": 0.9281, "step": 9833 }, { "epoch": 4.455822383325781, "grad_norm": 0.0749720173196952, "learning_rate": 2.852328212288944e-06, "loss": 0.9213, "step": 9834 }, { "epoch": 4.456275487086542, "grad_norm": 0.07433212672021772, "learning_rate": 2.8476374223036864e-06, "loss": 0.9292, "step": 9835 }, { "epoch": 4.456728590847304, "grad_norm": 0.08670412778810038, "learning_rate": 2.8429503502399194e-06, "loss": 0.9286, "step": 9836 }, { "epoch": 4.457181694608066, "grad_norm": 0.07218229596053886, "learning_rate": 2.838266996566681e-06, "loss": 0.9114, "step": 9837 }, { "epoch": 4.457634798368827, "grad_norm": 0.07282958533349417, "learning_rate": 2.8335873617526545e-06, "loss": 0.908, "step": 9838 }, { "epoch": 4.458087902129588, "grad_norm": 0.07539543291325852, "learning_rate": 2.8289114462661407e-06, "loss": 0.9135, "step": 9839 }, { "epoch": 4.458541005890349, "grad_norm": 0.0766773311700549, "learning_rate": 2.824239250575058e-06, "loss": 0.9319, "step": 9840 }, { "epoch": 4.45899410965111, "grad_norm": 0.07631198813546436, "learning_rate": 2.819570775146976e-06, "loss": 0.91, "step": 9841 }, { "epoch": 4.459447213411871, "grad_norm": 0.07240680573210138, "learning_rate": 2.8149060204490708e-06, "loss": 0.9369, "step": 9842 }, { "epoch": 4.459900317172632, "grad_norm": 0.0763479005161015, "learning_rate": 2.810244986948165e-06, "loss": 0.9231, "step": 9843 }, { "epoch": 4.4603534209333935, "grad_norm": 0.07508070049999542, "learning_rate": 2.8055876751106858e-06, "loss": 0.9212, "step": 9844 }, { "epoch": 4.460806524694155, "grad_norm": 0.07464884030539276, "learning_rate": 2.800934085402718e-06, "loss": 0.9306, "step": 9845 }, { "epoch": 4.461259628454917, "grad_norm": 0.0753599459779128, "learning_rate": 2.7962842182899376e-06, "loss": 0.9328, "step": 9846 }, { "epoch": 4.461712732215678, "grad_norm": 0.07894215552770868, "learning_rate": 2.7916380742376837e-06, "loss": 0.9468, "step": 9847 }, { "epoch": 4.462165835976439, "grad_norm": 0.07427708083714152, "learning_rate": 2.7869956537109e-06, "loss": 0.9067, "step": 9848 }, { "epoch": 4.4626189397372, "grad_norm": 0.07169753893562822, "learning_rate": 2.782356957174157e-06, "loss": 0.9125, "step": 9849 }, { "epoch": 4.463072043497961, "grad_norm": 0.07486384983878637, "learning_rate": 2.777721985091675e-06, "loss": 0.9443, "step": 9850 }, { "epoch": 4.463525147258722, "grad_norm": 0.07624835139679896, "learning_rate": 2.7730907379272685e-06, "loss": 0.9134, "step": 9851 }, { "epoch": 4.463978251019483, "grad_norm": 0.07493288238568198, "learning_rate": 2.768463216144417e-06, "loss": 0.9238, "step": 9852 }, { "epoch": 4.4644313547802446, "grad_norm": 0.07806775409374114, "learning_rate": 2.7638394202061847e-06, "loss": 0.9225, "step": 9853 }, { "epoch": 4.464884458541006, "grad_norm": 0.07690189016782575, "learning_rate": 2.7592193505753085e-06, "loss": 0.9209, "step": 9854 }, { "epoch": 4.465337562301767, "grad_norm": 0.07815893592885403, "learning_rate": 2.7546030077141117e-06, "loss": 0.9191, "step": 9855 }, { "epoch": 4.465790666062528, "grad_norm": 0.07999633354382822, "learning_rate": 2.749990392084567e-06, "loss": 0.9444, "step": 9856 }, { "epoch": 4.46624376982329, "grad_norm": 0.07727255546654559, "learning_rate": 2.7453815041482746e-06, "loss": 0.9298, "step": 9857 }, { "epoch": 4.466696873584051, "grad_norm": 0.07277033586451094, "learning_rate": 2.740776344366447e-06, "loss": 0.9056, "step": 9858 }, { "epoch": 4.467149977344812, "grad_norm": 0.07634390066777196, "learning_rate": 2.736174913199947e-06, "loss": 0.9143, "step": 9859 }, { "epoch": 4.467603081105573, "grad_norm": 0.08282324656318579, "learning_rate": 2.731577211109233e-06, "loss": 0.9175, "step": 9860 }, { "epoch": 4.4680561848663345, "grad_norm": 0.07494185285380667, "learning_rate": 2.7269832385544215e-06, "loss": 0.9239, "step": 9861 }, { "epoch": 4.468509288627096, "grad_norm": 0.0755776451505945, "learning_rate": 2.7223929959952333e-06, "loss": 0.9283, "step": 9862 }, { "epoch": 4.468962392387857, "grad_norm": 0.0727986854565848, "learning_rate": 2.71780648389103e-06, "loss": 0.9094, "step": 9863 }, { "epoch": 4.469415496148618, "grad_norm": 0.07877079504875283, "learning_rate": 2.7132237027007956e-06, "loss": 0.9367, "step": 9864 }, { "epoch": 4.469868599909379, "grad_norm": 0.07603922304726382, "learning_rate": 2.708644652883132e-06, "loss": 0.9134, "step": 9865 }, { "epoch": 4.47032170367014, "grad_norm": 0.07324784209927813, "learning_rate": 2.7040693348962778e-06, "loss": 0.9251, "step": 9866 }, { "epoch": 4.470774807430901, "grad_norm": 0.07527573728176702, "learning_rate": 2.699497749198101e-06, "loss": 0.9279, "step": 9867 }, { "epoch": 4.471227911191663, "grad_norm": 0.07405981667065248, "learning_rate": 2.6949298962460858e-06, "loss": 0.9208, "step": 9868 }, { "epoch": 4.471681014952424, "grad_norm": 0.07648439033822804, "learning_rate": 2.6903657764973414e-06, "loss": 0.9199, "step": 9869 }, { "epoch": 4.472134118713186, "grad_norm": 0.07216870708129564, "learning_rate": 2.6858053904086223e-06, "loss": 0.9249, "step": 9870 }, { "epoch": 4.472587222473947, "grad_norm": 0.07826103866595344, "learning_rate": 2.681248738436284e-06, "loss": 0.903, "step": 9871 }, { "epoch": 4.473040326234708, "grad_norm": 0.0757508234908076, "learning_rate": 2.676695821036335e-06, "loss": 0.9237, "step": 9872 }, { "epoch": 4.473493429995469, "grad_norm": 0.06636193454630317, "learning_rate": 2.6721466386643836e-06, "loss": 0.9013, "step": 9873 }, { "epoch": 4.47394653375623, "grad_norm": 0.07198902475256971, "learning_rate": 2.6676011917756885e-06, "loss": 0.9163, "step": 9874 }, { "epoch": 4.474399637516991, "grad_norm": 0.0764528287283689, "learning_rate": 2.6630594808251163e-06, "loss": 0.9145, "step": 9875 }, { "epoch": 4.474852741277752, "grad_norm": 0.07383645449521536, "learning_rate": 2.6585215062671575e-06, "loss": 0.9087, "step": 9876 }, { "epoch": 4.4753058450385135, "grad_norm": 0.07511318370237212, "learning_rate": 2.6539872685559555e-06, "loss": 0.9213, "step": 9877 }, { "epoch": 4.475758948799275, "grad_norm": 0.07762599653893215, "learning_rate": 2.649456768145249e-06, "loss": 0.9417, "step": 9878 }, { "epoch": 4.476212052560037, "grad_norm": 0.0749576716351577, "learning_rate": 2.644930005488422e-06, "loss": 0.9167, "step": 9879 }, { "epoch": 4.476665156320798, "grad_norm": 0.07519304278591246, "learning_rate": 2.640406981038468e-06, "loss": 0.9302, "step": 9880 }, { "epoch": 4.477118260081559, "grad_norm": 0.07184374366913023, "learning_rate": 2.635887695248034e-06, "loss": 0.9136, "step": 9881 }, { "epoch": 4.47757136384232, "grad_norm": 0.07213132840607515, "learning_rate": 2.631372148569358e-06, "loss": 0.919, "step": 9882 }, { "epoch": 4.478024467603081, "grad_norm": 0.07152689022969964, "learning_rate": 2.6268603414543313e-06, "loss": 0.9439, "step": 9883 }, { "epoch": 4.478477571363842, "grad_norm": 0.07480416241650753, "learning_rate": 2.6223522743544604e-06, "loss": 0.9325, "step": 9884 }, { "epoch": 4.478930675124603, "grad_norm": 0.07460766239219924, "learning_rate": 2.617847947720864e-06, "loss": 0.9273, "step": 9885 }, { "epoch": 4.4793837788853645, "grad_norm": 0.0728040996148146, "learning_rate": 2.6133473620043238e-06, "loss": 0.9112, "step": 9886 }, { "epoch": 4.479836882646126, "grad_norm": 0.07228642002891898, "learning_rate": 2.608850517655199e-06, "loss": 0.9279, "step": 9887 }, { "epoch": 4.480289986406887, "grad_norm": 0.07437810594527731, "learning_rate": 2.6043574151235218e-06, "loss": 0.9211, "step": 9888 }, { "epoch": 4.480743090167648, "grad_norm": 0.0696323429020626, "learning_rate": 2.5998680548589094e-06, "loss": 0.9348, "step": 9889 }, { "epoch": 4.48119619392841, "grad_norm": 0.07258441663634362, "learning_rate": 2.595382437310634e-06, "loss": 0.917, "step": 9890 }, { "epoch": 4.481649297689171, "grad_norm": 0.0721164779239824, "learning_rate": 2.5909005629275762e-06, "loss": 0.9307, "step": 9891 }, { "epoch": 4.482102401449932, "grad_norm": 0.07477212886153278, "learning_rate": 2.5864224321582444e-06, "loss": 0.9242, "step": 9892 }, { "epoch": 4.482555505210693, "grad_norm": 0.07394344705642245, "learning_rate": 2.581948045450786e-06, "loss": 0.9382, "step": 9893 }, { "epoch": 4.4830086089714545, "grad_norm": 0.07159324024267678, "learning_rate": 2.577477403252946e-06, "loss": 0.8972, "step": 9894 }, { "epoch": 4.483461712732216, "grad_norm": 0.0722238395724761, "learning_rate": 2.5730105060121303e-06, "loss": 0.9148, "step": 9895 }, { "epoch": 4.483914816492977, "grad_norm": 0.07629021029713269, "learning_rate": 2.5685473541753368e-06, "loss": 0.932, "step": 9896 }, { "epoch": 4.484367920253738, "grad_norm": 0.07280553878924619, "learning_rate": 2.5640879481892132e-06, "loss": 0.9261, "step": 9897 }, { "epoch": 4.484821024014499, "grad_norm": 0.07024104306206339, "learning_rate": 2.5596322885000157e-06, "loss": 0.9314, "step": 9898 }, { "epoch": 4.48527412777526, "grad_norm": 0.07260704230447049, "learning_rate": 2.555180375553641e-06, "loss": 0.9068, "step": 9899 }, { "epoch": 4.485727231536022, "grad_norm": 0.07141681007830891, "learning_rate": 2.5507322097955946e-06, "loss": 0.9304, "step": 9900 }, { "epoch": 4.486180335296783, "grad_norm": 0.0737749539296236, "learning_rate": 2.546287791671014e-06, "loss": 0.9406, "step": 9901 }, { "epoch": 4.486633439057544, "grad_norm": 0.06930094201230871, "learning_rate": 2.5418471216246674e-06, "loss": 0.9259, "step": 9902 }, { "epoch": 4.4870865428183055, "grad_norm": 0.07405669544570762, "learning_rate": 2.537410200100934e-06, "loss": 0.9051, "step": 9903 }, { "epoch": 4.487539646579067, "grad_norm": 0.07516259404542602, "learning_rate": 2.532977027543839e-06, "loss": 0.9413, "step": 9904 }, { "epoch": 4.487992750339828, "grad_norm": 0.06907998633674618, "learning_rate": 2.528547604397007e-06, "loss": 0.9203, "step": 9905 }, { "epoch": 4.488445854100589, "grad_norm": 0.07630709830443735, "learning_rate": 2.5241219311037134e-06, "loss": 0.9351, "step": 9906 }, { "epoch": 4.48889895786135, "grad_norm": 0.07649897107806083, "learning_rate": 2.5197000081068356e-06, "loss": 0.9154, "step": 9907 }, { "epoch": 4.489352061622111, "grad_norm": 0.07218734737513632, "learning_rate": 2.515281835848895e-06, "loss": 0.9293, "step": 9908 }, { "epoch": 4.489805165382872, "grad_norm": 0.07158246472529918, "learning_rate": 2.510867414772018e-06, "loss": 0.9163, "step": 9909 }, { "epoch": 4.490258269143634, "grad_norm": 0.07209447817857977, "learning_rate": 2.5064567453179666e-06, "loss": 0.9132, "step": 9910 }, { "epoch": 4.4907113729043955, "grad_norm": 0.07585651166690721, "learning_rate": 2.5020498279281393e-06, "loss": 0.9197, "step": 9911 }, { "epoch": 4.491164476665157, "grad_norm": 0.07068285107956189, "learning_rate": 2.4976466630435248e-06, "loss": 0.9153, "step": 9912 }, { "epoch": 4.491617580425918, "grad_norm": 0.06934639370713924, "learning_rate": 2.49324725110478e-06, "loss": 0.9017, "step": 9913 }, { "epoch": 4.492070684186679, "grad_norm": 0.07189184748818456, "learning_rate": 2.4888515925521483e-06, "loss": 0.9242, "step": 9914 }, { "epoch": 4.49252378794744, "grad_norm": 0.07042882266100853, "learning_rate": 2.484459687825522e-06, "loss": 0.9245, "step": 9915 }, { "epoch": 4.492976891708201, "grad_norm": 0.08395455664824032, "learning_rate": 2.480071537364408e-06, "loss": 0.9259, "step": 9916 }, { "epoch": 4.493429995468962, "grad_norm": 0.07177711508095902, "learning_rate": 2.4756871416079296e-06, "loss": 0.9116, "step": 9917 }, { "epoch": 4.493883099229723, "grad_norm": 0.07531937433434842, "learning_rate": 2.4713065009948565e-06, "loss": 0.9207, "step": 9918 }, { "epoch": 4.4943362029904845, "grad_norm": 0.06920824123732781, "learning_rate": 2.466929615963558e-06, "loss": 0.9106, "step": 9919 }, { "epoch": 4.494789306751246, "grad_norm": 0.06894692733182199, "learning_rate": 2.462556486952048e-06, "loss": 0.9451, "step": 9920 }, { "epoch": 4.495242410512008, "grad_norm": 0.07711578867155619, "learning_rate": 2.4581871143979497e-06, "loss": 0.9331, "step": 9921 }, { "epoch": 4.495695514272769, "grad_norm": 0.073485649374577, "learning_rate": 2.453821498738518e-06, "loss": 0.945, "step": 9922 }, { "epoch": 4.49614861803353, "grad_norm": 0.07310342204294142, "learning_rate": 2.449459640410634e-06, "loss": 0.9242, "step": 9923 }, { "epoch": 4.496601721794291, "grad_norm": 0.06935508190130978, "learning_rate": 2.4451015398507936e-06, "loss": 0.9137, "step": 9924 }, { "epoch": 4.497054825555052, "grad_norm": 0.07875190382029544, "learning_rate": 2.4407471974951193e-06, "loss": 0.9119, "step": 9925 }, { "epoch": 4.497507929315813, "grad_norm": 0.07122124345066415, "learning_rate": 2.436396613779368e-06, "loss": 0.908, "step": 9926 }, { "epoch": 4.4979610330765745, "grad_norm": 0.07508973008252062, "learning_rate": 2.4320497891388993e-06, "loss": 0.9208, "step": 9927 }, { "epoch": 4.498414136837336, "grad_norm": 0.07130698653417085, "learning_rate": 2.4277067240087294e-06, "loss": 0.9258, "step": 9928 }, { "epoch": 4.498867240598097, "grad_norm": 0.07656662459542284, "learning_rate": 2.4233674188234657e-06, "loss": 0.9243, "step": 9929 }, { "epoch": 4.499320344358858, "grad_norm": 0.07449003578554708, "learning_rate": 2.419031874017348e-06, "loss": 0.9168, "step": 9930 }, { "epoch": 4.499773448119619, "grad_norm": 0.07182235926460495, "learning_rate": 2.41470009002426e-06, "loss": 0.9173, "step": 9931 }, { "epoch": 4.500226551880381, "grad_norm": 0.06966807429275604, "learning_rate": 2.4103720672776733e-06, "loss": 0.9166, "step": 9932 }, { "epoch": 4.500679655641142, "grad_norm": 0.07645683727546461, "learning_rate": 2.406047806210725e-06, "loss": 0.9332, "step": 9933 }, { "epoch": 4.501132759401903, "grad_norm": 0.07536321604369674, "learning_rate": 2.401727307256132e-06, "loss": 0.9202, "step": 9934 }, { "epoch": 4.501585863162664, "grad_norm": 0.0719214533975997, "learning_rate": 2.397410570846277e-06, "loss": 0.9047, "step": 9935 }, { "epoch": 4.5020389669234255, "grad_norm": 0.0771183718354116, "learning_rate": 2.3930975974131345e-06, "loss": 0.9147, "step": 9936 }, { "epoch": 4.502492070684187, "grad_norm": 0.07289491445844655, "learning_rate": 2.3887883873883055e-06, "loss": 0.9265, "step": 9937 }, { "epoch": 4.502945174444948, "grad_norm": 0.07107870405870725, "learning_rate": 2.384482941203041e-06, "loss": 0.9114, "step": 9938 }, { "epoch": 4.503398278205709, "grad_norm": 0.07025872924851749, "learning_rate": 2.380181259288179e-06, "loss": 0.9028, "step": 9939 }, { "epoch": 4.50385138196647, "grad_norm": 0.07217229483972666, "learning_rate": 2.3758833420742144e-06, "loss": 0.9209, "step": 9940 }, { "epoch": 4.504304485727231, "grad_norm": 0.07038922978874586, "learning_rate": 2.3715891899912347e-06, "loss": 0.939, "step": 9941 }, { "epoch": 4.504757589487992, "grad_norm": 0.07698602998876548, "learning_rate": 2.3672988034689803e-06, "loss": 0.9425, "step": 9942 }, { "epoch": 4.505210693248754, "grad_norm": 0.07338120861296438, "learning_rate": 2.363012182936788e-06, "loss": 0.92, "step": 9943 }, { "epoch": 4.5056637970095155, "grad_norm": 0.0706361473655677, "learning_rate": 2.3587293288236392e-06, "loss": 0.9287, "step": 9944 }, { "epoch": 4.506116900770277, "grad_norm": 0.0681229443098156, "learning_rate": 2.354450241558124e-06, "loss": 0.9266, "step": 9945 }, { "epoch": 4.506570004531038, "grad_norm": 0.0740368543849176, "learning_rate": 2.3501749215684556e-06, "loss": 0.9333, "step": 9946 }, { "epoch": 4.507023108291799, "grad_norm": 0.07574329124962913, "learning_rate": 2.345903369282483e-06, "loss": 0.9398, "step": 9947 }, { "epoch": 4.50747621205256, "grad_norm": 0.0734722483601962, "learning_rate": 2.3416355851276597e-06, "loss": 0.9142, "step": 9948 }, { "epoch": 4.507929315813321, "grad_norm": 0.07310087037649698, "learning_rate": 2.3373715695310883e-06, "loss": 0.9073, "step": 9949 }, { "epoch": 4.508382419574082, "grad_norm": 0.0709339889733322, "learning_rate": 2.3331113229194635e-06, "loss": 0.9058, "step": 9950 }, { "epoch": 4.508835523334843, "grad_norm": 0.07049822795013455, "learning_rate": 2.328854845719133e-06, "loss": 0.9356, "step": 9951 }, { "epoch": 4.5092886270956045, "grad_norm": 0.07161743952456909, "learning_rate": 2.32460213835604e-06, "loss": 0.9115, "step": 9952 }, { "epoch": 4.509741730856366, "grad_norm": 0.07247175536071426, "learning_rate": 2.3203532012557606e-06, "loss": 0.9342, "step": 9953 }, { "epoch": 4.510194834617128, "grad_norm": 0.07226598285131576, "learning_rate": 2.3161080348435096e-06, "loss": 0.9106, "step": 9954 }, { "epoch": 4.510647938377889, "grad_norm": 0.07660941184450187, "learning_rate": 2.3118666395440938e-06, "loss": 0.9212, "step": 9955 }, { "epoch": 4.51110104213865, "grad_norm": 0.07190539523566497, "learning_rate": 2.3076290157819737e-06, "loss": 0.9347, "step": 9956 }, { "epoch": 4.511554145899411, "grad_norm": 0.07300097699511292, "learning_rate": 2.3033951639812057e-06, "loss": 0.9239, "step": 9957 }, { "epoch": 4.512007249660172, "grad_norm": 0.07265559294273356, "learning_rate": 2.2991650845654956e-06, "loss": 0.9008, "step": 9958 }, { "epoch": 4.512460353420933, "grad_norm": 0.07485581133354688, "learning_rate": 2.29493877795814e-06, "loss": 0.923, "step": 9959 }, { "epoch": 4.5129134571816945, "grad_norm": 0.06888498691404109, "learning_rate": 2.2907162445820895e-06, "loss": 0.9371, "step": 9960 }, { "epoch": 4.513366560942456, "grad_norm": 0.07216595196237124, "learning_rate": 2.2864974848598996e-06, "loss": 0.9351, "step": 9961 }, { "epoch": 4.513819664703217, "grad_norm": 0.07345391774383978, "learning_rate": 2.2822824992137393e-06, "loss": 0.9015, "step": 9962 }, { "epoch": 4.514272768463979, "grad_norm": 0.07704807318552657, "learning_rate": 2.2780712880654264e-06, "loss": 0.9294, "step": 9963 }, { "epoch": 4.514725872224739, "grad_norm": 0.07136886434535282, "learning_rate": 2.2738638518363754e-06, "loss": 0.9343, "step": 9964 }, { "epoch": 4.515178975985501, "grad_norm": 0.0756582778052642, "learning_rate": 2.2696601909476446e-06, "loss": 0.9048, "step": 9965 }, { "epoch": 4.515632079746262, "grad_norm": 0.07660663368361037, "learning_rate": 2.2654603058198932e-06, "loss": 0.8955, "step": 9966 }, { "epoch": 4.516085183507023, "grad_norm": 0.0762816135785089, "learning_rate": 2.2612641968734294e-06, "loss": 0.944, "step": 9967 }, { "epoch": 4.516538287267784, "grad_norm": 0.06940778836776351, "learning_rate": 2.257071864528144e-06, "loss": 0.9295, "step": 9968 }, { "epoch": 4.5169913910285455, "grad_norm": 0.0733566150243609, "learning_rate": 2.252883309203595e-06, "loss": 0.9342, "step": 9969 }, { "epoch": 4.517444494789307, "grad_norm": 0.0728196050988327, "learning_rate": 2.2486985313189312e-06, "loss": 0.9235, "step": 9970 }, { "epoch": 4.517897598550068, "grad_norm": 0.07022421836849452, "learning_rate": 2.244517531292929e-06, "loss": 0.9121, "step": 9971 }, { "epoch": 4.518350702310829, "grad_norm": 0.07071782406215157, "learning_rate": 2.240340309544e-06, "loss": 0.9206, "step": 9972 }, { "epoch": 4.51880380607159, "grad_norm": 0.07196212236925313, "learning_rate": 2.236166866490157e-06, "loss": 0.9429, "step": 9973 }, { "epoch": 4.519256909832352, "grad_norm": 0.0718123899965215, "learning_rate": 2.2319972025490567e-06, "loss": 0.9216, "step": 9974 }, { "epoch": 4.519710013593113, "grad_norm": 0.07559349478812996, "learning_rate": 2.2278313181379562e-06, "loss": 0.9268, "step": 9975 }, { "epoch": 4.520163117353874, "grad_norm": 0.06918707955960578, "learning_rate": 2.223669213673758e-06, "loss": 0.9411, "step": 9976 }, { "epoch": 4.5206162211146355, "grad_norm": 0.07345461335762857, "learning_rate": 2.2195108895729645e-06, "loss": 0.9383, "step": 9977 }, { "epoch": 4.521069324875397, "grad_norm": 0.07485807955032013, "learning_rate": 2.2153563462517092e-06, "loss": 0.9452, "step": 9978 }, { "epoch": 4.521522428636158, "grad_norm": 0.07008050773022118, "learning_rate": 2.211205584125753e-06, "loss": 0.9112, "step": 9979 }, { "epoch": 4.521975532396919, "grad_norm": 0.06999582554269831, "learning_rate": 2.207058603610466e-06, "loss": 0.9194, "step": 9980 }, { "epoch": 4.52242863615768, "grad_norm": 0.06736230362622862, "learning_rate": 2.2029154051208445e-06, "loss": 0.9081, "step": 9981 }, { "epoch": 4.522881739918441, "grad_norm": 0.07238958929281922, "learning_rate": 2.1987759890715087e-06, "loss": 0.9182, "step": 9982 }, { "epoch": 4.523334843679202, "grad_norm": 0.07454813501206056, "learning_rate": 2.194640355876705e-06, "loss": 0.9149, "step": 9983 }, { "epoch": 4.523787947439963, "grad_norm": 0.07224436244350112, "learning_rate": 2.1905085059502882e-06, "loss": 0.9081, "step": 9984 }, { "epoch": 4.524241051200725, "grad_norm": 0.07056458154880776, "learning_rate": 2.1863804397057464e-06, "loss": 0.9349, "step": 9985 }, { "epoch": 4.5246941549614865, "grad_norm": 0.07349850807844632, "learning_rate": 2.1822561575561842e-06, "loss": 0.9176, "step": 9986 }, { "epoch": 4.525147258722248, "grad_norm": 0.06911646661989694, "learning_rate": 2.1781356599143287e-06, "loss": 0.907, "step": 9987 }, { "epoch": 4.525600362483009, "grad_norm": 0.07404136879749676, "learning_rate": 2.174018947192522e-06, "loss": 0.9409, "step": 9988 }, { "epoch": 4.52605346624377, "grad_norm": 0.07303834551415125, "learning_rate": 2.1699060198027455e-06, "loss": 0.9322, "step": 9989 }, { "epoch": 4.526506570004531, "grad_norm": 0.07640380489409263, "learning_rate": 2.1657968781565766e-06, "loss": 0.9165, "step": 9990 }, { "epoch": 4.526959673765292, "grad_norm": 0.07421795192768439, "learning_rate": 2.161691522665228e-06, "loss": 0.9428, "step": 9991 }, { "epoch": 4.527412777526053, "grad_norm": 0.07457584180709281, "learning_rate": 2.157589953739536e-06, "loss": 0.9153, "step": 9992 }, { "epoch": 4.527865881286814, "grad_norm": 0.07067263309531165, "learning_rate": 2.1534921717899504e-06, "loss": 0.8947, "step": 9993 }, { "epoch": 4.528318985047576, "grad_norm": 0.06913870540566502, "learning_rate": 2.149398177226556e-06, "loss": 0.8994, "step": 9994 }, { "epoch": 4.528772088808337, "grad_norm": 0.07654689598913264, "learning_rate": 2.1453079704590295e-06, "loss": 0.9152, "step": 9995 }, { "epoch": 4.529225192569099, "grad_norm": 0.07006058281682402, "learning_rate": 2.141221551896706e-06, "loss": 0.9267, "step": 9996 }, { "epoch": 4.52967829632986, "grad_norm": 0.07402427271077207, "learning_rate": 2.1371389219485163e-06, "loss": 0.9382, "step": 9997 }, { "epoch": 4.530131400090621, "grad_norm": 0.07739039892970404, "learning_rate": 2.1330600810230083e-06, "loss": 0.9295, "step": 9998 }, { "epoch": 4.530584503851382, "grad_norm": 0.07368147000075231, "learning_rate": 2.128985029528372e-06, "loss": 0.9028, "step": 9999 }, { "epoch": 4.531037607612143, "grad_norm": 0.07039306281161443, "learning_rate": 2.1249137678724006e-06, "loss": 0.9298, "step": 10000 }, { "epoch": 4.531490711372904, "grad_norm": 0.06906321124671279, "learning_rate": 2.1208462964625244e-06, "loss": 0.9287, "step": 10001 }, { "epoch": 4.5319438151336655, "grad_norm": 0.07343842009713665, "learning_rate": 2.1167826157057724e-06, "loss": 0.934, "step": 10002 }, { "epoch": 4.532396918894427, "grad_norm": 0.07213971055041467, "learning_rate": 2.11272272600882e-06, "loss": 0.9259, "step": 10003 }, { "epoch": 4.532850022655188, "grad_norm": 0.07438695453071746, "learning_rate": 2.1086666277779333e-06, "loss": 0.9441, "step": 10004 }, { "epoch": 4.533303126415949, "grad_norm": 0.07081919393693907, "learning_rate": 2.1046143214190318e-06, "loss": 0.9309, "step": 10005 }, { "epoch": 4.53375623017671, "grad_norm": 0.07541274105752376, "learning_rate": 2.100565807337631e-06, "loss": 0.9165, "step": 10006 }, { "epoch": 4.534209333937472, "grad_norm": 0.07354345022782187, "learning_rate": 2.096521085938865e-06, "loss": 0.903, "step": 10007 }, { "epoch": 4.534662437698233, "grad_norm": 0.0735837795388908, "learning_rate": 2.0924801576275165e-06, "loss": 0.9336, "step": 10008 }, { "epoch": 4.535115541458994, "grad_norm": 0.07055431009041847, "learning_rate": 2.08844302280796e-06, "loss": 0.936, "step": 10009 }, { "epoch": 4.5355686452197554, "grad_norm": 0.0709338541882321, "learning_rate": 2.084409681884205e-06, "loss": 0.9149, "step": 10010 }, { "epoch": 4.536021748980517, "grad_norm": 0.07623760542780528, "learning_rate": 2.0803801352598672e-06, "loss": 0.9171, "step": 10011 }, { "epoch": 4.536474852741278, "grad_norm": 0.07548548333863339, "learning_rate": 2.0763543833382105e-06, "loss": 0.9147, "step": 10012 }, { "epoch": 4.536927956502039, "grad_norm": 0.0727291016532829, "learning_rate": 2.07233242652209e-06, "loss": 0.9385, "step": 10013 }, { "epoch": 4.5373810602628, "grad_norm": 0.07670364779122632, "learning_rate": 2.0683142652139844e-06, "loss": 0.929, "step": 10014 }, { "epoch": 4.537834164023561, "grad_norm": 0.07154468992810198, "learning_rate": 2.064299899816016e-06, "loss": 0.9294, "step": 10015 }, { "epoch": 4.538287267784322, "grad_norm": 0.07236591067673954, "learning_rate": 2.060289330729899e-06, "loss": 0.9196, "step": 10016 }, { "epoch": 4.538740371545083, "grad_norm": 0.0761693434813946, "learning_rate": 2.0562825583569924e-06, "loss": 0.9352, "step": 10017 }, { "epoch": 4.539193475305845, "grad_norm": 0.07237280781414285, "learning_rate": 2.052279583098247e-06, "loss": 0.9332, "step": 10018 }, { "epoch": 4.5396465790666065, "grad_norm": 0.07569770660552984, "learning_rate": 2.048280405354266e-06, "loss": 0.9476, "step": 10019 }, { "epoch": 4.540099682827368, "grad_norm": 0.07313943553410979, "learning_rate": 2.0442850255252454e-06, "loss": 0.9183, "step": 10020 }, { "epoch": 4.540552786588129, "grad_norm": 0.07329266298691024, "learning_rate": 2.0402934440110166e-06, "loss": 0.9217, "step": 10021 }, { "epoch": 4.54100589034889, "grad_norm": 0.06924178264767854, "learning_rate": 2.036305661211029e-06, "loss": 0.9109, "step": 10022 }, { "epoch": 4.541458994109651, "grad_norm": 0.07253578227314555, "learning_rate": 2.0323216775243403e-06, "loss": 0.9286, "step": 10023 }, { "epoch": 4.541912097870412, "grad_norm": 0.07267676835114711, "learning_rate": 2.0283414933496414e-06, "loss": 0.9388, "step": 10024 }, { "epoch": 4.542365201631173, "grad_norm": 0.07304142355262076, "learning_rate": 2.02436510908524e-06, "loss": 0.9389, "step": 10025 }, { "epoch": 4.542818305391934, "grad_norm": 0.07309477856471844, "learning_rate": 2.0203925251290625e-06, "loss": 0.921, "step": 10026 }, { "epoch": 4.5432714091526964, "grad_norm": 0.07043509825317304, "learning_rate": 2.016423741878648e-06, "loss": 0.9208, "step": 10027 }, { "epoch": 4.543724512913457, "grad_norm": 0.07797063786796436, "learning_rate": 2.012458759731173e-06, "loss": 0.9376, "step": 10028 }, { "epoch": 4.544177616674219, "grad_norm": 0.07252530683526565, "learning_rate": 2.0084975790834083e-06, "loss": 0.9237, "step": 10029 }, { "epoch": 4.54463072043498, "grad_norm": 0.07986029128529795, "learning_rate": 2.0045402003317703e-06, "loss": 0.9054, "step": 10030 }, { "epoch": 4.545083824195741, "grad_norm": 0.0707629322615962, "learning_rate": 2.0005866238722803e-06, "loss": 0.9144, "step": 10031 }, { "epoch": 4.545536927956502, "grad_norm": 0.07663358791486923, "learning_rate": 1.996636850100573e-06, "loss": 0.9207, "step": 10032 }, { "epoch": 4.545990031717263, "grad_norm": 0.07921431857824196, "learning_rate": 1.9926908794119228e-06, "loss": 0.9324, "step": 10033 }, { "epoch": 4.546443135478024, "grad_norm": 0.07382210348581561, "learning_rate": 1.988748712201205e-06, "loss": 0.9205, "step": 10034 }, { "epoch": 4.5468962392387855, "grad_norm": 0.07087025329437159, "learning_rate": 1.9848103488629223e-06, "loss": 0.9215, "step": 10035 }, { "epoch": 4.547349342999547, "grad_norm": 0.07906760471371467, "learning_rate": 1.9808757897911945e-06, "loss": 0.9213, "step": 10036 }, { "epoch": 4.547802446760308, "grad_norm": 0.07589114134647879, "learning_rate": 1.9769450353797783e-06, "loss": 0.933, "step": 10037 }, { "epoch": 4.54825555052107, "grad_norm": 0.07202175601220598, "learning_rate": 1.9730180860220027e-06, "loss": 0.9281, "step": 10038 }, { "epoch": 4.548708654281831, "grad_norm": 0.07169835223866133, "learning_rate": 1.9690949421108694e-06, "loss": 0.9386, "step": 10039 }, { "epoch": 4.549161758042592, "grad_norm": 0.0747199625662352, "learning_rate": 1.9651756040389625e-06, "loss": 0.9117, "step": 10040 }, { "epoch": 4.549614861803353, "grad_norm": 0.07463782161955809, "learning_rate": 1.9612600721985142e-06, "loss": 0.925, "step": 10041 }, { "epoch": 4.550067965564114, "grad_norm": 0.07366871217912535, "learning_rate": 1.957348346981354e-06, "loss": 0.9442, "step": 10042 }, { "epoch": 4.550521069324875, "grad_norm": 0.06920658430290035, "learning_rate": 1.9534404287789276e-06, "loss": 0.92, "step": 10043 }, { "epoch": 4.550974173085637, "grad_norm": 0.0797585880774633, "learning_rate": 1.9495363179823236e-06, "loss": 0.9266, "step": 10044 }, { "epoch": 4.551427276846398, "grad_norm": 0.0742781734867392, "learning_rate": 1.945636014982224e-06, "loss": 0.9386, "step": 10045 }, { "epoch": 4.551880380607159, "grad_norm": 0.07306737191083672, "learning_rate": 1.9417395201689528e-06, "loss": 0.9107, "step": 10046 }, { "epoch": 4.55233348436792, "grad_norm": 0.06891489867340449, "learning_rate": 1.9378468339324285e-06, "loss": 0.9159, "step": 10047 }, { "epoch": 4.552786588128681, "grad_norm": 0.06857610936471155, "learning_rate": 1.933957956662211e-06, "loss": 0.9211, "step": 10048 }, { "epoch": 4.553239691889443, "grad_norm": 0.0771372733068935, "learning_rate": 1.9300728887474607e-06, "loss": 0.939, "step": 10049 }, { "epoch": 4.553692795650204, "grad_norm": 0.07480643984325266, "learning_rate": 1.9261916305769724e-06, "loss": 0.9363, "step": 10050 }, { "epoch": 4.554145899410965, "grad_norm": 0.06842100910659334, "learning_rate": 1.922314182539151e-06, "loss": 0.9142, "step": 10051 }, { "epoch": 4.5545990031717265, "grad_norm": 0.07030894852645446, "learning_rate": 1.918440545022016e-06, "loss": 0.9349, "step": 10052 }, { "epoch": 4.555052106932488, "grad_norm": 0.07097567527257764, "learning_rate": 1.91457071841322e-06, "loss": 0.9154, "step": 10053 }, { "epoch": 4.555505210693249, "grad_norm": 0.07030066672070871, "learning_rate": 1.9107047031000147e-06, "loss": 0.9309, "step": 10054 }, { "epoch": 4.55595831445401, "grad_norm": 0.07437145825234705, "learning_rate": 1.9068424994692946e-06, "loss": 0.9224, "step": 10055 }, { "epoch": 4.556411418214771, "grad_norm": 0.07462680716491442, "learning_rate": 1.9029841079075461e-06, "loss": 0.914, "step": 10056 }, { "epoch": 4.556864521975532, "grad_norm": 0.07353998259066255, "learning_rate": 1.899129528800896e-06, "loss": 0.8994, "step": 10057 }, { "epoch": 4.557317625736293, "grad_norm": 0.07291263399665768, "learning_rate": 1.89527876253508e-06, "loss": 0.9115, "step": 10058 }, { "epoch": 4.557770729497054, "grad_norm": 0.06786524750256469, "learning_rate": 1.8914318094954432e-06, "loss": 0.9245, "step": 10059 }, { "epoch": 4.558223833257816, "grad_norm": 0.0757081199290287, "learning_rate": 1.887588670066971e-06, "loss": 0.9227, "step": 10060 }, { "epoch": 4.558676937018578, "grad_norm": 0.07252239385681883, "learning_rate": 1.8837493446342448e-06, "loss": 0.9077, "step": 10061 }, { "epoch": 4.559130040779339, "grad_norm": 0.06915113248474367, "learning_rate": 1.8799138335814859e-06, "loss": 0.9398, "step": 10062 }, { "epoch": 4.5595831445401, "grad_norm": 0.0720454336418754, "learning_rate": 1.8760821372925165e-06, "loss": 0.9285, "step": 10063 }, { "epoch": 4.560036248300861, "grad_norm": 0.07153527552356215, "learning_rate": 1.8722542561507851e-06, "loss": 0.9411, "step": 10064 }, { "epoch": 4.560489352061622, "grad_norm": 0.07112440958316678, "learning_rate": 1.8684301905393497e-06, "loss": 0.9112, "step": 10065 }, { "epoch": 4.560942455822383, "grad_norm": 0.07486595893394028, "learning_rate": 1.8646099408409046e-06, "loss": 0.9217, "step": 10066 }, { "epoch": 4.561395559583144, "grad_norm": 0.0706653906610295, "learning_rate": 1.8607935074377438e-06, "loss": 0.9214, "step": 10067 }, { "epoch": 4.5618486633439055, "grad_norm": 0.07083593711222314, "learning_rate": 1.8569808907117793e-06, "loss": 0.9222, "step": 10068 }, { "epoch": 4.562301767104667, "grad_norm": 0.07034762836884874, "learning_rate": 1.8531720910445683e-06, "loss": 0.9001, "step": 10069 }, { "epoch": 4.562754870865428, "grad_norm": 0.07335307055580903, "learning_rate": 1.8493671088172461e-06, "loss": 0.9304, "step": 10070 }, { "epoch": 4.56320797462619, "grad_norm": 0.0727930635793334, "learning_rate": 1.8455659444105968e-06, "loss": 0.9267, "step": 10071 }, { "epoch": 4.563661078386951, "grad_norm": 0.07249959910860188, "learning_rate": 1.8417685982050094e-06, "loss": 0.9289, "step": 10072 }, { "epoch": 4.564114182147712, "grad_norm": 0.07047406881923246, "learning_rate": 1.8379750705804956e-06, "loss": 0.8997, "step": 10073 }, { "epoch": 4.564567285908473, "grad_norm": 0.06992859239937463, "learning_rate": 1.834185361916676e-06, "loss": 0.9037, "step": 10074 }, { "epoch": 4.565020389669234, "grad_norm": 0.06912554794538286, "learning_rate": 1.8303994725927942e-06, "loss": 0.9091, "step": 10075 }, { "epoch": 4.565473493429995, "grad_norm": 0.06872343918002924, "learning_rate": 1.8266174029877247e-06, "loss": 0.9084, "step": 10076 }, { "epoch": 4.5659265971907566, "grad_norm": 0.07073364526274027, "learning_rate": 1.8228391534799384e-06, "loss": 0.9136, "step": 10077 }, { "epoch": 4.566379700951518, "grad_norm": 0.07427303345319612, "learning_rate": 1.8190647244475324e-06, "loss": 0.9177, "step": 10078 }, { "epoch": 4.566832804712279, "grad_norm": 0.0666131603284144, "learning_rate": 1.815294116268227e-06, "loss": 0.8995, "step": 10079 }, { "epoch": 4.56728590847304, "grad_norm": 0.06940508754647275, "learning_rate": 1.8115273293193558e-06, "loss": 0.916, "step": 10080 }, { "epoch": 4.567739012233801, "grad_norm": 0.06882504892374557, "learning_rate": 1.807764363977862e-06, "loss": 0.9011, "step": 10081 }, { "epoch": 4.568192115994563, "grad_norm": 0.06790247486000509, "learning_rate": 1.8040052206203285e-06, "loss": 0.9118, "step": 10082 }, { "epoch": 4.568645219755324, "grad_norm": 0.07029233886709288, "learning_rate": 1.80024989962293e-06, "loss": 0.9132, "step": 10083 }, { "epoch": 4.569098323516085, "grad_norm": 0.07001203297019254, "learning_rate": 1.7964984013614684e-06, "loss": 0.9215, "step": 10084 }, { "epoch": 4.5695514272768465, "grad_norm": 0.06773729264703961, "learning_rate": 1.792750726211372e-06, "loss": 0.9294, "step": 10085 }, { "epoch": 4.570004531037608, "grad_norm": 0.07263857069273316, "learning_rate": 1.7890068745476696e-06, "loss": 0.9368, "step": 10086 }, { "epoch": 4.570457634798369, "grad_norm": 0.07163076282651534, "learning_rate": 1.7852668467450352e-06, "loss": 0.9154, "step": 10087 }, { "epoch": 4.57091073855913, "grad_norm": 0.07110612079880593, "learning_rate": 1.7815306431777201e-06, "loss": 0.8999, "step": 10088 }, { "epoch": 4.571363842319891, "grad_norm": 0.07209706696709271, "learning_rate": 1.7777982642196301e-06, "loss": 0.9285, "step": 10089 }, { "epoch": 4.571816946080652, "grad_norm": 0.06781496562896536, "learning_rate": 1.7740697102442662e-06, "loss": 0.9094, "step": 10090 }, { "epoch": 4.572270049841414, "grad_norm": 0.07106279407449062, "learning_rate": 1.7703449816247565e-06, "loss": 0.9253, "step": 10091 }, { "epoch": 4.572723153602174, "grad_norm": 0.07034570640588293, "learning_rate": 1.766624078733843e-06, "loss": 0.9022, "step": 10092 }, { "epoch": 4.573176257362936, "grad_norm": 0.07282316330135354, "learning_rate": 1.7629070019438765e-06, "loss": 0.923, "step": 10093 }, { "epoch": 4.573629361123698, "grad_norm": 0.06903450503271007, "learning_rate": 1.7591937516268488e-06, "loss": 0.9246, "step": 10094 }, { "epoch": 4.574082464884459, "grad_norm": 0.06927322172020511, "learning_rate": 1.7554843281543377e-06, "loss": 0.9209, "step": 10095 }, { "epoch": 4.57453556864522, "grad_norm": 0.07498713969183796, "learning_rate": 1.751778731897562e-06, "loss": 0.9365, "step": 10096 }, { "epoch": 4.574988672405981, "grad_norm": 0.08378287877840943, "learning_rate": 1.7480769632273454e-06, "loss": 0.9225, "step": 10097 }, { "epoch": 4.575441776166742, "grad_norm": 0.07588794603418163, "learning_rate": 1.7443790225141377e-06, "loss": 0.923, "step": 10098 }, { "epoch": 4.575894879927503, "grad_norm": 0.07245717800793004, "learning_rate": 1.7406849101279943e-06, "loss": 0.9144, "step": 10099 }, { "epoch": 4.576347983688264, "grad_norm": 0.07379250803129113, "learning_rate": 1.7369946264385973e-06, "loss": 0.9031, "step": 10100 }, { "epoch": 4.5768010874490255, "grad_norm": 0.07461879933277162, "learning_rate": 1.7333081718152377e-06, "loss": 0.92, "step": 10101 }, { "epoch": 4.5772541912097875, "grad_norm": 0.07748657600890181, "learning_rate": 1.7296255466268385e-06, "loss": 0.909, "step": 10102 }, { "epoch": 4.577707294970549, "grad_norm": 0.06994474995134813, "learning_rate": 1.725946751241918e-06, "loss": 0.9193, "step": 10103 }, { "epoch": 4.57816039873131, "grad_norm": 0.07316148446291403, "learning_rate": 1.7222717860286176e-06, "loss": 0.9273, "step": 10104 }, { "epoch": 4.578613502492071, "grad_norm": 0.07465846426665174, "learning_rate": 1.7186006513547093e-06, "loss": 0.9153, "step": 10105 }, { "epoch": 4.579066606252832, "grad_norm": 0.07084045763870045, "learning_rate": 1.7149333475875663e-06, "loss": 0.9292, "step": 10106 }, { "epoch": 4.579519710013593, "grad_norm": 0.07364888356641043, "learning_rate": 1.7112698750941925e-06, "loss": 0.9317, "step": 10107 }, { "epoch": 4.579972813774354, "grad_norm": 0.07227383732836234, "learning_rate": 1.7076102342411838e-06, "loss": 0.9151, "step": 10108 }, { "epoch": 4.580425917535115, "grad_norm": 0.06832002855474278, "learning_rate": 1.7039544253947848e-06, "loss": 0.9081, "step": 10109 }, { "epoch": 4.5808790212958765, "grad_norm": 0.06933017411830666, "learning_rate": 1.7003024489208364e-06, "loss": 0.9171, "step": 10110 }, { "epoch": 4.581332125056638, "grad_norm": 0.07039147820164145, "learning_rate": 1.6966543051847929e-06, "loss": 0.9042, "step": 10111 }, { "epoch": 4.581785228817399, "grad_norm": 0.07192920150077733, "learning_rate": 1.693009994551744e-06, "loss": 0.9239, "step": 10112 }, { "epoch": 4.582238332578161, "grad_norm": 0.0702467496238943, "learning_rate": 1.6893695173863679e-06, "loss": 0.9398, "step": 10113 }, { "epoch": 4.582691436338922, "grad_norm": 0.07072445827202495, "learning_rate": 1.6857328740529944e-06, "loss": 0.9221, "step": 10114 }, { "epoch": 4.583144540099683, "grad_norm": 0.06978302529031936, "learning_rate": 1.6821000649155327e-06, "loss": 0.9454, "step": 10115 }, { "epoch": 4.583597643860444, "grad_norm": 0.06597806997960964, "learning_rate": 1.6784710903375457e-06, "loss": 0.9361, "step": 10116 }, { "epoch": 4.584050747621205, "grad_norm": 0.07006652577609149, "learning_rate": 1.6748459506821735e-06, "loss": 0.925, "step": 10117 }, { "epoch": 4.5845038513819665, "grad_norm": 0.0687586180011553, "learning_rate": 1.6712246463122062e-06, "loss": 0.9429, "step": 10118 }, { "epoch": 4.584956955142728, "grad_norm": 0.07204962657845011, "learning_rate": 1.6676071775900294e-06, "loss": 0.928, "step": 10119 }, { "epoch": 4.585410058903489, "grad_norm": 0.07279118768890688, "learning_rate": 1.6639935448776512e-06, "loss": 0.9204, "step": 10120 }, { "epoch": 4.58586316266425, "grad_norm": 0.06945871184340033, "learning_rate": 1.6603837485367025e-06, "loss": 0.9302, "step": 10121 }, { "epoch": 4.586316266425011, "grad_norm": 0.06816685305099546, "learning_rate": 1.6567777889284142e-06, "loss": 0.8893, "step": 10122 }, { "epoch": 4.586769370185772, "grad_norm": 0.07476231758122477, "learning_rate": 1.653175666413649e-06, "loss": 0.9067, "step": 10123 }, { "epoch": 4.587222473946534, "grad_norm": 0.069092550938969, "learning_rate": 1.6495773813528782e-06, "loss": 0.9015, "step": 10124 }, { "epoch": 4.587675577707295, "grad_norm": 0.07050085963573585, "learning_rate": 1.6459829341061916e-06, "loss": 0.9443, "step": 10125 }, { "epoch": 4.588128681468056, "grad_norm": 0.07276545290909928, "learning_rate": 1.6423923250332885e-06, "loss": 0.9008, "step": 10126 }, { "epoch": 4.5885817852288175, "grad_norm": 0.07027372276405289, "learning_rate": 1.6388055544934989e-06, "loss": 0.9342, "step": 10127 }, { "epoch": 4.589034888989579, "grad_norm": 0.06868694544043814, "learning_rate": 1.6352226228457534e-06, "loss": 0.9252, "step": 10128 }, { "epoch": 4.58948799275034, "grad_norm": 0.07238449591428645, "learning_rate": 1.631643530448601e-06, "loss": 0.9156, "step": 10129 }, { "epoch": 4.589941096511101, "grad_norm": 0.0727304279765048, "learning_rate": 1.6280682776602174e-06, "loss": 0.923, "step": 10130 }, { "epoch": 4.590394200271862, "grad_norm": 0.07233032273626042, "learning_rate": 1.624496864838374e-06, "loss": 0.941, "step": 10131 }, { "epoch": 4.590847304032623, "grad_norm": 0.06900310081704111, "learning_rate": 1.6209292923404873e-06, "loss": 0.9253, "step": 10132 }, { "epoch": 4.591300407793384, "grad_norm": 0.07060523396730804, "learning_rate": 1.6173655605235561e-06, "loss": 0.9218, "step": 10133 }, { "epoch": 4.5917535115541455, "grad_norm": 0.07282534808418001, "learning_rate": 1.613805669744224e-06, "loss": 0.9286, "step": 10134 }, { "epoch": 4.5922066153149075, "grad_norm": 0.06929117499885416, "learning_rate": 1.6102496203587304e-06, "loss": 0.9267, "step": 10135 }, { "epoch": 4.592659719075669, "grad_norm": 0.07086748312228056, "learning_rate": 1.6066974127229328e-06, "loss": 0.9178, "step": 10136 }, { "epoch": 4.59311282283643, "grad_norm": 0.069449235459445, "learning_rate": 1.6031490471923206e-06, "loss": 0.9221, "step": 10137 }, { "epoch": 4.593565926597191, "grad_norm": 0.07048991840435821, "learning_rate": 1.5996045241219738e-06, "loss": 0.9209, "step": 10138 }, { "epoch": 4.594019030357952, "grad_norm": 0.07133706878524863, "learning_rate": 1.5960638438666133e-06, "loss": 0.9313, "step": 10139 }, { "epoch": 4.594472134118713, "grad_norm": 0.07278019547910759, "learning_rate": 1.5925270067805554e-06, "loss": 0.8975, "step": 10140 }, { "epoch": 4.594925237879474, "grad_norm": 0.0680782629943217, "learning_rate": 1.588994013217744e-06, "loss": 0.9276, "step": 10141 }, { "epoch": 4.595378341640235, "grad_norm": 0.07290881095173597, "learning_rate": 1.585464863531727e-06, "loss": 0.9211, "step": 10142 }, { "epoch": 4.5958314454009965, "grad_norm": 0.06806158896031966, "learning_rate": 1.5819395580756803e-06, "loss": 0.9267, "step": 10143 }, { "epoch": 4.596284549161758, "grad_norm": 0.07385315293313977, "learning_rate": 1.578418097202392e-06, "loss": 0.9212, "step": 10144 }, { "epoch": 4.596737652922519, "grad_norm": 0.07125448323248028, "learning_rate": 1.5749004812642521e-06, "loss": 0.9097, "step": 10145 }, { "epoch": 4.597190756683281, "grad_norm": 0.07341014492453785, "learning_rate": 1.5713867106132853e-06, "loss": 0.9104, "step": 10146 }, { "epoch": 4.597643860444042, "grad_norm": 0.06914971210597927, "learning_rate": 1.5678767856011169e-06, "loss": 0.9152, "step": 10147 }, { "epoch": 4.598096964204803, "grad_norm": 0.07055387037857273, "learning_rate": 1.5643707065789992e-06, "loss": 0.9078, "step": 10148 }, { "epoch": 4.598550067965564, "grad_norm": 0.07126956071015049, "learning_rate": 1.5608684738977897e-06, "loss": 0.9264, "step": 10149 }, { "epoch": 4.599003171726325, "grad_norm": 0.06893200088535058, "learning_rate": 1.5573700879079678e-06, "loss": 0.9207, "step": 10150 }, { "epoch": 4.5994562754870865, "grad_norm": 0.07153819094330961, "learning_rate": 1.5538755489596224e-06, "loss": 0.9386, "step": 10151 }, { "epoch": 4.599909379247848, "grad_norm": 0.06891044934742836, "learning_rate": 1.5503848574024605e-06, "loss": 0.9118, "step": 10152 }, { "epoch": 4.600362483008609, "grad_norm": 0.06889501773632137, "learning_rate": 1.546898013585798e-06, "loss": 0.9378, "step": 10153 }, { "epoch": 4.60081558676937, "grad_norm": 0.06536474871383827, "learning_rate": 1.5434150178585827e-06, "loss": 0.9179, "step": 10154 }, { "epoch": 4.601268690530132, "grad_norm": 0.06825893904945687, "learning_rate": 1.539935870569358e-06, "loss": 0.9103, "step": 10155 }, { "epoch": 4.601721794290892, "grad_norm": 0.0679986079987173, "learning_rate": 1.5364605720662896e-06, "loss": 0.9287, "step": 10156 }, { "epoch": 4.602174898051654, "grad_norm": 0.07117389067329004, "learning_rate": 1.5329891226971616e-06, "loss": 0.9023, "step": 10157 }, { "epoch": 4.602628001812415, "grad_norm": 0.06902727526431245, "learning_rate": 1.5295215228093674e-06, "loss": 0.9285, "step": 10158 }, { "epoch": 4.603081105573176, "grad_norm": 0.06859481596920024, "learning_rate": 1.5260577727499225e-06, "loss": 0.932, "step": 10159 }, { "epoch": 4.6035342093339375, "grad_norm": 0.07045509635087147, "learning_rate": 1.5225978728654478e-06, "loss": 0.9217, "step": 10160 }, { "epoch": 4.603987313094699, "grad_norm": 0.06929030363178657, "learning_rate": 1.519141823502186e-06, "loss": 0.9081, "step": 10161 }, { "epoch": 4.60444041685546, "grad_norm": 0.06960934460418372, "learning_rate": 1.5156896250059893e-06, "loss": 0.9272, "step": 10162 }, { "epoch": 4.604893520616221, "grad_norm": 0.07054512971746492, "learning_rate": 1.5122412777223283e-06, "loss": 0.9357, "step": 10163 }, { "epoch": 4.605346624376982, "grad_norm": 0.07209033603490379, "learning_rate": 1.5087967819962913e-06, "loss": 0.9174, "step": 10164 }, { "epoch": 4.605799728137743, "grad_norm": 0.06788848012697977, "learning_rate": 1.5053561381725624e-06, "loss": 0.9157, "step": 10165 }, { "epoch": 4.606252831898505, "grad_norm": 0.07006163979033439, "learning_rate": 1.5019193465954706e-06, "loss": 0.9233, "step": 10166 }, { "epoch": 4.606705935659266, "grad_norm": 0.06945512513262107, "learning_rate": 1.4984864076089367e-06, "loss": 0.9085, "step": 10167 }, { "epoch": 4.6071590394200275, "grad_norm": 0.06806810957070157, "learning_rate": 1.4950573215565034e-06, "loss": 0.9112, "step": 10168 }, { "epoch": 4.607612143180789, "grad_norm": 0.07144298089440893, "learning_rate": 1.4916320887813273e-06, "loss": 0.929, "step": 10169 }, { "epoch": 4.60806524694155, "grad_norm": 0.0727947513799577, "learning_rate": 1.4882107096261833e-06, "loss": 0.9384, "step": 10170 }, { "epoch": 4.608518350702311, "grad_norm": 0.07053630978176682, "learning_rate": 1.4847931844334507e-06, "loss": 0.9176, "step": 10171 }, { "epoch": 4.608971454463072, "grad_norm": 0.06883812131698058, "learning_rate": 1.4813795135451225e-06, "loss": 0.9176, "step": 10172 }, { "epoch": 4.609424558223833, "grad_norm": 0.06958055057241286, "learning_rate": 1.4779696973028324e-06, "loss": 0.933, "step": 10173 }, { "epoch": 4.609877661984594, "grad_norm": 0.07255561365861295, "learning_rate": 1.474563736047787e-06, "loss": 0.9458, "step": 10174 }, { "epoch": 4.610330765745355, "grad_norm": 0.06858102810968822, "learning_rate": 1.4711616301208476e-06, "loss": 0.9201, "step": 10175 }, { "epoch": 4.6107838695061165, "grad_norm": 0.06970434215729482, "learning_rate": 1.4677633798624525e-06, "loss": 0.9441, "step": 10176 }, { "epoch": 4.6112369732668785, "grad_norm": 0.06994274373750435, "learning_rate": 1.4643689856126898e-06, "loss": 0.9231, "step": 10177 }, { "epoch": 4.61169007702764, "grad_norm": 0.07086475529158359, "learning_rate": 1.4609784477112298e-06, "loss": 0.9115, "step": 10178 }, { "epoch": 4.612143180788401, "grad_norm": 0.07033410892117735, "learning_rate": 1.4575917664973792e-06, "loss": 0.9077, "step": 10179 }, { "epoch": 4.612596284549162, "grad_norm": 0.07256364877979772, "learning_rate": 1.4542089423100537e-06, "loss": 0.9127, "step": 10180 }, { "epoch": 4.613049388309923, "grad_norm": 0.06689979319717959, "learning_rate": 1.4508299754877685e-06, "loss": 0.9051, "step": 10181 }, { "epoch": 4.613502492070684, "grad_norm": 0.06682186066192172, "learning_rate": 1.447454866368676e-06, "loss": 0.9133, "step": 10182 }, { "epoch": 4.613955595831445, "grad_norm": 0.07137121135539873, "learning_rate": 1.4440836152905234e-06, "loss": 0.9268, "step": 10183 }, { "epoch": 4.6144086995922065, "grad_norm": 0.06907715348335862, "learning_rate": 1.4407162225906856e-06, "loss": 0.9129, "step": 10184 }, { "epoch": 4.614861803352968, "grad_norm": 0.07259996830857245, "learning_rate": 1.4373526886061418e-06, "loss": 0.9153, "step": 10185 }, { "epoch": 4.615314907113729, "grad_norm": 0.07084022546120307, "learning_rate": 1.4339930136734891e-06, "loss": 0.9339, "step": 10186 }, { "epoch": 4.61576801087449, "grad_norm": 0.0705476545997012, "learning_rate": 1.4306371981289347e-06, "loss": 0.9139, "step": 10187 }, { "epoch": 4.616221114635252, "grad_norm": 0.07243573984305303, "learning_rate": 1.4272852423083116e-06, "loss": 0.9557, "step": 10188 }, { "epoch": 4.616674218396013, "grad_norm": 0.07437439971541457, "learning_rate": 1.4239371465470498e-06, "loss": 0.9374, "step": 10189 }, { "epoch": 4.617127322156774, "grad_norm": 0.07073083940535564, "learning_rate": 1.4205929111802008e-06, "loss": 0.9231, "step": 10190 }, { "epoch": 4.617580425917535, "grad_norm": 0.06993560624401572, "learning_rate": 1.417252536542435e-06, "loss": 0.9194, "step": 10191 }, { "epoch": 4.618033529678296, "grad_norm": 0.07047341517827004, "learning_rate": 1.4139160229680272e-06, "loss": 0.9034, "step": 10192 }, { "epoch": 4.6184866334390575, "grad_norm": 0.07463732866668156, "learning_rate": 1.4105833707908701e-06, "loss": 0.9511, "step": 10193 }, { "epoch": 4.618939737199819, "grad_norm": 0.07074657990550794, "learning_rate": 1.4072545803444703e-06, "loss": 0.914, "step": 10194 }, { "epoch": 4.61939284096058, "grad_norm": 0.07219312396471218, "learning_rate": 1.4039296519619483e-06, "loss": 0.9334, "step": 10195 }, { "epoch": 4.619845944721341, "grad_norm": 0.06938010589956224, "learning_rate": 1.400608585976042e-06, "loss": 0.8884, "step": 10196 }, { "epoch": 4.620299048482102, "grad_norm": 0.06977925568908465, "learning_rate": 1.3972913827190859e-06, "loss": 0.9264, "step": 10197 }, { "epoch": 4.620752152242863, "grad_norm": 0.07543051834209254, "learning_rate": 1.3939780425230497e-06, "loss": 0.9197, "step": 10198 }, { "epoch": 4.621205256003625, "grad_norm": 0.07272137507875452, "learning_rate": 1.3906685657194995e-06, "loss": 0.9382, "step": 10199 }, { "epoch": 4.621658359764386, "grad_norm": 0.06939191275727093, "learning_rate": 1.3873629526396326e-06, "loss": 0.9408, "step": 10200 }, { "epoch": 4.6221114635251475, "grad_norm": 0.07704402831085241, "learning_rate": 1.384061203614242e-06, "loss": 0.9434, "step": 10201 }, { "epoch": 4.622564567285909, "grad_norm": 0.07346825265694215, "learning_rate": 1.3807633189737435e-06, "loss": 0.9208, "step": 10202 }, { "epoch": 4.62301767104667, "grad_norm": 0.07602800566675054, "learning_rate": 1.3774692990481575e-06, "loss": 0.938, "step": 10203 }, { "epoch": 4.623470774807431, "grad_norm": 0.0698798290694584, "learning_rate": 1.374179144167136e-06, "loss": 0.8966, "step": 10204 }, { "epoch": 4.623923878568192, "grad_norm": 0.07088630630743054, "learning_rate": 1.3708928546599265e-06, "loss": 0.9223, "step": 10205 }, { "epoch": 4.624376982328953, "grad_norm": 0.07268043366998088, "learning_rate": 1.3676104308553906e-06, "loss": 0.8978, "step": 10206 }, { "epoch": 4.624830086089714, "grad_norm": 0.0711620748765938, "learning_rate": 1.3643318730820166e-06, "loss": 0.8993, "step": 10207 }, { "epoch": 4.625283189850475, "grad_norm": 0.07380950912442512, "learning_rate": 1.361057181667893e-06, "loss": 0.9226, "step": 10208 }, { "epoch": 4.6257362936112365, "grad_norm": 0.0711405290438703, "learning_rate": 1.3577863569407269e-06, "loss": 0.9306, "step": 10209 }, { "epoch": 4.6261893973719985, "grad_norm": 0.07013700194217569, "learning_rate": 1.3545193992278293e-06, "loss": 0.9251, "step": 10210 }, { "epoch": 4.62664250113276, "grad_norm": 0.07222677710691451, "learning_rate": 1.3512563088561436e-06, "loss": 0.9256, "step": 10211 }, { "epoch": 4.627095604893521, "grad_norm": 0.0692880756701092, "learning_rate": 1.347997086152204e-06, "loss": 0.9153, "step": 10212 }, { "epoch": 4.627548708654282, "grad_norm": 0.07135877400854969, "learning_rate": 1.3447417314421807e-06, "loss": 0.9197, "step": 10213 }, { "epoch": 4.628001812415043, "grad_norm": 0.06785502702095089, "learning_rate": 1.3414902450518353e-06, "loss": 0.9205, "step": 10214 }, { "epoch": 4.628454916175804, "grad_norm": 0.06766901684380397, "learning_rate": 1.338242627306552e-06, "loss": 0.9145, "step": 10215 }, { "epoch": 4.628908019936565, "grad_norm": 0.07292332083577231, "learning_rate": 1.3349988785313328e-06, "loss": 0.9119, "step": 10216 }, { "epoch": 4.629361123697326, "grad_norm": 0.07141856880929012, "learning_rate": 1.3317589990507806e-06, "loss": 0.9292, "step": 10217 }, { "epoch": 4.629814227458088, "grad_norm": 0.06761284788034849, "learning_rate": 1.328522989189125e-06, "loss": 0.9282, "step": 10218 }, { "epoch": 4.63026733121885, "grad_norm": 0.0737886964405458, "learning_rate": 1.3252908492701866e-06, "loss": 0.9229, "step": 10219 }, { "epoch": 4.63072043497961, "grad_norm": 0.070822882547468, "learning_rate": 1.3220625796174314e-06, "loss": 0.9299, "step": 10220 }, { "epoch": 4.631173538740372, "grad_norm": 0.07317203361254604, "learning_rate": 1.3188381805539074e-06, "loss": 0.9094, "step": 10221 }, { "epoch": 4.631626642501133, "grad_norm": 0.07099886018518581, "learning_rate": 1.3156176524022945e-06, "loss": 0.9417, "step": 10222 }, { "epoch": 4.632079746261894, "grad_norm": 0.07292527120986714, "learning_rate": 1.3124009954848682e-06, "loss": 0.9227, "step": 10223 }, { "epoch": 4.632532850022655, "grad_norm": 0.0686450063459621, "learning_rate": 1.3091882101235399e-06, "loss": 0.9195, "step": 10224 }, { "epoch": 4.632985953783416, "grad_norm": 0.07271700841301618, "learning_rate": 1.3059792966398122e-06, "loss": 0.9328, "step": 10225 }, { "epoch": 4.6334390575441775, "grad_norm": 0.07056498810247705, "learning_rate": 1.3027742553548062e-06, "loss": 0.905, "step": 10226 }, { "epoch": 4.633892161304939, "grad_norm": 0.07137904390828267, "learning_rate": 1.2995730865892654e-06, "loss": 0.9072, "step": 10227 }, { "epoch": 4.6343452650657, "grad_norm": 0.06732524486746601, "learning_rate": 1.2963757906635333e-06, "loss": 0.9287, "step": 10228 }, { "epoch": 4.634798368826461, "grad_norm": 0.07521074040764399, "learning_rate": 1.2931823678975719e-06, "loss": 0.9365, "step": 10229 }, { "epoch": 4.635251472587223, "grad_norm": 0.069221931196433, "learning_rate": 1.2899928186109523e-06, "loss": 0.9268, "step": 10230 }, { "epoch": 4.635704576347984, "grad_norm": 0.06805962188018473, "learning_rate": 1.2868071431228635e-06, "loss": 0.9099, "step": 10231 }, { "epoch": 4.636157680108745, "grad_norm": 0.06744333552864859, "learning_rate": 1.2836253417520994e-06, "loss": 0.9289, "step": 10232 }, { "epoch": 4.636610783869506, "grad_norm": 0.06969101164908506, "learning_rate": 1.280447414817072e-06, "loss": 0.9281, "step": 10233 }, { "epoch": 4.6370638876302674, "grad_norm": 0.06591155229230478, "learning_rate": 1.2772733626358068e-06, "loss": 0.926, "step": 10234 }, { "epoch": 4.637516991391029, "grad_norm": 0.06940716167534251, "learning_rate": 1.27410318552593e-06, "loss": 0.9254, "step": 10235 }, { "epoch": 4.63797009515179, "grad_norm": 0.06594675075177285, "learning_rate": 1.2709368838046943e-06, "loss": 0.9397, "step": 10236 }, { "epoch": 4.638423198912551, "grad_norm": 0.06589064645795167, "learning_rate": 1.2677744577889573e-06, "loss": 0.9258, "step": 10237 }, { "epoch": 4.638876302673312, "grad_norm": 0.06829923236134816, "learning_rate": 1.2646159077951991e-06, "loss": 0.9267, "step": 10238 }, { "epoch": 4.639329406434073, "grad_norm": 0.06748119208042604, "learning_rate": 1.2614612341394872e-06, "loss": 0.9449, "step": 10239 }, { "epoch": 4.639782510194834, "grad_norm": 0.06668448613918328, "learning_rate": 1.2583104371375244e-06, "loss": 0.9026, "step": 10240 }, { "epoch": 4.640235613955596, "grad_norm": 0.06936433741112312, "learning_rate": 1.2551635171046227e-06, "loss": 0.9366, "step": 10241 }, { "epoch": 4.640688717716357, "grad_norm": 0.06862634733903633, "learning_rate": 1.2520204743556952e-06, "loss": 0.911, "step": 10242 }, { "epoch": 4.6411418214771185, "grad_norm": 0.06727317074790276, "learning_rate": 1.2488813092052764e-06, "loss": 0.9182, "step": 10243 }, { "epoch": 4.64159492523788, "grad_norm": 0.0664261463201802, "learning_rate": 1.2457460219675066e-06, "loss": 0.9251, "step": 10244 }, { "epoch": 4.642048028998641, "grad_norm": 0.06877679665862654, "learning_rate": 1.2426146129561433e-06, "loss": 0.923, "step": 10245 }, { "epoch": 4.642501132759402, "grad_norm": 0.06886920390650791, "learning_rate": 1.239487082484554e-06, "loss": 0.9255, "step": 10246 }, { "epoch": 4.642954236520163, "grad_norm": 0.06944702623309257, "learning_rate": 1.2363634308657192e-06, "loss": 0.9249, "step": 10247 }, { "epoch": 4.643407340280924, "grad_norm": 0.06720479532980796, "learning_rate": 1.2332436584122243e-06, "loss": 0.9218, "step": 10248 }, { "epoch": 4.643860444041685, "grad_norm": 0.06852824473172439, "learning_rate": 1.230127765436282e-06, "loss": 0.8979, "step": 10249 }, { "epoch": 4.644313547802446, "grad_norm": 0.06650343380777947, "learning_rate": 1.2270157522496962e-06, "loss": 0.8991, "step": 10250 }, { "epoch": 4.644766651563208, "grad_norm": 0.06742488402547422, "learning_rate": 1.2239076191638977e-06, "loss": 0.9254, "step": 10251 }, { "epoch": 4.64521975532397, "grad_norm": 0.06632255055168428, "learning_rate": 1.2208033664899266e-06, "loss": 0.9166, "step": 10252 }, { "epoch": 4.645672859084731, "grad_norm": 0.06947272436399679, "learning_rate": 1.2177029945384233e-06, "loss": 0.9321, "step": 10253 }, { "epoch": 4.646125962845492, "grad_norm": 0.06636760564250044, "learning_rate": 1.214606503619664e-06, "loss": 0.906, "step": 10254 }, { "epoch": 4.646579066606253, "grad_norm": 0.06901570773292569, "learning_rate": 1.2115138940435078e-06, "loss": 0.9124, "step": 10255 }, { "epoch": 4.647032170367014, "grad_norm": 0.06994552995605856, "learning_rate": 1.208425166119449e-06, "loss": 0.9121, "step": 10256 }, { "epoch": 4.647485274127775, "grad_norm": 0.06962051534461403, "learning_rate": 1.2053403201565783e-06, "loss": 0.9503, "step": 10257 }, { "epoch": 4.647938377888536, "grad_norm": 0.06677384051956058, "learning_rate": 1.2022593564636043e-06, "loss": 0.9407, "step": 10258 }, { "epoch": 4.6483914816492975, "grad_norm": 0.06718261748907921, "learning_rate": 1.1991822753488447e-06, "loss": 0.9393, "step": 10259 }, { "epoch": 4.648844585410059, "grad_norm": 0.0658931322910124, "learning_rate": 1.1961090771202311e-06, "loss": 0.9299, "step": 10260 }, { "epoch": 4.64929768917082, "grad_norm": 0.0703807613497133, "learning_rate": 1.1930397620853128e-06, "loss": 0.9273, "step": 10261 }, { "epoch": 4.649750792931581, "grad_norm": 0.0737944635037005, "learning_rate": 1.1899743305512267e-06, "loss": 0.9399, "step": 10262 }, { "epoch": 4.650203896692343, "grad_norm": 0.07256772029430497, "learning_rate": 1.1869127828247539e-06, "loss": 0.9235, "step": 10263 }, { "epoch": 4.650657000453104, "grad_norm": 0.0676256908780608, "learning_rate": 1.1838551192122583e-06, "loss": 0.9087, "step": 10264 }, { "epoch": 4.651110104213865, "grad_norm": 0.07008826055534219, "learning_rate": 1.1808013400197437e-06, "loss": 0.9155, "step": 10265 }, { "epoch": 4.651563207974626, "grad_norm": 0.06765484927598073, "learning_rate": 1.1777514455527882e-06, "loss": 0.9262, "step": 10266 }, { "epoch": 4.652016311735387, "grad_norm": 0.06882322652940757, "learning_rate": 1.1747054361166144e-06, "loss": 0.9265, "step": 10267 }, { "epoch": 4.652469415496149, "grad_norm": 0.06880660089614303, "learning_rate": 1.1716633120160359e-06, "loss": 0.9094, "step": 10268 }, { "epoch": 4.65292251925691, "grad_norm": 0.06828843637129262, "learning_rate": 1.1686250735554939e-06, "loss": 0.9134, "step": 10269 }, { "epoch": 4.653375623017671, "grad_norm": 0.06811699160782517, "learning_rate": 1.165590721039025e-06, "loss": 0.94, "step": 10270 }, { "epoch": 4.653828726778432, "grad_norm": 0.06529273678900203, "learning_rate": 1.162560254770284e-06, "loss": 0.9547, "step": 10271 }, { "epoch": 4.654281830539193, "grad_norm": 0.07095494762524504, "learning_rate": 1.15953367505254e-06, "loss": 0.9389, "step": 10272 }, { "epoch": 4.654734934299954, "grad_norm": 0.0704121891488099, "learning_rate": 1.15651098218867e-06, "loss": 0.9427, "step": 10273 }, { "epoch": 4.655188038060716, "grad_norm": 0.07045380744960272, "learning_rate": 1.153492176481157e-06, "loss": 0.91, "step": 10274 }, { "epoch": 4.655641141821477, "grad_norm": 0.06669575174993907, "learning_rate": 1.1504772582321011e-06, "loss": 0.9068, "step": 10275 }, { "epoch": 4.6560942455822385, "grad_norm": 0.0711032091830435, "learning_rate": 1.1474662277432215e-06, "loss": 0.94, "step": 10276 }, { "epoch": 4.656547349343, "grad_norm": 0.0699787569698959, "learning_rate": 1.1444590853158278e-06, "loss": 0.9188, "step": 10277 }, { "epoch": 4.657000453103761, "grad_norm": 0.07330617784677017, "learning_rate": 1.1414558312508484e-06, "loss": 0.948, "step": 10278 }, { "epoch": 4.657453556864522, "grad_norm": 0.07200067570130303, "learning_rate": 1.1384564658488383e-06, "loss": 0.9352, "step": 10279 }, { "epoch": 4.657906660625283, "grad_norm": 0.07333905256865149, "learning_rate": 1.13546098940994e-06, "loss": 0.9434, "step": 10280 }, { "epoch": 4.658359764386044, "grad_norm": 0.0696072806582279, "learning_rate": 1.1324694022339267e-06, "loss": 0.9283, "step": 10281 }, { "epoch": 4.658812868146805, "grad_norm": 0.06865962311388807, "learning_rate": 1.1294817046201677e-06, "loss": 0.9339, "step": 10282 }, { "epoch": 4.659265971907567, "grad_norm": 0.07267501658901374, "learning_rate": 1.126497896867651e-06, "loss": 0.9289, "step": 10283 }, { "epoch": 4.6597190756683275, "grad_norm": 0.07260273155160935, "learning_rate": 1.1235179792749685e-06, "loss": 0.9426, "step": 10284 }, { "epoch": 4.66017217942909, "grad_norm": 0.07014526382528029, "learning_rate": 1.120541952140335e-06, "loss": 0.9303, "step": 10285 }, { "epoch": 4.660625283189851, "grad_norm": 0.07405642066652124, "learning_rate": 1.1175698157615656e-06, "loss": 0.9166, "step": 10286 }, { "epoch": 4.661078386950612, "grad_norm": 0.06775307110066223, "learning_rate": 1.1146015704360802e-06, "loss": 0.9225, "step": 10287 }, { "epoch": 4.661531490711373, "grad_norm": 0.07003416575829388, "learning_rate": 1.1116372164609301e-06, "loss": 0.9031, "step": 10288 }, { "epoch": 4.661984594472134, "grad_norm": 0.0684787284574271, "learning_rate": 1.108676754132758e-06, "loss": 0.9076, "step": 10289 }, { "epoch": 4.662437698232895, "grad_norm": 0.06833839835440783, "learning_rate": 1.105720183747825e-06, "loss": 0.936, "step": 10290 }, { "epoch": 4.662890801993656, "grad_norm": 0.06743715037153802, "learning_rate": 1.1027675056020048e-06, "loss": 0.8977, "step": 10291 }, { "epoch": 4.6633439057544175, "grad_norm": 0.07131164951876207, "learning_rate": 1.0998187199907773e-06, "loss": 0.9148, "step": 10292 }, { "epoch": 4.663797009515179, "grad_norm": 0.07201297764193115, "learning_rate": 1.0968738272092305e-06, "loss": 0.9132, "step": 10293 }, { "epoch": 4.664250113275941, "grad_norm": 0.06928173784823034, "learning_rate": 1.093932827552071e-06, "loss": 0.9324, "step": 10294 }, { "epoch": 4.664703217036702, "grad_norm": 0.07456430962306497, "learning_rate": 1.0909957213136102e-06, "loss": 0.9119, "step": 10295 }, { "epoch": 4.665156320797463, "grad_norm": 0.06876364056680172, "learning_rate": 1.0880625087877683e-06, "loss": 0.8922, "step": 10296 }, { "epoch": 4.665609424558224, "grad_norm": 0.06630051671012566, "learning_rate": 1.085133190268084e-06, "loss": 0.9223, "step": 10297 }, { "epoch": 4.666062528318985, "grad_norm": 0.06974041385882258, "learning_rate": 1.082207766047696e-06, "loss": 0.9388, "step": 10298 }, { "epoch": 4.666515632079746, "grad_norm": 0.07808502416650949, "learning_rate": 1.0792862364193612e-06, "loss": 0.9057, "step": 10299 }, { "epoch": 4.666968735840507, "grad_norm": 0.06882883917921602, "learning_rate": 1.076368601675437e-06, "loss": 0.9316, "step": 10300 }, { "epoch": 4.6674218396012686, "grad_norm": 0.06938632041903654, "learning_rate": 1.0734548621079121e-06, "loss": 0.9277, "step": 10301 }, { "epoch": 4.66787494336203, "grad_norm": 0.07127194768908755, "learning_rate": 1.0705450180083577e-06, "loss": 0.9185, "step": 10302 }, { "epoch": 4.668328047122791, "grad_norm": 0.06631902149368454, "learning_rate": 1.0676390696679718e-06, "loss": 0.9186, "step": 10303 }, { "epoch": 4.668781150883552, "grad_norm": 0.06863349648104257, "learning_rate": 1.0647370173775618e-06, "loss": 0.9261, "step": 10304 }, { "epoch": 4.669234254644314, "grad_norm": 0.069539548725146, "learning_rate": 1.0618388614275399e-06, "loss": 0.9092, "step": 10305 }, { "epoch": 4.669687358405075, "grad_norm": 0.06897553231391117, "learning_rate": 1.0589446021079363e-06, "loss": 0.9507, "step": 10306 }, { "epoch": 4.670140462165836, "grad_norm": 0.06973759332687134, "learning_rate": 1.0560542397083773e-06, "loss": 0.9198, "step": 10307 }, { "epoch": 4.670593565926597, "grad_norm": 0.0691972253567225, "learning_rate": 1.0531677745181158e-06, "loss": 0.9333, "step": 10308 }, { "epoch": 4.6710466696873585, "grad_norm": 0.06977763844243927, "learning_rate": 1.050285206826005e-06, "loss": 0.9361, "step": 10309 }, { "epoch": 4.67149977344812, "grad_norm": 0.07213659173694893, "learning_rate": 1.0474065369205077e-06, "loss": 0.926, "step": 10310 }, { "epoch": 4.671952877208881, "grad_norm": 0.06732810099029948, "learning_rate": 1.0445317650897002e-06, "loss": 0.9209, "step": 10311 }, { "epoch": 4.672405980969642, "grad_norm": 0.06932586954684725, "learning_rate": 1.0416608916212678e-06, "loss": 0.9302, "step": 10312 }, { "epoch": 4.672859084730403, "grad_norm": 0.07051582024599656, "learning_rate": 1.0387939168025096e-06, "loss": 0.9152, "step": 10313 }, { "epoch": 4.673312188491164, "grad_norm": 0.0666332826806688, "learning_rate": 1.035930840920316e-06, "loss": 0.9419, "step": 10314 }, { "epoch": 4.673765292251925, "grad_norm": 0.06822916052726953, "learning_rate": 1.0330716642612227e-06, "loss": 0.9326, "step": 10315 }, { "epoch": 4.674218396012687, "grad_norm": 0.06530304316397521, "learning_rate": 1.0302163871113335e-06, "loss": 0.9477, "step": 10316 }, { "epoch": 4.674671499773448, "grad_norm": 0.06986928933353845, "learning_rate": 1.027365009756398e-06, "loss": 0.9234, "step": 10317 }, { "epoch": 4.6751246035342096, "grad_norm": 0.07321886398085702, "learning_rate": 1.024517532481757e-06, "loss": 0.946, "step": 10318 }, { "epoch": 4.675577707294971, "grad_norm": 0.06990775757407267, "learning_rate": 1.021673955572351e-06, "loss": 0.9406, "step": 10319 }, { "epoch": 4.676030811055732, "grad_norm": 0.06870513835091771, "learning_rate": 1.0188342793127614e-06, "loss": 0.944, "step": 10320 }, { "epoch": 4.676483914816493, "grad_norm": 0.06585961856346682, "learning_rate": 1.015998503987148e-06, "loss": 0.9299, "step": 10321 }, { "epoch": 4.676937018577254, "grad_norm": 0.06505038984361645, "learning_rate": 1.0131666298793052e-06, "loss": 0.9111, "step": 10322 }, { "epoch": 4.677390122338015, "grad_norm": 0.06978243420770262, "learning_rate": 1.010338657272607e-06, "loss": 0.9373, "step": 10323 }, { "epoch": 4.677843226098776, "grad_norm": 0.07161456323135404, "learning_rate": 1.0075145864500757e-06, "loss": 0.9289, "step": 10324 }, { "epoch": 4.6782963298595375, "grad_norm": 0.06775641707502912, "learning_rate": 1.004694417694303e-06, "loss": 0.9225, "step": 10325 }, { "epoch": 4.678749433620299, "grad_norm": 0.07029135340715575, "learning_rate": 1.0018781512875297e-06, "loss": 0.929, "step": 10326 }, { "epoch": 4.679202537381061, "grad_norm": 0.0708458224893363, "learning_rate": 9.990657875115662e-07, "loss": 0.9169, "step": 10327 }, { "epoch": 4.679655641141822, "grad_norm": 0.07980521000811144, "learning_rate": 9.96257326647867e-07, "loss": 0.9055, "step": 10328 }, { "epoch": 4.680108744902583, "grad_norm": 0.07024167256703137, "learning_rate": 9.934527689774786e-07, "loss": 0.9173, "step": 10329 }, { "epoch": 4.680561848663344, "grad_norm": 0.07119319186311814, "learning_rate": 9.90652114781052e-07, "loss": 0.9078, "step": 10330 }, { "epoch": 4.681014952424105, "grad_norm": 0.0656788410898231, "learning_rate": 9.87855364338861e-07, "loss": 0.9168, "step": 10331 }, { "epoch": 4.681468056184866, "grad_norm": 0.06771608097715526, "learning_rate": 9.850625179307793e-07, "loss": 0.9265, "step": 10332 }, { "epoch": 4.681921159945627, "grad_norm": 0.06965426104632735, "learning_rate": 9.82273575836299e-07, "loss": 0.9124, "step": 10333 }, { "epoch": 4.6823742637063885, "grad_norm": 0.06681116539007115, "learning_rate": 9.794885383345075e-07, "loss": 0.9354, "step": 10334 }, { "epoch": 4.68282736746715, "grad_norm": 0.07046902052503061, "learning_rate": 9.767074057041205e-07, "loss": 0.9162, "step": 10335 }, { "epoch": 4.683280471227912, "grad_norm": 0.07078051702856959, "learning_rate": 9.739301782234434e-07, "loss": 0.9374, "step": 10336 }, { "epoch": 4.683733574988672, "grad_norm": 0.06940707897657969, "learning_rate": 9.711568561704053e-07, "loss": 0.9127, "step": 10337 }, { "epoch": 4.684186678749434, "grad_norm": 0.07141784269541698, "learning_rate": 9.683874398225357e-07, "loss": 0.9175, "step": 10338 }, { "epoch": 4.684639782510195, "grad_norm": 0.070058674435887, "learning_rate": 9.656219294569723e-07, "loss": 0.916, "step": 10339 }, { "epoch": 4.685092886270956, "grad_norm": 0.06822030965376583, "learning_rate": 9.628603253504765e-07, "loss": 0.9124, "step": 10340 }, { "epoch": 4.685545990031717, "grad_norm": 0.07138316468602927, "learning_rate": 9.601026277794001e-07, "loss": 0.9159, "step": 10341 }, { "epoch": 4.6859990937924785, "grad_norm": 0.0666745661516817, "learning_rate": 9.573488370197182e-07, "loss": 0.9148, "step": 10342 }, { "epoch": 4.68645219755324, "grad_norm": 0.06591549040140705, "learning_rate": 9.545989533469968e-07, "loss": 0.931, "step": 10343 }, { "epoch": 4.686905301314001, "grad_norm": 0.07144262024516171, "learning_rate": 9.518529770364427e-07, "loss": 0.9169, "step": 10344 }, { "epoch": 4.687358405074762, "grad_norm": 0.06875836230499252, "learning_rate": 9.491109083628313e-07, "loss": 0.9008, "step": 10345 }, { "epoch": 4.687811508835523, "grad_norm": 0.06882373263784888, "learning_rate": 9.463727476005835e-07, "loss": 0.9318, "step": 10346 }, { "epoch": 4.688264612596285, "grad_norm": 0.06556333405965312, "learning_rate": 9.436384950237065e-07, "loss": 0.9278, "step": 10347 }, { "epoch": 4.688717716357046, "grad_norm": 0.07412740477945712, "learning_rate": 9.409081509058216e-07, "loss": 0.927, "step": 10348 }, { "epoch": 4.689170820117807, "grad_norm": 0.06765822005078656, "learning_rate": 9.381817155201634e-07, "loss": 0.9118, "step": 10349 }, { "epoch": 4.689623923878568, "grad_norm": 0.06663999646098132, "learning_rate": 9.354591891395715e-07, "loss": 0.9064, "step": 10350 }, { "epoch": 4.6900770276393295, "grad_norm": 0.0677326227950703, "learning_rate": 9.327405720364991e-07, "loss": 0.8978, "step": 10351 }, { "epoch": 4.690530131400091, "grad_norm": 0.06778881653162569, "learning_rate": 9.300258644829951e-07, "loss": 0.9337, "step": 10352 }, { "epoch": 4.690983235160852, "grad_norm": 0.07621719125184238, "learning_rate": 9.273150667507358e-07, "loss": 0.9294, "step": 10353 }, { "epoch": 4.691436338921613, "grad_norm": 0.06567177308382262, "learning_rate": 9.246081791109929e-07, "loss": 0.9392, "step": 10354 }, { "epoch": 4.691889442682374, "grad_norm": 0.06918754782813592, "learning_rate": 9.219052018346519e-07, "loss": 0.9387, "step": 10355 }, { "epoch": 4.692342546443135, "grad_norm": 0.06950608200441816, "learning_rate": 9.192061351922033e-07, "loss": 0.927, "step": 10356 }, { "epoch": 4.692795650203896, "grad_norm": 0.07814237816738824, "learning_rate": 9.165109794537507e-07, "loss": 0.9078, "step": 10357 }, { "epoch": 4.693248753964658, "grad_norm": 0.06791948351589672, "learning_rate": 9.138197348890076e-07, "loss": 0.9218, "step": 10358 }, { "epoch": 4.6937018577254195, "grad_norm": 0.06946721062599598, "learning_rate": 9.111324017672874e-07, "loss": 0.9347, "step": 10359 }, { "epoch": 4.694154961486181, "grad_norm": 0.06993057863416484, "learning_rate": 9.084489803575213e-07, "loss": 0.9017, "step": 10360 }, { "epoch": 4.694608065246942, "grad_norm": 0.06713633059266355, "learning_rate": 9.057694709282417e-07, "loss": 0.9353, "step": 10361 }, { "epoch": 4.695061169007703, "grad_norm": 0.06668392553873272, "learning_rate": 9.03093873747598e-07, "loss": 0.9228, "step": 10362 }, { "epoch": 4.695514272768464, "grad_norm": 0.06858039872004676, "learning_rate": 9.004221890833453e-07, "loss": 0.9027, "step": 10363 }, { "epoch": 4.695967376529225, "grad_norm": 0.06996690316630304, "learning_rate": 8.97754417202834e-07, "loss": 0.909, "step": 10364 }, { "epoch": 4.696420480289986, "grad_norm": 0.06873783687679835, "learning_rate": 8.950905583730463e-07, "loss": 0.9297, "step": 10365 }, { "epoch": 4.696873584050747, "grad_norm": 0.06774388548530266, "learning_rate": 8.924306128605509e-07, "loss": 0.9501, "step": 10366 }, { "epoch": 4.6973266878115085, "grad_norm": 0.0652121094100893, "learning_rate": 8.897745809315395e-07, "loss": 0.9103, "step": 10367 }, { "epoch": 4.69777979157227, "grad_norm": 0.06920562584449232, "learning_rate": 8.871224628518082e-07, "loss": 0.9388, "step": 10368 }, { "epoch": 4.698232895333032, "grad_norm": 0.06634904820538938, "learning_rate": 8.844742588867628e-07, "loss": 0.9057, "step": 10369 }, { "epoch": 4.698685999093793, "grad_norm": 0.06541205560541268, "learning_rate": 8.818299693014088e-07, "loss": 0.9325, "step": 10370 }, { "epoch": 4.699139102854554, "grad_norm": 0.06717454014462783, "learning_rate": 8.791895943603657e-07, "loss": 0.9166, "step": 10371 }, { "epoch": 4.699592206615315, "grad_norm": 0.06758786267023213, "learning_rate": 8.765531343278666e-07, "loss": 0.8972, "step": 10372 }, { "epoch": 4.700045310376076, "grad_norm": 0.0734950338227525, "learning_rate": 8.73920589467745e-07, "loss": 0.9358, "step": 10373 }, { "epoch": 4.700498414136837, "grad_norm": 0.06641304306919102, "learning_rate": 8.712919600434521e-07, "loss": 0.9218, "step": 10374 }, { "epoch": 4.7009515178975985, "grad_norm": 0.06655419660407295, "learning_rate": 8.686672463180312e-07, "loss": 0.9144, "step": 10375 }, { "epoch": 4.70140462165836, "grad_norm": 0.07016430019311105, "learning_rate": 8.66046448554152e-07, "loss": 0.9121, "step": 10376 }, { "epoch": 4.701857725419121, "grad_norm": 0.06694865622017555, "learning_rate": 8.634295670140757e-07, "loss": 0.9276, "step": 10377 }, { "epoch": 4.702310829179882, "grad_norm": 0.06618962186312694, "learning_rate": 8.608166019596864e-07, "loss": 0.919, "step": 10378 }, { "epoch": 4.702763932940643, "grad_norm": 0.06667067741089933, "learning_rate": 8.582075536524681e-07, "loss": 0.9136, "step": 10379 }, { "epoch": 4.703217036701405, "grad_norm": 0.0676901591664189, "learning_rate": 8.556024223535097e-07, "loss": 0.929, "step": 10380 }, { "epoch": 4.703670140462166, "grad_norm": 0.06587598976881562, "learning_rate": 8.530012083235139e-07, "loss": 0.9144, "step": 10381 }, { "epoch": 4.704123244222927, "grad_norm": 0.06846776297761978, "learning_rate": 8.504039118227969e-07, "loss": 0.9116, "step": 10382 }, { "epoch": 4.704576347983688, "grad_norm": 0.0677904669590331, "learning_rate": 8.478105331112707e-07, "loss": 0.9197, "step": 10383 }, { "epoch": 4.7050294517444495, "grad_norm": 0.06532150433796262, "learning_rate": 8.452210724484567e-07, "loss": 0.929, "step": 10384 }, { "epoch": 4.705482555505211, "grad_norm": 0.06617547132064164, "learning_rate": 8.426355300934985e-07, "loss": 0.9313, "step": 10385 }, { "epoch": 4.705935659265972, "grad_norm": 0.0718383116374311, "learning_rate": 8.400539063051228e-07, "loss": 0.9417, "step": 10386 }, { "epoch": 4.706388763026733, "grad_norm": 0.07036130744859434, "learning_rate": 8.374762013416959e-07, "loss": 0.918, "step": 10387 }, { "epoch": 4.706841866787494, "grad_norm": 0.06474360951527569, "learning_rate": 8.349024154611584e-07, "loss": 0.9132, "step": 10388 }, { "epoch": 4.707294970548255, "grad_norm": 0.06749645817813074, "learning_rate": 8.323325489210865e-07, "loss": 0.9295, "step": 10389 }, { "epoch": 4.707748074309016, "grad_norm": 0.07063180269415549, "learning_rate": 8.297666019786476e-07, "loss": 0.9063, "step": 10390 }, { "epoch": 4.708201178069778, "grad_norm": 0.06797402043881329, "learning_rate": 8.272045748906188e-07, "loss": 0.9254, "step": 10391 }, { "epoch": 4.7086542818305395, "grad_norm": 0.06968453232196418, "learning_rate": 8.246464679133992e-07, "loss": 0.9186, "step": 10392 }, { "epoch": 4.709107385591301, "grad_norm": 0.06622113332926428, "learning_rate": 8.220922813029708e-07, "loss": 0.9021, "step": 10393 }, { "epoch": 4.709560489352062, "grad_norm": 0.06677167424054788, "learning_rate": 8.195420153149425e-07, "loss": 0.9277, "step": 10394 }, { "epoch": 4.710013593112823, "grad_norm": 0.0649285757434167, "learning_rate": 8.169956702045279e-07, "loss": 0.9157, "step": 10395 }, { "epoch": 4.710466696873584, "grad_norm": 0.07005358284804528, "learning_rate": 8.144532462265409e-07, "loss": 0.9356, "step": 10396 }, { "epoch": 4.710919800634345, "grad_norm": 0.06691389180476047, "learning_rate": 8.119147436354135e-07, "loss": 0.9118, "step": 10397 }, { "epoch": 4.711372904395106, "grad_norm": 0.06785376217861136, "learning_rate": 8.093801626851738e-07, "loss": 0.9236, "step": 10398 }, { "epoch": 4.711826008155867, "grad_norm": 0.06663797327778778, "learning_rate": 8.068495036294677e-07, "loss": 0.9399, "step": 10399 }, { "epoch": 4.712279111916629, "grad_norm": 0.0670307490424974, "learning_rate": 8.043227667215369e-07, "loss": 0.9437, "step": 10400 }, { "epoch": 4.71273221567739, "grad_norm": 0.07354930838027157, "learning_rate": 8.017999522142461e-07, "loss": 0.9131, "step": 10401 }, { "epoch": 4.713185319438152, "grad_norm": 0.07139854009616499, "learning_rate": 7.992810603600554e-07, "loss": 0.9261, "step": 10402 }, { "epoch": 4.713638423198913, "grad_norm": 0.07005082469319947, "learning_rate": 7.967660914110342e-07, "loss": 0.919, "step": 10403 }, { "epoch": 4.714091526959674, "grad_norm": 0.0691184784561677, "learning_rate": 7.942550456188614e-07, "loss": 0.9097, "step": 10404 }, { "epoch": 4.714544630720435, "grad_norm": 0.06758888860469418, "learning_rate": 7.91747923234829e-07, "loss": 0.9142, "step": 10405 }, { "epoch": 4.714997734481196, "grad_norm": 0.06679215015508093, "learning_rate": 7.892447245098211e-07, "loss": 0.9142, "step": 10406 }, { "epoch": 4.715450838241957, "grad_norm": 0.06725359574726802, "learning_rate": 7.867454496943483e-07, "loss": 0.9113, "step": 10407 }, { "epoch": 4.7159039420027185, "grad_norm": 0.06723214050234963, "learning_rate": 7.842500990385171e-07, "loss": 0.9227, "step": 10408 }, { "epoch": 4.71635704576348, "grad_norm": 0.07225939097025298, "learning_rate": 7.817586727920301e-07, "loss": 0.9298, "step": 10409 }, { "epoch": 4.716810149524241, "grad_norm": 0.06728001683455409, "learning_rate": 7.792711712042256e-07, "loss": 0.9186, "step": 10410 }, { "epoch": 4.717263253285003, "grad_norm": 0.07071520262557868, "learning_rate": 7.767875945240244e-07, "loss": 0.9216, "step": 10411 }, { "epoch": 4.717716357045764, "grad_norm": 0.06766744626343511, "learning_rate": 7.7430794299997e-07, "loss": 0.956, "step": 10412 }, { "epoch": 4.718169460806525, "grad_norm": 0.0672159205094876, "learning_rate": 7.718322168802017e-07, "loss": 0.9241, "step": 10413 }, { "epoch": 4.718622564567286, "grad_norm": 0.06564187097744026, "learning_rate": 7.693604164124769e-07, "loss": 0.9409, "step": 10414 }, { "epoch": 4.719075668328047, "grad_norm": 0.07105867584192876, "learning_rate": 7.668925418441442e-07, "loss": 0.9413, "step": 10415 }, { "epoch": 4.719528772088808, "grad_norm": 0.06648391164511844, "learning_rate": 7.644285934221795e-07, "loss": 0.9275, "step": 10416 }, { "epoch": 4.7199818758495695, "grad_norm": 0.07001913058630552, "learning_rate": 7.619685713931501e-07, "loss": 0.9362, "step": 10417 }, { "epoch": 4.720434979610331, "grad_norm": 0.07151032722487091, "learning_rate": 7.595124760032369e-07, "loss": 0.9247, "step": 10418 }, { "epoch": 4.720888083371092, "grad_norm": 0.06463288790826205, "learning_rate": 7.570603074982297e-07, "loss": 0.9175, "step": 10419 }, { "epoch": 4.721341187131853, "grad_norm": 0.06775787434824972, "learning_rate": 7.54612066123519e-07, "loss": 0.93, "step": 10420 }, { "epoch": 4.721794290892614, "grad_norm": 0.06523270852113328, "learning_rate": 7.521677521241088e-07, "loss": 0.9151, "step": 10421 }, { "epoch": 4.722247394653376, "grad_norm": 0.06512185808464238, "learning_rate": 7.497273657446036e-07, "loss": 0.9514, "step": 10422 }, { "epoch": 4.722700498414137, "grad_norm": 0.06465667741660265, "learning_rate": 7.472909072292256e-07, "loss": 0.9163, "step": 10423 }, { "epoch": 4.723153602174898, "grad_norm": 0.0655796041524138, "learning_rate": 7.448583768217932e-07, "loss": 0.9333, "step": 10424 }, { "epoch": 4.7236067059356595, "grad_norm": 0.0663540176849521, "learning_rate": 7.42429774765725e-07, "loss": 0.9158, "step": 10425 }, { "epoch": 4.724059809696421, "grad_norm": 0.0670449467811481, "learning_rate": 7.400051013040709e-07, "loss": 0.9123, "step": 10426 }, { "epoch": 4.724512913457182, "grad_norm": 0.0652754878489241, "learning_rate": 7.375843566794683e-07, "loss": 0.9273, "step": 10427 }, { "epoch": 4.724966017217943, "grad_norm": 0.06881653709089466, "learning_rate": 7.351675411341674e-07, "loss": 0.9358, "step": 10428 }, { "epoch": 4.725419120978704, "grad_norm": 0.06667587547222728, "learning_rate": 7.327546549100195e-07, "loss": 0.9271, "step": 10429 }, { "epoch": 4.725872224739465, "grad_norm": 0.06789009975397858, "learning_rate": 7.30345698248498e-07, "loss": 0.8984, "step": 10430 }, { "epoch": 4.726325328500226, "grad_norm": 0.06710002180528023, "learning_rate": 7.279406713906633e-07, "loss": 0.92, "step": 10431 }, { "epoch": 4.726778432260987, "grad_norm": 0.06732789346732519, "learning_rate": 7.255395745771943e-07, "loss": 0.908, "step": 10432 }, { "epoch": 4.727231536021749, "grad_norm": 0.06661739109302015, "learning_rate": 7.23142408048374e-07, "loss": 0.9176, "step": 10433 }, { "epoch": 4.7276846397825105, "grad_norm": 0.06604051282787678, "learning_rate": 7.207491720440952e-07, "loss": 0.9201, "step": 10434 }, { "epoch": 4.728137743543272, "grad_norm": 0.06779941692745348, "learning_rate": 7.183598668038505e-07, "loss": 0.9189, "step": 10435 }, { "epoch": 4.728590847304033, "grad_norm": 0.06798579035838698, "learning_rate": 7.159744925667511e-07, "loss": 0.9311, "step": 10436 }, { "epoch": 4.729043951064794, "grad_norm": 0.06643413471608638, "learning_rate": 7.135930495714948e-07, "loss": 0.9132, "step": 10437 }, { "epoch": 4.729497054825555, "grad_norm": 0.06759101114237105, "learning_rate": 7.112155380564068e-07, "loss": 0.9171, "step": 10438 }, { "epoch": 4.729950158586316, "grad_norm": 0.0657081435072299, "learning_rate": 7.088419582594075e-07, "loss": 0.9169, "step": 10439 }, { "epoch": 4.730403262347077, "grad_norm": 0.0650914220571887, "learning_rate": 7.064723104180227e-07, "loss": 0.9162, "step": 10440 }, { "epoch": 4.730856366107838, "grad_norm": 0.07049155241203393, "learning_rate": 7.041065947693959e-07, "loss": 0.9514, "step": 10441 }, { "epoch": 4.7313094698686, "grad_norm": 0.06998043387990338, "learning_rate": 7.017448115502623e-07, "loss": 0.9412, "step": 10442 }, { "epoch": 4.731762573629361, "grad_norm": 0.06771783995154508, "learning_rate": 6.99386960996975e-07, "loss": 0.9362, "step": 10443 }, { "epoch": 4.732215677390123, "grad_norm": 0.06761832529535346, "learning_rate": 6.970330433454919e-07, "loss": 0.9246, "step": 10444 }, { "epoch": 4.732668781150884, "grad_norm": 0.06510180078198208, "learning_rate": 6.946830588313714e-07, "loss": 0.9303, "step": 10445 }, { "epoch": 4.733121884911645, "grad_norm": 0.06480146349689218, "learning_rate": 6.923370076897807e-07, "loss": 0.91, "step": 10446 }, { "epoch": 4.733574988672406, "grad_norm": 0.06610759932508023, "learning_rate": 6.899948901554965e-07, "loss": 0.9151, "step": 10447 }, { "epoch": 4.734028092433167, "grad_norm": 0.06967585031132524, "learning_rate": 6.876567064629003e-07, "loss": 0.9432, "step": 10448 }, { "epoch": 4.734481196193928, "grad_norm": 0.07138843933411476, "learning_rate": 6.853224568459738e-07, "loss": 0.9395, "step": 10449 }, { "epoch": 4.7349342999546895, "grad_norm": 0.06996045273151955, "learning_rate": 6.829921415383211e-07, "loss": 0.9537, "step": 10450 }, { "epoch": 4.735387403715451, "grad_norm": 0.06534625895308374, "learning_rate": 6.806657607731382e-07, "loss": 0.9301, "step": 10451 }, { "epoch": 4.735840507476212, "grad_norm": 0.06784741723282241, "learning_rate": 6.783433147832252e-07, "loss": 0.9089, "step": 10452 }, { "epoch": 4.736293611236973, "grad_norm": 0.06889969235169864, "learning_rate": 6.760248038010009e-07, "loss": 0.9269, "step": 10453 }, { "epoch": 4.736746714997734, "grad_norm": 0.06636727925051626, "learning_rate": 6.737102280584795e-07, "loss": 0.9279, "step": 10454 }, { "epoch": 4.737199818758496, "grad_norm": 0.06759573666382385, "learning_rate": 6.713995877872892e-07, "loss": 0.903, "step": 10455 }, { "epoch": 4.737652922519257, "grad_norm": 0.0672188757322776, "learning_rate": 6.690928832186627e-07, "loss": 0.9015, "step": 10456 }, { "epoch": 4.738106026280018, "grad_norm": 0.06477437924162183, "learning_rate": 6.667901145834333e-07, "loss": 0.9232, "step": 10457 }, { "epoch": 4.738559130040779, "grad_norm": 0.06651291713177931, "learning_rate": 6.644912821120431e-07, "loss": 0.9065, "step": 10458 }, { "epoch": 4.739012233801541, "grad_norm": 0.06638309619970661, "learning_rate": 6.621963860345482e-07, "loss": 0.9314, "step": 10459 }, { "epoch": 4.739465337562302, "grad_norm": 0.06399755994587557, "learning_rate": 6.599054265806004e-07, "loss": 0.9152, "step": 10460 }, { "epoch": 4.739918441323063, "grad_norm": 0.06678281771871614, "learning_rate": 6.576184039794564e-07, "loss": 0.9305, "step": 10461 }, { "epoch": 4.740371545083824, "grad_norm": 0.06681597488095199, "learning_rate": 6.553353184599909e-07, "loss": 0.9284, "step": 10462 }, { "epoch": 4.740824648844585, "grad_norm": 0.06460897037787167, "learning_rate": 6.530561702506744e-07, "loss": 0.9102, "step": 10463 }, { "epoch": 4.741277752605347, "grad_norm": 0.06445320848248791, "learning_rate": 6.507809595795867e-07, "loss": 0.9094, "step": 10464 }, { "epoch": 4.741730856366107, "grad_norm": 0.06471335297216661, "learning_rate": 6.485096866744123e-07, "loss": 0.9254, "step": 10465 }, { "epoch": 4.742183960126869, "grad_norm": 0.06344689962619697, "learning_rate": 6.462423517624495e-07, "loss": 0.9309, "step": 10466 }, { "epoch": 4.7426370638876305, "grad_norm": 0.06517858145631149, "learning_rate": 6.439789550705833e-07, "loss": 0.9312, "step": 10467 }, { "epoch": 4.743090167648392, "grad_norm": 0.06641029209957747, "learning_rate": 6.41719496825326e-07, "loss": 0.9135, "step": 10468 }, { "epoch": 4.743543271409153, "grad_norm": 0.06504395464046249, "learning_rate": 6.394639772527855e-07, "loss": 0.9185, "step": 10469 }, { "epoch": 4.743996375169914, "grad_norm": 0.06958620019742331, "learning_rate": 6.372123965786703e-07, "loss": 0.9261, "step": 10470 }, { "epoch": 4.744449478930675, "grad_norm": 0.06885822203253013, "learning_rate": 6.349647550283111e-07, "loss": 0.9165, "step": 10471 }, { "epoch": 4.744902582691436, "grad_norm": 0.07923352231607134, "learning_rate": 6.327210528266303e-07, "loss": 0.932, "step": 10472 }, { "epoch": 4.745355686452197, "grad_norm": 0.06257159142309497, "learning_rate": 6.304812901981594e-07, "loss": 0.9071, "step": 10473 }, { "epoch": 4.745808790212958, "grad_norm": 0.06760510611536882, "learning_rate": 6.282454673670347e-07, "loss": 0.9459, "step": 10474 }, { "epoch": 4.7462618939737204, "grad_norm": 0.06636176513658468, "learning_rate": 6.260135845570059e-07, "loss": 0.9414, "step": 10475 }, { "epoch": 4.746714997734482, "grad_norm": 0.0638305797356701, "learning_rate": 6.23785641991419e-07, "loss": 0.9279, "step": 10476 }, { "epoch": 4.747168101495243, "grad_norm": 0.06572777639913373, "learning_rate": 6.215616398932245e-07, "loss": 0.9162, "step": 10477 }, { "epoch": 4.747621205256004, "grad_norm": 0.06449465419164756, "learning_rate": 6.193415784849909e-07, "loss": 0.9483, "step": 10478 }, { "epoch": 4.748074309016765, "grad_norm": 0.06783123185488384, "learning_rate": 6.17125457988883e-07, "loss": 0.93, "step": 10479 }, { "epoch": 4.748527412777526, "grad_norm": 0.06649318034488935, "learning_rate": 6.149132786266743e-07, "loss": 0.9399, "step": 10480 }, { "epoch": 4.748980516538287, "grad_norm": 0.06656453554786299, "learning_rate": 6.127050406197343e-07, "loss": 0.9194, "step": 10481 }, { "epoch": 4.749433620299048, "grad_norm": 0.06566272890266611, "learning_rate": 6.105007441890598e-07, "loss": 0.9174, "step": 10482 }, { "epoch": 4.7498867240598095, "grad_norm": 0.06545397551748051, "learning_rate": 6.083003895552253e-07, "loss": 0.9058, "step": 10483 }, { "epoch": 4.750339827820571, "grad_norm": 0.06489703096573365, "learning_rate": 6.061039769384369e-07, "loss": 0.9222, "step": 10484 }, { "epoch": 4.750792931581332, "grad_norm": 0.06693626116507306, "learning_rate": 6.039115065584922e-07, "loss": 0.9214, "step": 10485 }, { "epoch": 4.751246035342094, "grad_norm": 0.06932605927161238, "learning_rate": 6.017229786347889e-07, "loss": 0.9506, "step": 10486 }, { "epoch": 4.751699139102855, "grad_norm": 0.06741722327895144, "learning_rate": 5.995383933863475e-07, "loss": 0.9268, "step": 10487 }, { "epoch": 4.752152242863616, "grad_norm": 0.06614011448116254, "learning_rate": 5.973577510317797e-07, "loss": 0.9313, "step": 10488 }, { "epoch": 4.752605346624377, "grad_norm": 0.06482405525035113, "learning_rate": 5.951810517893108e-07, "loss": 0.9036, "step": 10489 }, { "epoch": 4.753058450385138, "grad_norm": 0.06496654460817664, "learning_rate": 5.930082958767624e-07, "loss": 0.9213, "step": 10490 }, { "epoch": 4.753511554145899, "grad_norm": 0.06440313040730417, "learning_rate": 5.908394835115738e-07, "loss": 0.9323, "step": 10491 }, { "epoch": 4.753964657906661, "grad_norm": 0.06650605558361433, "learning_rate": 5.886746149107758e-07, "loss": 0.9057, "step": 10492 }, { "epoch": 4.754417761667422, "grad_norm": 0.07016596628504729, "learning_rate": 5.865136902910218e-07, "loss": 0.9217, "step": 10493 }, { "epoch": 4.754870865428183, "grad_norm": 0.06535843084291161, "learning_rate": 5.843567098685476e-07, "loss": 0.9169, "step": 10494 }, { "epoch": 4.755323969188944, "grad_norm": 0.0685259383705515, "learning_rate": 5.822036738592162e-07, "loss": 0.9487, "step": 10495 }, { "epoch": 4.755777072949705, "grad_norm": 0.06638313794748445, "learning_rate": 5.80054582478482e-07, "loss": 0.9254, "step": 10496 }, { "epoch": 4.756230176710467, "grad_norm": 0.06545446599861295, "learning_rate": 5.779094359414128e-07, "loss": 0.9465, "step": 10497 }, { "epoch": 4.756683280471228, "grad_norm": 0.06877432686468882, "learning_rate": 5.757682344626814e-07, "loss": 0.8986, "step": 10498 }, { "epoch": 4.757136384231989, "grad_norm": 0.06630306819187144, "learning_rate": 5.736309782565519e-07, "loss": 0.9437, "step": 10499 }, { "epoch": 4.7575894879927505, "grad_norm": 0.06670147351616833, "learning_rate": 5.714976675369155e-07, "loss": 0.9092, "step": 10500 }, { "epoch": 4.758042591753512, "grad_norm": 0.0700050182774302, "learning_rate": 5.693683025172503e-07, "loss": 0.9192, "step": 10501 }, { "epoch": 4.758495695514273, "grad_norm": 0.06696018354074465, "learning_rate": 5.672428834106524e-07, "loss": 0.9223, "step": 10502 }, { "epoch": 4.758948799275034, "grad_norm": 0.06674810665097378, "learning_rate": 5.651214104298098e-07, "loss": 0.9285, "step": 10503 }, { "epoch": 4.759401903035795, "grad_norm": 0.06666045727663532, "learning_rate": 5.630038837870322e-07, "loss": 0.9274, "step": 10504 }, { "epoch": 4.759855006796556, "grad_norm": 0.06870124131344353, "learning_rate": 5.608903036942215e-07, "loss": 0.9257, "step": 10505 }, { "epoch": 4.760308110557317, "grad_norm": 0.0663932919831683, "learning_rate": 5.58780670362884e-07, "loss": 0.9401, "step": 10506 }, { "epoch": 4.760761214318078, "grad_norm": 0.06745632228578467, "learning_rate": 5.566749840041397e-07, "loss": 0.9408, "step": 10507 }, { "epoch": 4.76121431807884, "grad_norm": 0.06437441519485447, "learning_rate": 5.545732448287089e-07, "loss": 0.9158, "step": 10508 }, { "epoch": 4.761667421839602, "grad_norm": 0.07379849902449445, "learning_rate": 5.524754530469212e-07, "loss": 0.9556, "step": 10509 }, { "epoch": 4.762120525600363, "grad_norm": 0.06717702709719756, "learning_rate": 5.503816088686976e-07, "loss": 0.9384, "step": 10510 }, { "epoch": 4.762573629361124, "grad_norm": 0.06637668112387958, "learning_rate": 5.482917125035858e-07, "loss": 0.91, "step": 10511 }, { "epoch": 4.763026733121885, "grad_norm": 0.063589345730409, "learning_rate": 5.462057641607166e-07, "loss": 0.9208, "step": 10512 }, { "epoch": 4.763479836882646, "grad_norm": 0.06408608612250148, "learning_rate": 5.441237640488428e-07, "loss": 0.9123, "step": 10513 }, { "epoch": 4.763932940643407, "grad_norm": 0.06597112360356819, "learning_rate": 5.42045712376309e-07, "loss": 0.9186, "step": 10514 }, { "epoch": 4.764386044404168, "grad_norm": 0.06901464889837623, "learning_rate": 5.399716093510732e-07, "loss": 0.9189, "step": 10515 }, { "epoch": 4.7648391481649295, "grad_norm": 0.06922665843400269, "learning_rate": 5.379014551806982e-07, "loss": 0.903, "step": 10516 }, { "epoch": 4.765292251925691, "grad_norm": 0.06878287029515173, "learning_rate": 5.35835250072343e-07, "loss": 0.9317, "step": 10517 }, { "epoch": 4.765745355686452, "grad_norm": 0.06508644789627636, "learning_rate": 5.337729942327841e-07, "loss": 0.9038, "step": 10518 }, { "epoch": 4.766198459447214, "grad_norm": 0.06612678434021524, "learning_rate": 5.317146878683899e-07, "loss": 0.919, "step": 10519 }, { "epoch": 4.766651563207975, "grad_norm": 0.06630465070097576, "learning_rate": 5.296603311851467e-07, "loss": 0.9349, "step": 10520 }, { "epoch": 4.767104666968736, "grad_norm": 0.06383556836248869, "learning_rate": 5.276099243886368e-07, "loss": 0.923, "step": 10521 }, { "epoch": 4.767557770729497, "grad_norm": 0.06499115449959314, "learning_rate": 5.255634676840427e-07, "loss": 0.9347, "step": 10522 }, { "epoch": 4.768010874490258, "grad_norm": 0.06608497823868292, "learning_rate": 5.235209612761649e-07, "loss": 0.9028, "step": 10523 }, { "epoch": 4.768463978251019, "grad_norm": 0.06311147942087195, "learning_rate": 5.214824053693956e-07, "loss": 0.9178, "step": 10524 }, { "epoch": 4.7689170820117806, "grad_norm": 0.06844675180925798, "learning_rate": 5.194478001677449e-07, "loss": 0.9189, "step": 10525 }, { "epoch": 4.769370185772542, "grad_norm": 0.06231232919288694, "learning_rate": 5.174171458748101e-07, "loss": 0.9098, "step": 10526 }, { "epoch": 4.769823289533303, "grad_norm": 0.06401008896596383, "learning_rate": 5.153904426938194e-07, "loss": 0.9151, "step": 10527 }, { "epoch": 4.770276393294065, "grad_norm": 0.06500633450894126, "learning_rate": 5.133676908275754e-07, "loss": 0.9113, "step": 10528 }, { "epoch": 4.770729497054825, "grad_norm": 0.06450389410065994, "learning_rate": 5.113488904785069e-07, "loss": 0.9213, "step": 10529 }, { "epoch": 4.771182600815587, "grad_norm": 0.06468381699920629, "learning_rate": 5.093340418486348e-07, "loss": 0.944, "step": 10530 }, { "epoch": 4.771635704576348, "grad_norm": 0.11238472458982102, "learning_rate": 5.073231451395932e-07, "loss": 0.9241, "step": 10531 }, { "epoch": 4.772088808337109, "grad_norm": 0.06316728129794234, "learning_rate": 5.053162005526168e-07, "loss": 0.9016, "step": 10532 }, { "epoch": 4.7725419120978705, "grad_norm": 0.0658758635431507, "learning_rate": 5.033132082885406e-07, "loss": 0.9401, "step": 10533 }, { "epoch": 4.772995015858632, "grad_norm": 0.06921064603165178, "learning_rate": 5.013141685478173e-07, "loss": 0.9304, "step": 10534 }, { "epoch": 4.773448119619393, "grad_norm": 0.07117989728387349, "learning_rate": 4.993190815304871e-07, "loss": 0.93, "step": 10535 }, { "epoch": 4.773901223380154, "grad_norm": 0.06368926821267802, "learning_rate": 4.97327947436208e-07, "loss": 0.9343, "step": 10536 }, { "epoch": 4.774354327140915, "grad_norm": 0.0643849669798699, "learning_rate": 4.95340766464234e-07, "loss": 0.924, "step": 10537 }, { "epoch": 4.774807430901676, "grad_norm": 0.06395981500444241, "learning_rate": 4.933575388134282e-07, "loss": 0.9183, "step": 10538 }, { "epoch": 4.775260534662438, "grad_norm": 0.06321220598876649, "learning_rate": 4.913782646822585e-07, "loss": 0.9209, "step": 10539 }, { "epoch": 4.775713638423199, "grad_norm": 0.06797740200136831, "learning_rate": 4.894029442687931e-07, "loss": 0.9103, "step": 10540 }, { "epoch": 4.77616674218396, "grad_norm": 0.0665484933929014, "learning_rate": 4.874315777707095e-07, "loss": 0.9131, "step": 10541 }, { "epoch": 4.7766198459447216, "grad_norm": 0.0669827468661552, "learning_rate": 4.85464165385281e-07, "loss": 0.9302, "step": 10542 }, { "epoch": 4.777072949705483, "grad_norm": 0.06212684793487109, "learning_rate": 4.835007073093945e-07, "loss": 0.9002, "step": 10543 }, { "epoch": 4.777526053466244, "grad_norm": 0.06504846801856681, "learning_rate": 4.815412037395373e-07, "loss": 0.9243, "step": 10544 }, { "epoch": 4.777979157227005, "grad_norm": 0.0661532461534477, "learning_rate": 4.795856548718059e-07, "loss": 0.9343, "step": 10545 }, { "epoch": 4.778432260987766, "grad_norm": 0.06694369336960512, "learning_rate": 4.776340609018926e-07, "loss": 0.9342, "step": 10546 }, { "epoch": 4.778885364748527, "grad_norm": 0.06818255086553054, "learning_rate": 4.7568642202509895e-07, "loss": 0.9471, "step": 10547 }, { "epoch": 4.779338468509288, "grad_norm": 0.06633309443545388, "learning_rate": 4.737427384363269e-07, "loss": 0.9323, "step": 10548 }, { "epoch": 4.7797915722700495, "grad_norm": 0.06682649376888533, "learning_rate": 4.71803010330083e-07, "loss": 0.9406, "step": 10549 }, { "epoch": 4.7802446760308115, "grad_norm": 0.06788790217300926, "learning_rate": 4.6986723790049205e-07, "loss": 0.9227, "step": 10550 }, { "epoch": 4.780697779791573, "grad_norm": 0.0648051301351012, "learning_rate": 4.679354213412568e-07, "loss": 0.9276, "step": 10551 }, { "epoch": 4.781150883552334, "grad_norm": 0.06831433239234497, "learning_rate": 4.6600756084571154e-07, "loss": 0.9295, "step": 10552 }, { "epoch": 4.781603987313095, "grad_norm": 0.06587869979101128, "learning_rate": 4.640836566067686e-07, "loss": 0.9199, "step": 10553 }, { "epoch": 4.782057091073856, "grad_norm": 0.06529248166763559, "learning_rate": 4.6216370881696726e-07, "loss": 0.9287, "step": 10554 }, { "epoch": 4.782510194834617, "grad_norm": 0.06320132400724482, "learning_rate": 4.6024771766843837e-07, "loss": 0.9181, "step": 10555 }, { "epoch": 4.782963298595378, "grad_norm": 0.06419486877782223, "learning_rate": 4.5833568335291736e-07, "loss": 0.9269, "step": 10556 }, { "epoch": 4.783416402356139, "grad_norm": 0.067194100089928, "learning_rate": 4.56427606061749e-07, "loss": 0.9015, "step": 10557 }, { "epoch": 4.7838695061169005, "grad_norm": 0.06686540150159852, "learning_rate": 4.5452348598587823e-07, "loss": 0.9261, "step": 10558 }, { "epoch": 4.784322609877662, "grad_norm": 0.06710175508902338, "learning_rate": 4.5262332331585056e-07, "loss": 0.8954, "step": 10559 }, { "epoch": 4.784775713638423, "grad_norm": 0.06312731668069772, "learning_rate": 4.5072711824182493e-07, "loss": 0.8897, "step": 10560 }, { "epoch": 4.785228817399185, "grad_norm": 0.06283812952584403, "learning_rate": 4.488348709535606e-07, "loss": 0.9362, "step": 10561 }, { "epoch": 4.785681921159946, "grad_norm": 0.06560119507178713, "learning_rate": 4.4694658164040837e-07, "loss": 0.9192, "step": 10562 }, { "epoch": 4.786135024920707, "grad_norm": 0.0644876329420095, "learning_rate": 4.4506225049134597e-07, "loss": 0.9101, "step": 10563 }, { "epoch": 4.786588128681468, "grad_norm": 0.06510142140921123, "learning_rate": 4.431818776949337e-07, "loss": 0.9279, "step": 10564 }, { "epoch": 4.787041232442229, "grad_norm": 0.06660813646746164, "learning_rate": 4.4130546343935433e-07, "loss": 0.9329, "step": 10565 }, { "epoch": 4.7874943362029905, "grad_norm": 0.06543365362626818, "learning_rate": 4.394330079123732e-07, "loss": 0.8964, "step": 10566 }, { "epoch": 4.787947439963752, "grad_norm": 0.07128748580957545, "learning_rate": 4.375645113013782e-07, "loss": 0.924, "step": 10567 }, { "epoch": 4.788400543724513, "grad_norm": 0.06276633201854281, "learning_rate": 4.356999737933576e-07, "loss": 0.9175, "step": 10568 }, { "epoch": 4.788853647485274, "grad_norm": 0.06271466515292168, "learning_rate": 4.338393955748865e-07, "loss": 0.9198, "step": 10569 }, { "epoch": 4.789306751246035, "grad_norm": 0.06513327074153273, "learning_rate": 4.319827768321716e-07, "loss": 0.9195, "step": 10570 }, { "epoch": 4.789759855006796, "grad_norm": 0.061795110162998615, "learning_rate": 4.30130117751002e-07, "loss": 0.9084, "step": 10571 }, { "epoch": 4.790212958767558, "grad_norm": 0.0688094499275257, "learning_rate": 4.282814185167805e-07, "loss": 0.9445, "step": 10572 }, { "epoch": 4.790666062528319, "grad_norm": 0.06395700965851642, "learning_rate": 4.264366793145103e-07, "loss": 0.9177, "step": 10573 }, { "epoch": 4.79111916628908, "grad_norm": 0.06712989131556953, "learning_rate": 4.245959003287947e-07, "loss": 0.9268, "step": 10574 }, { "epoch": 4.7915722700498415, "grad_norm": 0.06315213922309312, "learning_rate": 4.227590817438465e-07, "loss": 0.9396, "step": 10575 }, { "epoch": 4.792025373810603, "grad_norm": 0.06235594678786397, "learning_rate": 4.2092622374347857e-07, "loss": 0.9141, "step": 10576 }, { "epoch": 4.792478477571364, "grad_norm": 0.06620517242786084, "learning_rate": 4.1909732651111755e-07, "loss": 0.93, "step": 10577 }, { "epoch": 4.792931581332125, "grad_norm": 0.06836601995569236, "learning_rate": 4.172723902297771e-07, "loss": 0.9207, "step": 10578 }, { "epoch": 4.793384685092886, "grad_norm": 0.06361597885837891, "learning_rate": 4.154514150820843e-07, "loss": 0.9065, "step": 10579 }, { "epoch": 4.793837788853647, "grad_norm": 0.06424095881600438, "learning_rate": 4.136344012502669e-07, "loss": 0.9203, "step": 10580 }, { "epoch": 4.794290892614408, "grad_norm": 0.06559233934409943, "learning_rate": 4.11821348916166e-07, "loss": 0.9317, "step": 10581 }, { "epoch": 4.7947439963751695, "grad_norm": 0.06857115612989265, "learning_rate": 4.100122582612098e-07, "loss": 0.942, "step": 10582 }, { "epoch": 4.7951971001359315, "grad_norm": 0.068089880248352, "learning_rate": 4.0820712946643583e-07, "loss": 0.9329, "step": 10583 }, { "epoch": 4.795650203896693, "grad_norm": 0.06467119205068678, "learning_rate": 4.06405962712495e-07, "loss": 0.9062, "step": 10584 }, { "epoch": 4.796103307657454, "grad_norm": 0.06456416430055693, "learning_rate": 4.046087581796254e-07, "loss": 0.9442, "step": 10585 }, { "epoch": 4.796556411418215, "grad_norm": 0.06341609905287578, "learning_rate": 4.028155160476921e-07, "loss": 0.9158, "step": 10586 }, { "epoch": 4.797009515178976, "grad_norm": 0.06491043821415328, "learning_rate": 4.0102623649612926e-07, "loss": 0.9141, "step": 10587 }, { "epoch": 4.797462618939737, "grad_norm": 0.06634736754439213, "learning_rate": 3.9924091970401145e-07, "loss": 0.9122, "step": 10588 }, { "epoch": 4.797915722700498, "grad_norm": 0.06314092223591218, "learning_rate": 3.9745956584999134e-07, "loss": 0.9222, "step": 10589 }, { "epoch": 4.798368826461259, "grad_norm": 0.062076100304809284, "learning_rate": 3.956821751123352e-07, "loss": 0.9064, "step": 10590 }, { "epoch": 4.7988219302220205, "grad_norm": 0.06365876091928363, "learning_rate": 3.9390874766890517e-07, "loss": 0.9281, "step": 10591 }, { "epoch": 4.7992750339827825, "grad_norm": 0.06418267017712813, "learning_rate": 3.921392836971771e-07, "loss": 0.9136, "step": 10592 }, { "epoch": 4.799728137743543, "grad_norm": 0.06670408735817503, "learning_rate": 3.903737833742316e-07, "loss": 0.943, "step": 10593 }, { "epoch": 4.800181241504305, "grad_norm": 0.06527918804840079, "learning_rate": 3.886122468767317e-07, "loss": 0.9516, "step": 10594 }, { "epoch": 4.800634345265066, "grad_norm": 0.06575074615941733, "learning_rate": 3.868546743809676e-07, "loss": 0.9013, "step": 10595 }, { "epoch": 4.801087449025827, "grad_norm": 0.06548208192715842, "learning_rate": 3.851010660628207e-07, "loss": 0.9065, "step": 10596 }, { "epoch": 4.801540552786588, "grad_norm": 0.06245417574749343, "learning_rate": 3.833514220977863e-07, "loss": 0.9174, "step": 10597 }, { "epoch": 4.801993656547349, "grad_norm": 0.0657629594232806, "learning_rate": 3.8160574266094653e-07, "loss": 0.9188, "step": 10598 }, { "epoch": 4.8024467603081105, "grad_norm": 0.0634473726894749, "learning_rate": 3.798640279269927e-07, "loss": 0.9229, "step": 10599 }, { "epoch": 4.802899864068872, "grad_norm": 0.06426379627712914, "learning_rate": 3.7812627807022995e-07, "loss": 0.9258, "step": 10600 }, { "epoch": 4.803352967829633, "grad_norm": 0.06351632428001636, "learning_rate": 3.763924932645591e-07, "loss": 0.9288, "step": 10601 }, { "epoch": 4.803806071590394, "grad_norm": 0.0658026988799371, "learning_rate": 3.7466267368347687e-07, "loss": 0.9351, "step": 10602 }, { "epoch": 4.804259175351156, "grad_norm": 0.06538943384639569, "learning_rate": 3.7293681950009376e-07, "loss": 0.9292, "step": 10603 }, { "epoch": 4.804712279111917, "grad_norm": 0.06171014427214293, "learning_rate": 3.712149308871249e-07, "loss": 0.911, "step": 10604 }, { "epoch": 4.805165382872678, "grad_norm": 0.06430802573255381, "learning_rate": 3.6949700801687694e-07, "loss": 0.9381, "step": 10605 }, { "epoch": 4.805618486633439, "grad_norm": 0.06248322526432875, "learning_rate": 3.677830510612701e-07, "loss": 0.9218, "step": 10606 }, { "epoch": 4.8060715903942, "grad_norm": 0.06593206814968948, "learning_rate": 3.66073060191825e-07, "loss": 0.9472, "step": 10607 }, { "epoch": 4.8065246941549615, "grad_norm": 0.06657627585058605, "learning_rate": 3.64367035579658e-07, "loss": 0.9242, "step": 10608 }, { "epoch": 4.806977797915723, "grad_norm": 0.06485713615287696, "learning_rate": 3.626649773954993e-07, "loss": 0.901, "step": 10609 }, { "epoch": 4.807430901676484, "grad_norm": 0.06559471916839818, "learning_rate": 3.609668858096749e-07, "loss": 0.9281, "step": 10610 }, { "epoch": 4.807884005437245, "grad_norm": 0.06374089561846356, "learning_rate": 3.5927276099212426e-07, "loss": 0.9192, "step": 10611 }, { "epoch": 4.808337109198006, "grad_norm": 0.06352450681570466, "learning_rate": 3.575826031123697e-07, "loss": 0.9144, "step": 10612 }, { "epoch": 4.808790212958767, "grad_norm": 0.06784799996865878, "learning_rate": 3.558964123395647e-07, "loss": 0.9141, "step": 10613 }, { "epoch": 4.809243316719529, "grad_norm": 0.069944728663095, "learning_rate": 3.542141888424322e-07, "loss": 0.9101, "step": 10614 }, { "epoch": 4.80969642048029, "grad_norm": 0.0654522018728329, "learning_rate": 3.525359327893352e-07, "loss": 0.9268, "step": 10615 }, { "epoch": 4.8101495242410515, "grad_norm": 0.06446505022745319, "learning_rate": 3.50861644348206e-07, "loss": 0.9282, "step": 10616 }, { "epoch": 4.810602628001813, "grad_norm": 0.06715739425451843, "learning_rate": 3.4919132368659956e-07, "loss": 0.9389, "step": 10617 }, { "epoch": 4.811055731762574, "grad_norm": 0.0655259734141923, "learning_rate": 3.475249709716755e-07, "loss": 0.9357, "step": 10618 }, { "epoch": 4.811508835523335, "grad_norm": 0.0661224206475727, "learning_rate": 3.458625863701759e-07, "loss": 0.9372, "step": 10619 }, { "epoch": 4.811961939284096, "grad_norm": 0.06255074813820793, "learning_rate": 3.4420417004846994e-07, "loss": 0.9021, "step": 10620 }, { "epoch": 4.812415043044857, "grad_norm": 0.0636583542863492, "learning_rate": 3.425497221725094e-07, "loss": 0.9201, "step": 10621 }, { "epoch": 4.812868146805618, "grad_norm": 0.06533417519035298, "learning_rate": 3.408992429078728e-07, "loss": 0.9377, "step": 10622 }, { "epoch": 4.813321250566379, "grad_norm": 0.0643174586119579, "learning_rate": 3.392527324197126e-07, "loss": 0.9181, "step": 10623 }, { "epoch": 4.8137743543271405, "grad_norm": 0.06492919784118337, "learning_rate": 3.376101908728124e-07, "loss": 0.9273, "step": 10624 }, { "epoch": 4.8142274580879025, "grad_norm": 0.0644329655934364, "learning_rate": 3.3597161843152983e-07, "loss": 0.9212, "step": 10625 }, { "epoch": 4.814680561848664, "grad_norm": 0.06331642107434685, "learning_rate": 3.3433701525985794e-07, "loss": 0.9437, "step": 10626 }, { "epoch": 4.815133665609425, "grad_norm": 0.06443051520650293, "learning_rate": 3.327063815213594e-07, "loss": 0.9138, "step": 10627 }, { "epoch": 4.815586769370186, "grad_norm": 0.06557970987776165, "learning_rate": 3.3107971737922793e-07, "loss": 0.9172, "step": 10628 }, { "epoch": 4.816039873130947, "grad_norm": 0.0633754440088966, "learning_rate": 3.294570229962357e-07, "loss": 0.9228, "step": 10629 }, { "epoch": 4.816492976891708, "grad_norm": 0.06362422689045408, "learning_rate": 3.278382985347817e-07, "loss": 0.8853, "step": 10630 }, { "epoch": 4.816946080652469, "grad_norm": 0.06271530813452952, "learning_rate": 3.262235441568473e-07, "loss": 0.9324, "step": 10631 }, { "epoch": 4.8173991844132305, "grad_norm": 0.06203239805152381, "learning_rate": 3.2461276002402344e-07, "loss": 0.9213, "step": 10632 }, { "epoch": 4.817852288173992, "grad_norm": 0.06407494789351749, "learning_rate": 3.2300594629750994e-07, "loss": 0.9345, "step": 10633 }, { "epoch": 4.818305391934753, "grad_norm": 0.06348055356723027, "learning_rate": 3.214031031381026e-07, "loss": 0.9202, "step": 10634 }, { "epoch": 4.818758495695514, "grad_norm": 0.06345090400908707, "learning_rate": 3.1980423070620216e-07, "loss": 0.9148, "step": 10635 }, { "epoch": 4.819211599456276, "grad_norm": 0.06544366469178592, "learning_rate": 3.1820932916180933e-07, "loss": 0.9198, "step": 10636 }, { "epoch": 4.819664703217037, "grad_norm": 0.06626225100464225, "learning_rate": 3.1661839866453435e-07, "loss": 0.9044, "step": 10637 }, { "epoch": 4.820117806977798, "grad_norm": 0.0673355313094945, "learning_rate": 3.150314393735787e-07, "loss": 0.9114, "step": 10638 }, { "epoch": 4.820570910738559, "grad_norm": 0.06322823948011655, "learning_rate": 3.134484514477576e-07, "loss": 0.9213, "step": 10639 }, { "epoch": 4.82102401449932, "grad_norm": 0.06405268821672845, "learning_rate": 3.118694350454821e-07, "loss": 0.9138, "step": 10640 }, { "epoch": 4.8214771182600815, "grad_norm": 0.06785088432037728, "learning_rate": 3.1029439032476794e-07, "loss": 0.9292, "step": 10641 }, { "epoch": 4.821930222020843, "grad_norm": 0.06433793996368771, "learning_rate": 3.087233174432358e-07, "loss": 0.8983, "step": 10642 }, { "epoch": 4.822383325781604, "grad_norm": 0.06608371269952064, "learning_rate": 3.071562165581021e-07, "loss": 0.9103, "step": 10643 }, { "epoch": 4.822836429542365, "grad_norm": 0.06325817864869537, "learning_rate": 3.05593087826197e-07, "loss": 0.9217, "step": 10644 }, { "epoch": 4.823289533303126, "grad_norm": 0.06547979891876514, "learning_rate": 3.04033931403942e-07, "loss": 0.9152, "step": 10645 }, { "epoch": 4.823742637063887, "grad_norm": 0.06396450314334994, "learning_rate": 3.0247874744735897e-07, "loss": 0.8995, "step": 10646 }, { "epoch": 4.824195740824649, "grad_norm": 0.062035203712774926, "learning_rate": 3.0092753611209225e-07, "loss": 0.9134, "step": 10647 }, { "epoch": 4.82464884458541, "grad_norm": 0.06449100538348744, "learning_rate": 2.993802975533644e-07, "loss": 0.9106, "step": 10648 }, { "epoch": 4.8251019483461715, "grad_norm": 0.06549145829203103, "learning_rate": 2.978370319260204e-07, "loss": 0.9389, "step": 10649 }, { "epoch": 4.825555052106933, "grad_norm": 0.06331977248708794, "learning_rate": 2.9629773938449237e-07, "loss": 0.9244, "step": 10650 }, { "epoch": 4.826008155867694, "grad_norm": 0.06262983084465631, "learning_rate": 2.9476242008281697e-07, "loss": 0.9099, "step": 10651 }, { "epoch": 4.826461259628455, "grad_norm": 0.06661729273102558, "learning_rate": 2.932310741746447e-07, "loss": 0.9189, "step": 10652 }, { "epoch": 4.826914363389216, "grad_norm": 0.06401828076967854, "learning_rate": 2.9170370181321296e-07, "loss": 0.9262, "step": 10653 }, { "epoch": 4.827367467149977, "grad_norm": 0.06533314019754617, "learning_rate": 2.901803031513817e-07, "loss": 0.8998, "step": 10654 }, { "epoch": 4.827820570910738, "grad_norm": 0.06244198383001009, "learning_rate": 2.886608783415845e-07, "loss": 0.9347, "step": 10655 }, { "epoch": 4.8282736746715, "grad_norm": 0.06493346719857926, "learning_rate": 2.871454275358909e-07, "loss": 0.95, "step": 10656 }, { "epoch": 4.8287267784322605, "grad_norm": 0.06603529907288863, "learning_rate": 2.8563395088594403e-07, "loss": 0.9159, "step": 10657 }, { "epoch": 4.8291798821930225, "grad_norm": 0.06509587999288363, "learning_rate": 2.8412644854300066e-07, "loss": 0.9135, "step": 10658 }, { "epoch": 4.829632985953784, "grad_norm": 0.06480927180495173, "learning_rate": 2.8262292065792673e-07, "loss": 0.9413, "step": 10659 }, { "epoch": 4.830086089714545, "grad_norm": 0.06435737489367, "learning_rate": 2.8112336738117976e-07, "loss": 0.9382, "step": 10660 }, { "epoch": 4.830539193475306, "grad_norm": 0.06529501831458476, "learning_rate": 2.7962778886282624e-07, "loss": 0.9629, "step": 10661 }, { "epoch": 4.830992297236067, "grad_norm": 0.0625087274287506, "learning_rate": 2.781361852525244e-07, "loss": 0.9051, "step": 10662 }, { "epoch": 4.831445400996828, "grad_norm": 0.06680350181638785, "learning_rate": 2.7664855669955025e-07, "loss": 0.9175, "step": 10663 }, { "epoch": 4.831898504757589, "grad_norm": 0.06446715626274786, "learning_rate": 2.7516490335277147e-07, "loss": 0.9254, "step": 10664 }, { "epoch": 4.83235160851835, "grad_norm": 0.06552290873594567, "learning_rate": 2.736852253606648e-07, "loss": 0.9283, "step": 10665 }, { "epoch": 4.832804712279112, "grad_norm": 0.08030020565703995, "learning_rate": 2.7220952287129397e-07, "loss": 0.9204, "step": 10666 }, { "epoch": 4.833257816039874, "grad_norm": 0.06346949321198954, "learning_rate": 2.707377960323454e-07, "loss": 0.9069, "step": 10667 }, { "epoch": 4.833710919800635, "grad_norm": 0.06832277897947889, "learning_rate": 2.6927004499109676e-07, "loss": 0.954, "step": 10668 }, { "epoch": 4.834164023561396, "grad_norm": 0.06344779922567391, "learning_rate": 2.678062698944306e-07, "loss": 0.9009, "step": 10669 }, { "epoch": 4.834617127322157, "grad_norm": 0.06361441332402423, "learning_rate": 2.6634647088882524e-07, "loss": 0.9358, "step": 10670 }, { "epoch": 4.835070231082918, "grad_norm": 0.06432136619860496, "learning_rate": 2.648906481203639e-07, "loss": 0.9215, "step": 10671 }, { "epoch": 4.835523334843679, "grad_norm": 0.06412659551453773, "learning_rate": 2.634388017347478e-07, "loss": 0.9407, "step": 10672 }, { "epoch": 4.83597643860444, "grad_norm": 0.06691027526297343, "learning_rate": 2.619909318772473e-07, "loss": 0.9363, "step": 10673 }, { "epoch": 4.8364295423652015, "grad_norm": 0.06302888602560017, "learning_rate": 2.605470386927733e-07, "loss": 0.9077, "step": 10674 }, { "epoch": 4.836882646125963, "grad_norm": 0.06376891926278772, "learning_rate": 2.591071223258057e-07, "loss": 0.9305, "step": 10675 }, { "epoch": 4.837335749886724, "grad_norm": 0.06587359200826365, "learning_rate": 2.57671182920447e-07, "loss": 0.9211, "step": 10676 }, { "epoch": 4.837788853647485, "grad_norm": 0.06404207559434735, "learning_rate": 2.562392206203912e-07, "loss": 0.9304, "step": 10677 }, { "epoch": 4.838241957408247, "grad_norm": 0.06360839068199159, "learning_rate": 2.5481123556894136e-07, "loss": 0.9085, "step": 10678 }, { "epoch": 4.838695061169008, "grad_norm": 0.06383047628807702, "learning_rate": 2.533872279089966e-07, "loss": 0.9149, "step": 10679 }, { "epoch": 4.839148164929769, "grad_norm": 0.06773119577013131, "learning_rate": 2.519671977830607e-07, "loss": 0.9167, "step": 10680 }, { "epoch": 4.83960126869053, "grad_norm": 0.0644423016778694, "learning_rate": 2.505511453332421e-07, "loss": 0.919, "step": 10681 }, { "epoch": 4.840054372451291, "grad_norm": 0.06379834953465462, "learning_rate": 2.491390707012409e-07, "loss": 0.9306, "step": 10682 }, { "epoch": 4.840507476212053, "grad_norm": 0.062065580967466366, "learning_rate": 2.477309740283795e-07, "loss": 0.9157, "step": 10683 }, { "epoch": 4.840960579972814, "grad_norm": 0.0643967745127361, "learning_rate": 2.4632685545555865e-07, "loss": 0.9361, "step": 10684 }, { "epoch": 4.841413683733575, "grad_norm": 0.06521384344058513, "learning_rate": 2.449267151232926e-07, "loss": 0.9278, "step": 10685 }, { "epoch": 4.841866787494336, "grad_norm": 0.06527305858091897, "learning_rate": 2.4353055317170025e-07, "loss": 0.9048, "step": 10686 }, { "epoch": 4.842319891255097, "grad_norm": 0.06674251591884577, "learning_rate": 2.4213836974050107e-07, "loss": 0.9169, "step": 10687 }, { "epoch": 4.842772995015858, "grad_norm": 0.06421483636831583, "learning_rate": 2.407501649690103e-07, "loss": 0.9318, "step": 10688 }, { "epoch": 4.84322609877662, "grad_norm": 0.0643998980053158, "learning_rate": 2.3936593899614334e-07, "loss": 0.944, "step": 10689 }, { "epoch": 4.843679202537381, "grad_norm": 0.06508450439798064, "learning_rate": 2.3798569196043397e-07, "loss": 0.9193, "step": 10690 }, { "epoch": 4.8441323062981425, "grad_norm": 0.06420906964502548, "learning_rate": 2.366094239999983e-07, "loss": 0.9374, "step": 10691 }, { "epoch": 4.844585410058904, "grad_norm": 0.06324571924382079, "learning_rate": 2.3523713525256618e-07, "loss": 0.9216, "step": 10692 }, { "epoch": 4.845038513819665, "grad_norm": 0.06321703363910781, "learning_rate": 2.3386882585546332e-07, "loss": 0.9011, "step": 10693 }, { "epoch": 4.845491617580426, "grad_norm": 0.06603383919073694, "learning_rate": 2.325044959456202e-07, "loss": 0.9221, "step": 10694 }, { "epoch": 4.845944721341187, "grad_norm": 0.06321656638932918, "learning_rate": 2.31144145659572e-07, "loss": 0.926, "step": 10695 }, { "epoch": 4.846397825101948, "grad_norm": 0.06202994057674223, "learning_rate": 2.2978777513344985e-07, "loss": 0.9009, "step": 10696 }, { "epoch": 4.846850928862709, "grad_norm": 0.06415938551224466, "learning_rate": 2.284353845029852e-07, "loss": 0.8977, "step": 10697 }, { "epoch": 4.84730403262347, "grad_norm": 0.06381505560339798, "learning_rate": 2.2708697390351864e-07, "loss": 0.9267, "step": 10698 }, { "epoch": 4.847757136384232, "grad_norm": 0.06327641870573843, "learning_rate": 2.2574254346999113e-07, "loss": 0.9194, "step": 10699 }, { "epoch": 4.848210240144994, "grad_norm": 0.06417875069017025, "learning_rate": 2.244020933369351e-07, "loss": 0.9059, "step": 10700 }, { "epoch": 4.848663343905755, "grad_norm": 0.06484192917899607, "learning_rate": 2.2306562363850093e-07, "loss": 0.9376, "step": 10701 }, { "epoch": 4.849116447666516, "grad_norm": 0.07423429530768434, "learning_rate": 2.2173313450842615e-07, "loss": 0.9193, "step": 10702 }, { "epoch": 4.849569551427277, "grad_norm": 0.06309583467617881, "learning_rate": 2.2040462608006186e-07, "loss": 0.9233, "step": 10703 }, { "epoch": 4.850022655188038, "grad_norm": 0.06621699378020103, "learning_rate": 2.1908009848634616e-07, "loss": 0.9337, "step": 10704 }, { "epoch": 4.850475758948799, "grad_norm": 0.0644032297884178, "learning_rate": 2.177595518598352e-07, "loss": 0.9185, "step": 10705 }, { "epoch": 4.85092886270956, "grad_norm": 0.06714631939092609, "learning_rate": 2.164429863326767e-07, "loss": 0.9341, "step": 10706 }, { "epoch": 4.8513819664703215, "grad_norm": 0.06342120869817867, "learning_rate": 2.1513040203662295e-07, "loss": 0.9526, "step": 10707 }, { "epoch": 4.851835070231083, "grad_norm": 0.06396789406536553, "learning_rate": 2.138217991030267e-07, "loss": 0.9292, "step": 10708 }, { "epoch": 4.852288173991844, "grad_norm": 0.06346483194356006, "learning_rate": 2.1251717766283653e-07, "loss": 0.9408, "step": 10709 }, { "epoch": 4.852741277752605, "grad_norm": 0.06384154039503334, "learning_rate": 2.1121653784662356e-07, "loss": 0.9556, "step": 10710 }, { "epoch": 4.853194381513367, "grad_norm": 0.06446006150616015, "learning_rate": 2.0991987978453253e-07, "loss": 0.9169, "step": 10711 }, { "epoch": 4.853647485274128, "grad_norm": 0.06366926276522249, "learning_rate": 2.0862720360632638e-07, "loss": 0.9171, "step": 10712 }, { "epoch": 4.854100589034889, "grad_norm": 0.06556968874259493, "learning_rate": 2.073385094413727e-07, "loss": 0.9186, "step": 10713 }, { "epoch": 4.85455369279565, "grad_norm": 0.06508610260092926, "learning_rate": 2.0605379741862608e-07, "loss": 0.9124, "step": 10714 }, { "epoch": 4.855006796556411, "grad_norm": 0.06322513212319167, "learning_rate": 2.0477306766665038e-07, "loss": 0.9121, "step": 10715 }, { "epoch": 4.855459900317173, "grad_norm": 0.06578226622702271, "learning_rate": 2.0349632031361867e-07, "loss": 0.9333, "step": 10716 }, { "epoch": 4.855913004077934, "grad_norm": 0.06425483892226437, "learning_rate": 2.0222355548729088e-07, "loss": 0.944, "step": 10717 }, { "epoch": 4.856366107838695, "grad_norm": 0.06625903531638289, "learning_rate": 2.0095477331503631e-07, "loss": 0.934, "step": 10718 }, { "epoch": 4.856819211599456, "grad_norm": 0.06425763634146468, "learning_rate": 1.9968997392382894e-07, "loss": 0.9423, "step": 10719 }, { "epoch": 4.857272315360218, "grad_norm": 0.06542872421033209, "learning_rate": 1.984291574402386e-07, "loss": 0.9163, "step": 10720 }, { "epoch": 4.857725419120978, "grad_norm": 0.06439469863455814, "learning_rate": 1.9717232399043551e-07, "loss": 0.9145, "step": 10721 }, { "epoch": 4.85817852288174, "grad_norm": 0.06318014669140558, "learning_rate": 1.959194737001946e-07, "loss": 0.9284, "step": 10722 }, { "epoch": 4.858631626642501, "grad_norm": 0.06544259498581338, "learning_rate": 1.9467060669489557e-07, "loss": 0.9418, "step": 10723 }, { "epoch": 4.8590847304032625, "grad_norm": 0.06347535929888574, "learning_rate": 1.9342572309950957e-07, "loss": 0.9186, "step": 10724 }, { "epoch": 4.859537834164024, "grad_norm": 0.06785554793314376, "learning_rate": 1.9218482303861696e-07, "loss": 0.9338, "step": 10725 }, { "epoch": 4.859990937924785, "grad_norm": 0.06555329951994293, "learning_rate": 1.909479066363984e-07, "loss": 0.9301, "step": 10726 }, { "epoch": 4.860444041685546, "grad_norm": 0.06078750658042468, "learning_rate": 1.8971497401663487e-07, "loss": 0.9196, "step": 10727 }, { "epoch": 4.860897145446307, "grad_norm": 0.06254819193213637, "learning_rate": 1.8848602530270765e-07, "loss": 0.9004, "step": 10728 }, { "epoch": 4.861350249207068, "grad_norm": 0.06238349812316565, "learning_rate": 1.8726106061760285e-07, "loss": 0.9262, "step": 10729 }, { "epoch": 4.861803352967829, "grad_norm": 0.06249670007582331, "learning_rate": 1.860400800839024e-07, "loss": 0.9243, "step": 10730 }, { "epoch": 4.862256456728591, "grad_norm": 0.06450734703154899, "learning_rate": 1.8482308382379743e-07, "loss": 0.9068, "step": 10731 }, { "epoch": 4.862709560489352, "grad_norm": 0.06400325927108025, "learning_rate": 1.836100719590661e-07, "loss": 0.9312, "step": 10732 }, { "epoch": 4.863162664250114, "grad_norm": 0.06505110752965369, "learning_rate": 1.8240104461110907e-07, "loss": 0.9336, "step": 10733 }, { "epoch": 4.863615768010875, "grad_norm": 0.06683559775493658, "learning_rate": 1.8119600190090515e-07, "loss": 0.9209, "step": 10734 }, { "epoch": 4.864068871771636, "grad_norm": 0.06765616691645916, "learning_rate": 1.799949439490556e-07, "loss": 0.9295, "step": 10735 }, { "epoch": 4.864521975532397, "grad_norm": 0.06423582645610783, "learning_rate": 1.7879787087574428e-07, "loss": 0.915, "step": 10736 }, { "epoch": 4.864975079293158, "grad_norm": 0.06381072416460916, "learning_rate": 1.776047828007732e-07, "loss": 0.9265, "step": 10737 }, { "epoch": 4.865428183053919, "grad_norm": 0.06493410436769022, "learning_rate": 1.7641567984353124e-07, "loss": 0.9365, "step": 10738 }, { "epoch": 4.86588128681468, "grad_norm": 0.06452224868072709, "learning_rate": 1.7523056212302104e-07, "loss": 0.9243, "step": 10739 }, { "epoch": 4.8663343905754415, "grad_norm": 0.061233419945346555, "learning_rate": 1.740494297578321e-07, "loss": 0.9005, "step": 10740 }, { "epoch": 4.866787494336203, "grad_norm": 0.06261417953198112, "learning_rate": 1.7287228286616774e-07, "loss": 0.9204, "step": 10741 }, { "epoch": 4.867240598096965, "grad_norm": 0.06299684335546887, "learning_rate": 1.7169912156582703e-07, "loss": 0.9175, "step": 10742 }, { "epoch": 4.867693701857726, "grad_norm": 0.06526031600867825, "learning_rate": 1.7052994597420935e-07, "loss": 0.9303, "step": 10743 }, { "epoch": 4.868146805618487, "grad_norm": 0.06519674097264895, "learning_rate": 1.6936475620831894e-07, "loss": 0.9213, "step": 10744 }, { "epoch": 4.868599909379248, "grad_norm": 0.0643098753102219, "learning_rate": 1.6820355238475583e-07, "loss": 0.9257, "step": 10745 }, { "epoch": 4.869053013140009, "grad_norm": 0.06493265199387666, "learning_rate": 1.6704633461973374e-07, "loss": 0.9177, "step": 10746 }, { "epoch": 4.86950611690077, "grad_norm": 0.06425796288119189, "learning_rate": 1.658931030290445e-07, "loss": 0.9128, "step": 10747 }, { "epoch": 4.869959220661531, "grad_norm": 0.06244178683839735, "learning_rate": 1.6474385772810242e-07, "loss": 0.9202, "step": 10748 }, { "epoch": 4.8704123244222925, "grad_norm": 0.06392479834449018, "learning_rate": 1.6359859883191776e-07, "loss": 0.9213, "step": 10749 }, { "epoch": 4.870865428183054, "grad_norm": 0.0625818922160902, "learning_rate": 1.6245732645509217e-07, "loss": 0.9061, "step": 10750 }, { "epoch": 4.871318531943815, "grad_norm": 0.06334756414292539, "learning_rate": 1.613200407118365e-07, "loss": 0.9229, "step": 10751 }, { "epoch": 4.871771635704576, "grad_norm": 0.06348975375767621, "learning_rate": 1.6018674171596637e-07, "loss": 0.9306, "step": 10752 }, { "epoch": 4.872224739465338, "grad_norm": 0.06352017582400248, "learning_rate": 1.5905742958088888e-07, "loss": 0.9252, "step": 10753 }, { "epoch": 4.872677843226099, "grad_norm": 0.06345778411513504, "learning_rate": 1.5793210441962025e-07, "loss": 0.9306, "step": 10754 }, { "epoch": 4.87313094698686, "grad_norm": 0.06153528605997023, "learning_rate": 1.5681076634477265e-07, "loss": 0.9101, "step": 10755 }, { "epoch": 4.873584050747621, "grad_norm": 0.06466130416961675, "learning_rate": 1.5569341546855855e-07, "loss": 0.9155, "step": 10756 }, { "epoch": 4.8740371545083825, "grad_norm": 0.06507324481959115, "learning_rate": 1.5458005190279956e-07, "loss": 0.9326, "step": 10757 }, { "epoch": 4.874490258269144, "grad_norm": 0.06345799840371058, "learning_rate": 1.5347067575890883e-07, "loss": 0.9176, "step": 10758 }, { "epoch": 4.874943362029905, "grad_norm": 0.06228008144625528, "learning_rate": 1.5236528714790422e-07, "loss": 0.9026, "step": 10759 }, { "epoch": 4.875396465790666, "grad_norm": 0.06443406344929173, "learning_rate": 1.512638861804039e-07, "loss": 0.9082, "step": 10760 }, { "epoch": 4.875849569551427, "grad_norm": 0.06587491624270422, "learning_rate": 1.5016647296662634e-07, "loss": 0.9147, "step": 10761 }, { "epoch": 4.876302673312188, "grad_norm": 0.06341381425916344, "learning_rate": 1.4907304761639928e-07, "loss": 0.9473, "step": 10762 }, { "epoch": 4.876755777072949, "grad_norm": 0.06153910053971111, "learning_rate": 1.47983610239133e-07, "loss": 0.9048, "step": 10763 }, { "epoch": 4.877208880833711, "grad_norm": 0.06336720143529805, "learning_rate": 1.4689816094386466e-07, "loss": 0.9308, "step": 10764 }, { "epoch": 4.877661984594472, "grad_norm": 0.061092725184436564, "learning_rate": 1.458166998392052e-07, "loss": 0.9258, "step": 10765 }, { "epoch": 4.8781150883552336, "grad_norm": 0.06402581653151547, "learning_rate": 1.4473922703337917e-07, "loss": 0.9339, "step": 10766 }, { "epoch": 4.878568192115995, "grad_norm": 0.06338463614548222, "learning_rate": 1.436657426342203e-07, "loss": 0.9117, "step": 10767 }, { "epoch": 4.879021295876756, "grad_norm": 0.06387076473128724, "learning_rate": 1.425962467491493e-07, "loss": 0.9128, "step": 10768 }, { "epoch": 4.879474399637517, "grad_norm": 0.06455203133175802, "learning_rate": 1.415307394851917e-07, "loss": 0.9308, "step": 10769 }, { "epoch": 4.879927503398278, "grad_norm": 0.0650274425347725, "learning_rate": 1.4046922094897775e-07, "loss": 0.9124, "step": 10770 }, { "epoch": 4.880380607159039, "grad_norm": 0.062007176779941026, "learning_rate": 1.3941169124673805e-07, "loss": 0.9149, "step": 10771 }, { "epoch": 4.8808337109198, "grad_norm": 0.06208768770162914, "learning_rate": 1.3835815048429902e-07, "loss": 0.9217, "step": 10772 }, { "epoch": 4.8812868146805615, "grad_norm": 0.06234250844971596, "learning_rate": 1.3730859876708746e-07, "loss": 0.9243, "step": 10773 }, { "epoch": 4.881739918441323, "grad_norm": 0.06494877603669517, "learning_rate": 1.3626303620013936e-07, "loss": 0.9331, "step": 10774 }, { "epoch": 4.882193022202085, "grad_norm": 0.06540392530805496, "learning_rate": 1.3522146288808657e-07, "loss": 0.9367, "step": 10775 }, { "epoch": 4.882646125962846, "grad_norm": 0.06380730029429878, "learning_rate": 1.3418387893516126e-07, "loss": 0.9182, "step": 10776 }, { "epoch": 4.883099229723607, "grad_norm": 0.06283419166134273, "learning_rate": 1.3315028444519596e-07, "loss": 0.9149, "step": 10777 }, { "epoch": 4.883552333484368, "grad_norm": 0.06503176686078274, "learning_rate": 1.3212067952162345e-07, "loss": 0.937, "step": 10778 }, { "epoch": 4.884005437245129, "grad_norm": 0.06328759278069314, "learning_rate": 1.310950642674813e-07, "loss": 0.921, "step": 10779 }, { "epoch": 4.88445854100589, "grad_norm": 0.06563119798659435, "learning_rate": 1.3007343878540302e-07, "loss": 0.9173, "step": 10780 }, { "epoch": 4.884911644766651, "grad_norm": 0.06150752665800771, "learning_rate": 1.290558031776268e-07, "loss": 0.9221, "step": 10781 }, { "epoch": 4.8853647485274125, "grad_norm": 0.06475278030801593, "learning_rate": 1.2804215754599115e-07, "loss": 0.9496, "step": 10782 }, { "epoch": 4.885817852288174, "grad_norm": 0.06417015873354272, "learning_rate": 1.2703250199192607e-07, "loss": 0.8991, "step": 10783 }, { "epoch": 4.886270956048936, "grad_norm": 0.06262726310270648, "learning_rate": 1.2602683661647964e-07, "loss": 0.9157, "step": 10784 }, { "epoch": 4.886724059809696, "grad_norm": 0.0647410561131551, "learning_rate": 1.2502516152028686e-07, "loss": 0.9106, "step": 10785 }, { "epoch": 4.887177163570458, "grad_norm": 0.06313952439064664, "learning_rate": 1.2402747680358763e-07, "loss": 0.9157, "step": 10786 }, { "epoch": 4.887630267331219, "grad_norm": 0.06365268726800531, "learning_rate": 1.23033782566222e-07, "loss": 0.9378, "step": 10787 }, { "epoch": 4.88808337109198, "grad_norm": 0.0612262178751123, "learning_rate": 1.2204407890763047e-07, "loss": 0.906, "step": 10788 }, { "epoch": 4.888536474852741, "grad_norm": 0.06417947159669292, "learning_rate": 1.210583659268627e-07, "loss": 0.9346, "step": 10789 }, { "epoch": 4.8889895786135025, "grad_norm": 0.06504819984871539, "learning_rate": 1.2007664372254647e-07, "loss": 0.9166, "step": 10790 }, { "epoch": 4.889442682374264, "grad_norm": 0.06340152604106428, "learning_rate": 1.1909891239294091e-07, "loss": 0.9123, "step": 10791 }, { "epoch": 4.889895786135025, "grad_norm": 0.06206427134446877, "learning_rate": 1.1812517203587892e-07, "loss": 0.9237, "step": 10792 }, { "epoch": 4.890348889895786, "grad_norm": 0.06199249547848501, "learning_rate": 1.1715542274880698e-07, "loss": 0.9211, "step": 10793 }, { "epoch": 4.890801993656547, "grad_norm": 0.06342136502014843, "learning_rate": 1.1618966462877635e-07, "loss": 0.902, "step": 10794 }, { "epoch": 4.891255097417309, "grad_norm": 0.06514223806668681, "learning_rate": 1.1522789777242082e-07, "loss": 0.9268, "step": 10795 }, { "epoch": 4.89170820117807, "grad_norm": 0.06237306120220695, "learning_rate": 1.1427012227600121e-07, "loss": 0.9192, "step": 10796 }, { "epoch": 4.892161304938831, "grad_norm": 0.06297160452552422, "learning_rate": 1.1331633823535193e-07, "loss": 0.9333, "step": 10797 }, { "epoch": 4.892614408699592, "grad_norm": 0.06393869537861939, "learning_rate": 1.1236654574592554e-07, "loss": 0.9152, "step": 10798 }, { "epoch": 4.8930675124603535, "grad_norm": 0.06274600825237284, "learning_rate": 1.1142074490277044e-07, "loss": 0.9122, "step": 10799 }, { "epoch": 4.893520616221115, "grad_norm": 0.06600931184483107, "learning_rate": 1.1047893580053537e-07, "loss": 0.9388, "step": 10800 }, { "epoch": 4.893973719981876, "grad_norm": 0.06529765085047151, "learning_rate": 1.0954111853346494e-07, "loss": 0.9279, "step": 10801 }, { "epoch": 4.894426823742637, "grad_norm": 0.06316305226163813, "learning_rate": 1.0860729319541297e-07, "loss": 0.9266, "step": 10802 }, { "epoch": 4.894879927503398, "grad_norm": 0.06287549845766628, "learning_rate": 1.0767745987982914e-07, "loss": 0.9461, "step": 10803 }, { "epoch": 4.895333031264159, "grad_norm": 0.06418893317523786, "learning_rate": 1.0675161867976347e-07, "loss": 0.9293, "step": 10804 }, { "epoch": 4.89578613502492, "grad_norm": 0.06454375410587218, "learning_rate": 1.058297696878663e-07, "loss": 0.916, "step": 10805 }, { "epoch": 4.896239238785682, "grad_norm": 0.06428782747594516, "learning_rate": 1.0491191299638381e-07, "loss": 0.9283, "step": 10806 }, { "epoch": 4.8966923425464435, "grad_norm": 0.06305715935388552, "learning_rate": 1.0399804869718033e-07, "loss": 0.9133, "step": 10807 }, { "epoch": 4.897145446307205, "grad_norm": 0.06333912272307754, "learning_rate": 1.030881768817027e-07, "loss": 0.9084, "step": 10808 }, { "epoch": 4.897598550067966, "grad_norm": 0.06462971004872918, "learning_rate": 1.0218229764099807e-07, "loss": 0.9379, "step": 10809 }, { "epoch": 4.898051653828727, "grad_norm": 0.0644376321707515, "learning_rate": 1.0128041106572728e-07, "loss": 0.8915, "step": 10810 }, { "epoch": 4.898504757589488, "grad_norm": 0.061798383881884224, "learning_rate": 1.0038251724614256e-07, "loss": 0.9198, "step": 10811 }, { "epoch": 4.898957861350249, "grad_norm": 0.06234974454938404, "learning_rate": 9.948861627209649e-08, "loss": 0.912, "step": 10812 }, { "epoch": 4.89941096511101, "grad_norm": 0.0623178392933292, "learning_rate": 9.859870823304196e-08, "loss": 0.9094, "step": 10813 }, { "epoch": 4.899864068871771, "grad_norm": 0.0621426783640164, "learning_rate": 9.771279321803662e-08, "loss": 0.908, "step": 10814 }, { "epoch": 4.9003171726325325, "grad_norm": 0.06256102453041999, "learning_rate": 9.683087131573842e-08, "loss": 0.9164, "step": 10815 }, { "epoch": 4.900770276393294, "grad_norm": 0.06794662359079338, "learning_rate": 9.595294261440125e-08, "loss": 0.9296, "step": 10816 }, { "epoch": 4.901223380154056, "grad_norm": 0.06370405796097106, "learning_rate": 9.507900720187924e-08, "loss": 0.9227, "step": 10817 }, { "epoch": 4.901676483914817, "grad_norm": 0.06621290291479677, "learning_rate": 9.420906516563133e-08, "loss": 0.926, "step": 10818 }, { "epoch": 4.902129587675578, "grad_norm": 0.0618517489812681, "learning_rate": 9.334311659271677e-08, "loss": 0.9209, "step": 10819 }, { "epoch": 4.902582691436339, "grad_norm": 0.06422340322461768, "learning_rate": 9.248116156978626e-08, "loss": 0.9155, "step": 10820 }, { "epoch": 4.9030357951971, "grad_norm": 0.06320256274776874, "learning_rate": 9.162320018310411e-08, "loss": 0.9231, "step": 10821 }, { "epoch": 4.903488898957861, "grad_norm": 0.06517590190422654, "learning_rate": 9.076923251852165e-08, "loss": 0.9141, "step": 10822 }, { "epoch": 4.9039420027186225, "grad_norm": 0.06323403489186281, "learning_rate": 8.99192586615083e-08, "loss": 0.9004, "step": 10823 }, { "epoch": 4.904395106479384, "grad_norm": 0.06178162980200867, "learning_rate": 8.907327869711158e-08, "loss": 0.9195, "step": 10824 }, { "epoch": 4.904848210240145, "grad_norm": 0.06361807303835323, "learning_rate": 8.823129270999709e-08, "loss": 0.9122, "step": 10825 }, { "epoch": 4.905301314000906, "grad_norm": 0.06607323696249351, "learning_rate": 8.739330078442632e-08, "loss": 0.8958, "step": 10826 }, { "epoch": 4.905754417761667, "grad_norm": 0.06220785592256636, "learning_rate": 8.65593030042522e-08, "loss": 0.931, "step": 10827 }, { "epoch": 4.906207521522429, "grad_norm": 0.06349324570500182, "learning_rate": 8.572929945293684e-08, "loss": 0.9496, "step": 10828 }, { "epoch": 4.90666062528319, "grad_norm": 0.06421762190254325, "learning_rate": 8.49032902135427e-08, "loss": 0.922, "step": 10829 }, { "epoch": 4.907113729043951, "grad_norm": 0.06577274506051908, "learning_rate": 8.408127536872812e-08, "loss": 0.9291, "step": 10830 }, { "epoch": 4.907566832804712, "grad_norm": 0.06558903903588979, "learning_rate": 8.326325500075616e-08, "loss": 0.9078, "step": 10831 }, { "epoch": 4.9080199365654735, "grad_norm": 0.06372129901813813, "learning_rate": 8.24492291914858e-08, "loss": 0.9239, "step": 10832 }, { "epoch": 4.908473040326235, "grad_norm": 0.062113695763217354, "learning_rate": 8.163919802238519e-08, "loss": 0.9018, "step": 10833 }, { "epoch": 4.908926144086996, "grad_norm": 0.06469111507965171, "learning_rate": 8.083316157450505e-08, "loss": 0.9356, "step": 10834 }, { "epoch": 4.909379247847757, "grad_norm": 0.06319856831360252, "learning_rate": 8.00311199285142e-08, "loss": 0.9126, "step": 10835 }, { "epoch": 4.909832351608518, "grad_norm": 0.06419521112740585, "learning_rate": 7.923307316467289e-08, "loss": 0.937, "step": 10836 }, { "epoch": 4.910285455369279, "grad_norm": 0.06501960734326781, "learning_rate": 7.843902136284609e-08, "loss": 0.9269, "step": 10837 }, { "epoch": 4.91073855913004, "grad_norm": 0.06505277861348295, "learning_rate": 7.764896460249027e-08, "loss": 0.9311, "step": 10838 }, { "epoch": 4.911191662890802, "grad_norm": 0.06251069256896424, "learning_rate": 7.686290296267552e-08, "loss": 0.9328, "step": 10839 }, { "epoch": 4.9116447666515635, "grad_norm": 0.06337156826687244, "learning_rate": 7.608083652205889e-08, "loss": 0.922, "step": 10840 }, { "epoch": 4.912097870412325, "grad_norm": 0.06625337202622045, "learning_rate": 7.530276535890668e-08, "loss": 0.9252, "step": 10841 }, { "epoch": 4.912550974173086, "grad_norm": 0.06409942324423443, "learning_rate": 7.452868955108105e-08, "loss": 0.9303, "step": 10842 }, { "epoch": 4.913004077933847, "grad_norm": 0.062118553029921864, "learning_rate": 7.375860917604893e-08, "loss": 0.9463, "step": 10843 }, { "epoch": 4.913457181694608, "grad_norm": 0.06561245126202532, "learning_rate": 7.299252431086868e-08, "loss": 0.9176, "step": 10844 }, { "epoch": 4.913910285455369, "grad_norm": 0.06411894350076194, "learning_rate": 7.223043503220784e-08, "loss": 0.9187, "step": 10845 }, { "epoch": 4.91436338921613, "grad_norm": 0.06319262713065325, "learning_rate": 7.147234141632986e-08, "loss": 0.913, "step": 10846 }, { "epoch": 4.914816492976891, "grad_norm": 0.0641577355911203, "learning_rate": 7.07182435390985e-08, "loss": 0.9471, "step": 10847 }, { "epoch": 4.915269596737653, "grad_norm": 0.0627133128196284, "learning_rate": 6.996814147597786e-08, "loss": 0.9301, "step": 10848 }, { "epoch": 4.915722700498414, "grad_norm": 0.06246220949026179, "learning_rate": 6.922203530202787e-08, "loss": 0.9354, "step": 10849 }, { "epoch": 4.916175804259176, "grad_norm": 0.06535003781249438, "learning_rate": 6.847992509192214e-08, "loss": 0.9235, "step": 10850 }, { "epoch": 4.916628908019937, "grad_norm": 0.06290408095871293, "learning_rate": 6.774181091992127e-08, "loss": 0.9271, "step": 10851 }, { "epoch": 4.917082011780698, "grad_norm": 0.06211003098817982, "learning_rate": 6.700769285988618e-08, "loss": 0.9081, "step": 10852 }, { "epoch": 4.917535115541459, "grad_norm": 0.06312417020291677, "learning_rate": 6.627757098528697e-08, "loss": 0.9221, "step": 10853 }, { "epoch": 4.91798821930222, "grad_norm": 0.06173791722301111, "learning_rate": 6.555144536918523e-08, "loss": 0.9161, "step": 10854 }, { "epoch": 4.918441323062981, "grad_norm": 0.0644242639728836, "learning_rate": 6.482931608425169e-08, "loss": 0.9428, "step": 10855 }, { "epoch": 4.9188944268237424, "grad_norm": 0.06408766251272627, "learning_rate": 6.411118320274413e-08, "loss": 0.9259, "step": 10856 }, { "epoch": 4.919347530584504, "grad_norm": 0.06226977182313163, "learning_rate": 6.339704679652947e-08, "loss": 0.9229, "step": 10857 }, { "epoch": 4.919800634345265, "grad_norm": 0.06285197773624239, "learning_rate": 6.268690693707503e-08, "loss": 0.9252, "step": 10858 }, { "epoch": 4.920253738106027, "grad_norm": 0.06521593378248607, "learning_rate": 6.198076369544836e-08, "loss": 0.9242, "step": 10859 }, { "epoch": 4.920706841866788, "grad_norm": 0.06305747773338531, "learning_rate": 6.127861714230854e-08, "loss": 0.9159, "step": 10860 }, { "epoch": 4.921159945627549, "grad_norm": 0.06236198768901989, "learning_rate": 6.058046734792822e-08, "loss": 0.9223, "step": 10861 }, { "epoch": 4.92161304938831, "grad_norm": 0.061569715752625424, "learning_rate": 5.988631438216708e-08, "loss": 0.9227, "step": 10862 }, { "epoch": 4.922066153149071, "grad_norm": 0.06529693501707878, "learning_rate": 5.919615831448955e-08, "loss": 0.9356, "step": 10863 }, { "epoch": 4.922519256909832, "grad_norm": 0.06368804387409353, "learning_rate": 5.850999921396483e-08, "loss": 0.9158, "step": 10864 }, { "epoch": 4.9229723606705935, "grad_norm": 0.06359159859504054, "learning_rate": 5.782783714925799e-08, "loss": 0.9102, "step": 10865 }, { "epoch": 4.923425464431355, "grad_norm": 0.06502353114868493, "learning_rate": 5.714967218863887e-08, "loss": 0.926, "step": 10866 }, { "epoch": 4.923878568192116, "grad_norm": 0.06305536465782992, "learning_rate": 5.6475504399964295e-08, "loss": 0.9218, "step": 10867 }, { "epoch": 4.924331671952877, "grad_norm": 0.06433905937133501, "learning_rate": 5.580533385070475e-08, "loss": 0.9364, "step": 10868 }, { "epoch": 4.924784775713638, "grad_norm": 0.06332005922354343, "learning_rate": 5.5139160607922124e-08, "loss": 0.927, "step": 10869 }, { "epoch": 4.9252378794744, "grad_norm": 0.06391168123303828, "learning_rate": 5.4476984738287556e-08, "loss": 0.9257, "step": 10870 }, { "epoch": 4.925690983235161, "grad_norm": 0.06231718249827397, "learning_rate": 5.381880630805914e-08, "loss": 0.9208, "step": 10871 }, { "epoch": 4.926144086995922, "grad_norm": 0.06361822571700214, "learning_rate": 5.3164625383113064e-08, "loss": 0.9021, "step": 10872 }, { "epoch": 4.9265971907566835, "grad_norm": 0.06225831277747331, "learning_rate": 5.2514442028903654e-08, "loss": 0.9245, "step": 10873 }, { "epoch": 4.927050294517445, "grad_norm": 0.06314336411012512, "learning_rate": 5.1868256310503294e-08, "loss": 0.908, "step": 10874 }, { "epoch": 4.927503398278206, "grad_norm": 0.06338586174820801, "learning_rate": 5.1226068292571374e-08, "loss": 0.9249, "step": 10875 }, { "epoch": 4.927956502038967, "grad_norm": 0.061706375294984614, "learning_rate": 5.058787803938092e-08, "loss": 0.9042, "step": 10876 }, { "epoch": 4.928409605799728, "grad_norm": 0.06376744619112859, "learning_rate": 4.9953685614787525e-08, "loss": 0.9028, "step": 10877 }, { "epoch": 4.928862709560489, "grad_norm": 0.06256570455367548, "learning_rate": 4.9323491082264864e-08, "loss": 0.9081, "step": 10878 }, { "epoch": 4.92931581332125, "grad_norm": 0.06471021781140712, "learning_rate": 4.8697294504873595e-08, "loss": 0.9069, "step": 10879 }, { "epoch": 4.929768917082011, "grad_norm": 0.06316820744364847, "learning_rate": 4.8075095945279147e-08, "loss": 0.9136, "step": 10880 }, { "epoch": 4.930222020842773, "grad_norm": 0.06333780770307451, "learning_rate": 4.7456895465747276e-08, "loss": 0.9129, "step": 10881 }, { "epoch": 4.9306751246035345, "grad_norm": 0.06283101063749282, "learning_rate": 4.6842693128144044e-08, "loss": 0.941, "step": 10882 }, { "epoch": 4.931128228364296, "grad_norm": 0.06320438566954441, "learning_rate": 4.62324889939314e-08, "loss": 0.9119, "step": 10883 }, { "epoch": 4.931581332125057, "grad_norm": 0.06330320831582087, "learning_rate": 4.562628312417161e-08, "loss": 0.8968, "step": 10884 }, { "epoch": 4.932034435885818, "grad_norm": 0.06360775060793897, "learning_rate": 4.5024075579536145e-08, "loss": 0.9175, "step": 10885 }, { "epoch": 4.932487539646579, "grad_norm": 0.06326453078039129, "learning_rate": 4.442586642028346e-08, "loss": 0.9171, "step": 10886 }, { "epoch": 4.93294064340734, "grad_norm": 0.06497256764077333, "learning_rate": 4.383165570628123e-08, "loss": 0.9385, "step": 10887 }, { "epoch": 4.933393747168101, "grad_norm": 0.06407508195979951, "learning_rate": 4.324144349698856e-08, "loss": 0.9385, "step": 10888 }, { "epoch": 4.933846850928862, "grad_norm": 0.06294313888223448, "learning_rate": 4.2655229851473746e-08, "loss": 0.9112, "step": 10889 }, { "epoch": 4.934299954689624, "grad_norm": 0.06198993278862499, "learning_rate": 4.207301482839654e-08, "loss": 0.9173, "step": 10890 }, { "epoch": 4.934753058450385, "grad_norm": 0.06324992191708864, "learning_rate": 4.149479848603033e-08, "loss": 0.9139, "step": 10891 }, { "epoch": 4.935206162211147, "grad_norm": 0.0633493800440037, "learning_rate": 4.0920580882226615e-08, "loss": 0.9147, "step": 10892 }, { "epoch": 4.935659265971908, "grad_norm": 0.06435093121050653, "learning_rate": 4.0350362074454974e-08, "loss": 0.95, "step": 10893 }, { "epoch": 4.936112369732669, "grad_norm": 0.06469218601786235, "learning_rate": 3.9784142119776434e-08, "loss": 0.9288, "step": 10894 }, { "epoch": 4.93656547349343, "grad_norm": 0.060810799314126356, "learning_rate": 3.9221921074852345e-08, "loss": 0.905, "step": 10895 }, { "epoch": 4.937018577254191, "grad_norm": 0.06332061593189667, "learning_rate": 3.8663698995948795e-08, "loss": 0.9288, "step": 10896 }, { "epoch": 4.937471681014952, "grad_norm": 0.06515219975938732, "learning_rate": 3.8109475938932214e-08, "loss": 0.9358, "step": 10897 }, { "epoch": 4.9379247847757135, "grad_norm": 0.06412030568708572, "learning_rate": 3.755925195925603e-08, "loss": 0.942, "step": 10898 }, { "epoch": 4.938377888536475, "grad_norm": 0.06432069905813197, "learning_rate": 3.7013027111982845e-08, "loss": 0.919, "step": 10899 }, { "epoch": 4.938830992297236, "grad_norm": 0.06380683154643174, "learning_rate": 3.64708014517845e-08, "loss": 0.9186, "step": 10900 }, { "epoch": 4.939284096057997, "grad_norm": 0.06377410372741008, "learning_rate": 3.5932575032910924e-08, "loss": 0.9227, "step": 10901 }, { "epoch": 4.939737199818758, "grad_norm": 0.06314241045764625, "learning_rate": 3.539834790923458e-08, "loss": 0.9188, "step": 10902 }, { "epoch": 4.94019030357952, "grad_norm": 0.062478960325547166, "learning_rate": 3.486812013420604e-08, "loss": 0.9185, "step": 10903 }, { "epoch": 4.940643407340281, "grad_norm": 0.062330400022996404, "learning_rate": 3.434189176089398e-08, "loss": 0.9381, "step": 10904 }, { "epoch": 4.941096511101042, "grad_norm": 0.06430750202551452, "learning_rate": 3.381966284195848e-08, "loss": 0.9226, "step": 10905 }, { "epoch": 4.941549614861803, "grad_norm": 0.06344235556213489, "learning_rate": 3.330143342965553e-08, "loss": 0.9079, "step": 10906 }, { "epoch": 4.942002718622565, "grad_norm": 0.06292903167644259, "learning_rate": 3.27872035758503e-08, "loss": 0.9231, "step": 10907 }, { "epoch": 4.942455822383326, "grad_norm": 0.0648256153148737, "learning_rate": 3.2276973331999414e-08, "loss": 0.9128, "step": 10908 }, { "epoch": 4.942908926144087, "grad_norm": 0.06448278261985789, "learning_rate": 3.177074274916869e-08, "loss": 0.9398, "step": 10909 }, { "epoch": 4.943362029904848, "grad_norm": 0.06343080771849584, "learning_rate": 3.126851187801094e-08, "loss": 0.9188, "step": 10910 }, { "epoch": 4.943815133665609, "grad_norm": 0.062263673937461114, "learning_rate": 3.0770280768788186e-08, "loss": 0.9296, "step": 10911 }, { "epoch": 4.944268237426371, "grad_norm": 0.06405317194250669, "learning_rate": 3.0276049471358316e-08, "loss": 0.9177, "step": 10912 }, { "epoch": 4.944721341187131, "grad_norm": 0.06316247736706312, "learning_rate": 2.978581803518399e-08, "loss": 0.9306, "step": 10913 }, { "epoch": 4.945174444947893, "grad_norm": 0.06556376467816681, "learning_rate": 2.929958650932374e-08, "loss": 0.9377, "step": 10914 }, { "epoch": 4.9456275487086545, "grad_norm": 0.0649374421383059, "learning_rate": 2.8817354942427544e-08, "loss": 0.9324, "step": 10915 }, { "epoch": 4.946080652469416, "grad_norm": 0.06236865722955991, "learning_rate": 2.8339123382763455e-08, "loss": 0.8993, "step": 10916 }, { "epoch": 4.946533756230177, "grad_norm": 0.065661237317549, "learning_rate": 2.7864891878186528e-08, "loss": 0.9171, "step": 10917 }, { "epoch": 4.946986859990938, "grad_norm": 0.06335245293672244, "learning_rate": 2.7394660476147695e-08, "loss": 0.9216, "step": 10918 }, { "epoch": 4.947439963751699, "grad_norm": 0.06558389818213431, "learning_rate": 2.692842922371153e-08, "loss": 0.9066, "step": 10919 }, { "epoch": 4.94789306751246, "grad_norm": 0.06355339349865306, "learning_rate": 2.646619816753404e-08, "loss": 0.9368, "step": 10920 }, { "epoch": 4.948346171273221, "grad_norm": 0.06245180039382253, "learning_rate": 2.6007967353867125e-08, "loss": 0.9106, "step": 10921 }, { "epoch": 4.948799275033982, "grad_norm": 0.06045597843030077, "learning_rate": 2.5553736828571874e-08, "loss": 0.9223, "step": 10922 }, { "epoch": 4.9492523787947444, "grad_norm": 0.06120358768141158, "learning_rate": 2.5103506637105257e-08, "loss": 0.899, "step": 10923 }, { "epoch": 4.949705482555506, "grad_norm": 0.06567678943490042, "learning_rate": 2.4657276824515687e-08, "loss": 0.9368, "step": 10924 }, { "epoch": 4.950158586316267, "grad_norm": 0.0634012081821103, "learning_rate": 2.421504743546521e-08, "loss": 0.9355, "step": 10925 }, { "epoch": 4.950611690077028, "grad_norm": 0.06283043507624785, "learning_rate": 2.377681851420288e-08, "loss": 0.9413, "step": 10926 }, { "epoch": 4.951064793837789, "grad_norm": 0.06371821798262166, "learning_rate": 2.3342590104586947e-08, "loss": 0.917, "step": 10927 }, { "epoch": 4.95151789759855, "grad_norm": 0.06425788278599362, "learning_rate": 2.2912362250075982e-08, "loss": 0.9188, "step": 10928 }, { "epoch": 4.951971001359311, "grad_norm": 0.06347303116344535, "learning_rate": 2.248613499371555e-08, "loss": 0.9376, "step": 10929 }, { "epoch": 4.952424105120072, "grad_norm": 0.064125814882883, "learning_rate": 2.2063908378160416e-08, "loss": 0.9287, "step": 10930 }, { "epoch": 4.9528772088808335, "grad_norm": 0.0608886100640434, "learning_rate": 2.1645682445670113e-08, "loss": 0.9194, "step": 10931 }, { "epoch": 4.953330312641595, "grad_norm": 0.06112132095221234, "learning_rate": 2.1231457238091168e-08, "loss": 0.911, "step": 10932 }, { "epoch": 4.953783416402356, "grad_norm": 0.06532208337342063, "learning_rate": 2.082123279687931e-08, "loss": 0.9384, "step": 10933 }, { "epoch": 4.954236520163118, "grad_norm": 0.06228952140458781, "learning_rate": 2.0415009163086143e-08, "loss": 0.9118, "step": 10934 }, { "epoch": 4.954689623923879, "grad_norm": 0.06434266480456274, "learning_rate": 2.001278637735915e-08, "loss": 0.9133, "step": 10935 }, { "epoch": 4.95514272768464, "grad_norm": 0.063669424311863, "learning_rate": 1.9614564479959464e-08, "loss": 0.9039, "step": 10936 }, { "epoch": 4.955595831445401, "grad_norm": 0.061769064326087815, "learning_rate": 1.9220343510726324e-08, "loss": 0.9098, "step": 10937 }, { "epoch": 4.956048935206162, "grad_norm": 0.0646462985354511, "learning_rate": 1.883012350911706e-08, "loss": 0.9328, "step": 10938 }, { "epoch": 4.956502038966923, "grad_norm": 0.06342637835554062, "learning_rate": 1.8443904514184874e-08, "loss": 0.9364, "step": 10939 }, { "epoch": 4.956955142727685, "grad_norm": 0.06394156820837972, "learning_rate": 1.8061686564569968e-08, "loss": 0.9197, "step": 10940 }, { "epoch": 4.957408246488446, "grad_norm": 0.06219612889725725, "learning_rate": 1.7683469698530632e-08, "loss": 0.9289, "step": 10941 }, { "epoch": 4.957861350249207, "grad_norm": 0.06412826459808069, "learning_rate": 1.730925395391214e-08, "loss": 0.9226, "step": 10942 }, { "epoch": 4.958314454009968, "grad_norm": 0.06333565367151446, "learning_rate": 1.6939039368164544e-08, "loss": 0.9176, "step": 10943 }, { "epoch": 4.958767557770729, "grad_norm": 0.06318719834507246, "learning_rate": 1.657282597833376e-08, "loss": 0.8951, "step": 10944 }, { "epoch": 4.959220661531491, "grad_norm": 0.06199826824042231, "learning_rate": 1.6210613821070475e-08, "loss": 0.9156, "step": 10945 }, { "epoch": 4.959673765292252, "grad_norm": 0.06672433534342648, "learning_rate": 1.5852402932616807e-08, "loss": 0.9223, "step": 10946 }, { "epoch": 4.960126869053013, "grad_norm": 0.061314512933030516, "learning_rate": 1.5498193348824076e-08, "loss": 0.8997, "step": 10947 }, { "epoch": 4.9605799728137745, "grad_norm": 0.06209143833021292, "learning_rate": 1.5147985105139483e-08, "loss": 0.9364, "step": 10948 }, { "epoch": 4.961033076574536, "grad_norm": 0.0625773432680176, "learning_rate": 1.4801778236610554e-08, "loss": 0.9405, "step": 10949 }, { "epoch": 4.961486180335297, "grad_norm": 0.062122724991739556, "learning_rate": 1.445957277787624e-08, "loss": 0.9311, "step": 10950 }, { "epoch": 4.961939284096058, "grad_norm": 0.0635202989319981, "learning_rate": 1.4121368763189147e-08, "loss": 0.8868, "step": 10951 }, { "epoch": 4.962392387856819, "grad_norm": 0.06332660004028276, "learning_rate": 1.3787166226388871e-08, "loss": 0.9218, "step": 10952 }, { "epoch": 4.96284549161758, "grad_norm": 0.06209871389253431, "learning_rate": 1.345696520091977e-08, "loss": 0.9146, "step": 10953 }, { "epoch": 4.963298595378341, "grad_norm": 0.06152544837067098, "learning_rate": 1.3130765719830962e-08, "loss": 0.9323, "step": 10954 }, { "epoch": 4.963751699139102, "grad_norm": 0.06285376094372774, "learning_rate": 1.2808567815763007e-08, "loss": 0.9459, "step": 10955 }, { "epoch": 4.964204802899864, "grad_norm": 0.06281150109902821, "learning_rate": 1.2490371520956779e-08, "loss": 0.9172, "step": 10956 }, { "epoch": 4.964657906660626, "grad_norm": 0.06269717621914075, "learning_rate": 1.2176176867257916e-08, "loss": 0.9253, "step": 10957 }, { "epoch": 4.965111010421387, "grad_norm": 0.07182997905997256, "learning_rate": 1.1865983886107935e-08, "loss": 0.9143, "step": 10958 }, { "epoch": 4.965564114182148, "grad_norm": 0.06194577329908563, "learning_rate": 1.155979260854867e-08, "loss": 0.928, "step": 10959 }, { "epoch": 4.966017217942909, "grad_norm": 0.061955777544919856, "learning_rate": 1.1257603065217838e-08, "loss": 0.9003, "step": 10960 }, { "epoch": 4.96647032170367, "grad_norm": 0.0654934029957693, "learning_rate": 1.0959415286362352e-08, "loss": 0.9243, "step": 10961 }, { "epoch": 4.966923425464431, "grad_norm": 0.06269059060204255, "learning_rate": 1.0665229301820568e-08, "loss": 0.9284, "step": 10962 }, { "epoch": 4.967376529225192, "grad_norm": 0.06312304560517173, "learning_rate": 1.0375045141026719e-08, "loss": 0.9306, "step": 10963 }, { "epoch": 4.9678296329859535, "grad_norm": 0.06394529819001332, "learning_rate": 1.008886283302868e-08, "loss": 0.9274, "step": 10964 }, { "epoch": 4.968282736746715, "grad_norm": 0.06279065685920274, "learning_rate": 9.80668240646132e-09, "loss": 0.9215, "step": 10965 }, { "epoch": 4.968735840507476, "grad_norm": 0.06571921472710365, "learning_rate": 9.528503889564277e-09, "loss": 0.9172, "step": 10966 }, { "epoch": 4.969188944268238, "grad_norm": 0.06306763797481822, "learning_rate": 9.25432731016862e-09, "loss": 0.9043, "step": 10967 }, { "epoch": 4.969642048028999, "grad_norm": 0.06396562094825858, "learning_rate": 8.984152695723503e-09, "loss": 0.9232, "step": 10968 }, { "epoch": 4.97009515178976, "grad_norm": 0.06342973818843041, "learning_rate": 8.717980073260635e-09, "loss": 0.9188, "step": 10969 }, { "epoch": 4.970548255550521, "grad_norm": 0.06385254025602995, "learning_rate": 8.455809469412046e-09, "loss": 0.9375, "step": 10970 }, { "epoch": 4.971001359311282, "grad_norm": 0.06290230625782096, "learning_rate": 8.19764091041897e-09, "loss": 0.9141, "step": 10971 }, { "epoch": 4.971454463072043, "grad_norm": 0.06355864818021026, "learning_rate": 7.943474422114073e-09, "loss": 0.9077, "step": 10972 }, { "epoch": 4.9719075668328045, "grad_norm": 0.06626858239860009, "learning_rate": 7.693310029939228e-09, "loss": 0.9208, "step": 10973 }, { "epoch": 4.972360670593566, "grad_norm": 0.06339234361804089, "learning_rate": 7.44714775891886e-09, "loss": 0.9236, "step": 10974 }, { "epoch": 4.972813774354327, "grad_norm": 0.060774535567143115, "learning_rate": 7.204987633691041e-09, "loss": 0.903, "step": 10975 }, { "epoch": 4.973266878115089, "grad_norm": 0.06361256901165725, "learning_rate": 6.966829678494158e-09, "loss": 0.9139, "step": 10976 }, { "epoch": 4.973719981875849, "grad_norm": 0.061511762983383254, "learning_rate": 6.7326739171535985e-09, "loss": 0.9266, "step": 10977 }, { "epoch": 4.974173085636611, "grad_norm": 0.06396914877559398, "learning_rate": 6.5025203731083894e-09, "loss": 0.9119, "step": 10978 }, { "epoch": 4.974626189397372, "grad_norm": 0.0625899210504785, "learning_rate": 6.276369069384558e-09, "loss": 0.9291, "step": 10979 }, { "epoch": 4.975079293158133, "grad_norm": 0.062047938446465827, "learning_rate": 6.0542200286173305e-09, "loss": 0.8996, "step": 10980 }, { "epoch": 4.9755323969188945, "grad_norm": 0.06316118251469398, "learning_rate": 5.8360732730378125e-09, "loss": 0.9258, "step": 10981 }, { "epoch": 4.975985500679656, "grad_norm": 0.06263455959065842, "learning_rate": 5.621928824472988e-09, "loss": 0.94, "step": 10982 }, { "epoch": 4.976438604440417, "grad_norm": 0.06321793679647143, "learning_rate": 5.4117867043590455e-09, "loss": 0.9215, "step": 10983 }, { "epoch": 4.976891708201178, "grad_norm": 0.06349462287475861, "learning_rate": 5.205646933719166e-09, "loss": 0.936, "step": 10984 }, { "epoch": 4.977344811961939, "grad_norm": 0.06173791716469971, "learning_rate": 5.0035095331857356e-09, "loss": 0.9265, "step": 10985 }, { "epoch": 4.9777979157227, "grad_norm": 0.06412865582886051, "learning_rate": 4.805374522987016e-09, "loss": 0.9433, "step": 10986 }, { "epoch": 4.978251019483462, "grad_norm": 0.06331974411286581, "learning_rate": 4.611241922947152e-09, "loss": 0.9348, "step": 10987 }, { "epoch": 4.978704123244223, "grad_norm": 0.06477313231355653, "learning_rate": 4.421111752495044e-09, "loss": 0.9225, "step": 10988 }, { "epoch": 4.979157227004984, "grad_norm": 0.0630443421469633, "learning_rate": 4.234984030664357e-09, "loss": 0.9202, "step": 10989 }, { "epoch": 4.9796103307657456, "grad_norm": 0.06375251943833789, "learning_rate": 4.0528587760713114e-09, "loss": 0.9454, "step": 10990 }, { "epoch": 4.980063434526507, "grad_norm": 0.06391455484747013, "learning_rate": 3.874736006945767e-09, "loss": 0.9416, "step": 10991 }, { "epoch": 4.980516538287268, "grad_norm": 0.06349770196218647, "learning_rate": 3.700615741113467e-09, "loss": 0.9338, "step": 10992 }, { "epoch": 4.980969642048029, "grad_norm": 0.06434919632238782, "learning_rate": 3.5304979959960293e-09, "loss": 0.9182, "step": 10993 }, { "epoch": 4.98142274580879, "grad_norm": 0.0668475331888571, "learning_rate": 3.3643827886242764e-09, "loss": 0.9214, "step": 10994 }, { "epoch": 4.981875849569551, "grad_norm": 0.06393783561737701, "learning_rate": 3.2022701356160256e-09, "loss": 0.9223, "step": 10995 }, { "epoch": 4.982328953330312, "grad_norm": 0.061703058179677306, "learning_rate": 3.0441600531938564e-09, "loss": 0.9294, "step": 10996 }, { "epoch": 4.9827820570910735, "grad_norm": 0.061216604974492744, "learning_rate": 2.890052557185108e-09, "loss": 0.9216, "step": 10997 }, { "epoch": 4.9832351608518355, "grad_norm": 0.06298753828070398, "learning_rate": 2.7399476630041167e-09, "loss": 0.9055, "step": 10998 }, { "epoch": 4.983688264612597, "grad_norm": 0.06360467797708795, "learning_rate": 2.5938453856788615e-09, "loss": 0.9279, "step": 10999 }, { "epoch": 4.984141368373358, "grad_norm": 0.0626596535293152, "learning_rate": 2.451745739828759e-09, "loss": 0.9092, "step": 11000 }, { "epoch": 4.984594472134119, "grad_norm": 0.06432202809158977, "learning_rate": 2.3136487396735465e-09, "loss": 0.9278, "step": 11001 }, { "epoch": 4.98504757589488, "grad_norm": 0.06381234846477402, "learning_rate": 2.179554399033279e-09, "loss": 0.932, "step": 11002 }, { "epoch": 4.985500679655641, "grad_norm": 0.06451053449591763, "learning_rate": 2.0494627313238923e-09, "loss": 0.9432, "step": 11003 }, { "epoch": 4.985953783416402, "grad_norm": 0.060858508655193745, "learning_rate": 1.923373749566082e-09, "loss": 0.9332, "step": 11004 }, { "epoch": 4.986406887177163, "grad_norm": 0.0639470317910377, "learning_rate": 1.8012874663808633e-09, "loss": 0.9119, "step": 11005 }, { "epoch": 4.9868599909379245, "grad_norm": 0.06229663229553552, "learning_rate": 1.6832038939806893e-09, "loss": 0.91, "step": 11006 }, { "epoch": 4.987313094698686, "grad_norm": 0.06290449099640562, "learning_rate": 1.569123044187215e-09, "loss": 0.931, "step": 11007 }, { "epoch": 4.987766198459447, "grad_norm": 0.06286394386156617, "learning_rate": 1.4590449284090924e-09, "loss": 0.9108, "step": 11008 }, { "epoch": 4.988219302220209, "grad_norm": 0.06454210165894984, "learning_rate": 1.3529695576730562e-09, "loss": 0.9284, "step": 11009 }, { "epoch": 4.98867240598097, "grad_norm": 0.06419667538429237, "learning_rate": 1.2508969425839568e-09, "loss": 0.9363, "step": 11010 }, { "epoch": 4.989125509741731, "grad_norm": 0.06322121631610331, "learning_rate": 1.1528270933647279e-09, "loss": 0.9307, "step": 11011 }, { "epoch": 4.989578613502492, "grad_norm": 0.06182972897654768, "learning_rate": 1.0587600198253001e-09, "loss": 0.9049, "step": 11012 }, { "epoch": 4.990031717263253, "grad_norm": 0.061520773291435696, "learning_rate": 9.686957313803647e-10, "loss": 0.9105, "step": 11013 }, { "epoch": 4.9904848210240145, "grad_norm": 0.06354155802784192, "learning_rate": 8.826342370404916e-10, "loss": 0.9233, "step": 11014 }, { "epoch": 4.990937924784776, "grad_norm": 0.06266530540505308, "learning_rate": 8.005755454210118e-10, "loss": 0.9106, "step": 11015 }, { "epoch": 4.991391028545537, "grad_norm": 0.06310770690345778, "learning_rate": 7.225196647286936e-10, "loss": 0.9171, "step": 11016 }, { "epoch": 4.991844132306298, "grad_norm": 0.06402040187453724, "learning_rate": 6.484666027839481e-10, "loss": 0.9303, "step": 11017 }, { "epoch": 4.992297236067059, "grad_norm": 0.06392331548270425, "learning_rate": 5.784163669853016e-10, "loss": 0.9357, "step": 11018 }, { "epoch": 4.99275033982782, "grad_norm": 0.06407856384993864, "learning_rate": 5.123689643538043e-10, "loss": 0.9218, "step": 11019 }, { "epoch": 4.993203443588582, "grad_norm": 0.06241999378598966, "learning_rate": 4.503244014975039e-10, "loss": 0.9254, "step": 11020 }, { "epoch": 4.993656547349343, "grad_norm": 0.06264069576359992, "learning_rate": 3.922826846158856e-10, "loss": 0.9302, "step": 11021 }, { "epoch": 4.994109651110104, "grad_norm": 0.06229595697172689, "learning_rate": 3.38243819530959e-10, "loss": 0.917, "step": 11022 }, { "epoch": 4.9945627548708655, "grad_norm": 0.06353574665395337, "learning_rate": 2.8820781163840796e-10, "loss": 0.9107, "step": 11023 }, { "epoch": 4.995015858631627, "grad_norm": 0.062497595048627515, "learning_rate": 2.421746659519997e-10, "loss": 0.8979, "step": 11024 }, { "epoch": 4.995468962392388, "grad_norm": 0.06396051674598557, "learning_rate": 2.0014438707693927e-10, "loss": 0.9032, "step": 11025 }, { "epoch": 4.995922066153149, "grad_norm": 0.06211454441055644, "learning_rate": 1.6211697921875158e-10, "loss": 0.9221, "step": 11026 }, { "epoch": 4.99637516991391, "grad_norm": 0.06131497174043291, "learning_rate": 1.280924461832811e-10, "loss": 0.9095, "step": 11027 }, { "epoch": 4.996828273674671, "grad_norm": 0.06386491075588138, "learning_rate": 9.807079137669207e-11, "loss": 0.9292, "step": 11028 }, { "epoch": 4.997281377435432, "grad_norm": 0.06327058927782486, "learning_rate": 7.205201780102755e-11, "loss": 0.9169, "step": 11029 }, { "epoch": 4.9977344811961935, "grad_norm": 0.06104884006585481, "learning_rate": 5.0036128063091214e-11, "loss": 0.9293, "step": 11030 }, { "epoch": 4.9981875849569555, "grad_norm": 0.06258212603181304, "learning_rate": 3.202312436556554e-11, "loss": 0.9395, "step": 11031 }, { "epoch": 4.998640688717717, "grad_norm": 0.06236557543899841, "learning_rate": 1.801300850701182e-11, "loss": 0.9096, "step": 11032 }, { "epoch": 4.999093792478478, "grad_norm": 0.0633166964327996, "learning_rate": 8.005781890751962e-12, "loss": 0.9321, "step": 11033 }, { "epoch": 4.999546896239239, "grad_norm": 0.06351645827119114, "learning_rate": 2.0014455248684726e-12, "loss": 0.9055, "step": 11034 }, { "epoch": 5.0, "grad_norm": 0.0641531063840526, "learning_rate": 0.0, "loss": 0.9064, "step": 11035 }, { "epoch": 5.0, "step": 11035, "total_flos": 8.047664906108928e+16, "train_loss": 0.3708416173562887, "train_runtime": 42068.7646, "train_samples_per_second": 134.287, "train_steps_per_second": 0.262 } ], "logging_steps": 1, "max_steps": 11035, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.047664906108928e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }